2 hours ago by gojomo

Something about this reminds me of the 'Winning Ticket Hypothesis' in artifical neural networks: that some 'random' initializations prime a network far better for later faster learning. From https://arxiv.org/abs/1803.03635 - the abstract:

Neural network pruning techniques can reduce the parameter counts of trained networks by over 90%, decreasing storage requirements and improving computational performance of inference without compromising accuracy. However, contemporary experience is that the sparse architectures produced by pruning are difficult to train from the start, which would similarly improve training performance.

We find that a standard pruning technique naturally uncovers subnetworks whose initializations made them capable of training effectively. Based on these results, we articulate the "lottery ticket hypothesis:" dense, randomly-initialized, feed-forward networks contain subnetworks ("winning tickets") that - when trained in isolation - reach test accuracy comparable to the original network in a similar number of iterations. The winning tickets we find have won the initialization lottery: their connections have initial weights that make training particularly effective.

We present an algorithm to identify winning tickets and a series of experiments that support the lottery ticket hypothesis and the importance of these fortuitous initializations. We consistently find winning tickets that are less than 10-20% of the size of several fully-connected and convolutional feed-forward architectures for MNIST and CIFAR10. Above this size, the winning tickets that we find learn faster than the original network and reach higher test accuracy.

Are these innate patterns of activity priming mammalian brains in a similar way?

an hour ago by teruakohatu

> Are these innate patterns of activity priming mammalian brains in a similar way?

A newborn foal can stand within 55 minutes of being born and can walk or run within 90 minutes. That is crazy fast training speed. Is it training initialization or is it a form of transfer learning?

Most human babies take 9-18 months to walk, and this is commonly attributed to the fact that human babies have immature brains that cannot mature due to birth canal constraints, which is constrained by pelvis size which is constrained by the need to walk upright.

I have wondered about how true this is. Maybe our brains are just so much more powerful, or capable of much deeper understanding, that we require a much lower initial learning rate.

The lottery ticket could explain why some children learn to walk at 6 months, while others are closer to 24 months, with no different in intelligence or motor skills in later life.

Edit: As some people below have correctly pointed out, human babies could not walk within an hour for physiological reasons, but nor do they exhibit the basic motor skills, spatial reasoning or image processing required for walking. Many human babies even struggle to feed for the first 24-48 hours or longer.

Meanwhile baboon babies can hold onto their mother from birth while the mother climbs trees.

a minute ago by jvanderbot

> Is it training initialization or is it a form of transfer learning?

Let's not ignore genetics, which pre-populate weights and subnetworks in the form of motion-control centers in the brain. Once you consider the gene-brain interaction over the course of evolution as part of your training, the huge data required for modern ML starts to make more sense.

an hour ago by chrisseaton

> Most human babies take 9-18 months to walk, and this is commonly attributed to the fact that human babies have immature brains

Have you ever seen a new-born baby's legs?

They aren't even remotely strong enough to walk, no matter what the brain tells them. It takes those months to develop leg muscles. New-borns don't have them.

It's absolutely not restricted by the brain size.

22 minutes ago by emmelaich

Well, the legs may be less developed since they're not useful because ... the brain size is not big enough?

Anyway, hard to conclude, just think your logic is missing a little.

24 minutes ago by outworlder

True. But the movements on all limbs are quite uncoordinated at first. It's also not restricted to muscle strength.

an hour ago by gojomo

Maybe given the length & complexity of a full human life, a bunch of brain factors actually work to slow specialization down to prevent early overfitting.

an hour ago by 3pt14159

> that we require a much lower initial learning rate.

The way I understand it is that is not so much required as not selected for. We could certainly be able to walk at one day old if evolutionary pressure made it necessary, but it prioritized other aspects in our cognitive development to survive so we don't. It's not as if we say "oh, horses can walk early and therefore are incapable of rational thought" we evaluate them on their own merits.

21 minutes ago by gota

I guess I can see some parallel the way you describe it, but it is not that tight of an analogy IMO. Correct me if I'm wrong but the Lottery Ticket Hypothesis mainly means that, for many of our tasks, we start with needlessly huge networks, and thus a good portion of the learning process is just the network learning to "cope with the noise" from the extraneous connections

Hence why the lottery ticket network (being a smaller version of the same initial network) can achieve similar accuracy (but not 'better') in much shorter time - because you took away all the confounding stuff. That kind of 'counterfactual' doesn't really translate to a biological domain, right?

Also - the baby brain is not really a 'pruned', smaller version of the adult brain, right? Neither is the fetus brain compared to the baby brain (which is more to the point of the article). So the analogy breaks down there too

I guess what the article describes is a bit like "pre-training" a full network with 'synthetic' experiences that quickly tune it for later. In that sense, I think its like these "dreams" are a 'distilled dataset' [1]. The question then is: how is this dataset being passed on from mother/father to child?

[1] https://arxiv.org/abs/1811.10959

Abstract:

> Model distillation aims to distill the knowledge of a complex model into a simpler one. In this paper, we consider an alternative formulation called dataset distillation: we keep the model fixed and instead attempt to distill the knowledge from a large training dataset into a small one. The idea is to synthesize a small number of data points that do not need to come from the correct data distribution, but will, when given to the learning algorithm as training data, approximate the model trained on the original data. For example, we show that it is possible to compress 60,000 MNIST training images into just 10 synthetic distilled images (one per class) and achieve close to original performance with only a few gradient descent steps, given a fixed network initialization. We evaluate our method in various initialization settings and with different learning objectives. Experiments on multiple datasets show the advantage of our approach compared to alternative methods.

12 minutes ago by optimalsolver

How hard would it be to find random networks that don't need to be trained at all?

39 minutes ago by emtel

It’s not just mammals. I’ve always been astounded that spiders know how to spin webs. Their mothers don’t teach them how!

14 minutes ago by gota

Or this, which is mind-blowing:

https://www.smithsonianmag.com/smart-news/dragonfly-undertak...

Insects that have large migration patterns - all of the individuals die before coming back to the initial spot; only their descendants continue the cycle. How?!

2 hours ago by khr

This article didn't link to the research article directly. Here it is for those interested:

https://science.sciencemag.org/content/373/6553/eabd0830/

an hour ago by axiom92

Similar: https://www.cell.com/patterns/fulltext/S2666-3899(21)00064-7

Dreams help the brain in generalizing the world. From the article:

"That is, dreams are a biological mechanism for increasing generalizability via the creation of corrupted sensory inputs from stochastic activity across the hierarchy of neural structures."

an hour ago by imbnwa

Plato grinning ear-to-ear somewhere right now

Freud also predicted this in that he postulated that hallucination was more primordial than actual sense-experience

an hour ago by sjg007

Do they do this in the dark and is it correlated with light input.. so is the eye/brain responding to stimulus even though the eyes aren't open yet?

an hour ago by rudyfink

I recall once hearing a talk that discussed, what I recall as, "test patterns" being run across the visual cortex prior to birth. Essentially, waves / patterns of stimulation would go across the optic structure that would be used to, again as I recall it, "calibrate" the visual system.

I dug a bit and was able to find this paper https://pubmed.ncbi.nlm.nih.gov/15289028/ , which seems to track my memory. It has a number of forward and backward citations if you are interested. Perhaps someone with far more than my vague recollection of a talk will step in though!

an hour ago by yellow_lead

The full title is pretty funny.

> Eyes wide shut: How newborn mammals dream the world they’re entering

Reference to the Kubrick film I assume.

2 hours ago by euske

I always wondered how our brain is pre-wired to recognize 3D objects right after a birth, but this article is saying that eyes are giving fake image-like input to our brain so that it can perform "pre-training"? Does this imply that our brain is truly a blank slate if these initial seeds aren't given?

2 minutes ago by Davidzheng

sorry but I don't think there is any sense in which the infant brain is a blank slate. Definitely not for vital functions (as seen by the "instinctual" abilities noted elsewhere in this thread, not in language/cognition, and IMO not even culturally/morally.

2 hours ago by fossuser

I don't think brains are pre-wired to recognize 3D objects.

There was an example case of a blind man with cataracts getting it fixed at around 50 years old. People were curious if he would be able to know if a sphere was round just by looking at it without touching it.

IIRC - not only could he not do that, but he couldn't visually interpret shadows (saw them as black splotches - didn't recognize depth), and was confused why objects got smaller as they were moving away.

I think the human visual system trains on a lot of visual input data, but it typically happens at the baby stage where you can't really interact.

an hour ago by echelon

Fascinating anecdote, and I'd love to see the report on this. I wonder what else they learned from such an interesting case study.

But I doubt it's 100% "no pre-wiring". What about gravity and musculature? Surely brains must know or train on these in some way before birth.

And what about fight or flight reflexes and presupposing monsters in the shadows? My understanding is that these had a primitive evolutionary basis and that we came hardwired to respond to certain stimuli.

an hour ago by fossuser

Yeah - I just meant specifically 3D visual modeling. I think some fears are known to be innate among apes - fear of the dark, snakes, and falling. Also paying attention to faces?

The cataracts thing was an article I saw on HN a few years ago, but don’t remember the title.

2 hours ago by dukeofdoom

Makes sense. Baby wildebeests has to learn to walk in minutes after birth. Hungry Lions are literally waiting for their lunch to be born.

Daily digest email

Get a daily email with the the top stories from Hacker News. No spam, unsubscribe at any time.