Why do neural networks need so many training examples to perform?What can we learn about the human brain from...
GRASS not working with QGIS 3.6
Constexpr if with a non-bool condition
Differences between prior distribution and prior predictive distribution?
What makes papers publishable in top-tier journals?
Salsa20 Implementation: Sum of 2 Words with Carries Suppressed
When do I have to declare that I want to twin my spell?
How do I draw the dashed lines as shown in this figure
Is this ordinary workplace experiences for a job in Software Engineering?
Why do cars have plastic shrouds over the engine?
general past possibility with COULD
Why do neural networks need so many training examples to perform?
Is there any risk in sharing info about technologies and products we use with a supplier?
Hilchos Shabbos English Sefer
How can I play a serial killer in a party of good PCs?
Is it a fallacy if someone claims they need an explanation for every word of your argument to the point where they don't understand common terms?
Crontab: Ubuntu running script (noob)
How to deal with possible delayed baggage?
Play Zip, Zap, Zop
Has any human ever had the choice to leave Earth permanently?
Can I announce prefix 161.117.25.0/24 even though I don't have all of /24 IPs?
How do I append a character to the end of every line in an Excel cell?
Do authors have to be politically correct in article-writing?
Why does magnet wire need to be insulated?
Avoid page break between paragraphs
Why do neural networks need so many training examples to perform?
What can we learn about the human brain from artificial neural networks?Incremental training of Neural NetworksDo Neural Networks need “compound” features?Can a neural network learn a functional, and its functional derivative?Training Neural Networks on variable length vectorsNeural networks: why do we randomize the training set?Deep networks vs shallow networks: why do we need depth?Training Neural Net with examples it misclassifiedOne big neural network or many small neural networks?Early stopping criteria when training neural networksWhy do neural networks need feature selection / engineering?
$begingroup$
A human child at age 2 needs around 5 instances of a car to be able to identify it with reasonable accuracy regardless of color, make, etc. When my son was 2, he was able to identify trams and trains, even though he had seen just a few. Since he was usually confusing one with each other, apparently his neural network was not trained enough, but still.
What is it that artificial neural networks are missing that prevent them from being able to learn way quicker? Is transfer learning an answer?
neural-networks neuroscience
$endgroup$
|
show 14 more comments
$begingroup$
A human child at age 2 needs around 5 instances of a car to be able to identify it with reasonable accuracy regardless of color, make, etc. When my son was 2, he was able to identify trams and trains, even though he had seen just a few. Since he was usually confusing one with each other, apparently his neural network was not trained enough, but still.
What is it that artificial neural networks are missing that prevent them from being able to learn way quicker? Is transfer learning an answer?
neural-networks neuroscience
$endgroup$
19
$begingroup$
Elephants might be a better example than cars. As others have noted, a child may have seen many cars before hearing the label, so if their mind already defines "natural kinds" it now has a label for one. However, a Western child indisputably develops a good elephant-classifying system on the basis of just a few data.
$endgroup$
– J.G.
2 days ago
40
$begingroup$
What makes you think that a human child’s brain works like a neural network?
$endgroup$
– Paul Wasilewski
2 days ago
9
$begingroup$
A NN can be shown an image of a car. Your child gets a full 3D movie from different perspectives, for several different types of car. Your child also likely has similar examples to distinguish a car from. For instance their baby stroller, toys, etc. Without those, I think your child would have needed more examples.
$endgroup$
– Stian Yttervik
2 days ago
15
$begingroup$
@MSalters In the sense of an Artificial Neural Network? Probably not.
$endgroup$
– Firebug
2 days ago
9
$begingroup$
"A human child at age 2 needs around 5 instances of a car to be able to identify it with reasonable accuracy" Such a child has had two full years of experience with things that aren't cars. I'm certain that plays a significant role.
$endgroup$
– DarthFennec
yesterday
|
show 14 more comments
$begingroup$
A human child at age 2 needs around 5 instances of a car to be able to identify it with reasonable accuracy regardless of color, make, etc. When my son was 2, he was able to identify trams and trains, even though he had seen just a few. Since he was usually confusing one with each other, apparently his neural network was not trained enough, but still.
What is it that artificial neural networks are missing that prevent them from being able to learn way quicker? Is transfer learning an answer?
neural-networks neuroscience
$endgroup$
A human child at age 2 needs around 5 instances of a car to be able to identify it with reasonable accuracy regardless of color, make, etc. When my son was 2, he was able to identify trams and trains, even though he had seen just a few. Since he was usually confusing one with each other, apparently his neural network was not trained enough, but still.
What is it that artificial neural networks are missing that prevent them from being able to learn way quicker? Is transfer learning an answer?
neural-networks neuroscience
neural-networks neuroscience
edited yesterday
smci
87211018
87211018
asked Feb 24 at 14:07
MarcinMarcin
405410
405410
19
$begingroup$
Elephants might be a better example than cars. As others have noted, a child may have seen many cars before hearing the label, so if their mind already defines "natural kinds" it now has a label for one. However, a Western child indisputably develops a good elephant-classifying system on the basis of just a few data.
$endgroup$
– J.G.
2 days ago
40
$begingroup$
What makes you think that a human child’s brain works like a neural network?
$endgroup$
– Paul Wasilewski
2 days ago
9
$begingroup$
A NN can be shown an image of a car. Your child gets a full 3D movie from different perspectives, for several different types of car. Your child also likely has similar examples to distinguish a car from. For instance their baby stroller, toys, etc. Without those, I think your child would have needed more examples.
$endgroup$
– Stian Yttervik
2 days ago
15
$begingroup$
@MSalters In the sense of an Artificial Neural Network? Probably not.
$endgroup$
– Firebug
2 days ago
9
$begingroup$
"A human child at age 2 needs around 5 instances of a car to be able to identify it with reasonable accuracy" Such a child has had two full years of experience with things that aren't cars. I'm certain that plays a significant role.
$endgroup$
– DarthFennec
yesterday
|
show 14 more comments
19
$begingroup$
Elephants might be a better example than cars. As others have noted, a child may have seen many cars before hearing the label, so if their mind already defines "natural kinds" it now has a label for one. However, a Western child indisputably develops a good elephant-classifying system on the basis of just a few data.
$endgroup$
– J.G.
2 days ago
40
$begingroup$
What makes you think that a human child’s brain works like a neural network?
$endgroup$
– Paul Wasilewski
2 days ago
9
$begingroup$
A NN can be shown an image of a car. Your child gets a full 3D movie from different perspectives, for several different types of car. Your child also likely has similar examples to distinguish a car from. For instance their baby stroller, toys, etc. Without those, I think your child would have needed more examples.
$endgroup$
– Stian Yttervik
2 days ago
15
$begingroup$
@MSalters In the sense of an Artificial Neural Network? Probably not.
$endgroup$
– Firebug
2 days ago
9
$begingroup$
"A human child at age 2 needs around 5 instances of a car to be able to identify it with reasonable accuracy" Such a child has had two full years of experience with things that aren't cars. I'm certain that plays a significant role.
$endgroup$
– DarthFennec
yesterday
19
19
$begingroup$
Elephants might be a better example than cars. As others have noted, a child may have seen many cars before hearing the label, so if their mind already defines "natural kinds" it now has a label for one. However, a Western child indisputably develops a good elephant-classifying system on the basis of just a few data.
$endgroup$
– J.G.
2 days ago
$begingroup$
Elephants might be a better example than cars. As others have noted, a child may have seen many cars before hearing the label, so if their mind already defines "natural kinds" it now has a label for one. However, a Western child indisputably develops a good elephant-classifying system on the basis of just a few data.
$endgroup$
– J.G.
2 days ago
40
40
$begingroup$
What makes you think that a human child’s brain works like a neural network?
$endgroup$
– Paul Wasilewski
2 days ago
$begingroup$
What makes you think that a human child’s brain works like a neural network?
$endgroup$
– Paul Wasilewski
2 days ago
9
9
$begingroup$
A NN can be shown an image of a car. Your child gets a full 3D movie from different perspectives, for several different types of car. Your child also likely has similar examples to distinguish a car from. For instance their baby stroller, toys, etc. Without those, I think your child would have needed more examples.
$endgroup$
– Stian Yttervik
2 days ago
$begingroup$
A NN can be shown an image of a car. Your child gets a full 3D movie from different perspectives, for several different types of car. Your child also likely has similar examples to distinguish a car from. For instance their baby stroller, toys, etc. Without those, I think your child would have needed more examples.
$endgroup$
– Stian Yttervik
2 days ago
15
15
$begingroup$
@MSalters In the sense of an Artificial Neural Network? Probably not.
$endgroup$
– Firebug
2 days ago
$begingroup$
@MSalters In the sense of an Artificial Neural Network? Probably not.
$endgroup$
– Firebug
2 days ago
9
9
$begingroup$
"A human child at age 2 needs around 5 instances of a car to be able to identify it with reasonable accuracy" Such a child has had two full years of experience with things that aren't cars. I'm certain that plays a significant role.
$endgroup$
– DarthFennec
yesterday
$begingroup$
"A human child at age 2 needs around 5 instances of a car to be able to identify it with reasonable accuracy" Such a child has had two full years of experience with things that aren't cars. I'm certain that plays a significant role.
$endgroup$
– DarthFennec
yesterday
|
show 14 more comments
9 Answers
9
active
oldest
votes
$begingroup$
There's a kind of goalpost moving that underlies this question. It used to be that NNs weren't very good at image recognition, so no one compared them to humans. Now that NNs are good at image tasks, suddenly it's the fault of NNs that they require a lot of training data to be comparable to children.
You can also turn this logic on its head. Suppose a child sees a number of cars the day that it's born. I wouldn't expect the child to be able to pick out a car the next day, or the next week, even though it's seen so many examples. Why are newborns so slow to learn? Because it takes a lot of exposure to the real world and, and the passage of time to change the child's neural pathways. For a neural network, we call this “training data,” but for a child we call it “growing up.”
$endgroup$
26
$begingroup$
To make it a bit more specific, a human child has already had years of training with tens of thousands of example allowing them to determining how objects look when viewed from different angles, how to identify their boundaries, the relationship between apparent size and actual size, and so on.
$endgroup$
– David Schwartz
2 days ago
18
$begingroup$
A child's brain is active inside the womb. The baby can identify their parents by sound, after the sound is filtered through water. A new-born baby had months of data to work with before they're born, but they still need years more before they can form a word, then couple more years before they can form a sentence, then couple more for a grammatically correct sentence, etc... learning is very complicated.
$endgroup$
– Nelson
2 days ago
4
$begingroup$
@EelcoHoogendoorn it explains the contrast 'child' versus 'neural network' that has been used in the question. The answer is that this is only an apparent contrast. Neural networks do not need that many examples at all, as kids get also many examples (but just in a different way) before they are able to recognize cars.
$endgroup$
– Martijn Weterings
yesterday
3
$begingroup$
@Nelson, I am not sure what the reason is for your comment, but you can change 'years' into 'year'. With 1 year kids speak words, with 2 years the first sentences are spoken, and with 3 years grammar, such as past tense and pronouns, becomes correctly used.
$endgroup$
– Martijn Weterings
yesterday
1
$begingroup$
@EelcoHoogendoorn I think the premise of the question is a case of reasoning from a faulty analogy, so directly address the analogy is responsive. Contrasting biological and artificial neural networks is also responsive, because the answer would outline how biological and artificial neural networks are most similar in their name (both contain the phrase "neural networks") but not similar in their essential characteristics, or at least the characteristics assumed by the question.
$endgroup$
– Sycorax
20 hours ago
|
show 4 more comments
$begingroup$
First of all, at age two, a child knows a lot about the world and actively applies this knowledge. A child does a lot of "transfer learning" by applying this knowledge to new concepts.
Second, before seeing those five "labeled" examples of cars, a child sees a lot of cars on the street, on TV, toy cars, etc., so also a lot of "unsupervised learning" happens beforehand.
Finally, neural networks have almost nothing in common with the human brain, so there's not much point in comparing them. Also notice that there are algorithms for one-shot learning, and pretty much research on it currently happens.
$endgroup$
2
$begingroup$
4th point, a child also has more than 100 million years of evolutionary selection towards learning efficiently/accurately.
$endgroup$
– csiz
11 hours ago
$begingroup$
@csiz People still believe in Darwin's hypothesis, eh?
$endgroup$
– Failed Scientist
4 hours ago
add a comment |
$begingroup$
One major aspect that I don't see in current answers is evolution.
A child's brain does not learn from scratch. It's similar to asking how deer and giraffe babies can walk a few minutes after birth. Because they are born with their brains already wired for this task. There is some fine-tuning needed of course, but the baby deer doesn't learn to walk from "random initialization".
Similarly, the fact that big moving objects exist and are important to keep track of is something we are born with.
So I think the presupposition of this question is simply false. Human neural networks had the opportunity to see tons of - maybe not cars but - moving, rotating 3D objects with difficult textures and shapes etc., but this happened through lots of generations and the learning took place by evolutionary algorithms, i.e. the ones whose brain was better structured for this task, could live to reproduce with higher chance, leaving the next generation with better and better brain wiring from the start.
$endgroup$
4
$begingroup$
Fun aside: there's evidence that when it comes to discriminating between different models of cars, we actually leverage the specialized facial recognition center of our brain. It's plausible that, while a child may not distinguish between different models, the implicit presence of a 'face' on a mobile object may cause cars to be categorized as a type of creature and therefore be favored to be identified by evolution, since recognizing mobile objects with faces is helpful to survival.
$endgroup$
– Dan Bryant
yesterday
4
$begingroup$
This answer addresses exactly what I was thinking. Children are not born as blank slates. They come with features that make some patterns easier to recognize, some things easier to learn, etc.
$endgroup$
– Eff
yesterday
1
$begingroup$
While animals that walk right out of the womb are indeed fascinating, such evolutionary hardwiring is thought to be at the very opposite extreme of human learning, which is thought to be the extreme of experience-driven learning in the natural world. Certainly cars will have left minimal evolutionary impact on the evolution of our brains.
$endgroup$
– Eelco Hoogendoorn
yesterday
2
$begingroup$
@EelcoHoogendoorn The ability to learn and understand the environment has been evolutionarily selected for. The brain has been set up by evolution to be extremely efficient at learning. The ability to connect the dots, see patterns, understand shapes and movement, makes inferences, etc.
$endgroup$
– Eff
yesterday
2
$begingroup$
This is a good point, but it's also true that as researchers come to understand this, they build NN's that have hard-coded structures that facilitate certain types of learning. Consider that a convolutional NN has hard coded receptive fields that greatly speed up learning / enhance performance on visual tasks. Those fields could be learned from scratch in a fully connected network, but it's much harder. @EelcoHoogendoorn, human brains are full of structure that facilitates learning.
$endgroup$
– gung♦
22 hours ago
|
show 5 more comments
$begingroup$
I don't know much about neural networks but I know a fair bit about babies.
Many 2 year olds have a lot of issues with how general words should be. For instance, it is quite common at that age for kids to use "dog" for any four legged animal. That's a more difficult distinction than "car" - just think how different a poodle looks from a great Dane, for instance and yet they are both "dog" while a cat is not.
And a child at 2 has seen many many more than 5 examples of "car". A kid sees dozens or even hundreds of examples of cars any time the family goes for a drive. And a lot of parents will comment "look at the car" a lot more than 5 times. But kids can also think in ways that they weren't told about. For instance, on the street the kid sees lots of things lined up. His dad says (of one) "look at the shiny car!" and the kid thinks "maybe all those other things lined up are also cars?"
$endgroup$
$begingroup$
Other examples: Taxi's, driving lesson cars, and police cars are the same. Whenever a car is red then it is a firetruck. Campervans are ambulances. A lorry with a loader crane becomes classified as an excavator. The bus that just passed by goes to the train station, so the next bus, which looks the same, must also be going to the train station. And seeing the moon during broad daylight is a very special event.
$endgroup$
– Martijn Weterings
yesterday
add a comment |
$begingroup$
This is an a fascinating question that I've pondered over a lot also, and can come up with a few explanations why.
- Neural networks work nothing like the brain. Backpropagation is unique to neural networks, and does not happen in the brain. In that sense, we just don't know the general learning algorithm in our brains. It could be electrical, it could be chemical, it could even be a combination of the two. Neural networks could be considered an inferior form of learning compared to our brains because of how simplified they are.
- If neural networks are indeed like our brain, then human babies undergo extensive "training" of the early layers, like feature extraction, in their early days. So their neural networks aren't really trained from scratch, but rather the last layer is retrained to add more and more classes and labels.
New contributor
$endgroup$
add a comment |
$begingroup$
A human child at age 2 needs around 5 instances of a car to be able to identify it with reasonable accuracy regardless of color, make, etc.
The concept of "instances" gets easily muddied. While a child may have seen 5 unique instances of a car, they have actually seen thousands of thousands of frames, in many differing environments. They have likely seen cars in other contexts. They also have an intuition for the physical world developed over their lifetime - some transfer learning probably happens here. Yet we wrap all of that up into "5 instances."
Meanwhile, every single frame/image you pass to a CNN is considered an "example." If you apply a consistent definition, both systems are really utilizing a much more similar amount of training data.
Also, I would like to note that convolutional neural networks - CNNs - are more useful in computer vision than ANNs, and in fact approach human performance in tasks like image classification. Deep learning is (probably) not a panacea, but it does perform admirably in this domain.
$endgroup$
add a comment |
$begingroup$
As pointed out by others, the data-efficiency of artificial neural networks varies quite substantially, depending on the details. As a matter of fact, there are many so called one-shot learning methods, that can solve the task of labelling trams with quite good accuracy, using only a single labelled sample.
One way to do this is by so-called transfer learning; a network trained on other labels is usually very effectively adaptable to new labels, since the hard work is breaking down the low level components of the image in a sensible way.
But we do not infact need such labeled data to perform such task; much like babies dont need nearly as much labeled data as the neural networs you are thinking of do.
For instance, one such unsupervised methods that I have also successfully applied in other contexts, is to take an unlabeled set of images, randomly rotate them, and train a network to predict which side of the image is 'up'. Without knowing what the visible objects are, or what they are called, this forces the network to learn a tremendous amount of structure about the images; and this can form an excellent basis for much more data-efficient subsequent labeled learning.
While it is true that artificial networks are quite different from real ones in probably meaningful ways, such as the absence of an obvious analogue of backpropagation, it is very probably true that real neural networks make use of the same tricks, of trying to learn the structure in the data implied by some simple priors.
One other example which almost certainly plays a role in animals and has also shown great promise in understanding video, is in the assumption that the future should be predictable from the past. Just by starting from that assumption, you can teach a neural network a whole lot. Or on a philosophical level, I am inclined to believe that this assumption underlies almost everything what we consider to be 'knowledge'.
I am not saying anything new here; but it is relatively new in the sense that these possibilities are too young to have found many applications yet, and do not yet have percolated down to the textbook understanding of 'what an ANN can do'. So to answer the OPs question; ANN's have already closed much of the gap that you describe.
$endgroup$
add a comment |
$begingroup$
One way to train a deep neural network is to treat it as a stack of auto-encoders (Restricted Boltzmann Machines).
In theory, an auto-encoder learns in an unsupervised manner: It takes arbitrary, unlabelled input data and processes it to generate output data. Then it takes that output data, and tries to regenerate its input data. It tweaks its nodes' parameters until it can come close to round-tripping its data. If you think about it, the auto-encoder is writing its own automated unit tests. In effect, it is turning its "unlabelled input data" into labelled data: The original data serves as a label for the round-tripped data.
After the layers of auto-encoders are trained, the neural network is fine-tuned using labelled data to perform its intended function. In effect, these are functional tests.
The original poster asks why a lot of data is needed to train an artificial neural network, and compares that to the allegedly low amount of training data needed by a two-year-old human. The original poster is comparing apples-to-oranges: The overall training process for the artificial neural net, versus the fine-tuning with labels for the two-year-old.
But in reality, the two-year old has been training its auto-encoders on random, self-labelled data for more than two years. Babies dream when they are in utero. (So do kittens.) Researchers have described these dreams as involving random neuron firings in the visual processing centers.
New contributor
$endgroup$
1
$begingroup$
Agreed; except that auto-encoders in practice are not very powerful tools at doing much unsupervised learning at all; everything we know points at there being more going on, so the phrasing 'the two-year old has been training its auto-encoders' should not be taken too literally I suppose.
$endgroup$
– Eelco Hoogendoorn
yesterday
add a comment |
$begingroup$
We don't learn to "see cars" until we learn to see
It takes quite a long time and lots of examples for a child to learn how to see objects as such. After that, a child can learn to identify a particular type of object from just a few examples. If you compare a two year old child with a learning system that literally starts from a blank slate, it's an apples and oranges comparison; at that age child has seen thousands of hours of "video footage".
In a similar manner, it takes artificial neural networks a lot of examples to learn "how to see" but after that it's possible to transfer that knowledge to new examples. Transfer learning is a whole domain of machine learning, and things like "one shot learning" are possible - you can build ANNs that will learn to identify new types of objects that it hasn't seen before from a single example, or to identify a particular person from a single photo of their face. But doing this initial "learning to see" part well requires quite a lot of data.
Furthermore, there's some evidence that not all training data is equal, namely, that data which you "choose" while learning is more effective than data that's simply provided to you. E.g. Held & Hein twin kitten experiment. https://www.lri.fr/~mbl/ENS/FONDIHM/2013/papers/about-HeldHein63.pdf
$endgroup$
add a comment |
Your Answer
StackExchange.ifUsing("editor", function () {
return StackExchange.using("mathjaxEditing", function () {
StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix) {
StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
});
});
}, "mathjax-editing");
StackExchange.ready(function() {
var channelOptions = {
tags: "".split(" "),
id: "65"
};
initTagRenderer("".split(" "), "".split(" "), channelOptions);
StackExchange.using("externalEditor", function() {
// Have to fire editor after snippets, if snippets enabled
if (StackExchange.settings.snippets.snippetsEnabled) {
StackExchange.using("snippets", function() {
createEditor();
});
}
else {
createEditor();
}
});
function createEditor() {
StackExchange.prepareEditor({
heartbeatType: 'answer',
autoActivateHeartbeat: false,
convertImagesToLinks: false,
noModals: true,
showLowRepImageUploadWarning: true,
reputationToPostImages: null,
bindNavPrevention: true,
postfix: "",
imageUploader: {
brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
allowUrls: true
},
onDemand: true,
discardSelector: ".discard-answer"
,immediatelyShowMarkdownHelp:true
});
}
});
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstats.stackexchange.com%2fquestions%2f394118%2fwhy-do-neural-networks-need-so-many-training-examples-to-perform%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
9 Answers
9
active
oldest
votes
9 Answers
9
active
oldest
votes
active
oldest
votes
active
oldest
votes
$begingroup$
There's a kind of goalpost moving that underlies this question. It used to be that NNs weren't very good at image recognition, so no one compared them to humans. Now that NNs are good at image tasks, suddenly it's the fault of NNs that they require a lot of training data to be comparable to children.
You can also turn this logic on its head. Suppose a child sees a number of cars the day that it's born. I wouldn't expect the child to be able to pick out a car the next day, or the next week, even though it's seen so many examples. Why are newborns so slow to learn? Because it takes a lot of exposure to the real world and, and the passage of time to change the child's neural pathways. For a neural network, we call this “training data,” but for a child we call it “growing up.”
$endgroup$
26
$begingroup$
To make it a bit more specific, a human child has already had years of training with tens of thousands of example allowing them to determining how objects look when viewed from different angles, how to identify their boundaries, the relationship between apparent size and actual size, and so on.
$endgroup$
– David Schwartz
2 days ago
18
$begingroup$
A child's brain is active inside the womb. The baby can identify their parents by sound, after the sound is filtered through water. A new-born baby had months of data to work with before they're born, but they still need years more before they can form a word, then couple more years before they can form a sentence, then couple more for a grammatically correct sentence, etc... learning is very complicated.
$endgroup$
– Nelson
2 days ago
4
$begingroup$
@EelcoHoogendoorn it explains the contrast 'child' versus 'neural network' that has been used in the question. The answer is that this is only an apparent contrast. Neural networks do not need that many examples at all, as kids get also many examples (but just in a different way) before they are able to recognize cars.
$endgroup$
– Martijn Weterings
yesterday
3
$begingroup$
@Nelson, I am not sure what the reason is for your comment, but you can change 'years' into 'year'. With 1 year kids speak words, with 2 years the first sentences are spoken, and with 3 years grammar, such as past tense and pronouns, becomes correctly used.
$endgroup$
– Martijn Weterings
yesterday
1
$begingroup$
@EelcoHoogendoorn I think the premise of the question is a case of reasoning from a faulty analogy, so directly address the analogy is responsive. Contrasting biological and artificial neural networks is also responsive, because the answer would outline how biological and artificial neural networks are most similar in their name (both contain the phrase "neural networks") but not similar in their essential characteristics, or at least the characteristics assumed by the question.
$endgroup$
– Sycorax
20 hours ago
|
show 4 more comments
$begingroup$
There's a kind of goalpost moving that underlies this question. It used to be that NNs weren't very good at image recognition, so no one compared them to humans. Now that NNs are good at image tasks, suddenly it's the fault of NNs that they require a lot of training data to be comparable to children.
You can also turn this logic on its head. Suppose a child sees a number of cars the day that it's born. I wouldn't expect the child to be able to pick out a car the next day, or the next week, even though it's seen so many examples. Why are newborns so slow to learn? Because it takes a lot of exposure to the real world and, and the passage of time to change the child's neural pathways. For a neural network, we call this “training data,” but for a child we call it “growing up.”
$endgroup$
26
$begingroup$
To make it a bit more specific, a human child has already had years of training with tens of thousands of example allowing them to determining how objects look when viewed from different angles, how to identify their boundaries, the relationship between apparent size and actual size, and so on.
$endgroup$
– David Schwartz
2 days ago
18
$begingroup$
A child's brain is active inside the womb. The baby can identify their parents by sound, after the sound is filtered through water. A new-born baby had months of data to work with before they're born, but they still need years more before they can form a word, then couple more years before they can form a sentence, then couple more for a grammatically correct sentence, etc... learning is very complicated.
$endgroup$
– Nelson
2 days ago
4
$begingroup$
@EelcoHoogendoorn it explains the contrast 'child' versus 'neural network' that has been used in the question. The answer is that this is only an apparent contrast. Neural networks do not need that many examples at all, as kids get also many examples (but just in a different way) before they are able to recognize cars.
$endgroup$
– Martijn Weterings
yesterday
3
$begingroup$
@Nelson, I am not sure what the reason is for your comment, but you can change 'years' into 'year'. With 1 year kids speak words, with 2 years the first sentences are spoken, and with 3 years grammar, such as past tense and pronouns, becomes correctly used.
$endgroup$
– Martijn Weterings
yesterday
1
$begingroup$
@EelcoHoogendoorn I think the premise of the question is a case of reasoning from a faulty analogy, so directly address the analogy is responsive. Contrasting biological and artificial neural networks is also responsive, because the answer would outline how biological and artificial neural networks are most similar in their name (both contain the phrase "neural networks") but not similar in their essential characteristics, or at least the characteristics assumed by the question.
$endgroup$
– Sycorax
20 hours ago
|
show 4 more comments
$begingroup$
There's a kind of goalpost moving that underlies this question. It used to be that NNs weren't very good at image recognition, so no one compared them to humans. Now that NNs are good at image tasks, suddenly it's the fault of NNs that they require a lot of training data to be comparable to children.
You can also turn this logic on its head. Suppose a child sees a number of cars the day that it's born. I wouldn't expect the child to be able to pick out a car the next day, or the next week, even though it's seen so many examples. Why are newborns so slow to learn? Because it takes a lot of exposure to the real world and, and the passage of time to change the child's neural pathways. For a neural network, we call this “training data,” but for a child we call it “growing up.”
$endgroup$
There's a kind of goalpost moving that underlies this question. It used to be that NNs weren't very good at image recognition, so no one compared them to humans. Now that NNs are good at image tasks, suddenly it's the fault of NNs that they require a lot of training data to be comparable to children.
You can also turn this logic on its head. Suppose a child sees a number of cars the day that it's born. I wouldn't expect the child to be able to pick out a car the next day, or the next week, even though it's seen so many examples. Why are newborns so slow to learn? Because it takes a lot of exposure to the real world and, and the passage of time to change the child's neural pathways. For a neural network, we call this “training data,” but for a child we call it “growing up.”
edited 2 days ago
answered 2 days ago
SycoraxSycorax
40.9k12104204
40.9k12104204
26
$begingroup$
To make it a bit more specific, a human child has already had years of training with tens of thousands of example allowing them to determining how objects look when viewed from different angles, how to identify their boundaries, the relationship between apparent size and actual size, and so on.
$endgroup$
– David Schwartz
2 days ago
18
$begingroup$
A child's brain is active inside the womb. The baby can identify their parents by sound, after the sound is filtered through water. A new-born baby had months of data to work with before they're born, but they still need years more before they can form a word, then couple more years before they can form a sentence, then couple more for a grammatically correct sentence, etc... learning is very complicated.
$endgroup$
– Nelson
2 days ago
4
$begingroup$
@EelcoHoogendoorn it explains the contrast 'child' versus 'neural network' that has been used in the question. The answer is that this is only an apparent contrast. Neural networks do not need that many examples at all, as kids get also many examples (but just in a different way) before they are able to recognize cars.
$endgroup$
– Martijn Weterings
yesterday
3
$begingroup$
@Nelson, I am not sure what the reason is for your comment, but you can change 'years' into 'year'. With 1 year kids speak words, with 2 years the first sentences are spoken, and with 3 years grammar, such as past tense and pronouns, becomes correctly used.
$endgroup$
– Martijn Weterings
yesterday
1
$begingroup$
@EelcoHoogendoorn I think the premise of the question is a case of reasoning from a faulty analogy, so directly address the analogy is responsive. Contrasting biological and artificial neural networks is also responsive, because the answer would outline how biological and artificial neural networks are most similar in their name (both contain the phrase "neural networks") but not similar in their essential characteristics, or at least the characteristics assumed by the question.
$endgroup$
– Sycorax
20 hours ago
|
show 4 more comments
26
$begingroup$
To make it a bit more specific, a human child has already had years of training with tens of thousands of example allowing them to determining how objects look when viewed from different angles, how to identify their boundaries, the relationship between apparent size and actual size, and so on.
$endgroup$
– David Schwartz
2 days ago
18
$begingroup$
A child's brain is active inside the womb. The baby can identify their parents by sound, after the sound is filtered through water. A new-born baby had months of data to work with before they're born, but they still need years more before they can form a word, then couple more years before they can form a sentence, then couple more for a grammatically correct sentence, etc... learning is very complicated.
$endgroup$
– Nelson
2 days ago
4
$begingroup$
@EelcoHoogendoorn it explains the contrast 'child' versus 'neural network' that has been used in the question. The answer is that this is only an apparent contrast. Neural networks do not need that many examples at all, as kids get also many examples (but just in a different way) before they are able to recognize cars.
$endgroup$
– Martijn Weterings
yesterday
3
$begingroup$
@Nelson, I am not sure what the reason is for your comment, but you can change 'years' into 'year'. With 1 year kids speak words, with 2 years the first sentences are spoken, and with 3 years grammar, such as past tense and pronouns, becomes correctly used.
$endgroup$
– Martijn Weterings
yesterday
1
$begingroup$
@EelcoHoogendoorn I think the premise of the question is a case of reasoning from a faulty analogy, so directly address the analogy is responsive. Contrasting biological and artificial neural networks is also responsive, because the answer would outline how biological and artificial neural networks are most similar in their name (both contain the phrase "neural networks") but not similar in their essential characteristics, or at least the characteristics assumed by the question.
$endgroup$
– Sycorax
20 hours ago
26
26
$begingroup$
To make it a bit more specific, a human child has already had years of training with tens of thousands of example allowing them to determining how objects look when viewed from different angles, how to identify their boundaries, the relationship between apparent size and actual size, and so on.
$endgroup$
– David Schwartz
2 days ago
$begingroup$
To make it a bit more specific, a human child has already had years of training with tens of thousands of example allowing them to determining how objects look when viewed from different angles, how to identify their boundaries, the relationship between apparent size and actual size, and so on.
$endgroup$
– David Schwartz
2 days ago
18
18
$begingroup$
A child's brain is active inside the womb. The baby can identify their parents by sound, after the sound is filtered through water. A new-born baby had months of data to work with before they're born, but they still need years more before they can form a word, then couple more years before they can form a sentence, then couple more for a grammatically correct sentence, etc... learning is very complicated.
$endgroup$
– Nelson
2 days ago
$begingroup$
A child's brain is active inside the womb. The baby can identify their parents by sound, after the sound is filtered through water. A new-born baby had months of data to work with before they're born, but they still need years more before they can form a word, then couple more years before they can form a sentence, then couple more for a grammatically correct sentence, etc... learning is very complicated.
$endgroup$
– Nelson
2 days ago
4
4
$begingroup$
@EelcoHoogendoorn it explains the contrast 'child' versus 'neural network' that has been used in the question. The answer is that this is only an apparent contrast. Neural networks do not need that many examples at all, as kids get also many examples (but just in a different way) before they are able to recognize cars.
$endgroup$
– Martijn Weterings
yesterday
$begingroup$
@EelcoHoogendoorn it explains the contrast 'child' versus 'neural network' that has been used in the question. The answer is that this is only an apparent contrast. Neural networks do not need that many examples at all, as kids get also many examples (but just in a different way) before they are able to recognize cars.
$endgroup$
– Martijn Weterings
yesterday
3
3
$begingroup$
@Nelson, I am not sure what the reason is for your comment, but you can change 'years' into 'year'. With 1 year kids speak words, with 2 years the first sentences are spoken, and with 3 years grammar, such as past tense and pronouns, becomes correctly used.
$endgroup$
– Martijn Weterings
yesterday
$begingroup$
@Nelson, I am not sure what the reason is for your comment, but you can change 'years' into 'year'. With 1 year kids speak words, with 2 years the first sentences are spoken, and with 3 years grammar, such as past tense and pronouns, becomes correctly used.
$endgroup$
– Martijn Weterings
yesterday
1
1
$begingroup$
@EelcoHoogendoorn I think the premise of the question is a case of reasoning from a faulty analogy, so directly address the analogy is responsive. Contrasting biological and artificial neural networks is also responsive, because the answer would outline how biological and artificial neural networks are most similar in their name (both contain the phrase "neural networks") but not similar in their essential characteristics, or at least the characteristics assumed by the question.
$endgroup$
– Sycorax
20 hours ago
$begingroup$
@EelcoHoogendoorn I think the premise of the question is a case of reasoning from a faulty analogy, so directly address the analogy is responsive. Contrasting biological and artificial neural networks is also responsive, because the answer would outline how biological and artificial neural networks are most similar in their name (both contain the phrase "neural networks") but not similar in their essential characteristics, or at least the characteristics assumed by the question.
$endgroup$
– Sycorax
20 hours ago
|
show 4 more comments
$begingroup$
First of all, at age two, a child knows a lot about the world and actively applies this knowledge. A child does a lot of "transfer learning" by applying this knowledge to new concepts.
Second, before seeing those five "labeled" examples of cars, a child sees a lot of cars on the street, on TV, toy cars, etc., so also a lot of "unsupervised learning" happens beforehand.
Finally, neural networks have almost nothing in common with the human brain, so there's not much point in comparing them. Also notice that there are algorithms for one-shot learning, and pretty much research on it currently happens.
$endgroup$
2
$begingroup$
4th point, a child also has more than 100 million years of evolutionary selection towards learning efficiently/accurately.
$endgroup$
– csiz
11 hours ago
$begingroup$
@csiz People still believe in Darwin's hypothesis, eh?
$endgroup$
– Failed Scientist
4 hours ago
add a comment |
$begingroup$
First of all, at age two, a child knows a lot about the world and actively applies this knowledge. A child does a lot of "transfer learning" by applying this knowledge to new concepts.
Second, before seeing those five "labeled" examples of cars, a child sees a lot of cars on the street, on TV, toy cars, etc., so also a lot of "unsupervised learning" happens beforehand.
Finally, neural networks have almost nothing in common with the human brain, so there's not much point in comparing them. Also notice that there are algorithms for one-shot learning, and pretty much research on it currently happens.
$endgroup$
2
$begingroup$
4th point, a child also has more than 100 million years of evolutionary selection towards learning efficiently/accurately.
$endgroup$
– csiz
11 hours ago
$begingroup$
@csiz People still believe in Darwin's hypothesis, eh?
$endgroup$
– Failed Scientist
4 hours ago
add a comment |
$begingroup$
First of all, at age two, a child knows a lot about the world and actively applies this knowledge. A child does a lot of "transfer learning" by applying this knowledge to new concepts.
Second, before seeing those five "labeled" examples of cars, a child sees a lot of cars on the street, on TV, toy cars, etc., so also a lot of "unsupervised learning" happens beforehand.
Finally, neural networks have almost nothing in common with the human brain, so there's not much point in comparing them. Also notice that there are algorithms for one-shot learning, and pretty much research on it currently happens.
$endgroup$
First of all, at age two, a child knows a lot about the world and actively applies this knowledge. A child does a lot of "transfer learning" by applying this knowledge to new concepts.
Second, before seeing those five "labeled" examples of cars, a child sees a lot of cars on the street, on TV, toy cars, etc., so also a lot of "unsupervised learning" happens beforehand.
Finally, neural networks have almost nothing in common with the human brain, so there's not much point in comparing them. Also notice that there are algorithms for one-shot learning, and pretty much research on it currently happens.
edited 14 mins ago
Peter Mortensen
19918
19918
answered 2 days ago
Tim♦Tim
58.4k9128221
58.4k9128221
2
$begingroup$
4th point, a child also has more than 100 million years of evolutionary selection towards learning efficiently/accurately.
$endgroup$
– csiz
11 hours ago
$begingroup$
@csiz People still believe in Darwin's hypothesis, eh?
$endgroup$
– Failed Scientist
4 hours ago
add a comment |
2
$begingroup$
4th point, a child also has more than 100 million years of evolutionary selection towards learning efficiently/accurately.
$endgroup$
– csiz
11 hours ago
$begingroup$
@csiz People still believe in Darwin's hypothesis, eh?
$endgroup$
– Failed Scientist
4 hours ago
2
2
$begingroup$
4th point, a child also has more than 100 million years of evolutionary selection towards learning efficiently/accurately.
$endgroup$
– csiz
11 hours ago
$begingroup$
4th point, a child also has more than 100 million years of evolutionary selection towards learning efficiently/accurately.
$endgroup$
– csiz
11 hours ago
$begingroup$
@csiz People still believe in Darwin's hypothesis, eh?
$endgroup$
– Failed Scientist
4 hours ago
$begingroup$
@csiz People still believe in Darwin's hypothesis, eh?
$endgroup$
– Failed Scientist
4 hours ago
add a comment |
$begingroup$
One major aspect that I don't see in current answers is evolution.
A child's brain does not learn from scratch. It's similar to asking how deer and giraffe babies can walk a few minutes after birth. Because they are born with their brains already wired for this task. There is some fine-tuning needed of course, but the baby deer doesn't learn to walk from "random initialization".
Similarly, the fact that big moving objects exist and are important to keep track of is something we are born with.
So I think the presupposition of this question is simply false. Human neural networks had the opportunity to see tons of - maybe not cars but - moving, rotating 3D objects with difficult textures and shapes etc., but this happened through lots of generations and the learning took place by evolutionary algorithms, i.e. the ones whose brain was better structured for this task, could live to reproduce with higher chance, leaving the next generation with better and better brain wiring from the start.
$endgroup$
4
$begingroup$
Fun aside: there's evidence that when it comes to discriminating between different models of cars, we actually leverage the specialized facial recognition center of our brain. It's plausible that, while a child may not distinguish between different models, the implicit presence of a 'face' on a mobile object may cause cars to be categorized as a type of creature and therefore be favored to be identified by evolution, since recognizing mobile objects with faces is helpful to survival.
$endgroup$
– Dan Bryant
yesterday
4
$begingroup$
This answer addresses exactly what I was thinking. Children are not born as blank slates. They come with features that make some patterns easier to recognize, some things easier to learn, etc.
$endgroup$
– Eff
yesterday
1
$begingroup$
While animals that walk right out of the womb are indeed fascinating, such evolutionary hardwiring is thought to be at the very opposite extreme of human learning, which is thought to be the extreme of experience-driven learning in the natural world. Certainly cars will have left minimal evolutionary impact on the evolution of our brains.
$endgroup$
– Eelco Hoogendoorn
yesterday
2
$begingroup$
@EelcoHoogendoorn The ability to learn and understand the environment has been evolutionarily selected for. The brain has been set up by evolution to be extremely efficient at learning. The ability to connect the dots, see patterns, understand shapes and movement, makes inferences, etc.
$endgroup$
– Eff
yesterday
2
$begingroup$
This is a good point, but it's also true that as researchers come to understand this, they build NN's that have hard-coded structures that facilitate certain types of learning. Consider that a convolutional NN has hard coded receptive fields that greatly speed up learning / enhance performance on visual tasks. Those fields could be learned from scratch in a fully connected network, but it's much harder. @EelcoHoogendoorn, human brains are full of structure that facilitates learning.
$endgroup$
– gung♦
22 hours ago
|
show 5 more comments
$begingroup$
One major aspect that I don't see in current answers is evolution.
A child's brain does not learn from scratch. It's similar to asking how deer and giraffe babies can walk a few minutes after birth. Because they are born with their brains already wired for this task. There is some fine-tuning needed of course, but the baby deer doesn't learn to walk from "random initialization".
Similarly, the fact that big moving objects exist and are important to keep track of is something we are born with.
So I think the presupposition of this question is simply false. Human neural networks had the opportunity to see tons of - maybe not cars but - moving, rotating 3D objects with difficult textures and shapes etc., but this happened through lots of generations and the learning took place by evolutionary algorithms, i.e. the ones whose brain was better structured for this task, could live to reproduce with higher chance, leaving the next generation with better and better brain wiring from the start.
$endgroup$
4
$begingroup$
Fun aside: there's evidence that when it comes to discriminating between different models of cars, we actually leverage the specialized facial recognition center of our brain. It's plausible that, while a child may not distinguish between different models, the implicit presence of a 'face' on a mobile object may cause cars to be categorized as a type of creature and therefore be favored to be identified by evolution, since recognizing mobile objects with faces is helpful to survival.
$endgroup$
– Dan Bryant
yesterday
4
$begingroup$
This answer addresses exactly what I was thinking. Children are not born as blank slates. They come with features that make some patterns easier to recognize, some things easier to learn, etc.
$endgroup$
– Eff
yesterday
1
$begingroup$
While animals that walk right out of the womb are indeed fascinating, such evolutionary hardwiring is thought to be at the very opposite extreme of human learning, which is thought to be the extreme of experience-driven learning in the natural world. Certainly cars will have left minimal evolutionary impact on the evolution of our brains.
$endgroup$
– Eelco Hoogendoorn
yesterday
2
$begingroup$
@EelcoHoogendoorn The ability to learn and understand the environment has been evolutionarily selected for. The brain has been set up by evolution to be extremely efficient at learning. The ability to connect the dots, see patterns, understand shapes and movement, makes inferences, etc.
$endgroup$
– Eff
yesterday
2
$begingroup$
This is a good point, but it's also true that as researchers come to understand this, they build NN's that have hard-coded structures that facilitate certain types of learning. Consider that a convolutional NN has hard coded receptive fields that greatly speed up learning / enhance performance on visual tasks. Those fields could be learned from scratch in a fully connected network, but it's much harder. @EelcoHoogendoorn, human brains are full of structure that facilitates learning.
$endgroup$
– gung♦
22 hours ago
|
show 5 more comments
$begingroup$
One major aspect that I don't see in current answers is evolution.
A child's brain does not learn from scratch. It's similar to asking how deer and giraffe babies can walk a few minutes after birth. Because they are born with their brains already wired for this task. There is some fine-tuning needed of course, but the baby deer doesn't learn to walk from "random initialization".
Similarly, the fact that big moving objects exist and are important to keep track of is something we are born with.
So I think the presupposition of this question is simply false. Human neural networks had the opportunity to see tons of - maybe not cars but - moving, rotating 3D objects with difficult textures and shapes etc., but this happened through lots of generations and the learning took place by evolutionary algorithms, i.e. the ones whose brain was better structured for this task, could live to reproduce with higher chance, leaving the next generation with better and better brain wiring from the start.
$endgroup$
One major aspect that I don't see in current answers is evolution.
A child's brain does not learn from scratch. It's similar to asking how deer and giraffe babies can walk a few minutes after birth. Because they are born with their brains already wired for this task. There is some fine-tuning needed of course, but the baby deer doesn't learn to walk from "random initialization".
Similarly, the fact that big moving objects exist and are important to keep track of is something we are born with.
So I think the presupposition of this question is simply false. Human neural networks had the opportunity to see tons of - maybe not cars but - moving, rotating 3D objects with difficult textures and shapes etc., but this happened through lots of generations and the learning took place by evolutionary algorithms, i.e. the ones whose brain was better structured for this task, could live to reproduce with higher chance, leaving the next generation with better and better brain wiring from the start.
answered 2 days ago
isarandiisarandi
40428
40428
4
$begingroup$
Fun aside: there's evidence that when it comes to discriminating between different models of cars, we actually leverage the specialized facial recognition center of our brain. It's plausible that, while a child may not distinguish between different models, the implicit presence of a 'face' on a mobile object may cause cars to be categorized as a type of creature and therefore be favored to be identified by evolution, since recognizing mobile objects with faces is helpful to survival.
$endgroup$
– Dan Bryant
yesterday
4
$begingroup$
This answer addresses exactly what I was thinking. Children are not born as blank slates. They come with features that make some patterns easier to recognize, some things easier to learn, etc.
$endgroup$
– Eff
yesterday
1
$begingroup$
While animals that walk right out of the womb are indeed fascinating, such evolutionary hardwiring is thought to be at the very opposite extreme of human learning, which is thought to be the extreme of experience-driven learning in the natural world. Certainly cars will have left minimal evolutionary impact on the evolution of our brains.
$endgroup$
– Eelco Hoogendoorn
yesterday
2
$begingroup$
@EelcoHoogendoorn The ability to learn and understand the environment has been evolutionarily selected for. The brain has been set up by evolution to be extremely efficient at learning. The ability to connect the dots, see patterns, understand shapes and movement, makes inferences, etc.
$endgroup$
– Eff
yesterday
2
$begingroup$
This is a good point, but it's also true that as researchers come to understand this, they build NN's that have hard-coded structures that facilitate certain types of learning. Consider that a convolutional NN has hard coded receptive fields that greatly speed up learning / enhance performance on visual tasks. Those fields could be learned from scratch in a fully connected network, but it's much harder. @EelcoHoogendoorn, human brains are full of structure that facilitates learning.
$endgroup$
– gung♦
22 hours ago
|
show 5 more comments
4
$begingroup$
Fun aside: there's evidence that when it comes to discriminating between different models of cars, we actually leverage the specialized facial recognition center of our brain. It's plausible that, while a child may not distinguish between different models, the implicit presence of a 'face' on a mobile object may cause cars to be categorized as a type of creature and therefore be favored to be identified by evolution, since recognizing mobile objects with faces is helpful to survival.
$endgroup$
– Dan Bryant
yesterday
4
$begingroup$
This answer addresses exactly what I was thinking. Children are not born as blank slates. They come with features that make some patterns easier to recognize, some things easier to learn, etc.
$endgroup$
– Eff
yesterday
1
$begingroup$
While animals that walk right out of the womb are indeed fascinating, such evolutionary hardwiring is thought to be at the very opposite extreme of human learning, which is thought to be the extreme of experience-driven learning in the natural world. Certainly cars will have left minimal evolutionary impact on the evolution of our brains.
$endgroup$
– Eelco Hoogendoorn
yesterday
2
$begingroup$
@EelcoHoogendoorn The ability to learn and understand the environment has been evolutionarily selected for. The brain has been set up by evolution to be extremely efficient at learning. The ability to connect the dots, see patterns, understand shapes and movement, makes inferences, etc.
$endgroup$
– Eff
yesterday
2
$begingroup$
This is a good point, but it's also true that as researchers come to understand this, they build NN's that have hard-coded structures that facilitate certain types of learning. Consider that a convolutional NN has hard coded receptive fields that greatly speed up learning / enhance performance on visual tasks. Those fields could be learned from scratch in a fully connected network, but it's much harder. @EelcoHoogendoorn, human brains are full of structure that facilitates learning.
$endgroup$
– gung♦
22 hours ago
4
4
$begingroup$
Fun aside: there's evidence that when it comes to discriminating between different models of cars, we actually leverage the specialized facial recognition center of our brain. It's plausible that, while a child may not distinguish between different models, the implicit presence of a 'face' on a mobile object may cause cars to be categorized as a type of creature and therefore be favored to be identified by evolution, since recognizing mobile objects with faces is helpful to survival.
$endgroup$
– Dan Bryant
yesterday
$begingroup$
Fun aside: there's evidence that when it comes to discriminating between different models of cars, we actually leverage the specialized facial recognition center of our brain. It's plausible that, while a child may not distinguish between different models, the implicit presence of a 'face' on a mobile object may cause cars to be categorized as a type of creature and therefore be favored to be identified by evolution, since recognizing mobile objects with faces is helpful to survival.
$endgroup$
– Dan Bryant
yesterday
4
4
$begingroup$
This answer addresses exactly what I was thinking. Children are not born as blank slates. They come with features that make some patterns easier to recognize, some things easier to learn, etc.
$endgroup$
– Eff
yesterday
$begingroup$
This answer addresses exactly what I was thinking. Children are not born as blank slates. They come with features that make some patterns easier to recognize, some things easier to learn, etc.
$endgroup$
– Eff
yesterday
1
1
$begingroup$
While animals that walk right out of the womb are indeed fascinating, such evolutionary hardwiring is thought to be at the very opposite extreme of human learning, which is thought to be the extreme of experience-driven learning in the natural world. Certainly cars will have left minimal evolutionary impact on the evolution of our brains.
$endgroup$
– Eelco Hoogendoorn
yesterday
$begingroup$
While animals that walk right out of the womb are indeed fascinating, such evolutionary hardwiring is thought to be at the very opposite extreme of human learning, which is thought to be the extreme of experience-driven learning in the natural world. Certainly cars will have left minimal evolutionary impact on the evolution of our brains.
$endgroup$
– Eelco Hoogendoorn
yesterday
2
2
$begingroup$
@EelcoHoogendoorn The ability to learn and understand the environment has been evolutionarily selected for. The brain has been set up by evolution to be extremely efficient at learning. The ability to connect the dots, see patterns, understand shapes and movement, makes inferences, etc.
$endgroup$
– Eff
yesterday
$begingroup$
@EelcoHoogendoorn The ability to learn and understand the environment has been evolutionarily selected for. The brain has been set up by evolution to be extremely efficient at learning. The ability to connect the dots, see patterns, understand shapes and movement, makes inferences, etc.
$endgroup$
– Eff
yesterday
2
2
$begingroup$
This is a good point, but it's also true that as researchers come to understand this, they build NN's that have hard-coded structures that facilitate certain types of learning. Consider that a convolutional NN has hard coded receptive fields that greatly speed up learning / enhance performance on visual tasks. Those fields could be learned from scratch in a fully connected network, but it's much harder. @EelcoHoogendoorn, human brains are full of structure that facilitates learning.
$endgroup$
– gung♦
22 hours ago
$begingroup$
This is a good point, but it's also true that as researchers come to understand this, they build NN's that have hard-coded structures that facilitate certain types of learning. Consider that a convolutional NN has hard coded receptive fields that greatly speed up learning / enhance performance on visual tasks. Those fields could be learned from scratch in a fully connected network, but it's much harder. @EelcoHoogendoorn, human brains are full of structure that facilitates learning.
$endgroup$
– gung♦
22 hours ago
|
show 5 more comments
$begingroup$
I don't know much about neural networks but I know a fair bit about babies.
Many 2 year olds have a lot of issues with how general words should be. For instance, it is quite common at that age for kids to use "dog" for any four legged animal. That's a more difficult distinction than "car" - just think how different a poodle looks from a great Dane, for instance and yet they are both "dog" while a cat is not.
And a child at 2 has seen many many more than 5 examples of "car". A kid sees dozens or even hundreds of examples of cars any time the family goes for a drive. And a lot of parents will comment "look at the car" a lot more than 5 times. But kids can also think in ways that they weren't told about. For instance, on the street the kid sees lots of things lined up. His dad says (of one) "look at the shiny car!" and the kid thinks "maybe all those other things lined up are also cars?"
$endgroup$
$begingroup$
Other examples: Taxi's, driving lesson cars, and police cars are the same. Whenever a car is red then it is a firetruck. Campervans are ambulances. A lorry with a loader crane becomes classified as an excavator. The bus that just passed by goes to the train station, so the next bus, which looks the same, must also be going to the train station. And seeing the moon during broad daylight is a very special event.
$endgroup$
– Martijn Weterings
yesterday
add a comment |
$begingroup$
I don't know much about neural networks but I know a fair bit about babies.
Many 2 year olds have a lot of issues with how general words should be. For instance, it is quite common at that age for kids to use "dog" for any four legged animal. That's a more difficult distinction than "car" - just think how different a poodle looks from a great Dane, for instance and yet they are both "dog" while a cat is not.
And a child at 2 has seen many many more than 5 examples of "car". A kid sees dozens or even hundreds of examples of cars any time the family goes for a drive. And a lot of parents will comment "look at the car" a lot more than 5 times. But kids can also think in ways that they weren't told about. For instance, on the street the kid sees lots of things lined up. His dad says (of one) "look at the shiny car!" and the kid thinks "maybe all those other things lined up are also cars?"
$endgroup$
$begingroup$
Other examples: Taxi's, driving lesson cars, and police cars are the same. Whenever a car is red then it is a firetruck. Campervans are ambulances. A lorry with a loader crane becomes classified as an excavator. The bus that just passed by goes to the train station, so the next bus, which looks the same, must also be going to the train station. And seeing the moon during broad daylight is a very special event.
$endgroup$
– Martijn Weterings
yesterday
add a comment |
$begingroup$
I don't know much about neural networks but I know a fair bit about babies.
Many 2 year olds have a lot of issues with how general words should be. For instance, it is quite common at that age for kids to use "dog" for any four legged animal. That's a more difficult distinction than "car" - just think how different a poodle looks from a great Dane, for instance and yet they are both "dog" while a cat is not.
And a child at 2 has seen many many more than 5 examples of "car". A kid sees dozens or even hundreds of examples of cars any time the family goes for a drive. And a lot of parents will comment "look at the car" a lot more than 5 times. But kids can also think in ways that they weren't told about. For instance, on the street the kid sees lots of things lined up. His dad says (of one) "look at the shiny car!" and the kid thinks "maybe all those other things lined up are also cars?"
$endgroup$
I don't know much about neural networks but I know a fair bit about babies.
Many 2 year olds have a lot of issues with how general words should be. For instance, it is quite common at that age for kids to use "dog" for any four legged animal. That's a more difficult distinction than "car" - just think how different a poodle looks from a great Dane, for instance and yet they are both "dog" while a cat is not.
And a child at 2 has seen many many more than 5 examples of "car". A kid sees dozens or even hundreds of examples of cars any time the family goes for a drive. And a lot of parents will comment "look at the car" a lot more than 5 times. But kids can also think in ways that they weren't told about. For instance, on the street the kid sees lots of things lined up. His dad says (of one) "look at the shiny car!" and the kid thinks "maybe all those other things lined up are also cars?"
answered 2 days ago
Peter Flom♦Peter Flom
75.8k11107209
75.8k11107209
$begingroup$
Other examples: Taxi's, driving lesson cars, and police cars are the same. Whenever a car is red then it is a firetruck. Campervans are ambulances. A lorry with a loader crane becomes classified as an excavator. The bus that just passed by goes to the train station, so the next bus, which looks the same, must also be going to the train station. And seeing the moon during broad daylight is a very special event.
$endgroup$
– Martijn Weterings
yesterday
add a comment |
$begingroup$
Other examples: Taxi's, driving lesson cars, and police cars are the same. Whenever a car is red then it is a firetruck. Campervans are ambulances. A lorry with a loader crane becomes classified as an excavator. The bus that just passed by goes to the train station, so the next bus, which looks the same, must also be going to the train station. And seeing the moon during broad daylight is a very special event.
$endgroup$
– Martijn Weterings
yesterday
$begingroup$
Other examples: Taxi's, driving lesson cars, and police cars are the same. Whenever a car is red then it is a firetruck. Campervans are ambulances. A lorry with a loader crane becomes classified as an excavator. The bus that just passed by goes to the train station, so the next bus, which looks the same, must also be going to the train station. And seeing the moon during broad daylight is a very special event.
$endgroup$
– Martijn Weterings
yesterday
$begingroup$
Other examples: Taxi's, driving lesson cars, and police cars are the same. Whenever a car is red then it is a firetruck. Campervans are ambulances. A lorry with a loader crane becomes classified as an excavator. The bus that just passed by goes to the train station, so the next bus, which looks the same, must also be going to the train station. And seeing the moon during broad daylight is a very special event.
$endgroup$
– Martijn Weterings
yesterday
add a comment |
$begingroup$
This is an a fascinating question that I've pondered over a lot also, and can come up with a few explanations why.
- Neural networks work nothing like the brain. Backpropagation is unique to neural networks, and does not happen in the brain. In that sense, we just don't know the general learning algorithm in our brains. It could be electrical, it could be chemical, it could even be a combination of the two. Neural networks could be considered an inferior form of learning compared to our brains because of how simplified they are.
- If neural networks are indeed like our brain, then human babies undergo extensive "training" of the early layers, like feature extraction, in their early days. So their neural networks aren't really trained from scratch, but rather the last layer is retrained to add more and more classes and labels.
New contributor
$endgroup$
add a comment |
$begingroup$
This is an a fascinating question that I've pondered over a lot also, and can come up with a few explanations why.
- Neural networks work nothing like the brain. Backpropagation is unique to neural networks, and does not happen in the brain. In that sense, we just don't know the general learning algorithm in our brains. It could be electrical, it could be chemical, it could even be a combination of the two. Neural networks could be considered an inferior form of learning compared to our brains because of how simplified they are.
- If neural networks are indeed like our brain, then human babies undergo extensive "training" of the early layers, like feature extraction, in their early days. So their neural networks aren't really trained from scratch, but rather the last layer is retrained to add more and more classes and labels.
New contributor
$endgroup$
add a comment |
$begingroup$
This is an a fascinating question that I've pondered over a lot also, and can come up with a few explanations why.
- Neural networks work nothing like the brain. Backpropagation is unique to neural networks, and does not happen in the brain. In that sense, we just don't know the general learning algorithm in our brains. It could be electrical, it could be chemical, it could even be a combination of the two. Neural networks could be considered an inferior form of learning compared to our brains because of how simplified they are.
- If neural networks are indeed like our brain, then human babies undergo extensive "training" of the early layers, like feature extraction, in their early days. So their neural networks aren't really trained from scratch, but rather the last layer is retrained to add more and more classes and labels.
New contributor
$endgroup$
This is an a fascinating question that I've pondered over a lot also, and can come up with a few explanations why.
- Neural networks work nothing like the brain. Backpropagation is unique to neural networks, and does not happen in the brain. In that sense, we just don't know the general learning algorithm in our brains. It could be electrical, it could be chemical, it could even be a combination of the two. Neural networks could be considered an inferior form of learning compared to our brains because of how simplified they are.
- If neural networks are indeed like our brain, then human babies undergo extensive "training" of the early layers, like feature extraction, in their early days. So their neural networks aren't really trained from scratch, but rather the last layer is retrained to add more and more classes and labels.
New contributor
New contributor
answered 2 days ago
sd2017sd2017
911
911
New contributor
New contributor
add a comment |
add a comment |
$begingroup$
A human child at age 2 needs around 5 instances of a car to be able to identify it with reasonable accuracy regardless of color, make, etc.
The concept of "instances" gets easily muddied. While a child may have seen 5 unique instances of a car, they have actually seen thousands of thousands of frames, in many differing environments. They have likely seen cars in other contexts. They also have an intuition for the physical world developed over their lifetime - some transfer learning probably happens here. Yet we wrap all of that up into "5 instances."
Meanwhile, every single frame/image you pass to a CNN is considered an "example." If you apply a consistent definition, both systems are really utilizing a much more similar amount of training data.
Also, I would like to note that convolutional neural networks - CNNs - are more useful in computer vision than ANNs, and in fact approach human performance in tasks like image classification. Deep learning is (probably) not a panacea, but it does perform admirably in this domain.
$endgroup$
add a comment |
$begingroup$
A human child at age 2 needs around 5 instances of a car to be able to identify it with reasonable accuracy regardless of color, make, etc.
The concept of "instances" gets easily muddied. While a child may have seen 5 unique instances of a car, they have actually seen thousands of thousands of frames, in many differing environments. They have likely seen cars in other contexts. They also have an intuition for the physical world developed over their lifetime - some transfer learning probably happens here. Yet we wrap all of that up into "5 instances."
Meanwhile, every single frame/image you pass to a CNN is considered an "example." If you apply a consistent definition, both systems are really utilizing a much more similar amount of training data.
Also, I would like to note that convolutional neural networks - CNNs - are more useful in computer vision than ANNs, and in fact approach human performance in tasks like image classification. Deep learning is (probably) not a panacea, but it does perform admirably in this domain.
$endgroup$
add a comment |
$begingroup$
A human child at age 2 needs around 5 instances of a car to be able to identify it with reasonable accuracy regardless of color, make, etc.
The concept of "instances" gets easily muddied. While a child may have seen 5 unique instances of a car, they have actually seen thousands of thousands of frames, in many differing environments. They have likely seen cars in other contexts. They also have an intuition for the physical world developed over their lifetime - some transfer learning probably happens here. Yet we wrap all of that up into "5 instances."
Meanwhile, every single frame/image you pass to a CNN is considered an "example." If you apply a consistent definition, both systems are really utilizing a much more similar amount of training data.
Also, I would like to note that convolutional neural networks - CNNs - are more useful in computer vision than ANNs, and in fact approach human performance in tasks like image classification. Deep learning is (probably) not a panacea, but it does perform admirably in this domain.
$endgroup$
A human child at age 2 needs around 5 instances of a car to be able to identify it with reasonable accuracy regardless of color, make, etc.
The concept of "instances" gets easily muddied. While a child may have seen 5 unique instances of a car, they have actually seen thousands of thousands of frames, in many differing environments. They have likely seen cars in other contexts. They also have an intuition for the physical world developed over their lifetime - some transfer learning probably happens here. Yet we wrap all of that up into "5 instances."
Meanwhile, every single frame/image you pass to a CNN is considered an "example." If you apply a consistent definition, both systems are really utilizing a much more similar amount of training data.
Also, I would like to note that convolutional neural networks - CNNs - are more useful in computer vision than ANNs, and in fact approach human performance in tasks like image classification. Deep learning is (probably) not a panacea, but it does perform admirably in this domain.
answered yesterday
spinodalspinodal
1586
1586
add a comment |
add a comment |
$begingroup$
As pointed out by others, the data-efficiency of artificial neural networks varies quite substantially, depending on the details. As a matter of fact, there are many so called one-shot learning methods, that can solve the task of labelling trams with quite good accuracy, using only a single labelled sample.
One way to do this is by so-called transfer learning; a network trained on other labels is usually very effectively adaptable to new labels, since the hard work is breaking down the low level components of the image in a sensible way.
But we do not infact need such labeled data to perform such task; much like babies dont need nearly as much labeled data as the neural networs you are thinking of do.
For instance, one such unsupervised methods that I have also successfully applied in other contexts, is to take an unlabeled set of images, randomly rotate them, and train a network to predict which side of the image is 'up'. Without knowing what the visible objects are, or what they are called, this forces the network to learn a tremendous amount of structure about the images; and this can form an excellent basis for much more data-efficient subsequent labeled learning.
While it is true that artificial networks are quite different from real ones in probably meaningful ways, such as the absence of an obvious analogue of backpropagation, it is very probably true that real neural networks make use of the same tricks, of trying to learn the structure in the data implied by some simple priors.
One other example which almost certainly plays a role in animals and has also shown great promise in understanding video, is in the assumption that the future should be predictable from the past. Just by starting from that assumption, you can teach a neural network a whole lot. Or on a philosophical level, I am inclined to believe that this assumption underlies almost everything what we consider to be 'knowledge'.
I am not saying anything new here; but it is relatively new in the sense that these possibilities are too young to have found many applications yet, and do not yet have percolated down to the textbook understanding of 'what an ANN can do'. So to answer the OPs question; ANN's have already closed much of the gap that you describe.
$endgroup$
add a comment |
$begingroup$
As pointed out by others, the data-efficiency of artificial neural networks varies quite substantially, depending on the details. As a matter of fact, there are many so called one-shot learning methods, that can solve the task of labelling trams with quite good accuracy, using only a single labelled sample.
One way to do this is by so-called transfer learning; a network trained on other labels is usually very effectively adaptable to new labels, since the hard work is breaking down the low level components of the image in a sensible way.
But we do not infact need such labeled data to perform such task; much like babies dont need nearly as much labeled data as the neural networs you are thinking of do.
For instance, one such unsupervised methods that I have also successfully applied in other contexts, is to take an unlabeled set of images, randomly rotate them, and train a network to predict which side of the image is 'up'. Without knowing what the visible objects are, or what they are called, this forces the network to learn a tremendous amount of structure about the images; and this can form an excellent basis for much more data-efficient subsequent labeled learning.
While it is true that artificial networks are quite different from real ones in probably meaningful ways, such as the absence of an obvious analogue of backpropagation, it is very probably true that real neural networks make use of the same tricks, of trying to learn the structure in the data implied by some simple priors.
One other example which almost certainly plays a role in animals and has also shown great promise in understanding video, is in the assumption that the future should be predictable from the past. Just by starting from that assumption, you can teach a neural network a whole lot. Or on a philosophical level, I am inclined to believe that this assumption underlies almost everything what we consider to be 'knowledge'.
I am not saying anything new here; but it is relatively new in the sense that these possibilities are too young to have found many applications yet, and do not yet have percolated down to the textbook understanding of 'what an ANN can do'. So to answer the OPs question; ANN's have already closed much of the gap that you describe.
$endgroup$
add a comment |
$begingroup$
As pointed out by others, the data-efficiency of artificial neural networks varies quite substantially, depending on the details. As a matter of fact, there are many so called one-shot learning methods, that can solve the task of labelling trams with quite good accuracy, using only a single labelled sample.
One way to do this is by so-called transfer learning; a network trained on other labels is usually very effectively adaptable to new labels, since the hard work is breaking down the low level components of the image in a sensible way.
But we do not infact need such labeled data to perform such task; much like babies dont need nearly as much labeled data as the neural networs you are thinking of do.
For instance, one such unsupervised methods that I have also successfully applied in other contexts, is to take an unlabeled set of images, randomly rotate them, and train a network to predict which side of the image is 'up'. Without knowing what the visible objects are, or what they are called, this forces the network to learn a tremendous amount of structure about the images; and this can form an excellent basis for much more data-efficient subsequent labeled learning.
While it is true that artificial networks are quite different from real ones in probably meaningful ways, such as the absence of an obvious analogue of backpropagation, it is very probably true that real neural networks make use of the same tricks, of trying to learn the structure in the data implied by some simple priors.
One other example which almost certainly plays a role in animals and has also shown great promise in understanding video, is in the assumption that the future should be predictable from the past. Just by starting from that assumption, you can teach a neural network a whole lot. Or on a philosophical level, I am inclined to believe that this assumption underlies almost everything what we consider to be 'knowledge'.
I am not saying anything new here; but it is relatively new in the sense that these possibilities are too young to have found many applications yet, and do not yet have percolated down to the textbook understanding of 'what an ANN can do'. So to answer the OPs question; ANN's have already closed much of the gap that you describe.
$endgroup$
As pointed out by others, the data-efficiency of artificial neural networks varies quite substantially, depending on the details. As a matter of fact, there are many so called one-shot learning methods, that can solve the task of labelling trams with quite good accuracy, using only a single labelled sample.
One way to do this is by so-called transfer learning; a network trained on other labels is usually very effectively adaptable to new labels, since the hard work is breaking down the low level components of the image in a sensible way.
But we do not infact need such labeled data to perform such task; much like babies dont need nearly as much labeled data as the neural networs you are thinking of do.
For instance, one such unsupervised methods that I have also successfully applied in other contexts, is to take an unlabeled set of images, randomly rotate them, and train a network to predict which side of the image is 'up'. Without knowing what the visible objects are, or what they are called, this forces the network to learn a tremendous amount of structure about the images; and this can form an excellent basis for much more data-efficient subsequent labeled learning.
While it is true that artificial networks are quite different from real ones in probably meaningful ways, such as the absence of an obvious analogue of backpropagation, it is very probably true that real neural networks make use of the same tricks, of trying to learn the structure in the data implied by some simple priors.
One other example which almost certainly plays a role in animals and has also shown great promise in understanding video, is in the assumption that the future should be predictable from the past. Just by starting from that assumption, you can teach a neural network a whole lot. Or on a philosophical level, I am inclined to believe that this assumption underlies almost everything what we consider to be 'knowledge'.
I am not saying anything new here; but it is relatively new in the sense that these possibilities are too young to have found many applications yet, and do not yet have percolated down to the textbook understanding of 'what an ANN can do'. So to answer the OPs question; ANN's have already closed much of the gap that you describe.
answered 2 days ago
Eelco HoogendoornEelco Hoogendoorn
856
856
add a comment |
add a comment |
$begingroup$
One way to train a deep neural network is to treat it as a stack of auto-encoders (Restricted Boltzmann Machines).
In theory, an auto-encoder learns in an unsupervised manner: It takes arbitrary, unlabelled input data and processes it to generate output data. Then it takes that output data, and tries to regenerate its input data. It tweaks its nodes' parameters until it can come close to round-tripping its data. If you think about it, the auto-encoder is writing its own automated unit tests. In effect, it is turning its "unlabelled input data" into labelled data: The original data serves as a label for the round-tripped data.
After the layers of auto-encoders are trained, the neural network is fine-tuned using labelled data to perform its intended function. In effect, these are functional tests.
The original poster asks why a lot of data is needed to train an artificial neural network, and compares that to the allegedly low amount of training data needed by a two-year-old human. The original poster is comparing apples-to-oranges: The overall training process for the artificial neural net, versus the fine-tuning with labels for the two-year-old.
But in reality, the two-year old has been training its auto-encoders on random, self-labelled data for more than two years. Babies dream when they are in utero. (So do kittens.) Researchers have described these dreams as involving random neuron firings in the visual processing centers.
New contributor
$endgroup$
1
$begingroup$
Agreed; except that auto-encoders in practice are not very powerful tools at doing much unsupervised learning at all; everything we know points at there being more going on, so the phrasing 'the two-year old has been training its auto-encoders' should not be taken too literally I suppose.
$endgroup$
– Eelco Hoogendoorn
yesterday
add a comment |
$begingroup$
One way to train a deep neural network is to treat it as a stack of auto-encoders (Restricted Boltzmann Machines).
In theory, an auto-encoder learns in an unsupervised manner: It takes arbitrary, unlabelled input data and processes it to generate output data. Then it takes that output data, and tries to regenerate its input data. It tweaks its nodes' parameters until it can come close to round-tripping its data. If you think about it, the auto-encoder is writing its own automated unit tests. In effect, it is turning its "unlabelled input data" into labelled data: The original data serves as a label for the round-tripped data.
After the layers of auto-encoders are trained, the neural network is fine-tuned using labelled data to perform its intended function. In effect, these are functional tests.
The original poster asks why a lot of data is needed to train an artificial neural network, and compares that to the allegedly low amount of training data needed by a two-year-old human. The original poster is comparing apples-to-oranges: The overall training process for the artificial neural net, versus the fine-tuning with labels for the two-year-old.
But in reality, the two-year old has been training its auto-encoders on random, self-labelled data for more than two years. Babies dream when they are in utero. (So do kittens.) Researchers have described these dreams as involving random neuron firings in the visual processing centers.
New contributor
$endgroup$
1
$begingroup$
Agreed; except that auto-encoders in practice are not very powerful tools at doing much unsupervised learning at all; everything we know points at there being more going on, so the phrasing 'the two-year old has been training its auto-encoders' should not be taken too literally I suppose.
$endgroup$
– Eelco Hoogendoorn
yesterday
add a comment |
$begingroup$
One way to train a deep neural network is to treat it as a stack of auto-encoders (Restricted Boltzmann Machines).
In theory, an auto-encoder learns in an unsupervised manner: It takes arbitrary, unlabelled input data and processes it to generate output data. Then it takes that output data, and tries to regenerate its input data. It tweaks its nodes' parameters until it can come close to round-tripping its data. If you think about it, the auto-encoder is writing its own automated unit tests. In effect, it is turning its "unlabelled input data" into labelled data: The original data serves as a label for the round-tripped data.
After the layers of auto-encoders are trained, the neural network is fine-tuned using labelled data to perform its intended function. In effect, these are functional tests.
The original poster asks why a lot of data is needed to train an artificial neural network, and compares that to the allegedly low amount of training data needed by a two-year-old human. The original poster is comparing apples-to-oranges: The overall training process for the artificial neural net, versus the fine-tuning with labels for the two-year-old.
But in reality, the two-year old has been training its auto-encoders on random, self-labelled data for more than two years. Babies dream when they are in utero. (So do kittens.) Researchers have described these dreams as involving random neuron firings in the visual processing centers.
New contributor
$endgroup$
One way to train a deep neural network is to treat it as a stack of auto-encoders (Restricted Boltzmann Machines).
In theory, an auto-encoder learns in an unsupervised manner: It takes arbitrary, unlabelled input data and processes it to generate output data. Then it takes that output data, and tries to regenerate its input data. It tweaks its nodes' parameters until it can come close to round-tripping its data. If you think about it, the auto-encoder is writing its own automated unit tests. In effect, it is turning its "unlabelled input data" into labelled data: The original data serves as a label for the round-tripped data.
After the layers of auto-encoders are trained, the neural network is fine-tuned using labelled data to perform its intended function. In effect, these are functional tests.
The original poster asks why a lot of data is needed to train an artificial neural network, and compares that to the allegedly low amount of training data needed by a two-year-old human. The original poster is comparing apples-to-oranges: The overall training process for the artificial neural net, versus the fine-tuning with labels for the two-year-old.
But in reality, the two-year old has been training its auto-encoders on random, self-labelled data for more than two years. Babies dream when they are in utero. (So do kittens.) Researchers have described these dreams as involving random neuron firings in the visual processing centers.
New contributor
edited 2 days ago
New contributor
answered 2 days ago
JasperJasper
1314
1314
New contributor
New contributor
1
$begingroup$
Agreed; except that auto-encoders in practice are not very powerful tools at doing much unsupervised learning at all; everything we know points at there being more going on, so the phrasing 'the two-year old has been training its auto-encoders' should not be taken too literally I suppose.
$endgroup$
– Eelco Hoogendoorn
yesterday
add a comment |
1
$begingroup$
Agreed; except that auto-encoders in practice are not very powerful tools at doing much unsupervised learning at all; everything we know points at there being more going on, so the phrasing 'the two-year old has been training its auto-encoders' should not be taken too literally I suppose.
$endgroup$
– Eelco Hoogendoorn
yesterday
1
1
$begingroup$
Agreed; except that auto-encoders in practice are not very powerful tools at doing much unsupervised learning at all; everything we know points at there being more going on, so the phrasing 'the two-year old has been training its auto-encoders' should not be taken too literally I suppose.
$endgroup$
– Eelco Hoogendoorn
yesterday
$begingroup$
Agreed; except that auto-encoders in practice are not very powerful tools at doing much unsupervised learning at all; everything we know points at there being more going on, so the phrasing 'the two-year old has been training its auto-encoders' should not be taken too literally I suppose.
$endgroup$
– Eelco Hoogendoorn
yesterday
add a comment |
$begingroup$
We don't learn to "see cars" until we learn to see
It takes quite a long time and lots of examples for a child to learn how to see objects as such. After that, a child can learn to identify a particular type of object from just a few examples. If you compare a two year old child with a learning system that literally starts from a blank slate, it's an apples and oranges comparison; at that age child has seen thousands of hours of "video footage".
In a similar manner, it takes artificial neural networks a lot of examples to learn "how to see" but after that it's possible to transfer that knowledge to new examples. Transfer learning is a whole domain of machine learning, and things like "one shot learning" are possible - you can build ANNs that will learn to identify new types of objects that it hasn't seen before from a single example, or to identify a particular person from a single photo of their face. But doing this initial "learning to see" part well requires quite a lot of data.
Furthermore, there's some evidence that not all training data is equal, namely, that data which you "choose" while learning is more effective than data that's simply provided to you. E.g. Held & Hein twin kitten experiment. https://www.lri.fr/~mbl/ENS/FONDIHM/2013/papers/about-HeldHein63.pdf
$endgroup$
add a comment |
$begingroup$
We don't learn to "see cars" until we learn to see
It takes quite a long time and lots of examples for a child to learn how to see objects as such. After that, a child can learn to identify a particular type of object from just a few examples. If you compare a two year old child with a learning system that literally starts from a blank slate, it's an apples and oranges comparison; at that age child has seen thousands of hours of "video footage".
In a similar manner, it takes artificial neural networks a lot of examples to learn "how to see" but after that it's possible to transfer that knowledge to new examples. Transfer learning is a whole domain of machine learning, and things like "one shot learning" are possible - you can build ANNs that will learn to identify new types of objects that it hasn't seen before from a single example, or to identify a particular person from a single photo of their face. But doing this initial "learning to see" part well requires quite a lot of data.
Furthermore, there's some evidence that not all training data is equal, namely, that data which you "choose" while learning is more effective than data that's simply provided to you. E.g. Held & Hein twin kitten experiment. https://www.lri.fr/~mbl/ENS/FONDIHM/2013/papers/about-HeldHein63.pdf
$endgroup$
add a comment |
$begingroup$
We don't learn to "see cars" until we learn to see
It takes quite a long time and lots of examples for a child to learn how to see objects as such. After that, a child can learn to identify a particular type of object from just a few examples. If you compare a two year old child with a learning system that literally starts from a blank slate, it's an apples and oranges comparison; at that age child has seen thousands of hours of "video footage".
In a similar manner, it takes artificial neural networks a lot of examples to learn "how to see" but after that it's possible to transfer that knowledge to new examples. Transfer learning is a whole domain of machine learning, and things like "one shot learning" are possible - you can build ANNs that will learn to identify new types of objects that it hasn't seen before from a single example, or to identify a particular person from a single photo of their face. But doing this initial "learning to see" part well requires quite a lot of data.
Furthermore, there's some evidence that not all training data is equal, namely, that data which you "choose" while learning is more effective than data that's simply provided to you. E.g. Held & Hein twin kitten experiment. https://www.lri.fr/~mbl/ENS/FONDIHM/2013/papers/about-HeldHein63.pdf
$endgroup$
We don't learn to "see cars" until we learn to see
It takes quite a long time and lots of examples for a child to learn how to see objects as such. After that, a child can learn to identify a particular type of object from just a few examples. If you compare a two year old child with a learning system that literally starts from a blank slate, it's an apples and oranges comparison; at that age child has seen thousands of hours of "video footage".
In a similar manner, it takes artificial neural networks a lot of examples to learn "how to see" but after that it's possible to transfer that knowledge to new examples. Transfer learning is a whole domain of machine learning, and things like "one shot learning" are possible - you can build ANNs that will learn to identify new types of objects that it hasn't seen before from a single example, or to identify a particular person from a single photo of their face. But doing this initial "learning to see" part well requires quite a lot of data.
Furthermore, there's some evidence that not all training data is equal, namely, that data which you "choose" while learning is more effective than data that's simply provided to you. E.g. Held & Hein twin kitten experiment. https://www.lri.fr/~mbl/ENS/FONDIHM/2013/papers/about-HeldHein63.pdf
edited yesterday
answered yesterday
PeterisPeteris
1964
1964
add a comment |
add a comment |
Thanks for contributing an answer to Cross Validated!
- Please be sure to answer the question. Provide details and share your research!
But avoid …
- Asking for help, clarification, or responding to other answers.
- Making statements based on opinion; back them up with references or personal experience.
Use MathJax to format equations. MathJax reference.
To learn more, see our tips on writing great answers.
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstats.stackexchange.com%2fquestions%2f394118%2fwhy-do-neural-networks-need-so-many-training-examples-to-perform%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
19
$begingroup$
Elephants might be a better example than cars. As others have noted, a child may have seen many cars before hearing the label, so if their mind already defines "natural kinds" it now has a label for one. However, a Western child indisputably develops a good elephant-classifying system on the basis of just a few data.
$endgroup$
– J.G.
2 days ago
40
$begingroup$
What makes you think that a human child’s brain works like a neural network?
$endgroup$
– Paul Wasilewski
2 days ago
9
$begingroup$
A NN can be shown an image of a car. Your child gets a full 3D movie from different perspectives, for several different types of car. Your child also likely has similar examples to distinguish a car from. For instance their baby stroller, toys, etc. Without those, I think your child would have needed more examples.
$endgroup$
– Stian Yttervik
2 days ago
15
$begingroup$
@MSalters In the sense of an Artificial Neural Network? Probably not.
$endgroup$
– Firebug
2 days ago
9
$begingroup$
"A human child at age 2 needs around 5 instances of a car to be able to identify it with reasonable accuracy" Such a child has had two full years of experience with things that aren't cars. I'm certain that plays a significant role.
$endgroup$
– DarthFennec
yesterday