by stablemap on 10/28/19, 4:46 AM with 86 comments
by nutanc on 10/28/19, 7:26 AM
That said, I have a small problem with the examples presented to say that already machines understand us :)
The article says 'For example, when I tell Siri “Call Carol” and it dials the correct number, you will have a hard time convincing me that Siri did not understand my request"
Let me try to take a shot at trying to explain that Siri did not "understand" your request.
Siri was waiting for a command and executed the best command that matched. Which is, make a phone call.
It did not understand what you meant because it did not take the whole environment into consideration. What if Carol was just in the other room. A human would maybe just shout "hey Carol, Thomas is asking you to come", instead of making a phone call.
If listening to a request and executing a command is understanding, then computers have been understanding us for a long time. Even without the latest advances in AI.
by js8 on 10/28/19, 7:14 AM
For example, take the classical AI knowledgebase fragment, "bird is animal that flies". If I ask example of bird, it can say "eagle", and exhibit some understanding. We can then probe further and ask for a bird which is not an eagle. If it says "bat" or "balloon", it exhibits that it still doesn't understand birds quite right.
In particular, if the description is nonsensical and thus impossible to understand, we cannot give any examples.
This idea was really inspired by the study, where they asked people to recognize nonsensical and profound sentences, describing certain situation. The profound are the ones where you can create a concrete instance of the situation.
by cjfd on 10/28/19, 9:03 AM
On the other hand, machines still perform actions that one could call 'stupid'. When alphago was losing in the fourth match against Lee Sedol it would play 'stupid' moves. These were, for instance, trivial threads that any somewhat accomplished amateur go player would recognize in an instant and answer correctly.
Humans, and also animals, have a hierarchy in their understanding of things. This maps on brain structure too. Evolution has added layers to the brain while keeping the existing structure. In this layered structure the lower parts are faster and more accurate but not as sophisticated. Stupidity arises because of a lack of layeredness so when the goal of winning the game is thwarted the top layer doesn't have any useful thing to do anymore and it falls back on a layer behind that. For alphago pretty much the only layer behind its very strong go engine is the rules of go. So, even when it is losing it will never play an illegal move but it will do otherwise trivially stupid things. For humans there is a layer between these things that prevents them from doing useless stuff. For living entities this is essential for survival. You can be forgetful of your dentist appointment but it is not possible to forget to let your heart beat. It seems that this problem could be mended by putting layers between the top level algorithm and most basic hardware level such that stupid stuff is preempted.
by modeless on 10/28/19, 6:11 AM
Google Search doesn't, but Google Assistant does. I posed the exact queries suggested by the article and the second query of simply the word "when" did give the correct answer (May 11 1997).
by BoppreH on 10/28/19, 8:13 AM
For example, imagine a system that has as input the picture of a human face in RAW format. If the system runs the picture through JPEG compression, for example, and returns something substantially smaller, it has shown some understanding of the input (color, spatial repetition, etc).
A more advanced system, with more understanding, may recognize it as a human face, and convert it to a template like the ones used for facial recognition. It doesn't care about individual pixels anymore, or the lighting, just general features of faces. It understands faces.
An even more advanced system may recognize the specific person and compress the whole thing to a few bits.
I would say that an OCR scanner understands the alphabet and how text is laid out, GPT-2 understands the relationship between words and how text is written. And a physics simulator understands basic physics because it can approximately compress a sequence of object movements into only initial conditions and small corrections.
Lossy compression makes this concept non-trivial to measure, but it's still a world's away from the normal philosophical arguments.
by stared on 10/28/19, 10:43 AM
by Nasrudith on 10/28/19, 12:38 PM
Human understanding has been wrong often enough, missing enough crucial context to be dangerously hillariously wrong even amongst the "experts" of the day who came closest.
The isn't some epistemological nilhism but to point out that understanding is incomplete for everyone and just because a given intelligence subset doesn't match with our assumptions doesn't mean it is wrong - although it also isn't always right.
by ilaksh on 10/28/19, 7:09 PM
There are projects doing video and text understanding. I think the trick to efficient generalization is to have the representations properly factored out somehow. Maybe things like capsule networks will help. Although that my guess is that to get really sort of componentized efficient understanding neural networks are not going to be the most effective way.
by avmich on 10/28/19, 3:14 PM
This sounds a bit like a studying for a test taking. What if we made a definition and then worked successfully to reach the state when, according to this definition, the system "understands". Can we expect to be satisfied with the result in general, outside of the definition?
The definition of understanding could be tricky, as history suggests. Other than "to understand is to translate into a form which is suitable for some use", there could be many definitions. Article itself brings examples of chess playing or truck driving which were considered good indicators, yet failed to satisfy us in some ways.
Maybe we should just keep redefining "understanding" as good as we can today, and changing it if needed, and work trying to create a system "good", not necessarily "passing the test"?
by YeGoblynQueenne on 10/28/19, 10:00 AM
But I have to disagree with this (because of course I do):
>> For example, when I tell Siri “Call Carol” and it dials the correct number, you will have a hard time convincing me that Siri did not understand my request.
That is a very common-sense and down-to-earth non-definition of intelligence: how can an entity that is answering a question correctly not "understand" the question?
I am going to quote Richard Feynman who encountered an example of this "how":
After a lot of investigation, I finally figured out that the students had memorized everything, but they didn’t know what anything meant. When they heard “light that is reflected from a medium with an index,” they didn’t know that it meant a material such as water. They didn’t know that the “direction of the light” is the direction in which you see something when you’re looking at it, and so on. Everything was entirely memorized, yet nothing had been translated into meaningful words. So if I asked, “What is Brewster’s Angle?” I’m going into the computer with the right keywords. But if I say, “Look at the water,” nothing happens – they don’t have anything under “Look at the water”!
https://v.cx/2010/04/feynman-brazil-education
In this (in?) famous passage Feynman is arguing that students of physics that he met in Brazil didn't know physics, even though they had memorised physics textbooks.
Feynman doesn't talk about "understanding". Rather he talks about "knowing" a subject. But his is also a very straight-forward definition of knowing: you can tell whether someone knows a subject if you ask them many questions from different angles and find that they can only answer the questions asked from one single angle.
So if I follow up "Siri, call Carol" with "Siri, what is a call" and Siri answers by calling Carol, I know that Siri doesn't know what a call is, probably doesn't know what a Carol is, or what a call-Carol is, and so that Siri doesn't have any understanding from a very common-sense point of view.
Not sure if this goes beyond the Chinese room argument though. Perhaps I'm just on a diffferent side of it than Thomas Dietterich.
by visarga on 10/28/19, 7:20 AM
I think the key ingredient is 'being in the game', that means, having a body, being in an environment with a purpose. Humans are by default playing this game called 'life', we have to understand otherwise we perish, or our genes perish.
It's not about symbolic vs connectionist, or qualia, or self consciousness. It's about being in the world, acting and observing the effects of actions, and having something to win or lose as a consequence of acting. This doesn't happen when training a neural net to recognise objects in images or doing translation. It's just a static dataset, a 'dead' world.
AI until now has had a hard time simulating agents or creating real robotic bodies - it's expensive, and the system learns slowly, and it's unstable. But progress happens. Until our AI agents get real hands and feet and a purpose they can't be in the world and develop true understanding, they are more like subsystems of the brain than the whole brain. We need to close the loop with the environment for true understanding.
by boyadjian on 10/28/19, 8:29 AM
by RaiseProfits on 10/28/19, 7:56 AM
by basicplus2 on 10/28/19, 7:05 AM
by igammarays on 10/28/19, 6:53 AM
by friendlybus on 10/28/19, 7:26 AM
The idea that a new self- sustaining meaning generation can arise out of the interlocking mechanisms of a computer is an interesting one. As we see self driven car CEOs describe some of the most advanced systems we have, requiring to be run in controlled environments and balking at the infinite complexity of real life, are we really building computer systems that are anything more than an incredibly sophisticated loop?