Skip to content

Sense from Chaos – Crossing the Semantic Gap

March 23, 2016

Edsger Dijkstra once said:

“The question of whether machines can think is about as relevant as the question of whether submarines can swim.”

This was the view most AI researchers held back in the 1960s and 1970s. Back then, many thought that general-purpose AI could be achieved purely through symbolic manipulation. That is, it was thought that we could build machines that, through purely logical reasoning, would derive a sufficient understanding of the world to reach and even exceed human intelligence. This kind of vision of AI is illustrated in classic sci-fi novels, and embodied in the infamous HAL 9000 computer of 2001: A Space Odyssey.

The quest to build symbolic AI led to much research, and impressive early successes. These successes lead to great optimism, and the belief that computers would be able to effectively handle human language, machine translation and vehicle driving before 1980. Symbolic reasoning is very adequate for reasoning about mathematics, or small closed systems with very well-defined properties, such as the game of chess. Unfortunately, multiple dead ends were quickly reached. It was found, as philosophers had predicted, that you could hardly build a machine capable of reasoning about the real world through purely symbolic means. The problem was one of grounding.

Symbols in a vacuum don’t mean anything. You can create an ontology in which you define the concept of a chair, and you can put logical statements in this ontology as to how people use chairs for sitting on, and that chairs are movable objects that behave according to Newtonian laws of physics. However, if you’re trying to build a robot that can see and manipulate chairs, your ontology is basically worthless. You have two HD cameras providing you with 20 million pixels of raw data at 60 frames a second, and no logical statements are sufficient to help you tell where a chair might be in all that noise. Your machine has the concept of a chair, but this concept exists only in some Platonic realm that is completely divorced from the real-world.

Ultimately, the grandiose predictions of early AI researchers proved much too ambitious, because the real world was much too difficult to cope with, or at least, much more difficult to cope with than mathematicians had hoped. This lead to the first AI winter, with funding rapidly drying up for what is now known as GOFAI (Good Old Fashioned AI) research. For at least two decades after this, AI became kind of a dirty word in computer science circles. Those who still pursued AI-related research renamed their field machine learning, so as to avoid any association with the overhyped goal that we might one day build machines with human-level intelligence. From then on, machine learning researchers left behind the lofty goals of AI, and instead focused on the basics: small narrowly-defined learning tasks where they knew they could make headway.

Neural networks are not new, but it’s only in the last few years that they have started to really shine, with deep learning. Advances in algorithms, access to large datasets and the unprecedented availability of computational resources have made it possible to scale this machine learning technique to networks with many layers of depth. Recently, some impressive feats have been achieved with deep neural networks, including object classification that exceeds human performance, and of course the much-discussed victory of the computer program AlphaGo over human Go champion Lee Sedol. Some amazing projects such as the neural artistic style transfer and and deep networks that learn to synthesize new images have also surfaced.

At the moment, universities are seeing a huge increase in student interest in deep neural networks, with classes sometimes tripling in size over previous years. Governments and industry alike are pouring billions into deep learning research. It’s undeniable: we’re in the middle of a machine learning boom. The optimism is so great that well-known researchers such as Yann Lecun and Yoshua Bengio are even daring to use the word AI again.There’s a lot of excitement (and some fear) about what deep neural networks will bring. People are becoming scared about robots taking human jobs, and the question of when computers will reach human-level intelligence is being asked.

To me, it really is a question of when. Provided that humanity doesn’t sink into a post-apocalyptic middle-age, there is no doubt in my mind that machines can and will reach human intelligence. I just don’t think that we’ll get to Artificial General Intelligence (AGI) in the way that most people think we will. Many seem to believe that we just haven’t come up with the right algorithm for human-equivalent intelligence, or that we just don’t have enough computational power. Clearly, to reach human-level intelligence, what we need is a deep neural network with a number of connections equivalent to that present in the human brain, right? I don’t think so. I don’t think that’s what we’re missing.

In my view, the human brain is not running some one algorithm. The human brain is not just a huge wad of neurons, a huge uniform neural network. The human brain is made of many different specialized components that do different things, connected together by a variety of pathways. Deep neural networks are awesome, they’re an amazing achievement, but they’re just one piece of the puzzle. What deep neural networks do, is solve the perceptual problem. Deep learning allows us to do what the human visual cortex does. You get a million pixels of color information as input, and you turn this raw mass of data into a few classes of output. In short, with deep neural networks, we can turn real-world data into symbols.

There is no doubt in my mind that an AGI has to contain some sort of connectionist component, such as a neural network within it. To make useful AI programs however, the logical thing to do seems to be to assemble multiple specialized pieces together. In my view, AlphaGo is a beautiful illustration of this. It melds together multiple deep neural networks, which are used to do things such as assigning a value to different Go board configurations, along with a Markov tree search algorithm for looking at possible future moves. AlphaGo is very much a marriage of GOFAI techniques together with the power of deep neural networks. Deep networks make sense of complex board configurations without the need for hard-written logical rules as to how individual Go stones should be counted. Deep networks do the more intuitive perceptual work, and good old fashioned tree search does the logical reasoning based on this data.

Deep neural networks bridge the semantic gap between classical computer systems, where symbolic entities are defined in absolute terms in databases, and the fuzziness of the real world, where exceptions are the norm, and nothing can be entirely captured by absolute rules. If I had to guess, I would speculate that robots of the future, as they get more intelligence, are going to feature increasingly complex AI architectures made of multiple specialized components. There will be deep networks that do vision, hearing, perceive human faces and facial expressions. There will also be modules that do planning, navigation, logical/ontological reasoning and reasoning by analogy. All of this AI technology that we’ve been developing for the last 60 years is going to need to come together into a whole. That is one of the next big challenges. Making AGI happen won’t just be scientific work, it will also pose a plethora of engineering challenges.

I think it’s possible to build useful robots and human-equivalent AI without understanding the human brain in depth. However, I think that when we do build such AI, we’re necessarily going to converge towards an architecture that does many things in ways that are actually quite similar to what happens, in computational terms, inside the human brain. Submarines may not swim, but in order to move forward, they still have to displace water. It’s said that humans can only keep about 7 items of data in their working memory. It wouldn’t surprise me all that much if, one day, neuroscientists manage to map out the human frontal cortex, and they discover that in this region of the brain, there are neurons laid out in a system that implements what can essentially be thought of as a small set of general-purpose registers, each able to store a neural vector representing a semantic concept. In other words, our ability to reason by analogy and to manipulate abstract concepts in our mind is probably embodied by neural machinery that serves to perform symbolic manipulations.

8 Comments
  1. Anonymous permalink

    What is the reason to create AGI?

    • The more intelligent machines are, the more able they are to deal with the real-world. If we want to be able to have say, useful household robots, we’re actually going to need them to be very intelligent. I believe that at some point, you kind of cross a threshold into AGI.

      The flipside is, AI is an enormously powerful tool. You just need to look at the amount of money being poured into AI research, how every big tech company is doing it, to see that this is essentially an AI arms race. AGI is likely to happen if only because everyone knows that intelligence is an extremely powerful force. Everyone wishes they had that kind of power under their own control, and they wish to have it before everybody else does.

      • Anonymous permalink

        You can already buy a useful household robot: google iRobot products and try for yourself, they obviously don’t need AGI exactly because they don’t need to be able to do anything beyond doing household chores – they are narrowly specialized. Why would you need intelligence for specialized tasks? You can, but is there need?

        • Yes, of course there is. If the robot was intelligent enough, you could tell it to drive the kids to school and go do your groceries. Maybe even ask it to design and build you new furniture. The smarter it is, the more useful.

        • Anonymous permalink

          Something like google car should be able to drive kids to school. And what if your groceries is furniture design? Why human is not good enough for it? Also design takes time, it’s more useful and simple to buy ready furniture.

  2. I’m not sure how useful this is, but one idea that I’ve been riffing for the last 15 years or so is the distinction between the representation and the thing being represented.

    A deep neural network is just an idea: it doesn’t really have any independent existence. They are, at heart, just complicated mathematical functions. They do seem to have some serious legs for classification tasks (and possibly more), but I would be surprised if they were sufficient for fully explaining (and-implementing) human-level cognition, perception, and behavior.

    Similarly, our model of the brain is just that, a model. I’m sure it’s possible, on some level, to understand the brain as a symbol-manipulating machine (depending on your definition of “symbol”). But I wonder if there might not be other explanations for our thoughts and behavior that don’t actually involve such representation. For example, most of the time when I’m programming (a symbolic activity if there ever was one), I only rarely feel like I’m actually shuffling and juggling concepts. For the most part I feel like code just comes out of me.

    Also, while I’m sure neural tissue is an essential ingredient to cognition, I wonder if there might not be other elements to it — might blood flow or hormones also be information channels of some kind? Maybe even other things that we haven’t yet developed the concepts for?

    • I definitely don’t think that just a large deep neural network is sufficient to implement cognition. Neural networks are pattern matching machines, perception engines. Functions from some input to some output. The best you can hope to get out of them, it seems, is a response to a stimuli. You can get a fairly nuanced and sophisticated response, maybe, but scaling up deep neural networks won’t get you sentience.

      You say you feel that code just comes out of you. That’s normal. For the most part, our cognition is a largely unconscious phenomenon. We’re not really aware of all that goes on inside. When we’re highly focused on a task, we can actually lose our sense of self for a little bit, as the inner monologue is silenced to make space for something else.

      Maybe symbol is not the right word when talking about the human brain. It’s probably more of a “concept-manipulating machine” than a “symbol manipulating machine”. I would think though, that in order to be able to implement higher-order logical reasoning, in order for us to be able to think of entities in relationship to one another, our brain probably has to implement some kind of a way to shuffle these “concepts”, to move them around in memory, to transfer them from working memory to long term memory, to feed them into creative analogy-producing circuits, and things of that sort.

      We’re obviously missing a lot of information about the brain. My hunch is that the more we dig, the more we’ll find out that there are different specialized pathways that do different things, that implement different aspects of reasoning and awareness. We’re probably going to be able to find way more analogies with the digital computer systems that we do know of. I think the human brain probably has neural circuits that serve as buffers, registers, control lines, gating circuits, multiplexers, etc.

      • noah permalink

        Great post :)

        My own brain makes me think that human-like intelligence* lies further down the road of compounded experience than towards symbolic manipulation. That is, if we want to replicate human capabilities, I think we have to greatly break down or completely discard symbolic manipulation. I think that we fool ourselves if we think the symbols in our brains are static or even consistent from one thought to the next.

        I believe we like to think of ourselves as logical, (despite repeated evidence that we are terrible at logic) so we come up with pictures of our minds that imply we are logical, systematic, and reasoned in our thinking. In actuality, we have thoughts arise and then after the fact try to fit our opinions and decisions into a framework.

        Cognitive dissonance is a strong motivator to bend reason to fit whatever we have already done, but post-hoc analysis *is* useful. We do sometimes rationally analyze our behavior after the fact and make corrections. I don’t believe human intelligence starts there however. I think what we consider rational thought is analogous to back-propagation; it allows us to make corrections to the underlying system that doesn’t follow logical rules.

        While neural nets may be “just math” I don’t think calling them logical lines up with what we humans mean by logic. The set of logical rules that let a human or machine recognize a face are very likely as numerous as the tuning of each neuron. In other words, I don’t think you can effectively reduce the network size of a machine or brain that can recognize faces and produce a set of rules that a human being can consciously comprehend.

        Ultimately, if we discover that you don’t need a foundation of symbolic manipulation rules for an AI, I will not be surprised. After all, logic came out of a bunch of neural networks that are not grounded in logic :)

        * I think it’s strange that we say computers are not more intelligent than us? They are way better at math, logic, and any deterministic problem solving you can think of. I know what we mean is that a computer can’t do a bunch of things that humans do without thinking (like observe when someone is upset,) but that’s rigging the game to favor ourselves. I think an outside non-human observer would have to conclude that humans and computers are good at different things, with neither one clearly more intelligent overall.

Leave a comment