Artificial intelligence: summoning the demon

We need to understand our own intelligence is competition for our artificial, not-quite intelligences.

Blue_Swirl_edward_musiak_Flickr

A few days ago, Elon Musk likened artificial intelligence (AI) to “summoning the demon.” As I’m sure you know, there are many stories in which someone summons a demon. As Musk said, they rarely turn out well.

There’s no question that Musk is an astute student of technology. But his reaction is misplaced. There are certainly reasons for concern, but they’re not Musk’s.

The problem with AI right now is that its achievements are greatly over-hyped. That’s not to say those achievements aren’t real, but they don’t mean what people think they mean. Researchers in deep learning are happy if they can recognize human faces with 80% accuracy. (I’m skeptical about claims that deep learning systems can reach 97.5% accuracy; I suspect that the problem has been constrained some way that makes it much easier. For example, asking “is there a face in this picture?” or “where is the face in this picture?” is much different from asking “what is in this picture?”) That’s a hard problem, a really hard problem. But humans recognize faces with nearly 100% accuracy. For a deep learning system, that’s an almost inconceivable goal. And 100% accuracy is orders of magnitude harder than 80% accuracy, or even 97.5%.

What kinds of applications can you build from technologies that are only accurate 80% of the time, or even 97.5% of the time? Quite a few. You might build an application that created dynamic travel guides from online photos. Or you might build an application that measures how long diners stay in a restaurant, how long it takes them to be served, whether they’re smiling, and other statistics. You might build an application that tries to identify who appears in your photos, as Facebook has. In all of these cases, an occasional error (or even a frequent error) isn’t a big deal. But you wouldn’t build, say, a face-recognition-based car alarm that was wrong 20% of the time — or even 2% of the time.

Similarly, much has been made of Google’s self-driving cars. That’s a huge technological achievement. But Google has always made it very clear that their cars rely on the accuracy of their highly detailed street view. As Peter Norvig has said, it’s a hard problem to pick a traffic light out of a scene and determine if it is red, yellow, or green. It is trivially easy to recognize the color of a traffic light that you already know is there. But keeping Google’s street view up to date isn’t simple. While the roads change infrequently, towns frequently add stop signs and traffic lights. Dealing with these changes to the map is extremely difficult, and only one of many challenges that remain to be solved: we know how to interpret traffic cones, we know how to think about cars or humans behaving erratically, we know what to do when the lane markings are covered by snow. That ability to think like a human when something unexpected happens makes a self-driving car a “moonshot” project. Humans certainly don’t perform perfectly when the unexpected happens, but we’re surprisingly good at it.

You wouldn’t build a face-recognition-based car alarm that was wrong 20% of the time.So, AI systems can do, with difficulty and partial accuracy, some of what humans do all the time without even thinking about it. I’d guess that we’re 20 to 50 years away from anything that’s more than a crude approximation to human intelligence. It’s not just that we need bigger and faster computers, which will be here sooner than we think. We don’t understand how human intelligence works at a fundamental level. (Though I wouldn’t assume that understanding the brain is a prerequisite for artificial intelligence.) That’s not a problem or a criticism, it’s just a statement of how difficult the problems are. And let’s not misunderstand the importance of what we’ve accomplished: this level of intelligence is already extremely useful. Computers don’t get tired, don’t get distracted, and don’t panic. (Well, not often.) They’re great for assisting or augmenting human intelligence, precisely because as an assistant, 100% accuracy isn’t required. We’ve had cars with computer-assisted parking for more than a decade, and they’ve gotten quite good. Larry Page has talked about wanting Google search to be like the Star Trek computer, which can understand context and anticipate what the humans wants. The humans remain firmly in control, though, whether we’re talking to the Star Trek computer or Google Now.

I’m not without concerns about the application of AI. First, I’m concerned about what happens when humans start relying on AI systems that really aren’t all that intelligent. AI researchers, in my experience, are fully aware of the limitations of their systems. But their customers aren’t. I’ve written about what happens when HR departments trust computer systems to screen resumes: you get some crude pattern matching that ends up rejecting many good candidates. Cathy O’Neil has written on several occasions about machine learning’s potential for dressing up prejudice as “science.”

The problem isn’t machine learning itself, but users who uncritically expect a machine to provide an oracular “answer,” and faulty models that are hidden from public view. In a not-yet published paper, DJ Patil and Hilary Mason suggest that you search Google for GPS and cliff; you might be surprised at the number of people who drive their cars off cliffs because the GPS told them to. I’m not surprised; a friend of mine owns a company that makes propellers for high-performance boats, and he’s told me similar stories about replacing the propellers for clients who run their boats into islands.

David Ferrucci and the other IBMers who built Watson understand that Watson’s potential in medical diagnosis isn’t to have the last word, or to replace a human doctor. It’s to be part of the conversation, offering diagnostic possibilities that the doctor hasn’t considered, and the reasons one might accept (or reject) those diagnoses. That’s a healthy and potentially important step forward in medical treatment, but do the doctors using an automated service to help make diagnoses understand that? Does our profit-crazed health system understand that? When will your health insurance policy say “you can only consult a doctor after the AI has failed”? Or “Doctors are a thing of the past, and if the AI is wrong 10% of the time, that’s acceptable; after all, your doctor wasn’t right all the time, anyway”? The problem isn’t the tool; it’s the application of the tool. More specifically, the problem is forgetting that an assistive technology is assistive, and assuming that it can be a complete stand-in for a human.

Second, I’m concerned about what happens if consumer-facing researchers get discouraged and leave the field. Although that’s not likely now, it wouldn’t be the first time that AI was abandoned after a wave of hype. If Google, Facebook, and IBM give up on their “moonshot” AI projects, what will be left? I have a thesis (which may eventually become a Radar post) that a technology’s future has a lot to do with its origins. Nuclear reactors were developed to build bombs, and as a consequence, promising technologies like Thorium reactors were abandoned. If you can’t make a bomb from it, what good is it?

The problem isn’t the tool; it’s the application of the tool.If I’m right, what are the implications for AI? I’m thrilled that Google and Facebook are experimenting with deep learning, that Google is building autonomous vehicles, and that IBM is experimenting with Watson. I’m thrilled because I have no doubt that similar work is going on in other labs, in other places, that we know nothing about. I don’t want the future of AI to be shortchanged because researchers hidden in government labs choose not to investigate ideas that don’t have military potential. And we do need a discussion about the role of AI in our lives: what are its limits, what applications are OK, what are unnecessarily intrusive, and what are just creepy. That conversation will never happen when the research takes place behind locked doors.

At the end of a long, glowing report about the state of AI, Kevin Kelly makes the point that every advance in AI, every time computers make some other achievement (playing chess, playing Jeopardy, inventing new recipes, maybe next year playing Go), we redefine the meaning of our own human intelligence. That sounds funny; I’m certainly suspicious when the rules of the game are changed every time it appears to be “won,” but who really wants to define human intelligence in terms of chess-playing ability? That definition leaves out most of what’s important in humanness.

Perhaps we need to understand our own intelligence is competition for our artificial, not-quite intelligences. And perhaps we will, as Kelly suggests, realize that maybe we don’t really want “artificial intelligence.” After all, human intelligence includes the ability to be wrong, or to be evasive, as Turing himself recognized. We want “artificial smartness”: something to assist and extend our cognitive capabilities, rather than replace them.

That brings us back to “summoning the demon,” and the one story that’s an exception to the rule. In Goethe’s Faust, Faust is admitted to heaven: not because he was a good person, but because he never ceased striving, never became complacent, never stopped trying to figure out what it means to be human. At the start, Faust mocks Mephistopheles, saying “What can a poor devil give me? When has your kind ever understood a Human Spirit in its highest striving?” (lines 1176-7, my translation). When he makes the deal, it isn’t the typical “give me everything I want, and you can take my soul”; it’s “When I lie on my bed satisfied, let it be over…when I say to the Moment ‘Stay! You are so beautiful,’ you can haul me off in chains” (1691-1702). At the end of this massive play, Faust is almost satisfied; he’s building an earthly paradise for those who strive for freedom every day, and dies saying “In anticipation of that blessedness, I now enjoy that highest Moment” (11585-6), even quoting the terms of his deal.

So, who’s won the bet? The demon or the summoner? Mephistopheles certainly thinks he has, but the angels differ, and take Faust’s soul to heaven, saying “Whoever spends himself striving, him we can save” (11936-7). Faust may be enjoying the moment, but it’s still in anticipation of a paradise that he hasn’t built. Mephistopheles fails at luring Faust into complacency; rather, he is the driving force behind his striving, a comic figure who never understands that by trying to drag Faust to hell, he was pushing him toward humanity. If AI, even in its underdeveloped state, can serve this function for us, calling up that demon will be well worth it.

Cropped image on article and category pages by edward musiak on Flickr, used under a Creative Commons license.

tags: ,
  • maxkhesin

    Elon’s comments are motivated by Nick Bostrom’s new book Superintelligence (http://www.amazon.com/Superintelligence-Dangers-Strategies-Nick-Bostrom/dp/0199678111). If anyone is interested in the details of the arguments, see there, also here http://intelligence.org/research/. Whether Deep Learning is going to produce a smarter than human agent is completely besides the point; even it this happens in the next 100 years (via some other technology) there are still very significant problems that need to be worked out to insure the eventually produced agent is a good rather than terrible thing for humanity.

  • Doug Gibbs

    Your right, the origin story is the only one that matters to the majority. AI will always be an evil technology. It became evil when Arnold said “Are you Sara Connors?”. There is no way to displace the story, even if it is wrong, with boring facts in a human mind.

  • PaulTopping

    The fact that current AI is nothing to be scared of says nothing about what will happen when true AI is achieved. Nor does any argument here convince me that we won’t get there. I share Elon Musk’s fear but I believe that it should not stop AI research. We just need to keep one hand on the power cord, ready to pull the plug when something inevitably goes wrong. Hmmm. I wonder if a human could react fast enough. Perhaps we better use AI for that. Hmmm. AI monitoring AI. What could go wrong with that?

    • maxkhesin

      See Bostrom’s book – I mentioned it above. He describes several containment strategies (and their weaknesses)

  • I recently blogged on why there will be no rise of the machines. Anyone who thinks otherwise doesn’t know how humans or machines work, or falls prey to the black box fallacy, in which someone who doesn’t a system concludes it’s magical and therefore capable of anything (see also: fears of nuclear power and radiation, and before that, fears of electricity, e.g. Mary Shelly’s Frankenstein).

    In short, human intelligence is a synonym for the efficacy of which humans are able to self-replicate in their native environment.

    As no one will be designing machines capable of self-replication anytime soon (if ever), machine intelligence will be directed at quite different goals, and the psychology of generalized AI will be quite foreign to the psychology of humans — and varied by the exact use case. There are an infinite many different “types” of intelligence corresponding with the infinite variety of goals and environments. Our own comes from the goal of self-replication combined with the environment in which we evolved.

    People imagine “true AI” or “strong AI” will be something like a human, which has evolved over the course of 4 billion years to become very good at self-replication. Even if it were possible to build such a system (which is no trivial undertaking — it took the universe 4 billion years, and the universe is a massively parallel quantum computer “simulating” quadrillions of possibilities in parallel), it would serve zero commercial value.

    Rather, “true AI” and “strong AI” of the future is likely to be directed at specific human needs, e.g. answering questions based on the entirety of human knowledge, diagnosing a patient given DNA analysis and past and present sensor and biochip readings, etc. The “circuits” that make up such AI will be built through feedback loops that encode our goals for them, while the environments will be whatever sorts of machines we have in the future (silicon, optical, mechanical, biological, quantum, etc.). Thus, the intelligence these machines have cannot be compared to our own and will not look anything like human intelligence, despite the fact that these “AIs” will exceed at every task imaginable (except the one we excel at: self-replication in our natural environment).

  • Daisy McIntyre

    I’m sorry, but it looks like you’re being overly dramatic.
    Usually software developers include risks of inaccuracy into the program. If the consequences of the mistakes are too high it will effect the final result. For example, the program can show the percentage of similarities and after certain amount it could put its answer into some sort of “gray zone”, kinda it’s PROBABLY right but be careful.

    I recently write my essay on this topic.

  • Make this big
    celebration of Christmas very special and loving with amazing Wine Hamper
    Basket best available at online store. Send Christmas Gifts to USA to the doorsteps of your dear ones and
    express your best love and affection.

    http://www.hampersusa.com/Christmas-Gifts-USA.asp

  • Our intelligence? nice post.
    Maybe in 100 years we’ll be smarter than now with the same natural behavior, our exemple just on http://www.escortbruxelles.be the capital of europe, and we know why.