By Eric Vandebroeck
and co-workers
Calling it, a Cargo
Cult Wired was one of the first high profile publications that pointed out
specific misconceptions surrounding AI, including that
intelligence is a single dimension. Most technical people tend to graph
intelligence how Nick
Bostrom does in
his book, Superintelligence, as a literal, single-dimension, linear
graph of increasing amplitude. At one end is the low intelligence of a small
animal; at the other end is the high intelligence, of, say, a genius, almost as
if intelligence were a sound level in decibels. Of course, it is then
straightforward to imagine the extension so that intelligence's loudness
continues to grow, eventually to exceed our own high intelligence and become a
super-loud intelligence, a roar! Way beyond us, and maybe even off the chart.
To which
Forbes added, we have fallen into this trap before. We have tried to turn
machines into automatons that mimic human behavior; these attempts failed.
Machines can do incredible things, but we need to design them in a way that
fits their strengths. Today, we struggle with building autonomous cars,
intelligent analysis tools, complex computer vision systems, but we are not the
first to automate our world. Rewind to the early 1800s, and the challenge of
the day was how to build a sewing machine. The world population was increasing,
and walking naked was not an option. There were just not enough free hands to
make the clothes to meet demand. We had to come up with a better solution for
making garments.
Extensive analysis
about AI was put out by UNESCO when it aptly stated that the success of
the term AI is sometimes
based on a misunderstanding,
Or more recently, MDN
explains how AI does not mean data will be fixed automatically, more data
will produce better outcomes, or that
the AI is ready to create out-of-the-box solutions.
Of course, the Cargo
cult idea is not entirely
true; AI is possible. As to that, the future of AI is a scientific unknown. The
myth of artificial intelligence is that its arrival is inevitable, and only a
matter of time, that we have already embarked on the path that will lead to
human-level AI and then superintelligence. We have not.
The path exists only in our imaginations. Yet the
inevitability of AI is so ingrained in popular discussion, promoted by media
pundits, thought leaders like Elon Musk, and even many AI scientists (though
certainly not all), that arguing against it is often taken as a form of
Luddism, or at the very least a shortsighted view of the future of technology
and a dangerous failure to prepare for a world of intelligent machines.
The science of AI has
uncovered a huge mystery at the heart of intelligence, which no one currently
knows how to solve. Proponents of AI have huge incentives
to minimize its known limitations. After all, AI is big
business, and it's increasingly dominant in culture. Yet, the possibilities
for future AI systems are limited by what we currently know about
intelligence's nature, whether we like it or not. And here we should say it
directly: all evidence suggests that human and machine intelligence are
radically different. The myth of AI insists that the differences are only
temporary and that more powerful systems will eventually erase them. Futurists
like Ray
Kurzweil and philosopher Nick Bostrom, prominent purveyors
of the myth, talk not only as if human-level Al were inevitable but also as if,
soon after its arrival, superintelligent machines would leave us far
behind. The scientific part of the myth assumes that we need only keep
“chipping away” at the challenge of general intelligence by making progress on
narrow intelligence feats, like playing games or recognizing images. This is a
profound mistake: success on narrow applications gets us not one step closer to
general intelligence. The inferences that systems require for general
intelligence, read a newspaper, hold a basic conversation, or become a helpmeet
like Rosie
the Robot in Tlic Jetsons, cannot be
programmed, learned, or engineered with our current knowledge of Al. As we
successfully apply simpler, narrow versions of intelligence that benefit from
faster computers and lots of data, we are not making incremental progress.
Rather, picking low-hanging fruit, Hie jumps to general “common sense” is
completely different. There’s no known path from one to the other. No algorithm
exists for general intelligence. And we have good reason to be skeptical that
such an algorithm will emerge through further efforts on deep learning systems
or any other approach popular today. Much more likely, it will require a major
scientific breakthrough, and no one currently has the slightest idea what such
a breakthrough would even look like, let alone the details of getting to it.
Mythology about Al is
bad because it covers up a scientific mystery in endless talk of ongoing
progress. The myth props up belief in inevitable success, but genuine respect
for science should bring us back to the drawing board. Pursuing why this
is not a good way to follow the smart money or even a
neutral stance. It is bad for science, and it is bad for us. Why? One reason
is that we are unlikely to get innovation if we choose to ignore a core
mystery rather than face up to it. A healthy culture for innovation emphasizes
exploring unknowns, not hyping extensions of existing methods, especially when
these methods are inadequate to take us much further.
Mythology about inevitable success in AI tends to extinguish the very culture
of invention necessary for real progress, with or without human-level Al. The
myth also encourages resignation to the creep of a machine-land, where the
genuine invention is sidelined in favor of futuristic talk advocating current
approaches, often from entrenched interests.
While we cannot prove
that AI overlords will not one day appear, we can give you a reason to discount
the prospects of that scenario seriously. For example, AI culture has
simplified ideas about people while expanding ideas about technology; This
began with its founder, Alan Turing, and involved understandable but
unfortunate simplifications one could call intelligence errors. Initial errors
were magnified into an ideology by Turing’s friend and statistician, I. J. Good
introduced the idea of 'ultra
intelligence' as the predictable result once human-level AI had been
achieved.
Between Turing and
Good, we see the modern myth of AI take shape. Its development has landed us in
an era of what I call technological kitsch, cheap imitations of deeper ideas
that cut off intelligent engagement and weaken our culture. Kitsch tells us how
to think and how to feel. Tire purveyors of kitsch benefit, while the consumers
of kitsch experience a loss. They, we, end up in a shallow world.
The only type of
inference, thinking, in other words, that will work for human-level AI (or
anything even close to it) is the one we don't know how to program or engineer.
The inference problem goes to the heart of the AI debate because it deals
directly with intelligence, in people or machines. Our knowledge of the various
inference types dates back to Aristotle and other
ancient Greeks and has been developed in logic and mathematics. The inference
is already described using formal, symbolic systems like computer programs, so
a very dear view of the project of engineering intelligence can be gained by
exploring inference. There are three types. Classic AI explored
one (deduction), modern AI explores another (induction). The third type
(abduction) makes for general intelligence, and, surprise, no one is working on
it at all. Finally, since each type of inference is distinct, meaning, one
type cannot be reduced to another, we know that failure to build AI systems
using the type of inference undergirding general intelligence will fail to
progress toward artificial general intelligence or
AGI.
The myth one could
argue has terrible consequences if taken seriously because it subverts science.
In particular, it erodes a culture of human
intelligence and invention, which is necessary for
the very breakthroughs we will need to understand our own future. Data science
(the application of a to “big data" is at best a prosthetic for human
ingenuity, which if used correctly can help us deal with our modern data
deluge." If used as a replacement for individual intelligence, it tends to
chew up investment without delivering results. We explain, in
particular, how the myth has negatively affected research in
neuroscience, among other recent scientific pursuits. The price we are paying
for the myth is too high. Since we have no good scientific reason to believe
the myth is true, and every reason to reject it for our own future flourishing, we need to rethink the discussion about Al
radically.
Turing had made his
reputation as a mathematician long before he began writing
about A I. In 1936; he published a
short mathematical paper on the precise meaning of "computer," which
at the time referred to a person working through a sequence of steps to get a
definite result (like performing a calculation). In this paper, he
replaced the human-computer with the idea of a machine doing the same work. The
paper ventured into difficult mathematics. But in its treatment of
machines, it did not refer to human thinking or the mind. Machines can run
automatically, Turing said, and the problems they solve do not require any
“external" help or intelligence. This external intelligence,
the human factor, is what mathematicians sometimes call “intuition.”
Turing’s 1956 work on
computing machines helped launch computer science as a discipline and was an
important contribution to mathematical logic. Still, Turing apparently thought
that his early definition missed something essential. In fact, the same idea
of the mind or human faculties assisting problem-solving appeared two years
later in his Ph.D. thesis, a clever but ultimately unsuccessful attempt to
bypass the Austrian-born mathematical logician Kurt's result Gödel.
Though his
language is framed for specialists, Turing is pointing out the obvious:
mathematicians typically select problems or “see" an interesting problem to
work on using some capacity that at least seems indivisible into
steps, and therefore not obviously amenable to computer programming.
Gödel, too, was
thinking about mechanical intelligence. Like Turing, he was obsessed with the
distinction between ingenuity (mechanics) and intuition (mind). His distinction
was essentially the same as Turing's, in a different language: proof versus
truth (or “proof-theory" versus “model-theory" in mathematics lingo).
Are the concepts of proof and truth, Gödel wondered, in the end, the same? If
so, mathematics and even science itself might be understood purely mechanically. Human thinking,
in this view, would be mechanical, too. The concept of AI, though the term
remained to be coined, hovered above the question. Is the mind’s intuition,
ability to grasp truth and meaning reducible to a machine, to computation?
This was Gödel's
question. In answering it, he ran into a snag that would soon make him
world-famous. In 1931, Gödel published two theorems of mathematical logic known
as his incompleteness theorems. In them, he demonstrated the inherent
limitations of all formal mathematical systems. It was a brilliant stroke.
Gödel showed unmistakably that mathematics, all of mathematics, with
certain straightforward assumptions, is, strictly speaking, not mechanical or
for- realizable. More specifically, Gödel proved that there must
exist some statements in any formal (mathematical or computational) system that
are True, with capital-T standing, yet not provable in the system itself using
any of its rules. The True statement can be recognized by a human mind but is
(probably) not provable by the system it's formulated in.
How did Gödel reach
this conclusion? Hie details are complicated and technical, but Gödel’s basic
idea is to treat a mathematical system complicated enough to do addition as a
system of meaning, almost like a natural language such as English or German. The
same applies to all more complicated systems. By treating it this way, we
enable the system to talk about itself. It can say about
itself, for instance, that it has certain limitations. This was Gödel's
insight.
Formal systems like
those in mathematics allow for the precise expression
of truth and falsehood. Typically, we establish a truth by using the
tools of proof, we use rules to prove something, so we know it’s definitely true. But are there true
statements that can't be proven? Can the mind know things the system
cannot? In the simple arithmetic case, we express truths by writing equations
like “2 + 1 = 4." Ordinary equations are true statements in the system of
arithmetic, and they are provable using the rules of arithmetic. Here, provable
equals true. Mathematicians before Gödel thought all of
the mathematics had this property. This implied that machines could crank out
all truths in different mathematical systems by simply applying the rules
correctly. It’s a beautiful idea. It's just not true.
Gödel hit upon the
rare but powerful property of self-reference. Mathematical versions of
self-referring expressions, such as “This statement is not provable in this
system,” can be constructed without breaking the mathematical systems' rules.
But the so-called self-referring “Gödel statements” introduce contradictions
into mathematics: if they are true, they are improvable. If they are false,
then because they say they are improvable, they are actually
true. True means false, and false means true, a contradiction.
Going back to the
concept of intuition, we humans can see that the Gödel statement is, in fact,
true, but because of Gödel's result, we also know that the rules of the system
can’t prove it, the system is in effect blind to something not covered by its rules.
Truth and provability pull apart. Perhaps mind and
machine do, as well. The purely formal system has limits, at any rate. It
cannot prove in its own language something true. In other words, we can see
something that the computer cannot.
Gödel’s result dealt
a massive blow to a popular idea at the time that all of mathematics could be
converted into rule-based operations, cranking out
mathematical truths one by one. The Zeitgeist was formalism,
not talk of minds, spirits, souls, and the like. The formalist movement in
mathematics signaled a broader turn by intellectuals toward scientific
materialism, and in particular, logical positivism, a movement dedicated to
eradicating traditional metaphysics tike Platonism, with its abstract Forms that
couldn't be observed with the senses, and traditional notions in religion like
the existence of God. The world was turning to the idea of precision machines,
in effect. And no one took up the formalist cause as vigorously as the German
mathematician David
Hilbert.
At the outset of the
twentieth century (before Gödel), David Hilbert had issued a challenge to the
mathematical world: show that all of the mathematics
rested on a secure foundation. Hilbert's worry was understandable. If the
purely formal rules of mathematics can’t prove any truths,
it's at least theoretically possible for mathematics to disguise contradictions
and nonsense. A contradiction buried somewhere in
mathematics ruins everything because, from a contradiction, anything can be
proven. Formalism then becomes useless.
Hilbert expressed all
formalists' dream to prove finally that mathematics is
a closed system governed only by rules. Truth is just "proof.’’
We acquire knowledge by simply tracing the "code” of proof and confirming
no rules were violated. The larger dream, thinly disguised, was really a
worldview, a picture of the universe as itself a mechanism.
Condensing the history of the AI myth
The story of
artificial intelligence starts with the ideas of someone who had
immense human intelligence: the computer pioneer Alan Turing.
In 1950 Turing
published a provocative paper, “Computing Machinery
and Intelligence," about the possibility of intelligent
machines. The paper was bold, coming when computers were new and
unimpressive by today’s standards. Slow, heavy pieces of hardware sped up
scientific calculations like code-breaking. After much
preparation, they could be fed physical equations and initial conditions and
crank out the radius of a nuclear blast. IBM quickly grasped their potential
for replacing humans doing calculations for businesses, like updating
spreadsheets. But viewing computers as "thinking" took
imagination.
Turing’s proposal was
based on a popular entertainment called the "imitation game," In the
original game, a man and a woman are hidden from view. A third person, the
interrogator, relays questions to one of them at a time and, by reading the answers,
attempts to determine which is the man and which the woman. 'The twist is that
the man has to try to deceive the interrogator while
the woman tries to assist him, making replies from either side suspect. Turing
replaced the man and woman with a computer and a human. Titus began what we now
call the Turing test: a computer and a human receive typed questions from
a human judge, and if the judge can't accurately identify which is the
computer, the computer wins. Turing argued that with such an outcome, we have
no good reason to define the machine as unintelligent, regardless of whether it
is human or not. Thus, the question of whether a machine has intelligence replaces
the question of whether it can truly think.
The Turing test is actually tough; no computer has ever passed it. Turing, of
course, didn't know this long-term result in 1950; however, by replacing pesky
philosophical questions about “consciousness” and "thinking” with a test
of observable output, he encouraged the view of AI as a legitimate science with
a well-defined aim.
In
the course of its short existence,
AI has undergone many changes. These can be summarized in six stages.
First
of all, in the euphoria of AI’s origins and
early successes, the researchers had given free rein to their imagination,
indulging in certain reckless pronouncements for which they were heavily
criticized later. For instance, in 1958, American political scientist and
economist Herbert A. Simon, who received the Nobel Prize in Economic Sciences
in 1978 – had declared that, within ten years, machines would become world
chess champions if they were not barred from international competitions.
By the mid-1960s,
progress seemed to be slow in coming. A 10-year-old child beat a computer at a
chess game in 1965, and a report commissioned by the US Senate in 1966
described the intrinsic limitations of machine translation. AI got bad press
for about a decade.
The work went on
nevertheless, but the research was given a new direction. It focused on the
psychology of memory and the mechanisms of understanding, with attempts to
simulate these on computers – and the role of knowledge in reasoning. This gave
rise to techniques for the semantic
representation of knowledge,
which developed considerably in the mid-1970s and led to expert systems
development, so-called because they use skilled specialists to reproduce their
thought processes. Expert systems raised enormous hopes in the early 1980s with
many applications, including medical diagnosis.
Technical
improvements led to the development of machine learning algorithms, which
allowed computers to accumulate knowledge and automatically reprogram
themselves using their own experiences.
This led to
industrial applications' development (fingerprint identification, speech
recognition, etc.), where AI, computer science, artificial
life, and other
disciplines were combined to produce hybrid systems.
Starting in the late 1990s,
AI was coupled with robotics and human-machine interfaces to produce
intelligent agents that suggested the presence of feelings and emotions. This
gave rise, among other things, to the calculation of emotions (affective
computing), which evaluates the reactions of subject feeling emotions and
reproduces them on a machine, and especially to the development of
conversational agents (chatbots).
Since 2010, machines'
power has made it possible to exploit enormous data quantities (big data) with deep learning techniques based on formal neural networks. A
range of very successful applications in several areas, including speech and
image recognition, natural language comprehension, and autonomous cars, leads
to an AI renaissance.
Some of the applications
Many achievements
using AI techniques surpass human capabilities, in 1997, a computer program
defeated the reigning world chess champion. More recently, in 2016, other
computer programs have beaten the world’s best Go [an ancient Chinese board
game] players and some top poker players. Computers are proving or helping to
prove mathematical theorems; knowledge is being automatically constructed from
huge masses of data, in terabytes (1012 bytes), or even petabytes
(1015 bytes), using machine learning techniques.
As a result, machines
can recognize speech and transcribe it, just like typists did in the past.
Computers can accurately identify faces or fingerprints from tens of millions
or understand texts written in natural languages. Using machine learning techniques,
cars drive themselves; machines are better than dermatologists at diagnosing
melanomas using photographs of skin moles taken with mobile phone
cameras; robots are fighting wars instead of humans, and factory production lines
are becoming increasingly automated.
Scientists are also
using AI techniques to determine certain biological macromolecules' function,
especially proteins and genomes, from their constituents' sequences, amino
acids for proteins, bases for genomes. All the sciences are undergoing a major
epistemological rupture with in silico experiments.
They are named so because they are carried out by computers from massive
quantities of data, using powerful processors whose cores are silicon. In this
way, they differ from in vivo experiments performed on
living matter, and above all, from in vitro experiments carried out
in glass test tubes.
Today, AI
applications affect almost all activity fields, particularly in banking,
insurance, health, and defense sectors. Several routine tasks are now
automated, transforming many trades and eventually eliminating some.
The point is that AI
requires improvements in historically confounding areas, like the link between
the brain and intelligence. Progress in areas with historical growth (such as
processing power and neural network size) does not guarantee ultra-intelligence
at all.
Betting against a
catastrophic end to the world has been right so far. On the other hand is Elan Musk, Kurzweil, et al. and they have some
impressive street cred. After all, Kurzweil rightly
predicted the successful timeline of the human genome project, and not many
would bet against Musk’s engineering teams achieve.
For updates click homepage here