Menu
Home Explore People Places Arts History Plants & Animals Science Life & Culture Technology
On this page

The history of artificial intelligence (AI) began in antiquity with stories of artificial beings, evolving through the invention of the programmable digital computer in the 1940s. The AI field was formally founded at a 1956 Dartmouth workshop, supported by the U.S. government. After setbacks in the 1970s and an AI winter in the 1990s, research revived with advances in machine learning and deep learning. The 2017 introduction of the transformer architecture sparked a new AI boom, producing powerful large language models such as ChatGPT. While AI promises great innovation, it also raises significant ethical concerns for society’s future.

Precursors

Mythical, fictional, and speculative precursors

Myth and legend

In Greek mythology, Talos was a creature made of bronze who acted as guardian for the island of Crete. He would throw boulders at the ships of invaders and would complete 3 circuits around the island's perimeter daily.4 According to pseudo-Apollodorus' Bibliotheke, Hephaestus forged Talos with the aid of a cyclops and presented the automaton as a gift to Minos.5 In the Argonautica, Jason and the Argonauts defeated Talos by removing a plug near his foot, causing the vital ichor to flow out from his body and rendering him lifeless.6

Pygmalion was a legendary king and sculptor of Greek mythology, famously represented in Ovid's Metamorphoses. In the 10th book of Ovid's narrative poem, Pygmalion becomes disgusted with women when he witnesses the way in which the Propoetides prostitute themselves. Despite this, he makes offerings at the temple of Venus asking the goddess to bring to him a woman just like a statue he carved.7

Medieval legends of artificial beings

In Of the Nature of Things, the Swiss alchemist Paracelsus describes a procedure that he claims can fabricate an "artificial man". By placing the "sperm of a man" in horse dung, and feeding it the "Arcanum of Mans blood" after 40 days, the concoction will become a living infant.8

The earliest written account regarding golem-making is found in the writings of Eleazar ben Judah of Worms in the early 13th century.9 During the Middle Ages, it was believed that the animation of a Golem could be achieved by insertion of a piece of paper with any of God's names on it, into the mouth of the clay figure.10 Unlike legendary automata like Brazen Heads,11 a Golem was unable to speak.12

Takwin, the artificial creation of life, was a frequent topic of Ismaili alchemical manuscripts, especially those attributed to Jabir ibn Hayyan. Islamic alchemists attempted to create a broad range of life through their work, ranging from plants to animals.13

In Faust: The Second Part of the Tragedy by Johann Wolfgang von Goethe, an alchemically fabricated homunculus, destined to live forever in the flask in which he was made, endeavors to be born into a full human body. Upon the initiation of this transformation, however, the flask shatters and the homunculus dies.14

Modern fiction

Main article: Artificial intelligence in fiction

By the 19th century, ideas about artificial men and thinking machines became a popular theme in fiction. Notable works like Mary Shelley's Frankenstein and Karel Čapek's R.U.R. (Rossum's Universal Robots)15 explored the concept of artificial life. Speculative essays, such as Samuel Butler's "Darwin among the Machines",16 and Edgar Allan Poe's "Maelzel's Chess Player"17 reflected society's growing interest in machines with artificial intelligence. AI remains a common topic in science fiction today.18

Automata

Main article: Automaton

Realistic humanoid automata were built by craftsman from many civilizations, including Yan Shi,19 Hero of Alexandria,20 Al-Jazari,21 Haroun al-Rashid,22 Jacques de Vaucanson,2324 Leonardo Torres y Quevedo,25 Pierre Jaquet-Droz and Wolfgang von Kempelen.2627

The oldest known automata were the sacred statues of ancient Egypt and Greece.2829 The faithful believed that craftsman had imbued these figures with very real minds, capable of wisdom and emotion—Hermes Trismegistus wrote that "by discovering the true nature of the gods, man has been able to reproduce it".30 English scholar Alexander Neckham asserted that the Ancient Roman poet Virgil had built a palace with automaton statues.31

During the early modern period, these legendary automata were said to possess the magical ability to answer questions put to them. The late medieval alchemist and proto-Protestant Roger Bacon was purported to have fabricated a brazen head, having developed a legend of having been a wizard.3233 These legends were similar to the Norse myth of the Head of Mímir. According to legend, Mímir was known for his intellect and wisdom, and was beheaded in the Æsir-Vanir War. Odin is said to have "embalmed" the head with herbs and spoke incantations over it such that Mímir's head remained able to speak wisdom to Odin. Odin then kept the head near him for counsel.34

Formal reasoning

Artificial intelligence is based on the assumption that the process of human thought can be mechanized. The study of mechanical—or "formal"—reasoning has a long history. Chinese, Indian and Greek philosophers all developed structured methods of formal deduction by the first millennium BCE. Their ideas were developed over the centuries by philosophers such as Aristotle (who gave a formal analysis of the syllogism),35 Euclid (whose Elements was a model of formal reasoning), al-Khwārizmī (who developed algebra and gave his name to the word algorithm) and European scholastic philosophers such as William of Ockham and Duns Scotus.36

Spanish philosopher Ramon Llull (1232–1315) developed several logical machines devoted to the production of knowledge by logical means;3738 Llull described his machines as mechanical entities that could combine basic and undeniable truths by simple logical operations, produced by the machine by mechanical meanings, in such ways as to produce all the possible knowledge.39 Llull's work had a great influence on Gottfried Leibniz, who redeveloped his ideas.40

In the 17th century, Leibniz, Thomas Hobbes and René Descartes explored the possibility that all rational thought could be made as systematic as algebra or geometry.41 Hobbes famously wrote in Leviathan: "For reason ... is nothing but reckoning, that is adding and subtracting".42 Leibniz envisioned a universal language of reasoning, the characteristica universalis, which would reduce argumentation to calculation so that "there would be no more need of disputation between two philosophers than between two accountants. For it would suffice to take their pencils in hand, down to their slates, and to say each other (with a friend as witness, if they liked): Let us calculate."43 These philosophers had begun to articulate the physical symbol system hypothesis that would become the guiding faith of AI research.

The study of mathematical logic provided the essential breakthrough that made artificial intelligence seem plausible. The foundations had been set by such works as Boole's The Laws of Thought and Frege's Begriffsschrift.44 Building on Frege's system, Russell and Whitehead presented a formal treatment of the foundations of mathematics in their masterpiece, the Principia Mathematica in 1913. Inspired by Russell's success, David Hilbert challenged mathematicians of the 1920s and 30s to answer this fundamental question: "can all of mathematical reasoning be formalized?"45 His question was answered by Gödel's incompleteness proof,46 Turing's machine47 and Church's Lambda calculus.48

Their answer was surprising in two ways. First, they proved that there were, in fact, limits to what mathematical logic could accomplish. But second (and more important for AI) their work suggested that, within these limits, any form of mathematical reasoning could be mechanized. The Church-Turing thesis implied that a mechanical device, shuffling symbols as simple as 0 and 1, could imitate any conceivable process of mathematical deduction.49 The key insight was the Turing machine—a simple theoretical construct that captured the essence of abstract symbol manipulation.50 This invention would inspire a handful of scientists to begin discussing the possibility of thinking machines.

Computer science

Main articles: History of computer hardware and History of computer science

Calculating machines were designed or built in antiquity and throughout history by many people, including Gottfried Leibniz,5152 Joseph Marie Jacquard,53 Charles Babbage,5455 Percy Ludgate,56 Leonardo Torres Quevedo,57 Vannevar Bush,58 and others. Ada Lovelace speculated that Babbage's machine was "a thinking or ... reasoning machine", but warned "It is desirable to guard against the possibility of exaggerated ideas that arise as to the powers" of the machine.5960

The first modern computers were the massive machines of the Second World War (such as Konrad Zuse's Z3, Alan Turing's Heath Robinson and Colossus, Atanasoff and Berry's ABC and ENIAC at the University of Pennsylvania).61 ENIAC was based on the theoretical foundation laid by Alan Turing and developed by John von Neumann,62 and proved to be the most influential.63

Birth of artificial intelligence (1941-56)

The earliest research into thinking machines was inspired by a confluence of ideas that became prevalent in the late 1930s, 1940s, and early 1950s. Recent research in neurology had shown that the brain was an electrical network of neurons that fired in all-or-nothing pulses. Norbert Wiener's cybernetics described control and stability in electrical networks. Claude Shannon's information theory described digital signals (i.e., all-or-nothing signals). Alan Turing's theory of computation showed that any form of computation could be described digitally. The close relationship between these ideas suggested that it might be possible to construct an "electronic brain".

In the 1940s and 50s, a handful of scientists from a variety of fields (mathematics, psychology, engineering, economics and political science) explored several research directions that would be vital to later AI research.64 Alan Turing was among the first people to seriously investigate the theoretical possibility of "machine intelligence".65 The field of "artificial intelligence research" was founded as an academic discipline in 1956.66

Turing Test

Main article: Turing test

In 1950 Turing published a landmark paper "Computing Machinery and Intelligence", in which he speculated about the possibility of creating machines that think.6768 In the paper, he noted that "thinking" is difficult to define and devised his famous Turing Test: If a machine could carry on a conversation (over a teleprinter) that was indistinguishable from a conversation with a human being, then it was reasonable to say that the machine was "thinking".69 This simplified version of the problem allowed Turing to argue convincingly that a "thinking machine" was at least plausible and the paper answered all the most common objections to the proposition.70 The Turing Test was the first serious proposal in the philosophy of artificial intelligence.

Neuroscience and Hebbian theory

Donald Hebb was a Canadian psychologist whose work laid the foundation for modern neuroscience, particularly in understanding learning, memory, and neural plasticity. His most influential book, The Organization of Behavior (1949), introduced the concept of Hebbian learning, often summarized as "cells that fire together wire together." 71

Hebb began formulating the foundational ideas for this book in the early 1940s, particularly during his time at the Yerkes Laboratories of Primate Biology from 1942 to 1947. He made extensive notes between June 1944 and March 1945 and sent a complete draft to his mentor Karl Lashley in 1946. The manuscript for The Organization of Behavior wasn’t published until 1949. The delay was due to various factors, including World War II and shifts in academic focus. By the time it was published, several of his peers had already published related ideas, making Hebb’s work seem less groundbreaking at first glance. However, his synthesis of psychological and neurophysiological principles became a cornerstone of neuroscience and machine learning. 72 73

Artificial neural networks

Walter Pitts and Warren McCulloch analyzed networks of idealized artificial neurons and showed how they might perform simple logical functions in 1943. They were the first to describe what later researchers would call a neural network.74 The paper was influenced by Turing's paper 'On Computable Numbers' from 1936 using similar two-state boolean 'neurons', but was the first to apply it to neuronal function.75 One of the students inspired by Pitts and McCulloch was Marvin Minsky who was a 24-year-old graduate student at the time. In 1951 Minsky and Dean Edmonds built the first neural net machine, the SNARC.76 Minsky would later become one of the most important leaders and innovators in AI.

Cybernetic robots

Experimental robots such as W. Grey Walter's turtles and the Johns Hopkins Beast, were built in the 1950s. These machines did not use computers, digital electronics or symbolic reasoning; they were controlled entirely by analog circuitry.77

Game AI

In 1951, using the Ferranti Mark 1 machine of the University of Manchester, Christopher Strachey wrote a checkers program78 and Dietrich Prinz wrote one for chess.79 Arthur Samuel's checkers program, the subject of his 1959 paper "Some Studies in Machine Learning Using the Game of Checkers", eventually achieved sufficient skill to challenge a respectable amateur.80 Samuel's program was among the first uses of what would later be called machine learning.81 Game AI would continue to be used as a measure of progress in AI throughout its history.

Symbolic reasoning and the Logic Theorist

Main article: Logic Theorist

When access to digital computers became possible in the mid-fifties, a few scientists instinctively recognized that a machine that could manipulate numbers could also manipulate symbols and that the manipulation of symbols could well be the essence of human thought. This was a new approach to creating thinking machines.8283

In 1955, Allen Newell and future Nobel Laureate Herbert A. Simon created the "Logic Theorist", with help from J. C. Shaw. The program would eventually prove 38 of the first 52 theorems in Russell and Whitehead's Principia Mathematica, and find new and more elegant proofs for some.84 Simon said that they had "solved the venerable mind/body problem, explaining how a system composed of matter can have the properties of mind."8586 The symbolic reasoning paradigm they introduced would dominate AI research and funding until the middle 90s, as well as inspire the cognitive revolution.

Dartmouth Workshop

Main article: Dartmouth workshop

The Dartmouth workshop of 1956 was a pivotal event that marked the formal inception of AI as an academic discipline.87 It was organized by Marvin Minsky and John McCarthy, with the support of two senior scientists Claude Shannon and Nathan Rochester of IBM. The proposal for the conference stated they intended to test the assertion that "every aspect of learning or any other feature of intelligence can be so precisely described that a machine can be made to simulate it".8889 The term "Artificial Intelligence" was introduced by John McCarthy at the workshop.90 The participants included Ray Solomonoff, Oliver Selfridge, Trenchard More, Arthur Samuel, Allen Newell and Herbert A. Simon, all of whom would create important programs during the first decades of AI research.9192 At the workshop Newell and Simon debuted the "Logic Theorist".93 The workshop was the moment that AI gained its name, its mission, its first major success and its key players, and is widely considered the birth of AI.94

Cognitive revolution

Main article: cognitive revolution

In the autumn of 1956, Newell and Simon also presented the Logic Theorist at a meeting of the Special Interest Group in Information Theory at the Massachusetts Institute of Technology (MIT). At the same meeting, Noam Chomsky discussed his generative grammar, and George Miller described his landmark paper "The Magical Number Seven, Plus or Minus Two". Miller wrote "I left the symposium with a conviction, more intuitive than rational, that experimental psychology, theoretical linguistics, and the computer simulation of cognitive processes were all pieces from a larger whole."9596

This meeting was the beginning of the "cognitive revolution"—an interdisciplinary paradigm shift in psychology, philosophy, computer science and neuroscience. It inspired the creation of the sub-fields of symbolic artificial intelligence, generative linguistics, cognitive science, cognitive psychology, cognitive neuroscience and the philosophical schools of computationalism and functionalism. All these fields used related tools to model the mind and results discovered in one field were relevant to the others.

The cognitive approach allowed researchers to consider "mental objects" like thoughts, plans, goals, facts or memories, often analyzed using high level symbols in functional networks. These objects had been forbidden as "unobservable" by earlier paradigms such as behaviorism.97 Symbolic mental objects would become the major focus of AI research and funding for the next several decades.

Early successes (1956-1974)

The programs developed in the years after the Dartmouth Workshop were, to most people, simply "astonishing":98 computers were solving algebra word problems, proving theorems in geometry and learning to speak English. Few at the time would have believed that such "intelligent" behavior by machines was possible at all.99100101 Researchers expressed an intense optimism in private and in print, predicting that a fully intelligent machine would be built in less than 20 years.102 Government agencies like the Defense Advanced Research Projects Agency (DARPA, then known as "ARPA") poured money into the field.103 Artificial Intelligence laboratories were set up at a number of British and US universities in the latter 1950s and early 1960s.104

Approaches

There were many successful programs and new directions in the late 50s and 1960s. Among the most influential were these:

Reasoning, planning and problem solving as search

Many early AI programs used the same basic algorithm. To achieve some goal (like winning a game or proving a theorem), they proceeded step by step towards it (by making a move or a deduction) as if searching through a maze, backtracking whenever they reached a dead end.105 The principal difficulty was that, for many problems, the number of possible paths through the "maze" was astronomical (a situation known as a "combinatorial explosion"). Researchers would reduce the search space by using heuristics that would eliminate paths that were unlikely to lead to a solution.106

Newell and Simon tried to capture a general version of this algorithm in a program called the "General Problem Solver".107108 Other "searching" programs were able to accomplish impressive tasks like solving problems in geometry and algebra, such as Herbert Gelernter's Geometry Theorem Prover (1958)109 and Symbolic Automatic Integrator (SAINT), written by Minsky's student James Slagle in 1961.110111 Other programs searched through goals and subgoals to plan actions, like the STRIPS system developed at Stanford to control the behavior of the robot Shakey.112

Natural language

An important goal of AI research is to allow computers to communicate in natural languages like English. An early success was Daniel Bobrow's program STUDENT, which could solve high school algebra word problems.113

A semantic net represents concepts (e.g. "house", "door") as nodes, and relations among concepts as links between the nodes (e.g. "has-a"). The first AI program to use a semantic net was written by Ross Quillian114 and the most successful (and controversial) version was Roger Schank's Conceptual dependency theory.115

Joseph Weizenbaum's ELIZA could carry out conversations that were so realistic that users occasionally were fooled into thinking they were communicating with a human being and not a computer program (see ELIZA effect). But in fact, ELIZA simply gave a canned response or repeated back what was said to it, rephrasing its response with a few grammar rules. ELIZA was the first chatbot.116117

Micro-worlds

In the late 60s, Marvin Minsky and Seymour Papert of the MIT AI Laboratory proposed that AI research should focus on artificially simple situations known as micro-worlds.118 They pointed out that in successful sciences like physics, basic principles were often best understood using simplified models like frictionless planes or perfectly rigid bodies. Much of the research focused on a "blocks world," which consists of colored blocks of various shapes and sizes arrayed on a flat surface.119

This paradigm led to innovative work in machine vision by Gerald Sussman, Adolfo Guzman, David Waltz (who invented "constraint propagation"), and especially Patrick Winston. At the same time, Minsky and Papert built a robot arm that could stack blocks, bringing the blocks world to life. Terry Winograd's SHRDLU could communicate in ordinary English sentences about the micro-world, plan operations and execute them.120

Perceptrons and early neural networks

Main article: Perceptron

In the 1960s funding was primarily directed towards laboratories researching symbolic AI, however several people still pursued research in neural networks.

The perceptron, a single-layer neural network was introduced in 1958 by Frank Rosenblatt121 (who had been a schoolmate of Marvin Minsky at the Bronx High School of Science).122 Like most AI researchers, he was optimistic about their power, predicting that a perceptron "may eventually be able to learn, make decisions, and translate languages."123 Rosenblatt was primarily funded by Office of Naval Research.124

Bernard Widrow and his student Ted Hoff built ADALINE (1960) and MADALINE (1962), which had up to 1000 adjustable weights.125126 A group at Stanford Research Institute led by Charles A. Rosen and Alfred E. (Ted) Brain built two neural network machines named MINOS I (1960) and II (1963), mainly funded by U.S. Army Signal Corps. MINOS II127 had 6600 adjustable weights,128 and was controlled with an SDS 910 computer in a configuration named MINOS III (1968), which could classify symbols on army maps, and recognize hand-printed characters on Fortran coding sheets.129130 Most of neural network research during this early period involved building and using bespoke hardware, rather than simulation on digital computers.131

However, partly due to lack of results and partly due to competition from symbolic AI research, the MINOS project ran out of funding in 1966. Rosenblatt failed to secure continued funding in the 1960s.132 In 1969, research came to a sudden halt with the publication of Minsky and Papert's 1969 book Perceptrons.133 It suggested that there were severe limitations to what perceptrons could do and that Rosenblatt's predictions had been grossly exaggerated. The effect of the book was that virtually no research was funded in connectionism for 10 years.134 The competition for government funding ended with the victory of symbolic AI approaches over neural networks.135136

Minsky (who had worked on SNARC) became a staunch objector to pure connectionist AI. Widrow (who had worked on ADALINE) turned to adaptive signal processing. The SRI group (which worked on MINOS) turned to symbolic AI and robotics.137138

The main problem was the inability to train multilayered networks (versions of backpropagation had already been used in other fields but it was unknown to these researchers).139140 The AI community became aware of backpropogation in the 80s,141 and, in the 21st century, neural networks would become enormously successful, fulfilling all of Rosenblatt's optimistic predictions. Rosenblatt did not live to see this, however, as he died in a boating accident in 1971.142

Optimism

The first generation of AI researchers made these predictions about their work:

  • 1958, H. A. Simon and Allen Newell: "within ten years a digital computer will be the world's chess champion" and "within ten years a digital computer will discover and prove an important new mathematical theorem."143
  • 1965, H. A. Simon: "machines will be capable, within twenty years, of doing any work a man can do."144
  • 1967, Marvin Minsky: "Within a generation... the problem of creating 'artificial intelligence' will substantially be solved."145
  • 1970, Marvin Minsky (in Life magazine): "In from three to eight years we will have a machine with the general intelligence of an average human being."146147

Financing

In June 1963, MIT received a $2.2 million grant from the newly created Advanced Research Projects Agency (ARPA, later known as DARPA). The money was used to fund project MAC which subsumed the "AI Group" founded by Minsky and McCarthy five years earlier. DARPA continued to provide $3 million each year until the 70s.148 DARPA made similar grants to Newell and Simon's program at Carnegie Mellon University and to Stanford University's AI Lab, founded by John McCarthy in 1963.149 Another important AI laboratory was established at Edinburgh University by Donald Michie in 1965.150 These four institutions would continue to be the main centers of AI research and funding in academia for many years.151152

The money was given with few strings attached: J. C. R. Licklider, then the director of ARPA, believed that his organization should "fund people, not projects!" and allowed researchers to pursue whatever directions might interest them.153 This created a freewheeling atmosphere at MIT that gave birth to the hacker culture,154 but this "hands off" approach did not last.

First AI Winter (1974–1980)

In the 1970s, AI was subject to critiques and financial setbacks. AI researchers had failed to appreciate the difficulty of the problems they faced. Their tremendous optimism had raised public expectations impossibly high, and when the promised results failed to materialize, funding targeted at AI was severely reduced.155 The lack of success indicated the techniques being used by AI researchers at the time were insufficient to achieve their goals.156157

These setbacks did not affect the growth and progress of the field, however. The funding cuts only impacted a handful of major laboratories158 and the critiques were largely ignored.159 General public interest in the field continued to grow,160 the number of researchers increased dramatically,161 and new ideas were explored in logic programming, commonsense reasoning and many other areas. Historian Thomas Haigh argued in 2023 that there was no winter,162 and AI researcher Nils Nilsson described this period as the most "exciting" time to work in AI.163

Problems

In the early seventies, the capabilities of AI programs were limited. Even the most impressive could only handle trivial versions of the problems they were supposed to solve;164 all the programs were, in some sense, "toys".165 AI researchers had begun to run into several limits that would be only conquered decades later, and others that still stymie the field in the 2020s:

  • Limited computer power: There was not enough memory or processing speed to accomplish anything truly useful.166 For example: Ross Quillian's successful work on natural language was demonstrated with a vocabulary of only 20 words, because that was all that would fit in memory.167 Hans Moravec argued in 1976 that computers were still millions of times too weak to exhibit intelligence. He suggested an analogy: artificial intelligence requires computer power in the same way that aircraft require horsepower. Below a certain threshold, it's impossible, but, as power increases, eventually it could become easy. "With enough horsepower," he wrote, "anything will fly".168169
  • Intractability and the combinatorial explosion: In 1972 Richard Karp (building on Stephen Cook's 1971 theorem) showed there are many problems that can only be solved in exponential time. Finding optimal solutions to these problems requires extraordinary amounts of computer time, except when the problems are trivial. This limitation applied to all symbolic AI programs that used search trees and meant that many of the "toy" solutions used by AI would never scale to useful systems.170171
  • Moravec's paradox: Early AI research had been very successful at getting computers to do "intelligent" tasks like proving theorems, solving geometry problems and playing chess. Their success at these intelligent tasks convinced them that the problem of intelligent behavior had been largely solved.172173 However, they utterly failed to make progress on "unintelligent" tasks like recognizing a face or crossing a room without bumping into anything.174175 By the 1980s, researchers would realize that symbolic reasoning was utterly unsuited for these perceptual and sensorimotor tasks and that there were limits to this approach.176
  • The breadth of commonsense knowledge: Many important artificial intelligence applications like vision or natural language require enormous amounts of information about the world: the program needs to have some idea of what it might be looking at or what it is talking about. This requires that the program know most of the same things about the world that a child does. Researchers soon discovered that this was a vast amount of information with billions of atomic facts. No one in 1970 could build a database large enough and no one knew how a program might learn so much information.177
  • Representing commonsense reasoning: A number of related problems178 appeared when researchers tried to represent commonsense reasoning using formal logic or symbols. Descriptions of very ordinary deductions tended to get longer and longer the more one worked on them, as more and more exceptions, clarifications and distinctions were required.179 However, when people thought about ordinary concepts they did not rely on precise definitions, rather they seemed to make hundreds of imprecise assumptions, correcting them when necessary using their entire body of commonsense knowledge. Gerald Sussman observed that "using precise language to describe essentially imprecise concepts doesn't make them any more precise."180

Decrease in funding

See also: AI winter

The agencies which funded AI research, such as the British government, DARPA and the National Research Council (NRC) became frustrated with the lack of progress and eventually cut off almost all funding for undirected AI research. The pattern began in 1966 when the Automatic Language Processing Advisory Committee (ALPAC) report criticized machine translation efforts. After spending $20 million, the NRC ended all support.181 In 1973, the Lighthill report on the state of AI research in the UK criticized the failure of AI to achieve its "grandiose objectives" and led to the dismantling of AI research in that country.182 (The report specifically mentioned the combinatorial explosion problem as a reason for AI's failings.)183184185 DARPA was deeply disappointed with researchers working on the Speech Understanding Research program at CMU and canceled an annual grant of $3 million.186187

Hans Moravec blamed the crisis on the unrealistic predictions of his colleagues. "Many researchers were caught up in a web of increasing exaggeration."188189 However, there was another issue: since the passage of the Mansfield Amendment in 1969, DARPA had been under increasing pressure to fund "mission-oriented direct research, rather than basic undirected research". Funding for the creative, freewheeling exploration that had gone on in the 60s would not come from DARPA, which instead directed money at specific projects with clear objectives, such as autonomous tanks and battle management systems.190191

The major laboratories (MIT, Stanford, CMU and Edinburgh) had been receiving generous support from their governments, and when it was withdrawn, these were the only places that were seriously impacted by the budget cuts. The thousands of researchers outside these institutions and the many more thousands that were joining the field were unaffected.192

Philosophical and ethical critiques

See also: Philosophy of artificial intelligence

Several philosophers had strong objections to the claims being made by AI researchers. One of the earliest was John Lucas, who argued that Gödel's incompleteness theorem showed that a formal system (such as a computer program) could never see the truth of certain statements, while a human being could.193 Hubert Dreyfus ridiculed the broken promises of the 1960s and critiqued the assumptions of AI, arguing that human reasoning actually involved very little "symbol processing" and a great deal of embodied, instinctive, unconscious "know how".194195 John Searle's Chinese Room argument, presented in 1980, attempted to show that a program could not be said to "understand" the symbols that it uses (a quality called "intentionality"). If the symbols have no meaning for the machine, Searle argued, then the machine can not be described as "thinking".196

These critiques were not taken seriously by AI researchers. Problems like intractability and commonsense knowledge seemed much more immediate and serious. It was unclear what difference "know how" or "intentionality" made to an actual computer program. MIT's Minsky said of Dreyfus and Searle "they misunderstand, and should be ignored."197 Dreyfus, who also taught at MIT, was given a cold shoulder: he later said that AI researchers "dared not be seen having lunch with me."198 Joseph Weizenbaum, the author of ELIZA, was also an outspoken critic of Dreyfus' positions, but he "deliberately made it plain that [his AI colleagues' treatment of Dreyfus] was not the way to treat a human being,"199 and was unprofessional and childish.200

Weizenbaum began to have serious ethical doubts about AI when Kenneth Colby wrote a "computer program which can conduct psychotherapeutic dialogue" based on ELIZA.201202203 Weizenbaum was disturbed that Colby saw a mindless program as a serious therapeutic tool. A feud began, and the situation was not helped when Colby did not credit Weizenbaum for his contribution to the program. In 1976, Weizenbaum published Computer Power and Human Reason which argued that the misuse of artificial intelligence has the potential to devalue human life.204

Logic at Stanford, CMU and Edinburgh

Logic was introduced into AI research as early as 1958, by John McCarthy in his Advice Taker proposal.205206 In 1963, J. Alan Robinson had discovered a simple method to implement deduction on computers, the resolution and unification algorithm.207 However, straightforward implementations, like those attempted by McCarthy and his students in the late 1960s, were especially intractable: the programs required astronomical numbers of steps to prove simple theorems.208209 A more fruitful approach to logic was developed in the 1970s by Robert Kowalski at the University of Edinburgh, and soon this led to the collaboration with French researchers Alain Colmerauer and Philippe Roussel [fr] who created the successful logic programming language Prolog.210 Prolog uses a subset of logic (Horn clauses, closely related to "rules" and "production rules") that permit tractable computation. Rules would continue to be influential, providing a foundation for Edward Feigenbaum's expert systems and the continuing work by Allen Newell and Herbert A. Simon that would lead to Soar and their unified theories of cognition.211

Critics of the logical approach noted, as Dreyfus had, that human beings rarely used logic when they solved problems. Experiments by psychologists like Peter Wason, Eleanor Rosch, Amos Tversky, Daniel Kahneman and others provided proof.212 McCarthy responded that what people do is irrelevant. He argued that what is really needed are machines that can solve problems—not machines that think as people do.213

MIT's "anti-logic" approach

Among the critics of McCarthy's approach were his colleagues across the country at MIT. Marvin Minsky, Seymour Papert and Roger Schank were trying to solve problems like "story understanding" and "object recognition" that required a machine to think like a person. In order to use ordinary concepts like "chair" or "restaurant" they had to make all the same illogical assumptions that people normally made. Unfortunately, imprecise concepts like these are hard to represent in logic. MIT chose instead to focus on writing programs that solved a given task without using high-level abstract definitions or general theories of cognition, and measured performance by iterative testing, rather than arguments from first principles. Schank described their "anti-logic" approaches as scruffy, as opposed to the neat paradigm used by McCarthy, Kowalski, Feigenbaum, Newell and Simon.214215

In 1975, in a seminal paper, Minsky noted that many of his fellow researchers were using the same kind of tool: a framework that captures all our common sense assumptions about something. For example, if we use the concept of a bird, there is a constellation of facts that immediately come to mind: we might assume that it flies, eats worms and so on (none of which are true for all birds). Minsky associated these assumptions with the general category and they could be inherited by the frames for subcategories and individuals, or over-ridden as necessary. He called these structures frames. Schank used a version of frames he called "scripts" to successfully answer questions about short stories in English.216 Frames would eventually be widely used in software engineering under the name object-oriented programming.

The logicians rose to the challenge. Pat Hayes claimed that "most of 'frames' is just a new syntax for parts of first-order logic." But he noted that "there are one or two apparently minor details which give a lot of trouble, however, especially defaults".217

Ray Reiter admitted that "conventional logics, such as first-order logic, lack the expressive power to adequately represent the knowledge required for reasoning by default".218 He proposed augmenting first-order logic with a closed world assumption that a conclusion holds (by default) if its contrary cannot be shown. He showed how such an assumption corresponds to the common sense assumption made in reasoning with frames. He also showed that it has its "procedural equivalent" as negation as failure in Prolog. The closed world assumption, as formulated by Reiter, "is not a first-order notion. (It is a meta notion.)"219 However, Keith Clark showed that negation as finite failure can be understood as reasoning implicitly with definitions in first-order logic including a unique name assumption that different terms denote different individuals.220

During the late 1970s and throughout the 1980s, a variety of logics and extensions of first-order logic were developed both for negation as failure in logic programming and for default reasoning more generally. Collectively, these logics have become known as non-monotonic logics.

Boom (1980–1987)

In the 1980s, a form of AI program called "expert systems" was adopted by corporations around the world and knowledge became the focus of mainstream AI research. Governments provided substantial funding, such as Japan's fifth generation computer project and the U.S. Strategic Computing Initiative. "Overall, the AI industry boomed from a few million dollars in 1980 to billions of dollars in 1988."221

Expert systems become widely used

An expert system is a program that answers questions or solves problems about a specific domain of knowledge, using logical rules that are derived from the knowledge of experts.222 The earliest examples were developed by Edward Feigenbaum and his students. Dendral, begun in 1965, identified compounds from spectrometer readings.223224 MYCIN, developed in 1972, diagnosed infectious blood diseases.225 They demonstrated the feasibility of the approach.

Expert systems restricted themselves to a small domain of specific knowledge (thus avoiding the commonsense knowledge problem)226 and their simple design made it relatively easy for programs to be built and then modified once they were in place. All in all, the programs proved to be useful: something that AI had not been able to achieve up to this point.227

In 1980, an expert system called R1 was completed at CMU for the Digital Equipment Corporation. It was an enormous success: it was saving the company 40 million dollars annually by 1986.228 Corporations around the world began to develop and deploy expert systems and by 1985 they were spending over a billion dollars on AI, most of it to in-house AI departments.229 An industry grew up to support them, including hardware companies like Symbolics and Lisp Machines and software companies such as IntelliCorp and Aion.230

Government funding increases

In 1981, the Japanese Ministry of International Trade and Industry set aside $850 million for the Fifth generation computer project. Their objectives were to write programs and build machines that could carry on conversations, translate languages, interpret pictures, and reason like human beings.231 Much to the chagrin of scruffies, they initially chose Prolog as the primary computer language for the project.232

Other countries responded with new programs of their own. The UK began the £350 million Alvey project.233 A consortium of American companies formed the Microelectronics and Computer Technology Corporation (or "MCC") to fund large scale projects in AI and information technology.234235 DARPA responded as well, founding the Strategic Computing Initiative and tripling its investment in AI between 1984 and 1988.236237

Knowledge revolution

The power of expert systems came from the expert knowledge they contained. They were part of a new direction in AI research that had been gaining ground throughout the 70s. "AI researchers were beginning to suspect—reluctantly, for it violated the scientific canon of parsimony—that intelligence might very well be based on the ability to use large amounts of diverse knowledge in different ways,"238 writes Pamela McCorduck. "[T]he great lesson from the 1970s was that intelligent behavior depended very much on dealing with knowledge, sometimes quite detailed knowledge, of a domain where a given task lay".239 Knowledge based systems and knowledge engineering became a major focus of AI research in the 1980s.240 It was hoped that vast databases would solve the commonsense knowledge problem and provide the support that commonsense reasoning required.

In the 1980s some researchers attempted to attack the commonsense knowledge problem directly, by creating a massive database that would contain all the mundane facts that the average person knows. Douglas Lenat, who started a database called Cyc, argued that there is no shortcut ― the only way for machines to know the meaning of human concepts is to teach them, one concept at a time, by hand.241

New directions in the 1980s

Although symbolic knowledge representation and logical reasoning produced useful applications in the 80s and received massive amounts of funding, it was still unable to solve problems in perception, robotics, learning and common sense. A small number of scientists and engineers began to doubt that the symbolic approach would ever be sufficient for these tasks and developed other approaches, such as "connectionism", robotics, "soft" computing and reinforcement learning. Nils Nilsson called these approaches "sub-symbolic".

Revival of neural networks: "connectionism"

In 1982, physicist John Hopfield was able to prove that a form of neural network (now called a "Hopfield net") could learn and process information, and provably converges after enough time under any fixed condition. It was a breakthrough, as it was previously thought that nonlinear networks would, in general, evolve chaotically.242 Around the same time, Geoffrey Hinton and David Rumelhart popularized a method for training neural networks called "backpropagation".243 These two developments helped to revive the exploration of artificial neural networks.244245

Neural networks, along with several other similar models, received widespread attention after the 1986 publication of the Parallel Distributed Processing, a two volume collection of papers edited by Rumelhart and psychologist James McClelland. The new field was christened "connectionism" and there was a considerable debate between advocates of symbolic AI and the "connectionists".246 Hinton called symbols the "luminous aether of AI" – that is, an unworkable and misleading model of intelligence.247 This was a direct attack on the principles that inspired the cognitive revolution.

Neural networks started to advance state of the art in some specialist areas such as protein structure prediction. Following pioneering work from Terry Sejnowski,248 cascading multilayer perceptrons such as PhD249 and PsiPred250 reached near-theoretical maximum accuracy in predicting secondary structure.

In 1990, Yann LeCun at Bell Labs used convolutional neural networks to recognize handwritten digits. The system was used widely in 90s, reading zip codes and personal checks. This was the first genuinely useful application of neural networks.251252

Robotics and embodied reason

Main articles: Nouvelle AI, behavior-based AI, situated AI, and embodied cognitive science

Rodney Brooks, Hans Moravec and others argued that, in order to show real intelligence, a machine needs to have a body — it needs to perceive, move, survive and deal with the world.253 Sensorimotor skills are essential to higher level skills such as commonsense reasoning. They can't be efficiently implemented using abstract symbolic reasoning, so AI should solve the problems of perception, mobility, manipulation and survival without using symbolic representation at all. These robotics researchers advocated building intelligence "from the bottom up".254

A precursor to this idea was David Marr, who had come to MIT in the late 1970s from a successful background in theoretical neuroscience to lead the group studying vision. He rejected all symbolic approaches (both McCarthy's logic and Minsky's frames), arguing that AI needed to understand the physical machinery of vision from the bottom up before any symbolic processing took place. (Marr's work would be cut short by leukemia in 1980.)255

In his 1990 paper "Elephants Don't Play Chess,"256 robotics researcher Brooks took direct aim at the physical symbol system hypothesis, arguing that symbols are not always necessary since "the world is its own best model. It is always exactly up to date. It always has every detail there is to be known. The trick is to sense it appropriately and often enough."257

In the 1980s and 1990s, many cognitive scientists also rejected the symbol processing model of the mind and argued that the body was essential for reasoning, a theory called the "embodied mind thesis".258

Soft computing and probabilistic reasoning

Soft computing uses methods that work with incomplete and imprecise information. They do not attempt to give precise, logical answers, but give results that are only "probably" correct. This allowed them to solve problems that precise symbolic methods could not handle. Press accounts often claimed these tools could "think like a human".259260

Judea Pearl's Probabilistic Reasoning in Intelligent Systems: Networks of Plausible Inference, an influential 1988 book261 brought probability and decision theory into AI.262 Fuzzy logic, developed by Lofti Zadeh in the 60s, began to be more widely used in AI and robotics. Evolutionary computation and artificial neural networks also handle imprecise information, and are classified as "soft". In the 90s and early 2000s many other soft computing tools were developed and put into use, including Bayesian networks,263 hidden Markov models,264 information theory and stochastic modeling. These tools in turn depended on advanced mathematical techniques such as classical optimization. For a time in the 1990s and early 2000s, these soft tools were studied by a subfield of AI called "computational intelligence".265

Reinforcement learning

Reinforcement learning266 gives an agent a reward every time it performs a desired action well, and may give negative rewards (or "punishments") when it performs poorly. It was described in the first half of the twentieth century by psychologists using animal models, such as Thorndike,267268 Pavlov269 and Skinner.270 In the 1950s, Alan Turing271272 and Arthur Samuel273 foresaw the role of reinforcement learning in AI.

A successful and influential research program was led by Richard Sutton and Andrew Barto beginning 1972. Their collaboration revolutionized the study of reinforcement learning and decision making over the four decades.274275 In 1988, Sutton described machine learning in terms of decision theory (i.e., the Markov decision process). This gave the subject a solid theoretical foundation and access to a large body of theoretical results developed in the field of operations research.276

Also in 1988, Sutton and Barto developed the "temporal difference" (TD) learning algorithm, where the agent is rewarded only when its predictions about the future show improvement. It significantly outperformed previous algorithms.277 TD-learning was used by Gerald Tesauro in 1992 in the program TD-Gammon, which played backgammon as well as the best human players. The program learned the game by playing against itself with zero prior knowledge.278 In an interesting case of interdisciplinary convergence, neurologists discovered in 1997 that the dopamine reward system in brains also uses a version of the TD-learning algorithm.279280281 TD learning would be become highly influential in the 21st century, used in both AlphaGo and AlphaZero.282

Second AI winter

The business community's fascination with AI rose and fell in the 1980s in the classic pattern of an economic bubble. As dozens of companies failed, the perception in the business world was that the technology was not viable.283 The damage to AI's reputation would last into the 21st century. Inside the field there was little agreement on the reasons for AI's failure to fulfill the dream of human level intelligence that had captured the imagination of the world in the 1960s. Together, all these factors helped to fragment AI into competing subfields focused on particular problems or approaches, sometimes even under new names that disguised the tarnished pedigree of "artificial intelligence".284

Over the next 20 years, AI consistently delivered working solutions to specific isolated problems. By the late 1990s, it was being used throughout the technology industry, although somewhat behind the scenes. The success was due to increasing computer power, by collaboration with other fields (such as mathematical optimization and statistics) and using the highest standards of scientific accountability. By 2000, AI had achieved some of its oldest goals. The field was both more cautious and more successful than it had ever been.

AI winter

The term "AI winter" was coined by researchers who had survived the funding cuts of 1974 when they became concerned that enthusiasm for expert systems had spiraled out of control and that disappointment would certainly follow.285 Their fears were well founded: in the late 1980s and early 1990s, AI suffered a series of financial setbacks.286

The first indication of a change in weather was the sudden collapse of the market for specialized AI hardware in 1987. Desktop computers from Apple and IBM had been steadily gaining speed and power and in 1987 they became more powerful than the more expensive Lisp machines made by Symbolics and others. There was no longer a good reason to buy them. An entire industry worth half a billion dollars was demolished overnight.287

Eventually the earliest successful expert systems, such as XCON, proved too expensive to maintain. They were difficult to update, they could not learn, and they were "brittle" (i.e., they could make grotesque mistakes when given unusual inputs). Expert systems proved useful, but only in a few special contexts.288

In the late 1980s, the Strategic Computing Initiative cut funding to AI "deeply and brutally". New leadership at DARPA had decided that AI was not "the next wave" and directed funds towards projects that seemed more likely to produce immediate results.289

By 1991, the impressive list of goals penned in 1981 for Japan's Fifth Generation Project had not been met. Indeed, some of them, like "carry on a casual conversation" would not be accomplished for another 30 years. As with other AI projects, expectations had run much higher than what was actually possible.290291

Over 300 AI companies had shut down, gone bankrupt, or been acquired by the end of 1993, effectively ending the first commercial wave of AI.292 In 1994, HP Newquist stated in The Brain Makers that "The immediate future of artificial intelligence—in its commercial form—seems to rest in part on the continued success of neural networks."293

AI behind the scenes

In the 1990s, algorithms originally developed by AI researchers began to appear as parts of larger systems. AI had solved a lot of very difficult problems294 and their solutions proved to be useful throughout the technology industry,295296 such as data mining, industrial robotics, logistics, speech recognition,297 banking software,298 medical diagnosis299 and Google's search engine.300301

The field of AI received little or no credit for these successes in the 1990s and early 2000s. Many of AI's greatest innovations have been reduced to the status of just another item in the tool chest of computer science.302 Nick Bostrom explains: "A lot of cutting edge AI has filtered into general applications, often without being called AI because once something becomes useful enough and common enough it's not labeled AI anymore."303

Many researchers in AI in the 1990s deliberately called their work by other names, such as informatics, knowledge-based systems, "cognitive systems" or computational intelligence. In part, this may have been because they considered their field to be fundamentally different from AI, but also the new names help to procure funding.304305306 In the commercial world at least, the failed promises of the AI Winter continued to haunt AI research into the 2000s, as the New York Times reported in 2005: "Computer scientists and software engineers avoided the term artificial intelligence for fear of being viewed as wild-eyed dreamers."307

Mathematical rigor, greater collaboration and a narrow focus

AI researchers began to develop and use sophisticated mathematical tools more than they ever had in the past.308309 Most of the new directions in AI relied heavily on mathematical models, including artificial neural networks, probabilistic reasoning, soft computing and reinforcement learning. In the 90s and 2000s, many other highly mathematical tools were adapted for AI. These tools were applied to machine learning, perception and mobility.

There was a widespread realization that many of the problems that AI needed to solve were already being worked on by researchers in fields like statistics, mathematics, electrical engineering, economics or operations research. The shared mathematical language allowed both a higher level of collaboration with more established and successful fields and the achievement of results which were measurable and provable; AI had become a more rigorous "scientific" discipline.

Another key reason for the success in the 90s was that AI researchers focussed on specific problems with verifiable solutions (an approach later derided as narrow AI). This provided useful tools in the present, rather than speculation about the future.

Intelligent agents

A new paradigm called "intelligent agents" became widely accepted during the 1990s.310311312 Although earlier researchers had proposed modular "divide and conquer" approaches to AI,313 the intelligent agent did not reach its modern form until Judea Pearl, Allen Newell, Leslie P. Kaelbling, and others brought concepts from decision theory and economics into the study of AI.314 When the economist's definition of a rational agent was married to computer science's definition of an object or module, the intelligent agent paradigm was complete.

An intelligent agent is a system that perceives its environment and takes actions which maximize its chances of success. By this definition, simple programs that solve specific problems are "intelligent agents", as are human beings and organizations of human beings, such as firms. The intelligent agent paradigm defines AI research as "the study of intelligent agents".315 This is a generalization of some earlier definitions of AI: it goes beyond studying human intelligence; it studies all kinds of intelligence.

The paradigm gave researchers license to study isolated problems and to disagree about methods, but still retain hope that their work could be combined into an agent architecture that would be capable of general intelligence.316

Milestones and Moore's law

On May 11, 1997, Deep Blue became the first computer chess-playing system to beat a reigning world chess champion, Garry Kasparov.317 In 2005, a Stanford robot won the DARPA Grand Challenge by driving autonomously for 131 miles along an unrehearsed desert trail. Two years later, a team from CMU won the DARPA Urban Challenge by autonomously navigating 55 miles in an urban environment while responding to traffic hazards and adhering to traffic laws.318

These successes were not due to some revolutionary new paradigm, but mostly on the tedious application of engineering skill and on the tremendous increase in the speed and capacity of computers by the 90s.319 In fact, Deep Blue's computer was 10 million times faster than the Ferranti Mark 1 that Christopher Strachey taught to play chess in 1951.320 This dramatic increase is measured by Moore's law, which predicts that the speed and memory capacity of computers doubles every two years. The fundamental problem of "raw computer power" was slowly being overcome.

Big data, deep learning, AGI (2005–2017)

In the first decades of the 21st century, access to large amounts of data (known as "big data"), cheaper and faster computers and advanced machine learning techniques were successfully applied to many problems throughout the economy. A turning point was the success of deep learning around 2012 which improved the performance of machine learning on many tasks, including image and video processing, text analysis, and speech recognition.321 Investment in AI increased along with its capabilities, and by 2016, the market for AI-related products, hardware, and software reached more than $8 billion, and the New York Times reported that interest in AI had reached a "frenzy".322

In 2002, Ben Goertzel and others became concerned that AI had largely abandoned its original goal of producing versatile, fully intelligent machines, and argued in favor of more direct research into artificial general intelligence. By the mid-2010s several companies and institutions had been founded to pursue Artificial General Intelligence (AGI), such as OpenAI and Google's DeepMind. During the same period, new insights into superintelligence raised concerns that AI was an existential threat. The risks and unintended consequences of AI technology became an area of serious academic research after 2016.

Big data and big machines

See also: List of datasets for machine-learning research

The success of machine learning in the 2000s depended on the availability of vast amounts of training data and faster computers.323 Russell and Norvig wrote that the "improvement in performance obtained by increasing the size of the data set by two or three orders of magnitude outweighs any improvement that can be made by tweaking the algorithm."324 Geoffrey Hinton recalled that back in the 90s, the problem was that "our labeled datasets were thousands of times too small. [And] our computers were millions of times too slow."325 This was no longer true by 2010.

The most useful data in the 2000s came from curated, labeled data sets created specifically for machine learning and AI. In 2007, a group at UMass Amherst released Labeled Faces in the Wild, an annotated set of images of faces that was widely used to train and test face recognition systems for the next several decades.326 Fei-Fei Li developed ImageNet, a database of three million images captioned by volunteers using the Amazon Mechanical Turk. Released in 2009, it was a useful body of training data and a benchmark for testing for the next generation of image processing systems.327328 Google released word2vec in 2013 as an open source resource. It used large amounts of data text scraped from the internet and word embedding to create a numeric vector to represent each word. Users were surprised at how well it was able to capture word meanings, for example, ordinary vector addition would give equivalences like China + River = Yangtze, London-England+France = Paris.329 This database in particular would be essential for the development of large language models in the late 2010s.

The explosive growth of the internet gave machine learning programs access to billions of pages of text and images that could be scraped. And, for specific problems, large privately held databases contained the relevant data. McKinsey Global Institute reported that "by 2009, nearly all sectors in the US economy had at least an average of 200 terabytes of stored data".330 This collection of information was known in the 2000s as big data.

In a Jeopardy! exhibition match in February 2011, IBM's question answering system Watson defeated the two best Jeopardy! champions, Brad Rutter and Ken Jennings, by a significant margin.331 Watson's expertise would have been impossible without the information available on the internet.332

Deep learning

Main article: Deep learning

In 2012, AlexNet, a deep learning model,333 developed by Alex Krizhevsky, won the ImageNet Large Scale Visual Recognition Challenge, with significantly fewer errors than the second-place winner.334335 Krizhevsky worked with Geoffrey Hinton at the University of Toronto.336 This was a turning point in machine learning: over the next few years dozens of other approaches to image recognition were abandoned in favor of deep learning.337

Deep learning uses a multi-layer perceptron. Although this architecture has been known since the 60s, getting it to work requires powerful hardware and large amounts of training data.338 Before these became available, improving performance of image processing systems required hand-crafted ad hoc features that were difficult to implement.339 Deep learning was simpler and more general.340

Deep learning was applied to dozens of problems over the next few years (such as speech recognition, machine translation, medical diagnosis, and game playing). In every case it showed enormous gains in performance.341 Investment and interest in AI boomed as a result.342

The alignment problem

It became fashionable in the 2000s to begin talking about the future of AI again and several popular books considered the possibility of superintelligent machines and what they might mean for human society. Some of this was optimistic (such as Ray Kurzweil's The Singularity is Near), but others warned that a sufficiently powerful AI was existential threat to humanity, such as Nick Bostrom and Eliezer Yudkowsky.343 The topic became widely covered in the press and many leading intellectuals and politicians commented on the issue.

AI programs in the 21st century are defined by their goals – the specific measures that they are designed to optimize. Nick Bostrom's influential 2005 book Superintelligence argued that, if one isn't careful about defining these goals, the machine may cause harm to humanity in the process of achieving a goal. Stuart J. Russell used the example of an intelligent robot that kills its owner to prevent it from being unplugged, reasoning "you can't fetch the coffee if you're dead".344 (This problem is known by the technical term "instrumental convergence".) The solution is to align the machine's goal function with the goals of its owner and humanity in general. Thus, the problem of mitigating the risks and unintended consequences of AI became known as "the value alignment problem" or AI alignment.345

At the same time, machine learning systems had begun to have disturbing unintended consequences. Cathy O'Neil explained how statistical algorithms had been among the causes of the 2008 economic crash,346 Julia Angwin of ProPublica argued that the COMPAS system used by the criminal justice system exhibited racial bias under some measures,347348 others showed that many machine learning systems exhibited some form of racial bias,349 and there were many other examples of dangerous outcomes that had resulted from machine learning systems.350

In 2016, the election of Donald Trump and the controversy over the COMPAS system illuminated several problems with the current technological infrastructure, including misinformation, social media algorithms designed to maximize engagement, the misuse of personal data and the trustworthiness of predictive models.351 Issues of fairness and unintended consequences became significantly more popular at AI conferences, publications vastly increased, funding became available, and many researchers re-focussed their careers on these issues. The value alignment problem became a serious field of academic study.352353

Artificial general intelligence research

In the early 2000s, several researchers became concerned that mainstream AI was too focused on "measurable performance in specific applications"354 (known as "narrow AI") and had abandoned AI's original goal of creating versatile, fully intelligent machines. An early critic was Nils Nilsson in 1995, and similar opinions were published by AI elder statesmen John McCarthy, Marvin Minsky, and Patrick Winston in 2007–2009. Minsky organized a symposium on "human-level AI" in 2004.355 Ben Goertzel adopted the term "artificial general intelligence" for the new sub-field, founding a journal and holding conferences beginning in 2008.356 The new field grew rapidly, buoyed by the continuing success of artificial neural networks and the hope that it was the key to AGI.

Several competing companies, laboratories and foundations were founded to develop AGI in the 2010s. DeepMind was founded in 2010 by three English scientists, Demis Hassabis, Shane Legg and Mustafa Suleyman, with funding from Peter Thiel and later Elon Musk. The founders and financiers were deeply concerned about AI safety and the existential risk of AI. DeepMind's founders had a personal connection with Yudkowsky and Musk was among those who was actively raising the alarm.357 Hassabis was both worried about the dangers of AGI and optimistic about its power; he hoped they could "solve AI, then solve everything else."358 The New York Times wrote in 2023 "At the heart of this competition is a brain-stretching paradox. The people who say they are most worried about AI are among the most determined to create it and enjoy its riches. They have justified their ambition with their strong belief that they alone can keep AI from endangering Earth."359

In 2012, Geoffrey Hinton (who been leading neural network research since the 80s) was approached by Baidu, which wanted to hire him and all his students for an enormous sum. Hinton decided to hold an auction and, at a Lake Tahoe AI conference, they sold themselves to Google for a price of $44 million. Hassabis took notice and sold DeepMind to Google in 2014, on the condition that it would not accept military contracts and would be overseen by an ethics board.360

Larry Page of Google, unlike Musk and Hassabis, was an optimist about the future of AI. Musk and Paige became embroiled in an argument about the risk of AGI at Musk's 2015 birthday party. They had been friends for decades but stopped speaking to each other shortly afterwards. Musk attended the one and only meeting of the DeepMind's ethics board, where it became clear that Google was uninterested in mitigating the harm of AGI. Frustrated by his lack of influence he founded OpenAI in 2015, enlisting Sam Altman to run it and hiring top scientists. OpenAI began as a non-profit, "free from the economic incentives that were driving Google and other corporations."361 Musk became frustrated again and left the company in 2018. OpenAI turned to Microsoft for continued financial support and Altman and OpenAI formed a for-profit version of the company with more than $1 billion in financing.362

In 2021, Dario Amodei and 14 other scientists left OpenAI over concerns that the company was putting profits above safety. They formed Anthropic, which soon had $6 billion in financing from Microsoft and Google.363

Large language models, AI boom (2017–present)

Main article: AI boom

The AI boom started with the initial development of key architectures and algorithms such as the transformer architecture in 2017, leading to the scaling and development of large language models exhibiting human-like traits of knowledge, attention and creativity. The new AI era began since 2020, with the public release of scaled large language models (LLMs) such as ChatGPT.364

Transformer architecture and large language models

Main article: Large language models

In 2017, the transformer architecture was proposed by Google researchers. It exploits an attention mechanism and became widely used in large language models.365

Large language models, based on the transformer, were developed by AGI companies: OpenAI released GPT-3 in 2020, and DeepMind released Gato in 2022. These are foundation models: they are trained on vast quantities of unlabeled data and can be adapted to a wide range of downstream tasks.

These models can discuss a huge number of topics and display general knowledge. The question naturally arises: are these models an example of artificial general intelligence? Bill Gates was skeptical of the new technology and the hype that surrounded AGI. However, Altman presented him with a live demo of ChatGPT4 passing an advanced biology test. Gates was convinced.366 In 2023, Microsoft Research tested the model with a large variety of tasks, and concluded that "it could reasonably be viewed as an early (yet still incomplete) version of an artificial general intelligence (AGI) system".367

In 2024, OpenAI o3, a type of advanced reasoning model developed by OpenAI was announced. On the Abstraction and Reasoning Corpus for Artificial General Intelligence (ARC-AGI) benchmark developed by François Chollet in 2019, the model achieved an unofficial score of 87.5% on the semi-private test, surpassing the typical human score of 84%. The benchmark is supposed to be a necessary, but not sufficient test for AGI. Speaking of the benchmark, Chollet has said "You’ll know AGI is here when the exercise of creating tasks that are easy for regular humans but hard for AI becomes simply impossible."368

AI boom

Main article: AI boom

Investment in AI grew exponentially after 2020, with venture capital funding for generative AI companies increasing dramatically. Total AI investments rose from $18 billion in 2014 to $119 billion in 2021, with generative AI accounting for approximately 30% of investments by 2023.369 According to metrics from 2017 to 2021, the United States outranked the rest of the world in terms of venture capital funding, number of startups, and AI patents granted.370 The commercial AI scene became dominated by American Big Tech companies, whose investments in this area surpassed those from U.S.-based venture capitalists.371 OpenAI's valuation reached $86 billion by early 2024,372 while NVIDIA's market capitalization surpassed $3.3 trillion by mid-2024, making it the world's largest company by market capitalization as the demand for AI-capable GPUs surged.373

15.ai, launched in March 2020374 by an anonymous MIT researcher,375376 was one of the earliest examples of generative AI gaining widespread public attention during the initial stages of the AI boom.377 The free web application demonstrated the ability to clone character voices using neural networks with minimal training data, requiring as little as 15 seconds of audio to reproduce a voice—a capability later corroborated by OpenAI in 2024.378 The service went viral on social media platforms in early 2021,379380 allowing users to generate speech for characters from popular media franchises, and became particularly notable for its pioneering role in popularizing AI voice synthesis for creative content and memes.381

Advent of AI for public use

ChatGPT was launched on November 30, 2022, marking a pivotal moment in artificial intelligence's public adoption. Within days of its release it went viral, gaining over 100 million users in two months and becoming the fastest-growing consumer software application in history.382 The chatbot's ability to engage in human-like conversations, write code, and generate creative content captured public imagination and led to rapid adoption across various sectors including education, business, and research.383 ChatGPT's success prompted unprecedented responses from major technology companies—Google declared a "code red" and rapidly launched Gemini (formerly known as Google Bard), while Microsoft incorporated the technology into Bing Chat.384

The rapid adoption of these AI technologies sparked intense debate about their implications. Notable AI researchers and industry leaders voiced both optimism and concern about the accelerating pace of development. In March 2023, over 20,000 signatories, including computer scientist Yoshua Bengio, Elon Musk, and Apple co-founder Steve Wozniak, signed an open letter calling for a pause in advanced AI development, citing "profound risks to society and humanity."385 However, other prominent researchers like Juergen Schmidhuber took a more optimistic view, emphasizing that the majority of AI research aims to make "human lives longer and healthier and easier."386

By mid-2024, however, the financial sector began to scrutinize AI companies more closely, particularly questioning their capacity to produce a return on investment commensurate with their massive valuations. Some prominent investors raised concerns about market expectations becoming disconnected from fundamental business realities. Jeremy Grantham, co-founder of GMO LLC, warned investors to "be quite careful" and drew parallels to previous technology-driven market bubbles.387 Similarly, Jeffrey Gundlach, CEO of DoubleLine Capital, explicitly compared the AI boom to the dot-com bubble of the late 1990s, suggesting that investor enthusiasm might be outpacing realistic near-term capabilities and revenue potential.388 These concerns were amplified by the substantial market capitalizations of AI-focused companies, many of which had yet to demonstrate sustainable profitability models.

In March 2024, Anthropic released the Claude 3 family of large language models, including Claude 3 Haiku, Sonnet, and Opus.389 The models demonstrated significant improvements in capabilities across various benchmarks, with Claude 3 Opus notably outperforming leading models from OpenAI and Google.390 In June 2024, Anthropic released Claude 3.5 Sonnet, which demonstrated improved performance compared to the larger Claude 3 Opus, particularly in areas such as coding, multistep workflows, and image analysis.391

2024 Nobel Prizes

In 2024, the Royal Swedish Academy of Sciences awarded Nobel Prizes in recognition of groundbreaking contributions to artificial intelligence. The recipients included:

Further Study and development of AI

In January 2025, OpenAI announced a new AI, ChatGPT-Gov, which would be specifically designed for US government agencies to use securely.393 Open AI said that agencies could utilize ChatGPT Gov on a Microsoft Azure cloud or Azure Government cloud, "on top of Microsoft’s Azure’s OpenAI Service." OpenAI's announcement stated that "Self-hosting ChatGPT Gov enables agencies to more easily manage their own security, privacy, and compliance requirements, such as stringent cybersecurity frameworks (IL5, CJIS, ITAR, FedRAMP High). Additionally, we believe this infrastructure will expedite internal authorization of OpenAI’s tools for the handling of non-public sensitive data."394

Robotic Integration and Practical Applications of Artificial Intelligence (2025–present)

Advanced artificial intelligence (AI) systems, capable of understanding and responding to human dialogue with high accuracy, have matured to enable seamless integration with robotics, transforming industries such as manufacturing, household automation, healthcare, public services, and materials research.395 Applications of artificial intelligence also accelerates scientific research through advanced data analysis and hypothesis generation.396 Countries including China, the United States, and Japan have invested significantly in policies and funding to deploy AI-powered robots, addressing labor shortages, boosting innovation, and enhancing efficiency, while implementing regulatory frameworks to ensure ethical and safe development.397

China

The year 2025 has been heralded as the "Year of AI Robotics," marking a pivotal moment in the seamless integration of artificial intelligence (AI) and robotics. In 2025, China invested approximately 730 billion yuan (roughly $100 billion USD) to advance AI and robotics in smart manufacturing and healthcare.398399 The "14th Five-Year Plan" (2021–2025) prioritized service robots, with AI systems enabling robots to perform complex tasks like assisting in surgeries or automating factory assembly lines.400 For example, AI-powered humanoid robots in Chinese hospitals can interpret patient requests, deliver supplies, and assist nurses with routine tasks, demonstrating that existing AI conversational capabilities are robust enough for practical robotic applications. Starting in September 2025, China mandated labeling of AI-generated content to ensure transparency and public trust in these technologies.401

United States

In January 2025, a significant development in AI infrastructure investment occurred with the formation of Stargate LLC. The joint venture, created by OpenAI, SoftBank, Oracle, and MGX, announced plans to invest US$500 billion in AI infrastructure across the United States by 2029, starting with US$100 billion, in order to support the re-industrialization of the United States and provide a strategic capability to protect the national security of America and its allies.402 The venture was formally announced by U.S. President Donald Trump on January 21, 2025, with SoftBank CEO Masayoshi Son appointed as chairman.403404

The U.S. government allocated approximately $2 billion to integrate AI and robotics in manufacturing and logistics, leveraging AI's ability to process natural language and execute user instructions in 2025.405 State governments supplemented this with funding for service robots, such as those deployed in warehouses to fulfill verbal commands for inventory management or in eldercare facilities to respond to residents' requests for assistance.406 These applications highlight that merging advanced AI, already proficient in human interaction, with robotic hardware is a practical step forward. Some funds were directed to defense, including Lethal autonomous weapon and Military robot. In January 2025, Executive Order 14179 established an "AI Action Plan" to accelerate innovation and deployment of these technologies.407

Impact

In the 2020s, increased investments in AI by governments and organizations worldwide have accelerated the advancement of artificial intelligence, driving scientific breakthroughs, boosting workforce productivity, and transforming industries through the automation of complex tasks.408 By seamlessly integrating advanced AI systems into various sectors, these developments are poised to revolutionize smart manufacturing and service industries, fundamentally transforming everyday life.

See also

Notes

.

References

  1. Kaplan & Haenlein 2018. - Kaplan A, Haenlein M (2018), "Siri, Siri in my Hand, who's the Fairest in the Land? On the Interpretations, Illustrations and Implications of Artificial Intelligence", Business Horizons, 62: 15–25, doi:10.1016/j.bushor.2018.08.004, S2CID 158433736 https://doi.org/10.1016%2Fj.bushor.2018.08.004

  2. Newquist 1994, pp. 143–156. - Newquist HP (1994), The Brain Makers: Genius, Ego, And Greed in the Quest For Machines That Think, New York: Macmillan/SAMS, ISBN 978-0-9885937-1-8, OCLC 313139906 https://search.worldcat.org/oclc/313139906

  3. Newquist 1994, pp. 144–152. - Newquist HP (1994), The Brain Makers: Genius, Ego, And Greed in the Quest For Machines That Think, New York: Macmillan/SAMS, ISBN 978-0-9885937-1-8, OCLC 313139906 https://search.worldcat.org/oclc/313139906

  4. Rhodios 2007, Book 4, the Talos episode. - Rhodios A (2007). The Argonautika : Expanded Edition. University of California Press. p. 355. ISBN 978-0-520-93439-9. OCLC 811491744. https://search.worldcat.org/oclc/811491744

  5. Bibliotheke 1.9.26 /wiki/Bibliotheca_(Pseudo-Apollodorus)

  6. Rhodios 2007. - Rhodios A (2007). The Argonautika : Expanded Edition. University of California Press. p. 355. ISBN 978-0-520-93439-9. OCLC 811491744. https://search.worldcat.org/oclc/811491744

  7. Morford 2007. - Morford M (2007). Classical mythology. Oxford: Oxford University Press. p. 184. ISBN 978-0-19-085164-4. OCLC 1102437035. https://search.worldcat.org/oclc/1102437035

  8. Linden 2003. - Linden SJ (2003). The alchemy reader : from Hermes Trismegistus to Isaac Newton. New York: Cambridge University Press. pp. Ch. 18. ISBN 0-521-79234-7. OCLC 51210362. https://search.worldcat.org/oclc/51210362

  9. Kressel 2015. - Kressel M (1 October 2015). "36 Days of Judaic Myth: Day 24, The Golem of Prague 2015". Matthew Kressel. Retrieved 15 March 2020. https://www.matthewkressel.net/2015/10/01/36-days-of-judaic-myth-day-24-the-golem-of-prague/

  10. Jewish Encyclopedia, GOLEM. - "GOLEM", The Jewish Encyclopedia, retrieved 15 March 2020 http://www.jewishencyclopedia.com/articles/6777-golem

  11. Newquist 1994, p. 38. - Newquist HP (1994), The Brain Makers: Genius, Ego, And Greed in the Quest For Machines That Think, New York: Macmillan/SAMS, ISBN 978-0-9885937-1-8, OCLC 313139906 https://search.worldcat.org/oclc/313139906

  12. Talmud, Sanhedrin 65b. - "Sanhedrin 65b". www.sefaria.org. Retrieved 15 March 2020. https://www.sefaria.org/Sanhedrin.65b?lang=bi

  13. O'Connor 1994. - O'Connor KM (1994), The alchemical creation of life (takwin) and other concepts of Genesis in medieval Islam, University of Pennsylvania, pp. 1–435, retrieved 10 January 2007 http://repository.upenn.edu/dissertations/AAI9503804

  14. Goethe 1890. - Goethe JW (1890). Faust; a tragedy. Translated, in the original metres ... by Bayard Taylor. Authorised ed., published by special arrangement with Mrs. Bayard Taylor. With a biographical introd. London Ward, Lock. https://archive.org/details/fausttragedytran00goetuoft

  15. McCorduck 2004, pp. 17–25. - McCorduck P (2004), Machines Who Think (2nd ed.), Natick, MA: A. K. Peters, Ltd., ISBN 978-1-56881-205-2, OCLC 52197627 https://search.worldcat.org/oclc/52197627

  16. Butler 1863. - Butler S (13 June 1863), "Darwin Among the Machines", The Press, Christchurch, New Zealand, retrieved 10 October 2008 https://nzetc.victoria.ac.nz/tm/scholarly/tei-ButFir-t1-g1-t1-g1-t4-body.html

  17. Newquist 1994, p. 65. - Newquist HP (1994), The Brain Makers: Genius, Ego, And Greed in the Quest For Machines That Think, New York: Macmillan/SAMS, ISBN 978-0-9885937-1-8, OCLC 313139906 https://search.worldcat.org/oclc/313139906

  18. Cave & Dihal 2019. - Cave S, Dihal K (2019). "Hopes and fears for intelligent machines in fiction and reality". Nature Machine Intelligence. 1 (2): 74–78. doi:10.1038/s42256-019-0020-9. ISSN 2522-5839. S2CID 150700981. https://www.nature.com/articles/s42256-019-0020-9

  19. Needham 1986, p. 53. - Needham J (1986). Science and Civilization in China: Volume 2. Taipei: Caves Books Ltd.

  20. McCorduck 2004, p. 6. - McCorduck P (2004), Machines Who Think (2nd ed.), Natick, MA: A. K. Peters, Ltd., ISBN 978-1-56881-205-2, OCLC 52197627 https://search.worldcat.org/oclc/52197627

  21. Nick 2005. - Nick M (2005), Al Jazari: The Ingenious 13th Century Muslim Mechanic, Al Shindagah, retrieved 16 October 2008 http://www.alshindagah.com/marapr2005/jaziri.html

  22. McCorduck 2004, p. 10. - McCorduck P (2004), Machines Who Think (2nd ed.), Natick, MA: A. K. Peters, Ltd., ISBN 978-1-56881-205-2, OCLC 52197627 https://search.worldcat.org/oclc/52197627

  23. Newquist 1994, p. 40. - Newquist HP (1994), The Brain Makers: Genius, Ego, And Greed in the Quest For Machines That Think, New York: Macmillan/SAMS, ISBN 978-0-9885937-1-8, OCLC 313139906 https://search.worldcat.org/oclc/313139906

  24. McCorduck 2004, p. 16. - McCorduck P (2004), Machines Who Think (2nd ed.), Natick, MA: A. K. Peters, Ltd., ISBN 978-1-56881-205-2, OCLC 52197627 https://search.worldcat.org/oclc/52197627

  25. McCorduck 2004, pp. 59–62. - McCorduck P (2004), Machines Who Think (2nd ed.), Natick, MA: A. K. Peters, Ltd., ISBN 978-1-56881-205-2, OCLC 52197627 https://search.worldcat.org/oclc/52197627

  26. McCorduck 2004, p. 17. - McCorduck P (2004), Machines Who Think (2nd ed.), Natick, MA: A. K. Peters, Ltd., ISBN 978-1-56881-205-2, OCLC 52197627 https://search.worldcat.org/oclc/52197627

  27. Levitt 2000. - Levitt GM (2000), The Turk, Chess Automaton, Jefferson, N.C.: McFarland, ISBN 978-0-7864-0778-1

  28. Newquist 1994, p. 30. - Newquist HP (1994), The Brain Makers: Genius, Ego, And Greed in the Quest For Machines That Think, New York: Macmillan/SAMS, ISBN 978-0-9885937-1-8, OCLC 313139906 https://search.worldcat.org/oclc/313139906

  29. Crevier 1993, p. 1. - Crevier D (1993). AI: The Tumultuous Search for Artificial Intelligence. New York, NY: BasicBooks. ISBN 0-465-02997-3.

  30. Quoted in McCorduck 2004, p. 8. - McCorduck P (2004), Machines Who Think (2nd ed.), Natick, MA: A. K. Peters, Ltd., ISBN 978-1-56881-205-2, OCLC 52197627 https://search.worldcat.org/oclc/52197627

  31. Cave, Dihal & Dillon 2020, p. 56. - Cave S, Dihal K, Dillon S (2020). AI Narratives: A History of Imaginative Thinking about Intelligent Machines. Oxford University Press. ISBN 978-0-19-884666-6. Retrieved 2 May 2023. https://books.google.com/books?id=T53SDwAAQBAJ&pg=PA56

  32. Butler 1979. - Butler EM (1979) [1948]. The myth of the magus. London: Cambridge University Press. ISBN 0-521-22564-7. OCLC 5063114. https://search.worldcat.org/oclc/5063114

  33. Porterfield 2006, p. 136. - Porterfield A (2006). The Protestant Experience in America. American religious experience. Greenwood Press. p. 136. ISBN 978-0-313-32801-5. Retrieved 15 May 2023. https://books.google.com/books?id=V9VM9NEsqXwC&pg=PA136

  34. Hollander 1991. - Hollander LM (1991) [1964]. Heimskringla; history of the kings of Norway. Austin: Published for the American-Scandinavian Foundation by the University of Texas Press. ISBN 0-292-73061-6. OCLC 638953. https://search.worldcat.org/oclc/638953

  35. Russell & Norvig 2021, pp. 6 & 7. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  36. Berlinski 2000. - Berlinski D (2000), The Advent of the Algorithm, Harcourt Books, ISBN 978-0-15-601391-8, OCLC 46890682 https://archive.org/details/adventofalgorith0000berl

  37. Carreras y Artau 2018. - Carreras y Artau T (2018) [1939], Historia de la filosofía española. Filosofía cristiana de los siglos XIII al XV (in Spanish), vol. 1, Madrid: Forgotten Books, ISBN 9781390433708

  38. Russell & Norvig 2021, p. 6. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  39. Bonner 2007. - Bonner A (2007), The Art and Logic of Ramón Llull: A User's Guide, Brill, ISBN 978-9004163256

  40. Bonner 1985, pp. 57–71. - Bonner A (1985). "Llull's Influence: The History of Lullism". Doctor Illuminatus. A Ramon Llull Reader. Princeton University Press.

  41. 17th century mechanism and AI: McCorduck 2004, pp. 37–46 Russell & Norvig 2021, p. 6 Buchanan 2005, p. 53 - McCorduck P (2004), Machines Who Think (2nd ed.), Natick, MA: A. K. Peters, Ltd., ISBN 978-1-56881-205-2, OCLC 52197627 https://search.worldcat.org/oclc/52197627

  42. Hobbes and AI: Russell & Norvig 2021, p. 6 McCorduck 2004, p. 42 Hobbes 1651, chapter 5 - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  43. Leibniz and AI: McCorduck 2004, p. 41 Russell & Norvig 2021, p. 6} Berlinski 2000, p. 12 Buchanan 2005, p. 53 - McCorduck P (2004), Machines Who Think (2nd ed.), Natick, MA: A. K. Peters, Ltd., ISBN 978-1-56881-205-2, OCLC 52197627 https://search.worldcat.org/oclc/52197627

  44. Russell & Norvig 2021, p. 8. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  45. Berlinski 2000. - Berlinski D (2000), The Advent of the Algorithm, Harcourt Books, ISBN 978-0-15-601391-8, OCLC 46890682 https://archive.org/details/adventofalgorith0000berl

  46. Russell & Norvig 2021, p. 9. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  47. Russell & Norvig 2021, p. 9. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  48. The Lambda calculus was especially important to AI, since it was an inspiration for Lisp (the most important programming language used in 20th century AI).[46] /wiki/Lambda_calculus

  49. Russell & Norvig 2021, p. 9. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  50. The Turing machine: Newquist 1994, p. 56 McCorduck 2004, pp. 63–64 Crevier 1993, pp. 22–24 Russell & Norvig 2021, p. 9 and see Turing 1936–1937 /wiki/Turing_machine

  51. Russell & Norvig 2021, p. 6. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  52. Couturat 1901. - Couturat L (1901), La Logique de Leibniz

  53. Russell & Norvig 2021, p. 15. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  54. Russell & Norvig 2021, p. 15. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  55. Newquist 1994, p. 67. - Newquist HP (1994), The Brain Makers: Genius, Ego, And Greed in the Quest For Machines That Think, New York: Macmillan/SAMS, ISBN 978-0-9885937-1-8, OCLC 313139906 https://search.worldcat.org/oclc/313139906

  56. Randall (1982, pp. 4–5); Byrne (2012); Mulvihill (2012) - Randall B (1982), "From Analytical Engine to Electronic Digital Computer: The Contributions of Ludgate, Torres, and Bush", fano.co.uk, retrieved 29 October 2018 http://www.fano.co.uk/ludgate/

  57. Randall (1982, pp. 6, 11–13); Quevedo (1914); Quevedo (1915) - Randall B (1982), "From Analytical Engine to Electronic Digital Computer: The Contributions of Ludgate, Torres, and Bush", fano.co.uk, retrieved 29 October 2018 http://www.fano.co.uk/ludgate/

  58. Randall 1982, pp. 13, 16–17. - Randall B (1982), "From Analytical Engine to Electronic Digital Computer: The Contributions of Ludgate, Torres, and Bush", fano.co.uk, retrieved 29 October 2018 http://www.fano.co.uk/ludgate/

  59. Quoted in Russell & Norvig (2021, p. 15) - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  60. Menabrea & Lovelace 1843. - Menabrea LF, Lovelace A (1843), "Sketch of the Analytical Engine Invented by Charles Babbage", Scientific Memoirs, 3, retrieved 29 August 2008 http://www.fourmilab.ch/babbage/sketch.html

  61. Russell & Norvig 2021, p. 14. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  62. McCorduck 2004, pp. 76–80. - McCorduck P (2004), Machines Who Think (2nd ed.), Natick, MA: A. K. Peters, Ltd., ISBN 978-1-56881-205-2, OCLC 52197627 https://search.worldcat.org/oclc/52197627

  63. Russell & Norvig 2021, p. 14. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  64. AI's immediate predecessors: McCorduck 2004, pp. 51–57, 80–107 Crevier 1993, pp. 27–32 Russell & Norvig 2021, pp. 9, 11, 15–17, 981–984 Moravec 1988, p. 3 Cordeschi 2002, Chap. 5 - McCorduck P (2004), Machines Who Think (2nd ed.), Natick, MA: A. K. Peters, Ltd., ISBN 978-1-56881-205-2, OCLC 52197627 https://search.worldcat.org/oclc/52197627

  65. Copeland 2004. - Copeland J( (2004). The Essential Turing: the ideas that gave birth to the computer age. Oxford: Clarendon Press. ISBN 0-19-825079-7.

  66. Dartmouth workshop: McCorduck 2004, pp. 111–136 Crevier 1993, pp. 49–51 Russell & Norvig 2021, p. 18 Newquist 1994, pp. 91–112 /wiki/Dartmouth_workshop

  67. Turing Test, Computing Machinery and Intelligence: McCorduck 2004, pp. 70–72, Crevier 1993, pp. 22−25, Russell & Norvig 2021, pp. 18, 981–984, Haugeland 1985, pp. 6–9, Cordeschi 2002, pp. 170–176. See also Turing 1950 /wiki/Turing_Test

  68. Alan Turing was thinking about machine intelligence at least as early as 1941, when he circulated a paper on machine intelligence which could be the earliest paper in the field of AI — although it is now lost. His 1950 paper was followed by three radio broadcasts on AI by Turing, the two lectures 'Intelligent Machinery, A Heretical Theory' and 'Can Digital Computers Think?' and the panel discussion 'Can Automatic Calculating Machines be Said to Think?'[60]

  69. Newquist 1994, pp. 92–98. - Newquist HP (1994), The Brain Makers: Genius, Ego, And Greed in the Quest For Machines That Think, New York: Macmillan/SAMS, ISBN 978-0-9885937-1-8, OCLC 313139906 https://search.worldcat.org/oclc/313139906

  70. Russell & Norvig 2021, p. 981. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  71. "Donald Hebb". https://thedecisionlab.com/thinkers/neuroscience/donald-hebb

  72. Brown RE (2020). "Donald O. Hebb and the Organization of Behavior: 17 years in the writing". Molecular Brain. 13 (1): 55. doi:10.1186/s13041-020-00567-8. PMC 7137474. PMID 32252813. https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7137474

  73. https://ojs.library.dal.ca/nsis/article/viewFile/nsis44-1brown/3549 https://ojs.library.dal.ca/nsis/article/viewFile/nsis44-1brown/3549

  74. Pitts & McCullough: McCorduck 2004, pp. 51–57, 88–94 Crevier 1993, p. 30 Russell & Norvig 2021, p. 17 Cordeschi 2002, Chap. 5 Piccinini 2004 See also: McCulloch & Pitts 1943 - McCorduck P (2004), Machines Who Think (2nd ed.), Natick, MA: A. K. Peters, Ltd., ISBN 978-1-56881-205-2, OCLC 52197627 https://search.worldcat.org/oclc/52197627

  75. Copeland 2004. - Copeland J( (2004). The Essential Turing: the ideas that gave birth to the computer age. Oxford: Clarendon Press. ISBN 0-19-825079-7.

  76. SNARC: McCorduck 2004, p. 102 Crevier 1993, pp. 34–35 Russell & Norvig 2021, p. 17 /wiki/Stochastic_Neural_Analog_Reinforcement_Calculator

  77. Turtles and Johns Hopkins Beast: McCorduck 2004, p. 98 Crevier 1993, pp. 27–28 Moravec 1988, p. 3 Cordeschi 2002, Chap. 5 /wiki/Turtle_(robot)

  78. Russell & Norvig 2021, p. 17. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  79. Copeland 1999. - Copeland J (1999). "A Brief History of Computing". AlanTuring.net. http://www.alanturing.net/turing_archive/pages/Reference%20Articles/BriefHistofComp.html

  80. Schaeffer 1997, Chapter 6. - Schaeffer J (1997). One Jump Ahead:: Challenging Human Supremacy in Checkers. Springer. ISBN 978-0-387-76575-4.

  81. Russell & Norvig 2021, p. 17, p=19. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  82. McCorduck 2004, pp. 137–170. - McCorduck P (2004), Machines Who Think (2nd ed.), Natick, MA: A. K. Peters, Ltd., ISBN 978-1-56881-205-2, OCLC 52197627 https://search.worldcat.org/oclc/52197627

  83. Crevier 1993, pp. 44–47. - Crevier D (1993). AI: The Tumultuous Search for Artificial Intelligence. New York, NY: BasicBooks. ISBN 0-465-02997-3.

  84. Logic Theorist: McCorduck 2004, pp. 123–125 Crevier 1993, pp. 44–46 Russell & Norvig 2021, p. 18 /wiki/Logic_Theorist

  85. Quoted in Crevier 1993, p. 46 and Russell & Norvig 2021, p. 18 - Crevier D (1993). AI: The Tumultuous Search for Artificial Intelligence. New York, NY: BasicBooks. ISBN 0-465-02997-3.

  86. This was an early statement of the philosophical position John Searle would later call the "Strong AI hypothesis": that machines can contain minds just as human bodies do. /wiki/John_Searle

  87. Dartmouth workshop: McCorduck 2004, pp. 111–136 Crevier 1993, pp. 49–51 Russell & Norvig 2021, p. 18 Newquist 1994, pp. 91–112 /wiki/Dartmouth_workshop

  88. McCarthy et al. 1955. - McCarthy J, Minsky M, Rochester N, Shannon C (31 August 1955), A Proposal for the Dartmouth Summer Research Project on Artificial Intelligence, archived from the original on 30 September 2008, retrieved 16 October 2008 https://web.archive.org/web/20080930164306/http://www-formal.stanford.edu/jmc/history/dartmouth/dartmouth.html

  89. Daniel Crevier wrote "[the proposal] later became known as the 'physical symbol systems hypothesis'".[81] The physical symbol system hypothesis was articulated and named by Newell and Simon in their paper on GPS.[82] It includes a more specific definition of a "machine" as an agent that manipulates symbols. /wiki/Daniel_Crevier

  90. "I won't swear and I hadn't seen it before," McCarthy told Pamela McCorduck in 1979.[83] However, McCarthy also stated unequivocally "I came up with the term" in a CNET interview.[84] The term was chosen by McCarthy to avoid associations with cybernetics and the influence of Norbert Wiener. "[O]ne of the reasons for inventing the term "artificial intelligence" was to escape association with "cybernetics". Its concentration on analog feedback seemed misguided, and I wished to avoid having either to accept Norbert (not Robert) Wiener as a guru or having to argue with him.".[85] /wiki/Pamela_McCorduck

  91. McCorduck 2004, pp. 129–130. - McCorduck P (2004), Machines Who Think (2nd ed.), Natick, MA: A. K. Peters, Ltd., ISBN 978-1-56881-205-2, OCLC 52197627 https://search.worldcat.org/oclc/52197627

  92. Pamela McCorduck discusses how the Dartmouth conference alumni dominated the first two decades of AI research, calling them the "invisible college".[86]

  93. McCorduck 2004, p. 125. - McCorduck P (2004), Machines Who Think (2nd ed.), Natick, MA: A. K. Peters, Ltd., ISBN 978-1-56881-205-2, OCLC 52197627 https://search.worldcat.org/oclc/52197627

  94. Daniel Crevier wrote "the conference is generally recognized as the official birthdate of the new science."[88] /wiki/Daniel_Crevier

  95. Miller 2003. - Miller G (2003). "The cognitive revolution: a historical perspective" (PDF). Trends in Cognitive Sciences. 7 (3): 141–144. doi:10.1016/s1364-6613(03)00029-9. PMID 12639696. https://www.cs.princeton.edu/~rit/geo/Miller.pdf

  96. Russell & Norvig 2021, p. 14. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  97. There were a few psychologists who avoided behaviorism and embraced a cognitive approach before it was fashionable, such as Frederic Bartlett and Kenneth Craig[90] /wiki/Frederic_Bartlett

  98. Russell and Norvig wrote "it was astonishing whenever a computer did anything remotely clever."[91] AI founder John McCarthy called this the "Look, Ma, no hands!" era.[92]

  99. Crevier 1993, pp. 52–107. - Crevier D (1993). AI: The Tumultuous Search for Artificial Intelligence. New York, NY: BasicBooks. ISBN 0-465-02997-3.

  100. Moravec 1988, p. 9. - Moravec H (1988), Mind Children, Harvard University Press, ISBN 978-0-674-57618-6, OCLC 245755104 https://archive.org/details/mindchildren00hans

  101. Russell & Norvig 2021, p. 18. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  102. McCorduck 2004, p. 218; Newquist 1994, pp. 91–112; Crevier 1993, pp. 108–109 - McCorduck P (2004), Machines Who Think (2nd ed.), Natick, MA: A. K. Peters, Ltd., ISBN 978-1-56881-205-2, OCLC 52197627 https://search.worldcat.org/oclc/52197627

  103. Crevier 1993, pp. 52–107; Moravec 1988, p. 9 - Crevier D (1993). AI: The Tumultuous Search for Artificial Intelligence. New York, NY: BasicBooks. ISBN 0-465-02997-3.

  104. Copeland 2004. - Copeland J( (2004). The Essential Turing: the ideas that gave birth to the computer age. Oxford: Clarendon Press. ISBN 0-19-825079-7.

  105. State space search and problem solving: Russell & Norvig 2021, Chpt: 3-6 /wiki/State_space_search

  106. McCorduck 2004, p. 246. - McCorduck P (2004), Machines Who Think (2nd ed.), Natick, MA: A. K. Peters, Ltd., ISBN 978-1-56881-205-2, OCLC 52197627 https://search.worldcat.org/oclc/52197627

  107. McCorduck 2004, pp. 245–250. - McCorduck P (2004), Machines Who Think (2nd ed.), Natick, MA: A. K. Peters, Ltd., ISBN 978-1-56881-205-2, OCLC 52197627 https://search.worldcat.org/oclc/52197627

  108. Russell & Norvig 2021, pp. 19, 106. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  109. Russell & Norvig 2021, p. 19. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  110. Crevier 1993, pp. 51–58, 65–66. - Crevier D (1993). AI: The Tumultuous Search for Artificial Intelligence. New York, NY: BasicBooks. ISBN 0-465-02997-3.

  111. Russell & Norvig 2021, p. 20. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  112. STRIPS and Shakey: Russell & Norvig 2021, p. 20 McCorduck 2004, pp. 268–271 Crevier 1993, pp. 95–96 Newquist 1994, pp. 148–156 Moravec 1988, pp. 14–15 /wiki/Stanford_Research_Institute_Problem_Solver

  113. McCorduck 2004, p. 286, Crevier 1993, pp. 76–79, Russell & Norvig 2021, p. 20 - McCorduck P (2004), Machines Who Think (2nd ed.), Natick, MA: A. K. Peters, Ltd., ISBN 978-1-56881-205-2, OCLC 52197627 https://search.worldcat.org/oclc/52197627

  114. Crevier 1993, pp. 79–83. - Crevier D (1993). AI: The Tumultuous Search for Artificial Intelligence. New York, NY: BasicBooks. ISBN 0-465-02997-3.

  115. Crevier 1993, pp. 164–172. - Crevier D (1993). AI: The Tumultuous Search for Artificial Intelligence. New York, NY: BasicBooks. ISBN 0-465-02997-3.

  116. McCorduck 2004, pp. 291–296. - McCorduck P (2004), Machines Who Think (2nd ed.), Natick, MA: A. K. Peters, Ltd., ISBN 978-1-56881-205-2, OCLC 52197627 https://search.worldcat.org/oclc/52197627

  117. Crevier 1993, pp. 134–139. - Crevier D (1993). AI: The Tumultuous Search for Artificial Intelligence. New York, NY: BasicBooks. ISBN 0-465-02997-3.

  118. This avoided the commonsense knowledge problem, discussed below.

  119. Blocks world: McCorduck 2004, pp. 299–305 Crevier 1993, pp. 83–102 Russell & Norvig 2021, p. 20 Copeland 2000 /wiki/Blocks_world

  120. Blocks world: McCorduck 2004, pp. 299–305 Crevier 1993, pp. 83–102 Russell & Norvig 2021, p. 20 Copeland 2000 /wiki/Blocks_world

  121. Perceptrons in the 60s: Russell & Norvig 2021, p. 21 Crevier 1993, pp. 102–105 McCorduck 2004, pp. 104–107 Schmidhuber 2022 /wiki/Perceptron

  122. Crevier 1993, p. 102. - Crevier D (1993). AI: The Tumultuous Search for Artificial Intelligence. New York, NY: BasicBooks. ISBN 0-465-02997-3.

  123. Quoted in Crevier 1993, p. 102 - Crevier D (1993). AI: The Tumultuous Search for Artificial Intelligence. New York, NY: BasicBooks. ISBN 0-465-02997-3.

  124. Rosenblatt 1962. - Rosenblatt F (1962), Principles of neurodynamics: Perceptrons and the theory of brain mechanisms, vol. 55, Washington DC: Spartan books

  125. Russell & Norvig 2021, pp. 20–21. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  126. Widrow & Lehr 1990. - Widrow B, Lehr M (September 1990). "30 years of adaptive neural networks: perceptron, Madaline, and backpropagation". Proceedings of the IEEE. 78 (9): 1415–1442. doi:10.1109/5.58323. S2CID 195704643. https://ieeexplore.ieee.org/document/58323

  127. Rosen, Nilsson & Adams 1965. - Rosen CA, Nilsson NJ, Adams MB (8 January 1965). "A research and development program in applications of intelligent automata to reconnaissance-phase I. (Proposal for Research SRI No. ESU 65-1)" (PDF). Stanford Research Institute. Archived from the original (PDF) on 16 March 2006. https://web.archive.org/web/20060316081320/http://www.ai.sri.com/pubs/files/rosen65-esu65-1tech.pdf

  128. Nilsson 1984. - Nilsson NJ (1984). "The SRI Artificial Intelligence Center: A Brief History" (PDF). Artificial Intelligence Center, SRI International. Archived from the original (PDF) on 10 August 2022. https://web.archive.org/web/20220810142945/https://www.sri.com/wp-content/uploads/2021/12/635.pdf

  129. Hart et al. 2003. - Hart PE, Nilsson NJ, Perrault R, Mitchell T, Kulikowski CA, Leake DB (15 March 2003). "In Memoriam: Charles Rosen, Norman Nielsen, and Saul Amarel". AI Magazine. 24 (1): 6. doi:10.1609/aimag.v24i1.1683. ISSN 2371-9621. https://ojs.aaai.org/aimagazine/index.php/aimagazine/article/view/1683

  130. Nielson 2005. - Nielson DL (1 January 2005). "Chapter 4: The Life and Times of a Successful SRI Laboratory: Artificial Intelligence and Robotics" (PDF). A HERITAGE OF INNOVATION SRI's First Half Century (1st ed.). SRI International. ISBN 978-0-9745208-0-3. https://www.sri.com/wp-content/uploads/2022/08/A-heritage-of-innovation-The-Life-and-Times-of-a-Successful-SRI-Laboratory-Artificial-Intelligence-and-Robotics.pdf

  131. The hardware diversity was particularly clear in the different technologies used in implementing the adjustable weights. The perceptron machines and the SNARC used potentiometers moved by electric motors. ADALINE used memistors adjusted by electroplating, though they also used simulations on an IBM 1620 computer. The MINOS machines used ferrite cores with multiple holes in them that could be individually blocked, with the degree of blockage representing the weights.[121] /wiki/Stochastic_Neural_Analog_Reinforcement_Calculator

  132. Olazaran Rodriguez 1991. - Olazaran Rodriguez JM (1991). A historical sociology of neural network research] (PDF) (Thesis). University of Edinburgh. Archived from the original (PDF) on 11 November 2022. https://web.archive.org/web/20221111165150/https://era.ed.ac.uk/bitstream/handle/1842/20075/Olazaran-RodriguezJM_1991redux.pdf?sequence=1&isAllowed=y

  133. Minsky & Papert 1969. - Minsky M, Papert S (1969), Perceptrons: An Introduction to Computational Geometry, The MIT Press, ISBN 978-0-262-63111-2, OCLC 16924756 https://archive.org/details/perceptronsintro00mins

  134. Russell & Norvig 2021, p. 22. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  135. Nielson 2005. - Nielson DL (1 January 2005). "Chapter 4: The Life and Times of a Successful SRI Laboratory: Artificial Intelligence and Robotics" (PDF). A HERITAGE OF INNOVATION SRI's First Half Century (1st ed.). SRI International. ISBN 978-0-9745208-0-3. https://www.sri.com/wp-content/uploads/2022/08/A-heritage-of-innovation-The-Life-and-Times-of-a-Successful-SRI-Laboratory-Artificial-Intelligence-and-Robotics.pdf

  136. Olazaran Rodriguez 1991. - Olazaran Rodriguez JM (1991). A historical sociology of neural network research] (PDF) (Thesis). University of Edinburgh. Archived from the original (PDF) on 11 November 2022. https://web.archive.org/web/20221111165150/https://era.ed.ac.uk/bitstream/handle/1842/20075/Olazaran-RodriguezJM_1991redux.pdf?sequence=1&isAllowed=y

  137. Nielson 2005. - Nielson DL (1 January 2005). "Chapter 4: The Life and Times of a Successful SRI Laboratory: Artificial Intelligence and Robotics" (PDF). A HERITAGE OF INNOVATION SRI's First Half Century (1st ed.). SRI International. ISBN 978-0-9745208-0-3. https://www.sri.com/wp-content/uploads/2022/08/A-heritage-of-innovation-The-Life-and-Times-of-a-Successful-SRI-Laboratory-Artificial-Intelligence-and-Robotics.pdf

  138. Olazaran Rodriguez 1991. - Olazaran Rodriguez JM (1991). A historical sociology of neural network research] (PDF) (Thesis). University of Edinburgh. Archived from the original (PDF) on 11 November 2022. https://web.archive.org/web/20221111165150/https://era.ed.ac.uk/bitstream/handle/1842/20075/Olazaran-RodriguezJM_1991redux.pdf?sequence=1&isAllowed=y

  139. Schmidhuber 2022. - Schmidhuber J (2022). "Annotated History of Modern AI and Deep Learning". https://people.idsia.ch/~juergen/

  140. Russell & Norvig 2021, p. 22. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  141. Russell & Norvig 2021, p. 24. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  142. Crevier 1993, p. 105. - Crevier D (1993). AI: The Tumultuous Search for Artificial Intelligence. New York, NY: BasicBooks. ISBN 0-465-02997-3.

  143. Simon & Newell 1958, pp. 7−8 quoted in Crevier 1993, p. 108.Murgia 2023. - Simon HA, Newell A (1958), "Heuristic Problem Solving: The Next Advance in Operations Research", Operations Research, 6: 1–10, doi:10.1287/opre.6.1.1 https://doi.org/10.1287%2Fopre.6.1.1

  144. Simon 1965, p. 96 quoted in Crevier 1993, p. 109 - Simon HA (1965), The Shape of Automation for Men and Management, New York: Harper & Row

  145. Minsky 1967, p. 2 quoted in Crevier 1993, p. 109 - Minsky M (1967), Computation: Finite and Infinite Machines, Englewood Cliffs, N.J.: Prentice-Hall

  146. Darrach 1970. - Darrach B (20 November 1970), "Meet Shaky, the First Electronic Person", Life Magazine, pp. 58–68

  147. Minsky strongly believes he was misquoted.[131][132]

  148. Crevier 1993, pp. 64–65. - Crevier D (1993). AI: The Tumultuous Search for Artificial Intelligence. New York, NY: BasicBooks. ISBN 0-465-02997-3.

  149. Crevier 1993, p. 94. - Crevier D (1993). AI: The Tumultuous Search for Artificial Intelligence. New York, NY: BasicBooks. ISBN 0-465-02997-3.

  150. Howe 1994. - Howe J (November 1994), Artificial Intelligence at Edinburgh University: a Perspective, retrieved 30 August 2007 http://www.inf.ed.ac.uk/about/AIhistory.html

  151. Crevier 1993, p. 51. - Crevier D (1993). AI: The Tumultuous Search for Artificial Intelligence. New York, NY: BasicBooks. ISBN 0-465-02997-3.

  152. McCorduck also notes that funding was mostly under the direction of alumni of the Dartmouth workshop of 1956.[137] /wiki/Dartmouth_workshop

  153. Crevier 1993, p. 65. - Crevier D (1993). AI: The Tumultuous Search for Artificial Intelligence. New York, NY: BasicBooks. ISBN 0-465-02997-3.

  154. Crevier 1993, pp. 68–71; Turkle 1984 - Crevier D (1993). AI: The Tumultuous Search for Artificial Intelligence. New York, NY: BasicBooks. ISBN 0-465-02997-3.

  155. Crevier 1993, pp. 163–196. - Crevier D (1993). AI: The Tumultuous Search for Artificial Intelligence. New York, NY: BasicBooks. ISBN 0-465-02997-3.

  156. Dreyfus 1972. - Dreyfus H (1972), What Computers Can't Do, New York: MIT Press, ISBN 978-0-06-090613-9, OCLC 5056816 https://search.worldcat.org/oclc/5056816

  157. Lighthill 1973. - Lighthill PS (1973), "Artificial Intelligence: A General Survey", Artificial Intelligence: a paper symposium, Science Research Council

  158. Haigh 2023. - Haigh T (December 2023). "There Was No 'First AI Winter'". Communications of the ACM. 66 (12): 35–39. doi:10.1145/3625833. ISSN 0001-0782. https://dl.acm.org/doi/10.1145/3625833

  159. Crevier 1993, p. 143. - Crevier D (1993). AI: The Tumultuous Search for Artificial Intelligence. New York, NY: BasicBooks. ISBN 0-465-02997-3.

  160. Haigh 2023. - Haigh T (December 2023). "There Was No 'First AI Winter'". Communications of the ACM. 66 (12): 35–39. doi:10.1145/3625833. ISSN 0001-0782. https://dl.acm.org/doi/10.1145/3625833

  161. Haigh 2023. - Haigh T (December 2023). "There Was No 'First AI Winter'". Communications of the ACM. 66 (12): 35–39. doi:10.1145/3625833. ISSN 0001-0782. https://dl.acm.org/doi/10.1145/3625833

  162. Haigh 2023. - Haigh T (December 2023). "There Was No 'First AI Winter'". Communications of the ACM. 66 (12): 35–39. doi:10.1145/3625833. ISSN 0001-0782. https://dl.acm.org/doi/10.1145/3625833

  163. Nilsson 2009, p. 1. - Nilsson N (30 October 2009). The Quest for Artificial Intelligence. Cambridge University Press. ISBN 978-0-52-112293-1.

  164. Russell and Norvig wrote "in almost all cases, these early systems failed on more difficult tasks."[146]

  165. Crevier 1993, p. 146. - Crevier D (1993). AI: The Tumultuous Search for Artificial Intelligence. New York, NY: BasicBooks. ISBN 0-465-02997-3.

  166. Bruce Buchanan wrote: "Early programs were necessarily limited in scope by the size and speed of memory"[148] /wiki/Bruce_Buchanan

  167. Crevier 1993, pp. 146–148. - Crevier D (1993). AI: The Tumultuous Search for Artificial Intelligence. New York, NY: BasicBooks. ISBN 0-465-02997-3.

  168. Moravec 1976. - Moravec H (1976), The Role of Raw Power in Intelligence, archived from the original on 3 March 2016, retrieved 16 October 2008 https://web.archive.org/web/20160303232511/http://www.frc.ri.cmu.edu/users/hpm/project.archive/general.articles/1975/Raw.Power.html

  169. History would prove Moravec right about applications like computer vision. Moravec estimated that simply matching the edge and motion detection capabilities of the human retina in real time would require a general-purpose computer capable of 1000 million instructions per second (MIPS). In 1976, the fastest supercomputer, the $8 million Cray-1 was only capable of 130 MIPS, and a typical desktop computer had 1 MIPS. As of 2011, practical computer vision applications require 10,000 to 1,000,000 MIPS.[151] /wiki/Edge_detection

  170. Russell & Norvig 2021, p. 21. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  171. Lighthill 1973. - Lighthill PS (1973), "Artificial Intelligence: A General Survey", Artificial Intelligence: a paper symposium, Science Research Council

  172. McCorduck 2004, p. 456. - McCorduck P (2004), Machines Who Think (2nd ed.), Natick, MA: A. K. Peters, Ltd., ISBN 978-1-56881-205-2, OCLC 52197627 https://search.worldcat.org/oclc/52197627

  173. Brooks 2002. - Brooks R (2002), Flesh and Machines, Pantheon Books

  174. McCorduck 2004, p. 456. - McCorduck P (2004), Machines Who Think (2nd ed.), Natick, MA: A. K. Peters, Ltd., ISBN 978-1-56881-205-2, OCLC 52197627 https://search.worldcat.org/oclc/52197627

  175. Moravec 1988, pp. 15–16. - Moravec H (1988), Mind Children, Harvard University Press, ISBN 978-0-674-57618-6, OCLC 245755104 https://archive.org/details/mindchildren00hans

  176. Brooks 2002. - Brooks R (2002), Flesh and Machines, Pantheon Books

  177. Commonsense knowledge: McCorduck 2004, pp. 300 & 421 Crevier 1993, pp. 113–114 Moravec 1988, p. 13 Lenat & Guha 1989, (Introduction) - McCorduck P (2004), Machines Who Think (2nd ed.), Natick, MA: A. K. Peters, Ltd., ISBN 978-1-56881-205-2, OCLC 52197627 https://search.worldcat.org/oclc/52197627

  178. Such as the frame, ramification and qualification problems, as well as the difficulty of default reasoning and word-sense disambiguation. /wiki/Frame_problem

  179. Russell and Norvig write: "[M]any of the concepts we name in language fail, on closer inspection, to have the logically defined necessary and sufficient conditions that early AI researchers hoped to capture in axiomatic form."[125]

  180. Quoted in Crevier 1993, p. 175 - Crevier D (1993). AI: The Tumultuous Search for Artificial Intelligence. New York, NY: BasicBooks. ISBN 0-465-02997-3.

  181. ALPAC: McCorduck 2004, pp. 280–281 Crevier 1993, p. 110 Russell & Norvig 2021, p. 21 NRC 1999, under "Success in Speech Recognition". /wiki/ALPAC

  182. Lighthill report: Crevier 1993, p. 117 Howe 1994 Lighthill 1973 /wiki/Lighthill_report

  183. Lighthill 1973. - Lighthill PS (1973), "Artificial Intelligence: A General Survey", Artificial Intelligence: a paper symposium, Science Research Council

  184. Russell & Norvig 2021, p. 21. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  185. John McCarthy wrote in response that "the combinatorial explosion problem has been recognized in AI from the beginning"[159] /wiki/John_McCarthy_(computer_scientist)

  186. Crevier 1993, pp. 115–116. - Crevier D (1993). AI: The Tumultuous Search for Artificial Intelligence. New York, NY: BasicBooks. ISBN 0-465-02997-3.

  187. This account is based on Crevier 1993, pp. 115–116. Other views include McCorduck 2004, pp. 306–313 and NRC 1999 under "Success in Speech Recognition". - Crevier D (1993). AI: The Tumultuous Search for Artificial Intelligence. New York, NY: BasicBooks. ISBN 0-465-02997-3.

  188. Crevier 1993, p. 115. - Crevier D (1993). AI: The Tumultuous Search for Artificial Intelligence. New York, NY: BasicBooks. ISBN 0-465-02997-3.

  189. Moravec explains, "Their initial promises to DARPA had been much too optimistic. Of course, what they delivered stopped considerably short of that. But they felt they couldn't in their next proposal promise less than in the first one, so they promised more."[161]

  190. NRC 1999, under "Shift to Applied Research Increases Investment.". - NRC (1999), "Developments in Artificial Intelligence", Funding a Revolution: Government Support for Computing Research, National Academy Press, ISBN 978-0-309-06278-7, OCLC 246584055 https://archive.org/details/fundingrevolutio00nati

  191. While the autonomous tank was a failure, the battle management system (called "DART") proved to be enormously successful, saving billions in the first Gulf War, repaying the investment and justifying the DARPA's pragmatic policy, at least as far as DARPA was concerned.[163] /wiki/Dynamic_Analysis_and_Replanning_Tool

  192. Haigh 2023. - Haigh T (December 2023). "There Was No 'First AI Winter'". Communications of the ACM. 66 (12): 35–39. doi:10.1145/3625833. ISSN 0001-0782. https://dl.acm.org/doi/10.1145/3625833

  193. Lucas and Penrose' critique of AI: Crevier 1993, p. 22 Russell & Norvig 2021, pp. 983–984 Hofstadter 1999, pp. 471–477 Lucas original argument: Lucas 1961 - Crevier D (1993). AI: The Tumultuous Search for Artificial Intelligence. New York, NY: BasicBooks. ISBN 0-465-02997-3.

  194. "Know-how" is Dreyfus' term. Dreyfus makes a distinction between "knowing how" and "knowing that", a modern version of Heidegger's distinction of ready-to-hand and present-at-hand.[165] /wiki/Heidegger

  195. Dreyfus' critique of artificial intelligence: McCorduck 2004, pp. 211–239 Crevier 1993, pp. 120–132 Russell & Norvig 2021, pp. 981–982 Dreyfus' version: Dreyfus 1965 Dreyfus 1972 Dreyfus & Dreyfus 1986 /wiki/Dreyfus%27_critique_of_artificial_intelligence

  196. Searle's critique of AI: McCorduck 2004, pp. 443–445 Crevier 1993, pp. 269–271 Russell & Norvig 2021, pp. 985–986 Searle's version: Searle 1980 - McCorduck P (2004), Machines Who Think (2nd ed.), Natick, MA: A. K. Peters, Ltd., ISBN 978-1-56881-205-2, OCLC 52197627 https://search.worldcat.org/oclc/52197627

  197. Quoted in Crevier 1993, p. 143 - Crevier D (1993). AI: The Tumultuous Search for Artificial Intelligence. New York, NY: BasicBooks. ISBN 0-465-02997-3.

  198. Quoted in Crevier 1993, p. 122 - Crevier D (1993). AI: The Tumultuous Search for Artificial Intelligence. New York, NY: BasicBooks. ISBN 0-465-02997-3.

  199. Weizenbaum said: "I became the only member of the AI community to be seen eating lunch with Dreyfus. And I deliberately made it plain that theirs was not the way to treat a human being."[170]

  200. Newquist 1994, p. 276. - Newquist HP (1994), The Brain Makers: Genius, Ego, And Greed in the Quest For Machines That Think, New York: Macmillan/SAMS, ISBN 978-0-9885937-1-8, OCLC 313139906 https://search.worldcat.org/oclc/313139906

  201. Colby, Watt & Gilbert 1966, p. 148. - Colby KM, Watt JB, Gilbert JP (1966), "A Computer Method of Psychotherapy: Preliminary Communication", The Journal of Nervous and Mental Disease, vol. 142, no. 2, pp. 148–152, doi:10.1097/00005053-196602000-00005, PMID 5936301, S2CID 36947398 https://exhibits.stanford.edu/feigenbaum/catalog/hk334rq4790

  202. Weizenbaum 1976, pp. 5, 6. - Weizenbaum J (1976), Computer Power and Human Reason, W.H. Freeman & Company, ISBN 978-0-14-022535-8, OCLC 10952283 https://search.worldcat.org/oclc/10952283

  203. Colby and his colleagues later also developed chatterbot-like "computer simulations of paranoid processes (PARRY)" to "make intelligible paranoid processes in explicit symbol processing terms."[174] /wiki/Chatterbot

  204. Weizenbaum's critique of AI: McCorduck 2004, pp. 356–373 Crevier 1993, pp. 132–144 Russell & Norvig 2021, p. 1001 and see Weizenbaum 1976 - McCorduck P (2004), Machines Who Think (2nd ed.), Natick, MA: A. K. Peters, Ltd., ISBN 978-1-56881-205-2, OCLC 52197627 https://search.worldcat.org/oclc/52197627

  205. McCorduck 2004, p. 51. - McCorduck P (2004), Machines Who Think (2nd ed.), Natick, MA: A. K. Peters, Ltd., ISBN 978-1-56881-205-2, OCLC 52197627 https://search.worldcat.org/oclc/52197627

  206. Russell & Norvig 2021, p. 19. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  207. Russell & Norvig 2021, p. 19. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  208. McCorduck 2004, p. 51. - McCorduck P (2004), Machines Who Think (2nd ed.), Natick, MA: A. K. Peters, Ltd., ISBN 978-1-56881-205-2, OCLC 52197627 https://search.worldcat.org/oclc/52197627

  209. Crevier 1993, pp. 190–192. - Crevier D (1993). AI: The Tumultuous Search for Artificial Intelligence. New York, NY: BasicBooks. ISBN 0-465-02997-3.

  210. Crevier 1993, pp. 193–196. - Crevier D (1993). AI: The Tumultuous Search for Artificial Intelligence. New York, NY: BasicBooks. ISBN 0-465-02997-3.

  211. Crevier 1993, pp. 145–149, 258–63. - Crevier D (1993). AI: The Tumultuous Search for Artificial Intelligence. New York, NY: BasicBooks. ISBN 0-465-02997-3.

  212. Wason & Shapiro (1966) showed that people do poorly on completely abstract problems, but if the problem is restated to allow the use of intuitive social intelligence, performance dramatically improves. (See Wason selection task) Kahneman, Slovic & Tversky (1982) have shown that people are terrible at elementary problems that involve uncertain reasoning. (See list of cognitive biases for several examples). Eleanor Rosch's work is described in Lakoff 1987. Kahnmann published a more general theory of symbolic cognition and other kinds of thinking in his book Thinking Fast and Slow (2011) - Wason PC, Shapiro D (1966). "Reasoning". In Foss, B. M. (ed.). New horizons in psychology. Harmondsworth: Penguin. Retrieved 18 November 2019. https://archive.org/details/newhorizonsinpsy0000foss

  213. An early example of McCarthy's position was in the journal Science where he said "This is AI, so we don't care if it's psychologically real" (Kolata 1982), and he recently reiterated his position at the AI@50 conference where he said "Artificial intelligence is not, by definition, simulation of human intelligence" (Maker 2006). /wiki/John_McCarthy_(computer_scientist)

  214. Neats vs. scruffies: McCorduck 2004, pp. 421–424 (who picks up the state of the debate in 1984). Crevier 1993, p. 168 (who documents Schank's original use of the term). Russell & Norvig 2021, pp. 19–20 (who describe MIT's approach as "anti-logic") /wiki/Neats_vs._scruffies

  215. Another aspect of the conflict was called "the procedural/declarative distinction" but did not prove to be influential in later AI research.

  216. Frame (artificial intelligence): McCorduck 2004, pp. 305–306 Crevier 1993, pp. 170–173, 246 Russell & Norvig 2021, p. 23. Minsky's frame paper: Minsky 1974. /wiki/Frame_(artificial_intelligence)

  217. Hayes 1981. - Hayes P (1981). "The logic of frames". In Kaufmann M (ed.). Readings in artificial intelligence. pp. 451–458.

  218. Reiter 1978. - Reiter R (1978). "On reasoning by default". American Journal of Computational Linguistics: 29–37.

  219. Reiter 1978. - Reiter R (1978). "On reasoning by default". American Journal of Computational Linguistics: 29–37.

  220. Clark 1977. - Clark K (1977). "Negation as Failure". Logic and Data Bases. Boston, MA: Springer US. pp. 293–322. doi:10.1007/978-1-4684-3384-5_11. ISBN 978-1-4684-3386-9. https://doi.org/10.1007%2F978-1-4684-3384-5_11

  221. Russell & Norvig 2021, p. 24. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  222. Expert systems: Crevier 1993, pp. 148–159 Newquist 1994, p. 271 Russell & Norvig 2021, pp. 22–24 /wiki/Expert_system

  223. McCorduck 2004, pp. 327–335. - McCorduck P (2004), Machines Who Think (2nd ed.), Natick, MA: A. K. Peters, Ltd., ISBN 978-1-56881-205-2, OCLC 52197627 https://search.worldcat.org/oclc/52197627

  224. Russell & Norvig 2021, p. 22. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  225. Russell & Norvig 2021, p. 24. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  226. Russell & Norvig 2021, p. 22. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  227. Crevier 1993, pp. 158–159. - Crevier D (1993). AI: The Tumultuous Search for Artificial Intelligence. New York, NY: BasicBooks. ISBN 0-465-02997-3.

  228. Crevier 1993, p. 198. - Crevier D (1993). AI: The Tumultuous Search for Artificial Intelligence. New York, NY: BasicBooks. ISBN 0-465-02997-3.

  229. Newquist 1994, p. 259. - Newquist HP (1994), The Brain Makers: Genius, Ego, And Greed in the Quest For Machines That Think, New York: Macmillan/SAMS, ISBN 978-0-9885937-1-8, OCLC 313139906 https://search.worldcat.org/oclc/313139906

  230. Commercial expert systems: McCorduck 2004, pp. 434–435 Crevier 1993, pp. 161–162, 197–203 {{Harvnb|Russell|Norvig|20 Newquist 1994, p. 275 - McCorduck P (2004), Machines Who Think (2nd ed.), Natick, MA: A. K. Peters, Ltd., ISBN 978-1-56881-205-2, OCLC 52197627 https://search.worldcat.org/oclc/52197627

  231. Fifth generation computer: McCorduck 2004, pp. 436–441 Newquist 1994, pp. 231–240 Crevier 1993, p. 211 Russell & Norvig 2021, p. 23 Feigenbaum & McCorduck 1983 /wiki/Fifth_generation_computer

  232. Crevier 1993, p. 195. - Crevier D (1993). AI: The Tumultuous Search for Artificial Intelligence. New York, NY: BasicBooks. ISBN 0-465-02997-3.

  233. Russell & Norvig 2021, p. 23. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  234. Crevier 1993, p. 240. - Crevier D (1993). AI: The Tumultuous Search for Artificial Intelligence. New York, NY: BasicBooks. ISBN 0-465-02997-3.

  235. Russell & Norvig 2021, p. 23. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  236. McCorduck 2004, pp. 426–432. - McCorduck P (2004), Machines Who Think (2nd ed.), Natick, MA: A. K. Peters, Ltd., ISBN 978-1-56881-205-2, OCLC 52197627 https://search.worldcat.org/oclc/52197627

  237. NRC 1999, under "Shift to Applied Research Increases Investment". - NRC (1999), "Developments in Artificial Intelligence", Funding a Revolution: Government Support for Computing Research, National Academy Press, ISBN 978-0-309-06278-7, OCLC 246584055 https://archive.org/details/fundingrevolutio00nati

  238. McCorduck 2004, p. 299. - McCorduck P (2004), Machines Who Think (2nd ed.), Natick, MA: A. K. Peters, Ltd., ISBN 978-1-56881-205-2, OCLC 52197627 https://search.worldcat.org/oclc/52197627

  239. McCorduck 2004, p. 421. - McCorduck P (2004), Machines Who Think (2nd ed.), Natick, MA: A. K. Peters, Ltd., ISBN 978-1-56881-205-2, OCLC 52197627 https://search.worldcat.org/oclc/52197627

  240. Knowledge revolution: McCorduck 2004, pp. 266–276, 298–300, 314, 421 Newquist 1994, pp. 255–267 Russell & Norvig 2021, p. 23 - McCorduck P (2004), Machines Who Think (2nd ed.), Natick, MA: A. K. Peters, Ltd., ISBN 978-1-56881-205-2, OCLC 52197627 https://search.worldcat.org/oclc/52197627

  241. Cyc and ontological engineering McCorduck 2004, p. 489 Crevier 1993, pp. 239–243 Newquist 1994, pp. 431–455 Russell & Norvig 2021, pp. 314−316 Lenat & Guha 1989 /wiki/Cyc

  242. Sejnowski 2018. - Sejnowski TJ (23 October 2018). The Deep Learning Revolution (1st ed.). Cambridge, Massachusetts London, England: The MIT Press. pp. 93–94. ISBN 978-0-262-03803-4.

  243. Versions of backpropagation had been developed in several fields, most directly as the reverse mode of automatic differentiation published by Seppo Linnainmaa (1970). It was applied to neural networks in the 1970s by Paul Werbos.[124] /wiki/Automatic_differentiation

  244. Russell & Norvig 2021, p. 24. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  245. Crevier 1993, pp. 214–215. - Crevier D (1993). AI: The Tumultuous Search for Artificial Intelligence. New York, NY: BasicBooks. ISBN 0-465-02997-3.

  246. Russell & Norvig 2021, p. 24. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  247. Russell & Norvig 2021, p. 24. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  248. Qian N, Sejnovski TJ (1988). "Predicting the secondary structure of globular proteins using neural network models". Journal of Molecular Biology. 202 (4): 865–884. doi:10.1016/0022-2836(88)90564-5. https://www.sciencedirect.com/science/article/pii/0022283688905645

  249. Rost B, Sander C (1993). "Improved prediction of protein secondary structure by use of sequence profiles and neural networks". Proceedings of the National Academy of Sciences. 90 (16): 7558–7562. doi:10.1073/pnas.90.16.7558. PMC 47181. https://www.pnas.org/doi/10.1073/pnas.90.16.7558

  250. McGuffin LJ, Bryson K, Jones DT (2000). "The PSIPRED protein structure prediction server". Bioinformatics. 16 (4): 404–405. doi:10.1093/bioinformatics/16.4.404. https://academic.oup.com/bioinformatics/article/16/4/404/187312

  251. Russell & Norvig 2021, p. 26. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  252. Christian 2020, pp. 21–22. - Christian B (2020). The Alignment Problem: Machine learning and human values. W. W. Norton & Company. ISBN 978-0-393-86833-3. OCLC 1233266753. https://search.worldcat.org/oclc/1233266753

  253. McCorduck 2004, pp. 454–462. - McCorduck P (2004), Machines Who Think (2nd ed.), Natick, MA: A. K. Peters, Ltd., ISBN 978-1-56881-205-2, OCLC 52197627 https://search.worldcat.org/oclc/52197627

  254. Hans Moravec wrote: "I am confident that this bottom-up route to artificial intelligence will one date meet the traditional top-down route more than half way, ready to provide the real world competence and the commonsense knowledge that has been so frustratingly elusive in reasoning programs. Fully intelligent machines will result when the metaphorical golden spike is driven uniting the two efforts."[209] /wiki/Hans_Moravec

  255. Crevier 1993, pp. 183–190. - Crevier D (1993). AI: The Tumultuous Search for Artificial Intelligence. New York, NY: BasicBooks. ISBN 0-465-02997-3.

  256. Brooks 1990. - Brooks RA (1990). "Elephants Don't Play Chess" (PDF). Robotics and Autonomous Systems. 6 (1–2): 3–15. doi:10.1016/S0921-8890(05)80025-9. http://people.csail.mit.edu/brooks/papers/elephants.pdf

  257. Brooks 1990, p. 3. - Brooks RA (1990). "Elephants Don't Play Chess" (PDF). Robotics and Autonomous Systems. 6 (1–2): 3–15. doi:10.1016/S0921-8890(05)80025-9. http://people.csail.mit.edu/brooks/papers/elephants.pdf

  258. See, for example, Lakoff & Johnson 1999 - Lakoff G, Johnson M (1999). Philosophy in the flesh: The embodied mind and its challenge to western thought. Basic Books. ISBN 978-0-465-05674-3. https://www.basicbooks.com/titles/george-lakoff/philosophy-in-the-flesh/9780465056743/

  259. Pollack 1984. - Pollack A (11 October 1984). "Technology; Fuzzy Logic For Computers". The New York Times. https://www.nytimes.com/1984/10/11/business/technology-fuzzy-logic-for-computers.html

  260. Pollack 1989. - Pollack A (2 April 1989). "Fuzzy Computer Theory: How to Mimic the Mind?". The New York Times. https://www.nytimes.com/1989/04/02/us/fuzzy-computer-theory-how-to-mimic-the-mind.html

  261. Pearl 1988. - Pearl J (1988), Probabilistic Reasoning in Intelligent Systems: Networks of Plausible Inference, San Mateo, California: Morgan Kaufmann, ISBN 978-1-55860-479-7, OCLC 249625842 https://search.worldcat.org/oclc/249625842

  262. Russell & Norvig 2021, p. 25. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  263. Russell & Norvig 2021, p. 25. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  264. Russell & Norvig 2021, p. 25. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  265. Poole, Mackworth & Goebel 1998. - Poole D, Mackworth A, Goebel R (1998), Computational Intelligence: A Logical Approach, Oxford University Press., ISBN 978-0-19-510270-3 https://archive.org/details/computationalint00pool

  266. Russell & Norvig 2021, Section 23. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  267. Christian 2020, pp. 120–124. - Christian B (2020). The Alignment Problem: Machine learning and human values. W. W. Norton & Company. ISBN 978-0-393-86833-3. OCLC 1233266753. https://search.worldcat.org/oclc/1233266753

  268. Russell & Norvig 2021, p. 819. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  269. Christian 2020, p. 124. - Christian B (2020). The Alignment Problem: Machine learning and human values. W. W. Norton & Company. ISBN 978-0-393-86833-3. OCLC 1233266753. https://search.worldcat.org/oclc/1233266753

  270. Christian 2020, pp. 152–156. - Christian B (2020). The Alignment Problem: Machine learning and human values. W. W. Norton & Company. ISBN 978-0-393-86833-3. OCLC 1233266753. https://search.worldcat.org/oclc/1233266753

  271. Russell & Norvig 2021, p. 819. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  272. Christian 2020, p. 125. - Christian B (2020). The Alignment Problem: Machine learning and human values. W. W. Norton & Company. ISBN 978-0-393-86833-3. OCLC 1233266753. https://search.worldcat.org/oclc/1233266753

  273. Russell & Norvig 2021, p. 819. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  274. Christian 2020, pp. 127–129. - Christian B (2020). The Alignment Problem: Machine learning and human values. W. W. Norton & Company. ISBN 978-0-393-86833-3. OCLC 1233266753. https://search.worldcat.org/oclc/1233266753

  275. Russell & Norvig 2021, pp. 25, 820. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  276. Russell & Norvig 2021, pp. 25, 820. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  277. Christian 2020, p. 140. - Christian B (2020). The Alignment Problem: Machine learning and human values. W. W. Norton & Company. ISBN 978-0-393-86833-3. OCLC 1233266753. https://search.worldcat.org/oclc/1233266753

  278. Christian 2020, p. 141. - Christian B (2020). The Alignment Problem: Machine learning and human values. W. W. Norton & Company. ISBN 978-0-393-86833-3. OCLC 1233266753. https://search.worldcat.org/oclc/1233266753

  279. Christian 2020, p. ?. - Christian B (2020). The Alignment Problem: Machine learning and human values. W. W. Norton & Company. ISBN 978-0-393-86833-3. OCLC 1233266753. https://search.worldcat.org/oclc/1233266753

  280. Russell & Norvig 2021, p. 820. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  281. Schultz, Dayan & Montague 1997. - Schultz W, Dayan P, Montague PR (14 March 1997). "A Neural Substrate of Prediction and Reward". Science. 275 (5306): 1593–1599. doi:10.1126/science.275.5306.1593. PMID 9054347. https://doi.org/10.1126%2Fscience.275.5306.1593

  282. Russell & Norvig 2021, p. 822. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  283. Newquist 1994, pp. 501, 511. - Newquist HP (1994), The Brain Makers: Genius, Ego, And Greed in the Quest For Machines That Think, New York: Macmillan/SAMS, ISBN 978-0-9885937-1-8, OCLC 313139906 https://search.worldcat.org/oclc/313139906

  284. McCorduck 2004, p. 424. - McCorduck P (2004), Machines Who Think (2nd ed.), Natick, MA: A. K. Peters, Ltd., ISBN 978-1-56881-205-2, OCLC 52197627 https://search.worldcat.org/oclc/52197627

  285. AI winter was first used as the title of a seminar on the subject for the Association for the Advancement of Artificial Intelligence.[235] /wiki/AI_winter

  286. Russell & Norvig 2021, p. 24. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  287. Lisp machine crisis: Newquist 1994, pp. 359–379 McCorduck 2004, p. 435 Crevier 1993, pp. 209–210 /wiki/Lisp_machine

  288. Expert systems failure (and the reason for it): Russell & Norvig 2021, p. 24 (inability to handle uncertain reasoning or to learn) McCorduck 2004, p. 435 (institutional issues) Newquist 1994, pp. 258–283 (limited deployment after development) Crevier 1993, pp. 204–208 (the difficulty of truth maintenance, i.e., learning and updating) Lenat & Guha 1989, Introduction (brittleness and the inability to handle extensive qualification.) /wiki/Expert_system

  289. McCorduck 2004, pp. 430–431. - McCorduck P (2004), Machines Who Think (2nd ed.), Natick, MA: A. K. Peters, Ltd., ISBN 978-1-56881-205-2, OCLC 52197627 https://search.worldcat.org/oclc/52197627

  290. End of the Fifth generation computer initiative: McCorduck 2004, p. 441 Crevier 1993, p. 212 Newquist 1994, p. 476 /wiki/Fifth_generation_computer

  291. McCorduck writes "Two and a half decades later, we can see that the Japanese didn't quite meet all of those ambitious goals."[240]

  292. Newquist 1994, p. 440. - Newquist HP (1994), The Brain Makers: Genius, Ego, And Greed in the Quest For Machines That Think, New York: Macmillan/SAMS, ISBN 978-0-9885937-1-8, OCLC 313139906 https://search.worldcat.org/oclc/313139906

  293. Newquist 1994, p. 440. - Newquist HP (1994), The Brain Makers: Genius, Ego, And Greed in the Quest For Machines That Think, New York: Macmillan/SAMS, ISBN 978-0-9885937-1-8, OCLC 313139906 https://search.worldcat.org/oclc/313139906

  294. See Applications of artificial intelligence § Computer science /wiki/Applications_of_artificial_intelligence#Computer_science

  295. NRC 1999, Artificial Intelligence in the 90s. - NRC (1999), "Developments in Artificial Intelligence", Funding a Revolution: Government Support for Computing Research, National Academy Press, ISBN 978-0-309-06278-7, OCLC 246584055 https://archive.org/details/fundingrevolutio00nati

  296. Kurzweil 2005, p. 264. - Kurzweil R (2005), The Singularity is Near, Viking Press, ISBN 978-0-14-303788-0, OCLC 71826177 https://search.worldcat.org/oclc/71826177

  297. The Economist 2007. - The Economist (7 June 2007), "Are You Talking to Me?", The Economist, retrieved 16 October 2008 http://www.economist.com/science/tq/displaystory.cfm?story_id=9249338

  298. CNN 2006. - "AI set to exceed human brain power", CNN.com, 26 July 2006, retrieved 16 October 2007 http://www.cnn.com/2006/TECH/science/07/24/ai.bostrom/

  299. CNN 2006. - "AI set to exceed human brain power", CNN.com, 26 July 2006, retrieved 16 October 2007 http://www.cnn.com/2006/TECH/science/07/24/ai.bostrom/

  300. Olsen 2004. - Olsen S (10 May 2004), Newsmaker: Google's man behind the curtain, CNET, retrieved 17 October 2008 http://news.cnet.com/Googles-man-behind-the-curtain/2008-1024_3-5208228.html

  301. Olsen 2006. - Olsen S (18 August 2006), Spying an intelligent search engine, CNET, retrieved 17 October 2008 http://news.cnet.com/Spying-an-intelligent-search-engine/2100-1032_3-6107048.html

  302. AI effect, AI behind the scenes in the 90s & 2000s: McCorduck 2004, p. 423 Kurzweil 2005, p. 265 Hofstadter 1999, p. 601 Newquist 1994, p. 445 /wiki/AI_effect

  303. CNN 2006. - "AI set to exceed human brain power", CNN.com, 26 July 2006, retrieved 16 October 2007 http://www.cnn.com/2006/TECH/science/07/24/ai.bostrom/

  304. The Economist 2007. - The Economist (7 June 2007), "Are You Talking to Me?", The Economist, retrieved 16 October 2008 http://www.economist.com/science/tq/displaystory.cfm?story_id=9249338

  305. Tascarella 2006. - Tascarella P (14 August 2006), "Robotics firms find fundraising struggle, with venture capital shy", Pittsburgh Business Times, retrieved 15 March 2016 http://www.bizjournals.com/pittsburgh/stories/2006/08/14/focus3.html?b=1155528000%5E1329573

  306. Newquist 1994, p. 532. - Newquist HP (1994), The Brain Makers: Genius, Ego, And Greed in the Quest For Machines That Think, New York: Macmillan/SAMS, ISBN 978-0-9885937-1-8, OCLC 313139906 https://search.worldcat.org/oclc/313139906

  307. Markoff 2005. - Markoff J (14 October 2005), "Behind Artificial Intelligence, a Squadron of Bright Real People", The New York Times, retrieved 16 October 2008 https://www.nytimes.com/2005/10/14/technology/14artificial.html?_r=1&ei=5070&en=11ab55edb7cead5e&ex=1185940800&adxnnl=1&adxnnlx=1185805173-o7WsfW7qaP0x5/NUs1cQCQ&oref=slogin

  308. McCorduck 2004, pp. 486–487. - McCorduck P (2004), Machines Who Think (2nd ed.), Natick, MA: A. K. Peters, Ltd., ISBN 978-1-56881-205-2, OCLC 52197627 https://search.worldcat.org/oclc/52197627

  309. Russell & Norvig 2021, pp. 24–25. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  310. McCorduck 2004, pp. 471–478. - McCorduck P (2004), Machines Who Think (2nd ed.), Natick, MA: A. K. Peters, Ltd., ISBN 978-1-56881-205-2, OCLC 52197627 https://search.worldcat.org/oclc/52197627

  311. Russell & Norvig 2021, chpt. 2. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  312. Russell and Norvig wrote "The whole-agent view is now widely accepted."[256]

  313. Carl Hewitt's Actor model anticipated the modern definition of intelligent agents. (Hewitt, Bishop & Steiger 1973) Both John Doyle (Doyle 1983) and Marvin Minsky's popular classic The Society of Mind (Minsky 1986) used the word "agent". Other "modular" proposals included Rodney Brook's subsumption architecture, object-oriented programming and others. /wiki/Carl_Hewitt

  314. Russell & Norvig 2021, p. 61. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  315. This is how the most widely used textbooks of the 21st century define artificial intelligence, such as Russell and Norvig, 2021; Padgham and Winikoff, 2004; Jones, 2007; Poole and Mackworth, 2017.[256]

  316. McCorduck 2004, p. 478. - McCorduck P (2004), Machines Who Think (2nd ed.), Natick, MA: A. K. Peters, Ltd., ISBN 978-1-56881-205-2, OCLC 52197627 https://search.worldcat.org/oclc/52197627

  317. McCorduck 2004, pp. 480–483. - McCorduck P (2004), Machines Who Think (2nd ed.), Natick, MA: A. K. Peters, Ltd., ISBN 978-1-56881-205-2, OCLC 52197627 https://search.worldcat.org/oclc/52197627

  318. Russell & Norvig 2021, p. 28. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  319. Ray Kurzweil wrote that the improvement in computer chess "is governed only by the brute force expansion of computer hardware."[260] /wiki/Ray_Kurzweil

  320. Cycle time of Ferranti Mark 1 was 1.2 milliseconds, which is arguably equivalent to about 833 flops. Deep Blue ran at 11.38 gigaflops (and this does not even take into account Deep Blue's special-purpose hardware for chess). Very approximately, these differ by a factor of 107.[citation needed] /wiki/Ferranti_Mark_1

  321. LeCun, Bengio & Hinton 2015. - LeCun Y, Bengio Y, Hinton G (2015). "Deep learning" (PDF). Nature. 521 (7553): 436–444. Bibcode:2015Natur.521..436L. doi:10.1038/nature14539. PMID 26017442. S2CID 3074096. https://hal.science/hal-04206682/file/Lecun2015.pdf

  322. Lohr 2016. - Lohr S (17 October 2016), "IBM Is Counting on Its Bet on Watson, and Paying Big Money for It", New York Times https://www.nytimes.com/2016/10/17/technology/ibm-is-counting-on-its-bet-on-watson-and-paying-big-money-for-it.html?emc=edit_th_20161017&nl=todaysheadlines&nlid=62816440

  323. Russell & Norvig 2021, pp. 26–27. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  324. Russell & Norvig 2021, p. 26. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  325. Quoted in Christian 2020, p. 22 - Christian B (2020). The Alignment Problem: Machine learning and human values. W. W. Norton & Company. ISBN 978-0-393-86833-3. OCLC 1233266753. https://search.worldcat.org/oclc/1233266753

  326. Christian 2020, p. 31. - Christian B (2020). The Alignment Problem: Machine learning and human values. W. W. Norton & Company. ISBN 978-0-393-86833-3. OCLC 1233266753. https://search.worldcat.org/oclc/1233266753

  327. Christian 2020, pp. 22–23. - Christian B (2020). The Alignment Problem: Machine learning and human values. W. W. Norton & Company. ISBN 978-0-393-86833-3. OCLC 1233266753. https://search.worldcat.org/oclc/1233266753

  328. Russell & Norvig 2021, p. 26. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  329. Christian 2020, p. 6. - Christian B (2020). The Alignment Problem: Machine learning and human values. W. W. Norton & Company. ISBN 978-0-393-86833-3. OCLC 1233266753. https://search.worldcat.org/oclc/1233266753

  330. McKinsey & Co 2011. - "Big data: The next frontier for innovation, competition, and productivity". McKinsey.com. 1 May 2011. https://www.mckinsey.com/capabilities/mckinsey-digital/our-insights/big-data-the-next-frontier-for-innovation

  331. Markoff 2011. - Markoff J (16 February 2011). "On 'Jeopardy!' Watson Win Is All but Trivial". The New York Times. https://www.nytimes.com/2011/02/17/science/17jeopardy-watson.html

  332. Russell & Norvig 2021, p. 26. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  333. AlexNet had 650,000 neurons and trained using ImageNet, augmented with reversed, cropped and tinted images. The model also used Geoffrey Hinton's dropout technique and a rectified linear output function, both relatively new developments at the time.[270] /wiki/AlexNet

  334. Christian 2020, p. 24. - Christian B (2020). The Alignment Problem: Machine learning and human values. W. W. Norton & Company. ISBN 978-0-393-86833-3. OCLC 1233266753. https://search.worldcat.org/oclc/1233266753

  335. Russell & Norvig 2021, p. 26. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  336. Several other laboratories had developed systems that, like AlexNet, used GPU chips and performed nearly as well as AlexNet,[124] but AlexNet proved to be the most influential.

  337. Russell & Norvig 2021, pp. 26–27. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  338. Russell & Norvig 2021, p. 27. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  339. Russell & Norvig 2021, p. 27. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  340. See History of AI § The problems above, where Hans Moravec predicted that raw power would eventually make AI "easy". /wiki/History_of_AI#The_problems

  341. Russell & Norvig 2021, pp. 26–27. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  342. Russell & Norvig 2021, pp. 26–27. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  343. Russell & Norvig 2021, pp. 33, 1004. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  344. Russell 2020. - Russell SJ (2020). Human compatible: Artificial intelligence and the problem of control. Penguin Random House. ISBN 9780525558637. OCLC 1113410915. https://www.penguinrandomhouse.com/books/566677/human-compatible-by-stuart-russell/

  345. Russell & Norvig 2021, pp. 5, 33, 1002–1003. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  346. O'Neill 2016. - O'Neill C (6 September 2016). Weapons of Math Destruction: How Big Data Increases Inequality and Threatens Democracy. Crown. ISBN 978-0553418811.

  347. Christian 2020, pp. 60–61. - Christian B (2020). The Alignment Problem: Machine learning and human values. W. W. Norton & Company. ISBN 978-0-393-86833-3. OCLC 1233266753. https://search.worldcat.org/oclc/1233266753

  348. Later research showed that there was no way for system to avoid a measurable racist bias -- fixing one form of bias would necessarily introduce another.[278]

  349. Christian 2020, pp. 6–7, 25. - Christian B (2020). The Alignment Problem: Machine learning and human values. W. W. Norton & Company. ISBN 978-0-393-86833-3. OCLC 1233266753. https://search.worldcat.org/oclc/1233266753

  350. A short summary of topics would include privacy, surveillance, copyright, misinformation and deep fakes, filter bubbles and partisanship, algorithmic bias, misleading results that go undetected without algorithmic transparency, the right to an explanation, misuse of autonomous weapons and technological unemployment. See Artificial intelligence § Ethics /wiki/Privacy

  351. Christian 2020, p. 67. - Christian B (2020). The Alignment Problem: Machine learning and human values. W. W. Norton & Company. ISBN 978-0-393-86833-3. OCLC 1233266753. https://search.worldcat.org/oclc/1233266753

  352. Christian 2020, pp. 67, 73, 117. - Christian B (2020). The Alignment Problem: Machine learning and human values. W. W. Norton & Company. ISBN 978-0-393-86833-3. OCLC 1233266753. https://search.worldcat.org/oclc/1233266753

  353. Brian Christian wrote "ProPublica's study [of COMPAS in 2015] legitimated concepts like fairness as valid topics for research"[282] /wiki/Brian_Christian

  354. Russell & Norvig 2021, p. 32. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  355. Russell & Norvig 2021, p. 32. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  356. Russell & Norvig 2021, p. 33. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  357. Metz et al. 2023. - Metz C, Weise K, Grant N, Isaac M (3 December 2023). "Ego, Fear and Money: How the A.I. Fuse Was Lit". The New York Times. https://www.nytimes.com/2023/12/03/technology/ai-openai-musk-page-altman.html

  358. Russell & Norvig 2021, p. 31. - Russell SJ, Norvig P (2021). Artificial Intelligence: A Modern Approach (4th ed.). Hoboken: Pearson. ISBN 978-0-13-461099-3. LCCN 20190474. https://lccn.loc.gov/20190474

  359. Metz et al. 2023. - Metz C, Weise K, Grant N, Isaac M (3 December 2023). "Ego, Fear and Money: How the A.I. Fuse Was Lit". The New York Times. https://www.nytimes.com/2023/12/03/technology/ai-openai-musk-page-altman.html

  360. Metz et al. 2023. - Metz C, Weise K, Grant N, Isaac M (3 December 2023). "Ego, Fear and Money: How the A.I. Fuse Was Lit". The New York Times. https://www.nytimes.com/2023/12/03/technology/ai-openai-musk-page-altman.html

  361. Metz et al. 2023. - Metz C, Weise K, Grant N, Isaac M (3 December 2023). "Ego, Fear and Money: How the A.I. Fuse Was Lit". The New York Times. https://www.nytimes.com/2023/12/03/technology/ai-openai-musk-page-altman.html

  362. Metz et al. 2023. - Metz C, Weise K, Grant N, Isaac M (3 December 2023). "Ego, Fear and Money: How the A.I. Fuse Was Lit". The New York Times. https://www.nytimes.com/2023/12/03/technology/ai-openai-musk-page-altman.html

  363. Metz et al. 2023. - Metz C, Weise K, Grant N, Isaac M (3 December 2023). "Ego, Fear and Money: How the A.I. Fuse Was Lit". The New York Times. https://www.nytimes.com/2023/12/03/technology/ai-openai-musk-page-altman.html

  364. AI boom: Marr 2023 Clark 2023 Gates 2023 Lee 2024 /wiki/AI_boom

  365. Simon & Newell 1958, pp. 7−8 quoted in Crevier 1993, p. 108.Murgia 2023. - Simon HA, Newell A (1958), "Heuristic Problem Solving: The Next Advance in Operations Research", Operations Research, 6: 1–10, doi:10.1287/opre.6.1.1 https://doi.org/10.1287%2Fopre.6.1.1

  366. Metz et al. 2023. - Metz C, Weise K, Grant N, Isaac M (3 December 2023). "Ego, Fear and Money: How the A.I. Fuse Was Lit". The New York Times. https://www.nytimes.com/2023/12/03/technology/ai-openai-musk-page-altman.html

  367. Bubeck et al. 2023. - Bubeck S, Chandrasekaran V, Eldan R, Gehrke J, Horvitz E, Kamar E, Lee P, Lee YT, Li Y, Lundberg S, Nori H, Palangi H, Ribeiro MT, Zhang Y (22 March 2023). "Sparks of Artificial General Intelligence: Early experiments with GPT-4". arXiv:2303.12712 [cs.CL]. https://arxiv.org/abs/2303.12712

  368. Hsu J. "OpenAI's o3 model aced a test of AI reasoning – but it's still not AGI". New Scientist. Retrieved 29 December 2024. https://www.newscientist.com/article/2462000-openais-o3-model-aced-a-test-of-ai-reasoning-but-its-still-not-agi/

  369. "1.1 Generative AI stats on fundraising for AI companies by year (Copy) (update)". datawrapper.dwcdn.net. Retrieved 23 April 2024. https://datawrapper.dwcdn.net/PgFO7/7/

  370. Frank M (22 September 2023). "US Leadership in Artificial Intelligence Can Shape the 21st Century Global Order". The Diplomat. Retrieved 8 December 2023. https://thediplomat.com/2023/09/us-leadership-in-artificial-intelligence-can-shape-the-21st-century-global-order/

  371. Hammond G (27 December 2023). "Big Tech is spending more than VC firms on AI startups". Ars Technica. https://arstechnica.com/ai/2023/12/big-tech-is-spending-more-than-vc-firms-on-ai-startups/

  372. Metz C, Mickle T (16 February 2024). "OpenAI Completes Deal That Values the Company at $80 Billion". The New York Times.

  373. Hur K (19 June 2024). "Nvidia surpasses Microsoft to become the largest public company in the world". CNN. Retrieved 19 June 2024. https://edition.cnn.com/2024/06/18/markets/nvidia-largest-public-company/index.html

  374. Ng A (1 April 2020). "Voice Cloning for the Masses". DeepLearning.AI. Archived from the original on 28 December 2024. Retrieved 22 December 2024. https://www.deeplearning.ai/the-batch/voice-cloning-for-the-masses/

  375. Chandraseta R (21 January 2021). "Generate Your Favourite Characters' Voice Lines using Machine Learning". Towards Data Science. Archived from the original on 21 January 2021. Retrieved 18 December 2024. https://towardsdatascience.com/generate-your-favourite-characters-voice-lines-using-machine-learning-c0939270c0c6

  376. Temitope Y (10 December 2024). "15.ai Creator reveals journey from MIT Project to internet phenomenon". The Guardian. Archived from the original on 28 December 2024. Retrieved 25 December 2024. https://guardian.ng/technology/15-ai-creator-reveals-journey-from-mit-project-to-internet-phenomenon/

  377. Anirudh VK (18 March 2023). "Deepfakes Are Elevating Meme Culture, But At What Cost?". Analytics India Magazine. Archived from the original on 26 December 2024. Retrieved 18 December 2024. While AI voice memes have been around in some form since '15.ai' launched in 2020, [...] https://analyticsindiamag.com/ai-origins-evolution/deepfakes-are-elevating-meme-culture-but-at-what-cost/

  378. "Navigating the Challenges and Opportunities of Synthetic Voices". OpenAI. 9 March 2024. Archived from the original on 25 November 2024. Retrieved 18 December 2024. https://openai.com/index/navigating-the-challenges-and-opportunities-of-synthetic-voices/

  379. Clayton N (19 January 2021). "Make the cast of TF2 recite old memes with this AI text-to-speech tool". PC Gamer. Archived from the original on 19 January 2021. Retrieved 18 December 2024. https://www.pcgamer.com/make-the-cast-of-tf2-recite-old-memes-with-this-ai-text-to-speech-tool

  380. Kurosawa Y (19 January 2021). "ゲームキャラ音声読み上げソフト「15.ai」公開中。『Undertale』や『Portal』のキャラに好きなセリフを言ってもらえる" [Game Character Voice Reading Software "15.ai" Now Available. Get Characters from Undertale and Portal to Say Your Desired Lines]. AUTOMATON (in Japanese). Archived from the original on 19 January 2021. Retrieved 18 December 2024. https://automaton-media.com/articles/newsjp/20210119-149494/

  381. Wright S (21 March 2023). "Why Biden, Trump, and Obama Arguing Over Video Games Is YouTube's New Obsession". Inverse. Archived from the original on 20 December 2024. Retrieved 18 December 2024. AI voice tools used to create "audio deepfakes" have existed for years in one form or another, with 15.ai being a notable example. https://www.inverse.com/gaming/youtube-ai-presidential-gaming-debates

  382. Milmo D (2 December 2023). "ChatGPT reaches 100 million users two months after launch". The Guardian.

  383. Roumeliotis KI, Tselikas ND (2023). "ChatGPT and Open-AI Models: A Preliminary Review". Future Internet. 15 (6): 192. doi:10.3390/fi15060192. https://doi.org/10.3390%2Ffi15060192

  384. Weise K, Metz C, Grant N, Isaac M (5 December 2023). "Inside the A.I. Arms Race That Changed Silicon Valley Forever". The New York Times.

  385. Hurst L (30 March 2023). "'Profound risk to humanity': Tech leaders call for 'pause' on advanced AI development". Euronews.

  386. Taylor J (7 May 2023). "Rise of artificial intelligence is inevitable but should not be feared, 'father of AI' says". The Guardian. https://www.theguardian.com/technology/2023/may/07/rise-of-artificial-intelligence-is-inevitable-but-should-not-be-feared-father-of-ai-says

  387. Wollman Rusoff J (1 February 2024). "Jeremy Grantham Warns Investors to 'Be Quite Careful' as Big Risks Loom". https://www.thinkadvisor.com/2024/02/01/jeremy-grantham-where-to-invest-in-2024/

  388. Mohamed T (23 March 2024). "Elite investor Jeffrey Gundlach compares the AI boom in stocks to the dot-com bubble — and warns of economic pain". Business Insider. https://www.businessinsider.com/stock-market-bubble-crash-gundlach-ai-fed-economy-gross-hussman-2024-3?op=1

  389. "Introducing the next generation of Claude". Retrieved 4 March 2024. https://www.anthropic.com/news/claude-3-family

  390. Nuñez M (4 March 2024). "Anthropic unveils Claude 3, surpassing GPT-4 and Gemini Ultra in benchmark tests". Venture Beat. Retrieved 9 April 2024. https://venturebeat.com/ai/anthropic-unveils-claude-3-claims-new-standard-for-intelligence/

  391. Pierce D (20 June 2024). "Anthropic has a fast new AI model — and a clever new way to interact with chatbots". The Verge. Retrieved 22 June 2024. https://www.theverge.com/2024/6/20/24181961/anthropic-claude-35-sonnet-model-ai-launch

  392. Nobel Prizes for 2024, official Nobel website. https://www.nobelprize.org/all-nobel-prizes-2024/

  393. ChatGPT Gov, ChatGPT Gov is designed to streamline government agencies’ access to OpenAI’s frontier models, OpnAI website. https://openai.com/global-affairs/introducing-chatgpt-gov/Introducing

  394. ChatGPT Gov, ChatGPT Gov is designed to streamline government agencies’ access to OpenAI’s frontier models, OpnAI website. https://openai.com/global-affairs/introducing-chatgpt-gov/Introducing

  395. "Elon Musk thinks robots are a $10 trillion business. He's got some competition from China". CNN. 25 March 2025. Retrieved 31 May 2025. https://edition.cnn.com/2025/03/25/tech/china-robots-market-competitiveness-intl-hnk

  396. "Scientific discovery in the age of artificial intelligence". Nature. 2 August 2023. Retrieved 31 May 2025. https://www.nature.com/articles/s41586-023-06221-2

  397. "AI Action Summit". AI Action Summit. 11 February 2025. Retrieved 31 May 2025. https://onu.delegfrance.org/ai-action-summit-10-and-11-february-2025

  398. "Dark Factories and the Future of Work: How AI-Driven Automation is Reshaping Manufacturing". unite.ai. 26 March 2025. Retrieved 30 May 2025. https://www.unite.ai/dark-factories-and-the-future-of-work-how-ai-driven-automation-is-reshaping-manufacturing/

  399. "Elon Musk thinks robots are a $10 trillion business. He's got some competition from China". CNN. 25 March 2025. Retrieved 30 May 2025. https://edition.cnn.com/2025/03/25/tech/china-robots-market-competitiveness-intl-hnk

  400. "The Chinese Humanoid Robot AI Market - Investor Opportunities". China Briefing. 7 April 2025. Retrieved 30 May 2025. https://www.china-briefing.com/news/chinese-humanoid-robot-market-opportunities/

  401. "人工智能生成合成内容标识办法". 17 March 2025. Retrieved 30 May 2025. https://www.yulelaw.com/law/law/getContentDetail?id=bb2e08f59ade44afa9822334440fb0ba&classify=3

  402. "Announcing The Stargate Project". Open AI. 21 January 2025. Retrieved 30 May 2025. https://openai.com/index/announcing-the-stargate-project/

  403. "OpenAI, SoftBank, Oracle to invest US$500 BILLION in AI, Trump says". Reuters. Retrieved 22 January 2025. https://www.reuters.com/technology/artificial-intelligence/openai-softbank-oracle-invest-500-bln-ai-trump-says-2025-01-21/

  404. Zakrzewski C (24 January 2025). "OpenAI and other tech titans worked on Stargate deal months before Trump won". The Washington Post. https://www.washingtonpost.com/politics/2025/01/24/stargate-trump-ai-altman-musk/

  405. "Humanoid Robots to Become the Next US-China Battleground". Business Wire. 24 February 2025. Retrieved 30 May 2025. https://www.trendforce.com/presscenter/news/20250224-12481.html

  406. "US robotics companies push for national strategy, including a central office, to compete with China". The Boston Globe. 28 March 2025. Retrieved 30 May 2025. https://apnews.com/article/united-states-robotics-competition-china-tech-702796f1584fe1920e5fd86f15a99b4f

  407. "Removing Barriers to American Leadership in Artificial Intelligence". The White House. 24 January 2025. Retrieved 30 May 2025. https://www.whitehouse.gov/presidential-actions/2025/01/removing-barriers-to-american-leadership-in-artificial-intelligence/

  408. "China's AI-powered humanoid robots aim to transform manufacturing". Reuters. 13 May 2025. Retrieved 30 May 2025. https://www.reuters.com/world/china/chinas-ai-powered-humanoid-robots-aim-transform-manufacturing-2025-05-13/