Progress in artificial intelligence

From AviationSafetyX Wiki
Jump to navigation Jump to search

Template:Artificial intelligence

Progress in machine classification of images
The error rate of AI by year. Red line - the error rate of a trained human on a particular task.

Progress in artificial intelligence (AI) refers to the advances, milestones, and breakthroughs that have been achieved in the field of artificial intelligence over time. AI is a multidisciplinary branch of computer science that aims to create machines and systems capable of performing tasks that typically require human intelligence. AI applications have been used in a wide range of fields including medical diagnosis, finance, robotics, law, video games, agriculture, and scientific discovery. However, many AI applications are not perceived as AI: "A lot of cutting-edge AI has filtered into general applications, often without being called AI because once something becomes useful enough and common enough it's not labeled AI anymore."[1][2] "Many thousands of AI applications are deeply embedded in the infrastructure of every industry."[3] In the late 1990s and early 2000s, AI technology became widely used as elements of larger systems,[3][4] but the field was rarely credited for these successes at the time.

Kaplan and Haenlein structure artificial intelligence along three evolutionary stages:

  1. Artificial narrow intelligence – AI capable only of specific tasks;
  2. Artificial general intelligence – AI with ability in several areas, and able to autonomously solve problems they were never even designed for;
  3. Artificial superintelligence – AI capable of general tasks, including scientific creativity, social skills, and general wisdom.[2]

To allow comparison with human performance, artificial intelligence can be evaluated on constrained and well-defined problems. Such tests have been termed subject-matter expert Turing tests. Also, smaller problems provide more achievable goals and there are an ever-increasing number of positive results.

Humans still substantially outperform both GPT-4 and models trained on the ConceptARC benchmark that scored 60% on most, and 77% on one category, while humans 91% on all and 97% on one category.[5]

Current performance in specific areas[edit | edit source]

Game Champion year[6] Legal states (log10)[7] Game tree complexity (log10)[7] Game of perfect information? Ref.
Draughts (checkers) 1994 21 31 Perfect [8]
Othello (reversi) 1997 28 58 Perfect [9]
Chess 1997 46 123 Perfect
Scrabble 2006 [10]
Shogi 2017 71 226 Perfect [11]
Go 2017 172 360 Perfect
2p no-limit hold 'em 2017 Imperfect [12]
StarCraft - 270+ Imperfect [13]
StarCraft II 2019 Imperfect [14]

There are many useful abilities that can be described as showing some form of intelligence. This gives better insight into the comparative success of artificial intelligence in different areas.

AI, like electricity or the steam engine, is a general-purpose technology. There is no consensus on how to characterize which tasks AI tends to excel at.[15] Some versions of Moravec's paradox observe that humans are more likely to outperform machines in areas such as physical dexterity that have been the direct target of natural selection.[16] While projects such as AlphaZero have succeeded in generating their own knowledge from scratch, many other machine learning projects require large training datasets.[17][18] Researcher Andrew Ng has suggested, as a "highly imperfect rule of thumb", that "almost anything a typical human can do with less than one second of mental thought, we can probably now or in the near future automate using AI."[19]

Games provide a high-profile benchmark for assessing rates of progress; many games have a large professional player base and a well-established competitive rating system. AlphaGo brought the era of classical board-game benchmarks to a close when Artificial Intelligence proved their competitive edge over humans in 2016. Deep Mind's AlphaGo AI software program defeated the world's best professional Go Player Lee Sedol.[20] Games of imperfect knowledge provide new challenges to AI in the area of game theory; the most prominent milestone in this area was brought to a close by Libratus' poker victory in 2017.[21][22] E-sports continue to provide additional benchmarks; Facebook AI, Deepmind, and others have engaged with the popular StarCraft franchise of videogames.[23][24]

Broad classes of outcome for an AI test may be given as:

  • optimal: it is not possible to perform better (note: some of these entries were solved by humans)
  • super-human: performs better than all humans
  • high-human: performs better than most humans
  • par-human: performs similarly to most humans
  • sub-human: performs worse than most humans

Optimal[edit | edit source]

Super-human[edit | edit source]

High-human[edit | edit source]

Par-human[edit | edit source]

Sub-human[edit | edit source]

<templatestyles src="Module:Message box/ambox.css"></templatestyles>

Proposed tests of artificial intelligence[edit | edit source]

In his famous Turing test, Alan Turing picked language, the defining feature of human beings, for its basis.[66] The Turing test is now considered too exploitable to be a meaningful benchmark.[67]

The Feigenbaum test, proposed by the inventor of expert systems, tests a machine's knowledge and expertise about a specific subject.[68] A paper by Jim Gray of Microsoft in 2003 suggested extending the Turing test to speech understanding, speaking and recognizing objects and behavior.[69]

Proposed "universal intelligence" tests aim to compare how well machines, humans, and even non-human animals perform on problem sets that are generic as possible. At an extreme, the test suite can contain every possible problem, weighted by Kolmogorov complexity; however, these problem sets tend to be dominated by impoverished pattern-matching exercises where a tuned AI can easily exceed human performance levels.[70][71][72][73][74]

Exams[edit | edit source]

According to OpenAI, in 2023 ChatGPT GPT-4 scored the 90th percentile on the Uniform Bar Exam. On the SATs, GPT-4 scored the 89th percentile on math, and the 93rd percentile in Reading & Writing. On the GREs, it scored on the 54th percentile on the writing test, 88th percentile on the quantitative section, and 99th percentile on the verbal section. It scored in the 99th to 100th percentile on the 2020 USA Biology Olympiad semifinal exam. It scored a perfect "5" on several AP exams.[75]

Independent researchers found in 2023 that ChatGPT GPT-3.5 "performed at or near the passing threshold" for the three parts of the United States Medical Licensing Examination. GPT-3.5 was also assessed to attain a low, but passing, grade from exams for four law school courses at the University of Minnesota.[75] GPT-4 passed a text-based radiology board–style examination.[76][77]

Competitions[edit | edit source]

<templatestyles src="Module:Hatnote/styles.css"></templatestyles>

Many competitions and prizes, such as the Imagenet Challenge, promote research in artificial intelligence. The most common areas of competition include general machine intelligence, conversational behavior, data-mining, robotic cars, and robot soccer as well as conventional games.[78]

Past and current predictions[edit | edit source]

An expert poll around 2016, conducted by Katja Grace of the Future of Humanity Institute and associates, gave median estimates of 3 years for championship Angry Birds, 4 years for the World Series of Poker, and 6 years for StarCraft. On more subjective tasks, the poll gave 6 years for folding laundry as well as an average human worker, 7–10 years for expertly answering 'easily Googleable' questions, 8 years for average speech transcription, 9 years for average telephone banking, and 11 years for expert songwriting, but over 30 years for writing a New York Times bestseller or winning the Putnam math competition.[79][80][81]

Chess[edit | edit source]

Deep Blue at the Computer History Museum

An AI defeated a grandmaster in a regulation tournament game for the first time in 1988; rebranded as Deep Blue, it beat the reigning human world chess champion in 1997 (see Deep Blue versus Garry Kasparov).[82]

Estimates when computers would exceed humans at Chess
Year prediction made Predicted year Number of years Predictor Contemporaneous source
1957 1967 or sooner 10 or less Herbert A. Simon, economist[83]
1990 2000 or sooner 10 or less Ray Kurzweil, futurist Age of Intelligent Machines[84]

Go[edit | edit source]

AlphaGo defeated a European Go champion in October 2015, and Lee Sedol in March 2016, one of the world's top players (see AlphaGo versus Lee Sedol). According to Scientific American and other sources, most observers had expected superhuman Computer Go performance to be at least a decade away.[85][86][87]

Estimates when computers would exceed humans at Go
Year prediction made Predicted year Number of years Predictor Affiliation Contemporaneous source
1997 2100 or later 103 or more Piet Hutt, physicist and Go fan Institute for Advanced Study New York Times[88][89]
2007 2017 or sooner 10 or less Feng-Hsiung Hsu, Deep Blue lead Microsoft Research Asia IEEE Spectrum[90][91]
2014 2024 10 Rémi Coulom, Computer Go programmer CrazyStone Wired[91][92]

Human-level artificial general intelligence (AGI)[edit | edit source]

AI pioneer and economist Herbert A. Simon inaccurately predicted in 1965: "Machines will be capable, within twenty years, of doing any work a man can do". Similarly, in 1970 Marvin Minsky wrote that "Within a generation... the problem of creating artificial intelligence will substantially be solved."[93]

Four polls conducted in 2012 and 2013 suggested that the median estimate among experts for when AGI would arrive was 2040 to 2050, depending on the poll.[94][95]

The Grace poll around 2016 found results varied depending on how the question was framed. Respondents asked to estimate "when unaided machines can accomplish every task better and more cheaply than human workers" gave an aggregated median answer of 45 years and a 10% chance of it occurring within 9 years. Other respondents asked to estimate "when all occupations are fully automatable. That is, when for any occupation, machines could be built to carry out the task better and more cheaply than human workers" estimated a median of 122 years and a 10% probability of 20 years. The median response for when "AI researcher" could be fully automated was around 90 years. No link was found between seniority and optimism, but Asian researchers were much more optimistic than North American researchers on average; Asians predicted 30 years on average for "accomplish every task", compared with the 74 years predicted by North Americans.[79][80][81]

Estimates of when AGI will arrive
Year prediction made Predicted year Number of years Predictor Contemporaneous source
1965 1985 or sooner 20 or less Herbert A. Simon The shape of automation for men and management[93][96]
1993 2023 or sooner 30 or less Vernor Vinge, science fiction writer "The Coming Technological Singularity"[97]
1995 2040 or sooner 45 or less Hans Moravec, robotics researcher Wired[98]
2008 Never / Distant future[note 1] Gordon E. Moore, inventor of Moore's Law IEEE Spectrum[99]
2017 2029 12 Ray Kurzweil Interview[100]

See also[edit | edit source]

References[edit | edit source]


Notes[edit | edit source]


External links[edit | edit source]

  1. AI set to exceed human brain power Archived 2008-02-19 at the Wayback Machine CNN.com (July 26, 2006)
  2. 2.0 2.1
  3. 3.0 3.1 Kurtzweil 2005, p. 264
  4. under "Artificial Intelligence in the 90s"
  5. Approximate year AI started beating top human experts
  6. 7.0 7.1
  7. 9.0 9.1 www.othello-club.de.  Retrieved 2018-07-15 from berg.earthlingz.de
  8. 10.0 10.1
  9. 11.0 11.1
  10. 12.0 12.1
  11. God's Number is 20.  Retrieved 2011-08-07 from link
  12. The Week in Chess 771.  Retrieved 2018-07-15 from theweekinchess.com
  13. Zor Winner in an Exciting Photo Finish.  Arno Nickel.  (May 2017)  Innovative Solutions.  Retrieved 2018-07-17 from www.infinitychess.com
  14. The Arimaa Challenge.  Retrieved 2018-07-15 from arimaa.com
  15. Proverb: The probabilistic cruciverbalist. By Greg A. Keim, Noam Shazeer, Michael L. Littman, Sushant Agarwal, Catherine M. Cheves, Joseph Fitzgerald, Jason Grosland, Fan Jiang, Shannon Pollard, and Karl Weinmeister. 1999. In Proceedings of the Sixteenth National Conference on Artificial Intelligence, 710-717. Menlo Park, Calif.: AAAI Press.
  16. Bethe, P. M. (2009). The state of automated bridge play.
  17. AlphaStar: Mastering the Real-Time Strategy Game StarCraft II.  (24 January 2019)  Retrieved 2022-07-19 from link
  18. Suphx: The World Best Mahjong AI.  Retrieved 2022-07-19 from Microsoft
  19. Microsoft researchers say their newest deep learning system beats humans -- and Google - VentureBeat - Big Data - by Jordan Novet.  (2015-02-10)  Retrieved 2017-09-08 from link
  20. Man Versus Machine: Who Wins When It Comes to Facial Recognition?.  (2018-12-03)  Retrieved 2022-07-20 from link
  21. 56.0 56.1 56.2 Zhang, D., Mishra, S., Brynjolfsson, E., Etchemendy, J., Ganguli, D., Grosz, B., ... & Perrault, R. (2021). The AI index 2021 annual report. AI Index (Stanford University). arXiv preprint arXiv:2103.06312.
  22. 58.0 58.1 58.2
  23. Brynjolfsson, E., & Mitchell, T. (2017). What can machine learning do? Workforce implications. Science, 358(6370), 1530-1534.
  24. Nie, W., Yu, Z., Mao, L., Patel, A. B., Zhu, Y., & Anandkumar, A. (2020). Bongard-logo: A new benchmark for human-level concept learning and reasoning. Advances in Neural Information Processing Systems, 33, 16468-16480.
  25. Template:Turing 1950
  26. International licensing under an endogenous tariff in vertically-related markets.  Andy Wang Kuang-Cheng.  (2023)  Retrieved 2023-04-23 from Journal of Economics
  27. 75.0 75.1
  28. ILSVRC2017.  Retrieved 2018-11-06 from image-net.org
  29. 79.0 79.1
  30. 80.0 80.1
  31. 81.0 81.1 Grace, K., Salvatier, J., Dafoe, A., Zhang, B., & Evans, O. (2017). When will AI exceed human performance? Evidence from AI experts. arXiv preprint arXiv:1705.08807.
  32. 91.0 91.1
  33. 93.0 93.1
  34. Müller, V. C., & Bostrom, N. (2016). Future progress in artificial intelligence: A survey of expert opinion. In Fundamental issues of artificial intelligence (pp. 555-572). Springer, Cham.
  35. Muehlhauser, L., & Salamon, A. (2012). Intelligence explosion: Evidence and import. In Singularity Hypotheses (pp. 15-42). Springer, Berlin, Heidelberg.


Cite error: <ref> tags exist for a group named "note", but no corresponding <references group="note"/> tag was found