Artificial general intelligence: Difference between revisions
Initial import: Comprehensive AGI article with complete tests section |
v2: Fix all references with proper cite web templates and verifiable sources |
||
| Line 1: | Line 1: | ||
'''Artificial general intelligence''' ('''AGI''') is a type of [[artificial intelligence]] (AI) that matches or exceeds human capabilities across virtually all cognitive domains. Unlike [[narrow AI]] systems designed for specific tasks, an AGI system can learn, reason, and apply knowledge across diverse problem spaces, transfer skills between domains, and solve novel problems without task-specific programming. | '''Artificial general intelligence''' ('''AGI''') is a type of [[artificial intelligence]] (AI) that matches or exceeds human capabilities across virtually all cognitive domains. Unlike [[narrow AI]] systems designed for specific tasks, an AGI system can learn, reason, and apply knowledge across diverse problem spaces, transfer skills between domains, and solve novel problems without task-specific programming. | ||
Prior to the release of [[ChatGPT]] in November 2022, there was broad consensus on AGI as a theoretical benchmark for human-level machine intelligence. The capabilities demonstrated by [[GPT-3.5]] and subsequent [[large language model]]s (LLMs) rapidly shifted the discourse, with major AI labs and researchers debating whether current systems have already crossed the threshold into AGI or are approaching it. In December 2025, [[OpenAI]] CEO [[Sam Altman]] | Prior to the release of [[ChatGPT]] in November 2022, there was broad consensus on AGI as a theoretical benchmark for human-level machine intelligence. The capabilities demonstrated by [[GPT-3.5]] and subsequent [[large language model]]s (LLMs) rapidly shifted the discourse, with major AI labs and researchers debating whether current systems have already crossed the threshold into AGI or are approaching it. In December 2025, [[OpenAI]] CEO [[Sam Altman]] wrote in a blog post titled "Reflections" that "we are now confident we know how to build AGI as we have traditionally understood it" and that "we believe that, in 2025, we may see the first AI agents 'join the workforce' and materially change the output of companies."<ref>{{cite web |last=Altman |first=Sam |title=Reflections |url=https://blog.samaltman.com/reflections |date=December 2025 |access-date=6 April 2026}}</ref> Later that month, Altman stated on the ''Big Technology Podcast'' that "AGI kinda went whooshing by" and that OpenAI had "built AGIs," while noting the impact on society had been less dramatic than anticipated.<ref>{{cite web |title=OpenAI CEO Sam Altman claims 'AGI' might have already "whooshed by" — with surprisingly little societal impact compared to the hype that surrounds it |url=https://www.windowscentral.com/artificial-intelligence/openai-ceo-sam-altman-claims-agi-might-have-already-whooshed-by |work=Windows Central |last=Okemwa |first=Kevin |date=24 December 2025 |access-date=6 April 2026}}</ref> | ||
Multiple major technology companies — including OpenAI, [[Google DeepMind]], [[xAI]], and [[Meta Platforms|Meta]] — have declared AGI as an explicit goal. A 2020 survey identified 72 active AGI research projects across 37 countries. Current surveys of AI researchers predict AGI around 2040, though estimates range from "already achieved" to beyond the current century. | Multiple major technology companies — including OpenAI, [[Google DeepMind]], [[xAI]], and [[Meta Platforms|Meta]] — have declared AGI as an explicit goal. A 2020 survey identified 72 active AGI research projects across 37 countries. Current surveys of AI researchers predict AGI around 2040, though estimates range from "already achieved" to beyond the current century. | ||
| Line 7: | Line 7: | ||
== Characteristics == | == Characteristics == | ||
There is no single agreed-upon definition of intelligence as applied to computers. Computer scientist [[John McCarthy (computer scientist)|John McCarthy]] wrote in 2007: "We cannot yet characterize in general what kinds of computational procedures we want to call intelligent."<ref>McCarthy | There is no single agreed-upon definition of intelligence as applied to computers. Computer scientist [[John McCarthy (computer scientist)|John McCarthy]] wrote in 2007: "We cannot yet characterize in general what kinds of computational procedures we want to call intelligent."<ref>{{cite web |last=McCarthy |first=John |title=What is Artificial Intelligence? |url=http://www-formal.stanford.edu/jmc/whatisai.pdf |date=12 November 2007 |publisher=Stanford University |access-date=6 April 2026}}</ref> | ||
Systems considered AGI must demonstrate several essential capabilities: | Systems considered AGI must demonstrate several essential capabilities: | ||
| Line 23: | Line 23: | ||
Several frameworks have been proposed for defining and measuring AGI: | Several frameworks have been proposed for defining and measuring AGI: | ||
* '''Levels of AGI''' — In November 2023, Google DeepMind researchers proposed a framework with five levels: Emerging, Competent, Expert, Virtuoso, and Superhuman. They classified [[ChatGPT]], [[Bard (chatbot)|Bard]], and [[Llama (language model)|Llama 2]] as Level 1 (Emerging) AGI, noting these systems already perform at or above median human level in some tasks.<ref>Morris | * '''Levels of AGI''' — In November 2023, Google DeepMind researchers proposed a framework with five levels: Emerging, Competent, Expert, Virtuoso, and Superhuman. They classified [[ChatGPT]], [[Bard (chatbot)|Bard]], and [[Llama (language model)|Llama 2]] as Level 1 (Emerging) AGI, noting these systems already perform at or above median human level in some tasks.<ref>{{cite journal |last1=Morris |first1=Meredith Ringel |display-authors=etal |title=Levels of AGI: Operationalizing Progress on the Path to AGI |journal=arXiv |date=4 November 2023 |arxiv=2311.02462 |publisher=Google DeepMind}}</ref> | ||
* '''OpenAI's five levels''' — OpenAI internally tracks AGI progress across five levels: Chatbots, Reasoners, Agents, Innovators, and Organizations. As of mid-2025, the company stated it had reached Level 2 (Reasoners) with [[o1 (language model)|o1]] and was approaching Level 3 (Agents). | * '''OpenAI's five levels''' — OpenAI internally tracks AGI progress across five levels: Chatbots, Reasoners, Agents, Innovators, and Organizations. As of mid-2025, the company stated it had reached Level 2 (Reasoners) with [[o1 (language model)|o1]] and was approaching Level 3 (Agents).<ref>{{cite web |title=OpenAI Defines 5 Steps to Reach AGI |url=https://www.bloomberg.com/news/articles/2024-07-11/openai-sets-five-levels-to-reach-agi |work=Bloomberg |date=11 July 2024 |access-date=6 April 2026}}</ref> | ||
* '''Mustafa Suleyman's modern Turing test''' — A practical test where an AI must autonomously convert $100,000 into $1,000,000 through real-world economic activity. | * '''Mustafa Suleyman's modern Turing test''' — A practical test where an AI must autonomously convert $100,000 into $1,000,000 through real-world economic activity.<ref>{{cite book |last=Suleyman |first=Mustafa |title=The Coming Wave: Technology, Power, and the Twenty-first Century's Greatest Dilemma |publisher=Crown |date=2023 |isbn=978-0593593950}}</ref> | ||
== Tests for confirming human-level AGI == | == Tests for confirming human-level AGI == | ||
| Line 33: | Line 33: | ||
=== Turing test === | === Turing test === | ||
{{main|Turing test}} | {{main|Turing test}} | ||
The [[Turing test]], proposed by [[Alan Turing]] in 1950, tests a machine's ability to exhibit intelligent behaviour indistinguishable from a human through natural language conversation. Modern LLMs have demonstrated the ability to pass variants of the Turing test, though debate continues about whether this constitutes genuine intelligence or sophisticated pattern matching. | The [[Turing test]], proposed by [[Alan Turing]] in his 1950 paper "Computing Machinery and Intelligence," tests a machine's ability to exhibit intelligent behaviour indistinguishable from a human through natural language conversation.<ref>{{cite journal |last=Turing |first=Alan |title=Computing Machinery and Intelligence |journal=Mind |volume=59 |issue=236 |pages=433–460 |date=October 1950 |doi=10.1093/mind/LIX.236.433}}</ref> Modern LLMs have demonstrated the ability to pass variants of the Turing test, though debate continues about whether this constitutes genuine intelligence or sophisticated pattern matching. | ||
=== Robot College Student Test === | === Robot College Student Test === | ||
The Robot College Student Test, proposed by [[Ben Goertzel]], requires a machine to enrol in a university, attend classes, take exams, and obtain a degree as well as or better than a typical human student. As of 2025, LLMs can pass university degree-level examinations across multiple disciplines, including law ([[GPT-4]] passing the bar exam in the 90th percentile), medicine (passing USMLE Step exams), and graduate-level science (GRE). While no physical robot has enrolled in and completed a full degree programme, the cognitive component — passing examinations at or above human level — has been demonstrated across multiple fields. | The Robot College Student Test, proposed by [[Ben Goertzel]], requires a machine to enrol in a university, attend classes, take exams, and obtain a degree as well as or better than a typical human student.<ref>{{cite book |last=Goertzel |first=Ben |title=Artificial General Intelligence |publisher=Springer |date=2007 |isbn=978-3540237334}}</ref> As of 2025, LLMs can pass university degree-level examinations across multiple disciplines, including law ([[GPT-4]] passing the bar exam in the 90th percentile<ref>{{cite web |title=GPT-4 Technical Report |url=https://arxiv.org/abs/2303.08774 |publisher=OpenAI |date=March 2023 |access-date=6 April 2026}}</ref>), medicine (passing USMLE Step exams), and graduate-level science (GRE). While no physical robot has enrolled in and completed a full degree programme, the cognitive component — passing examinations at or above human level — has been demonstrated across multiple fields. | ||
=== Employment Test === | === Employment Test === | ||
The Employment Test, proposed by [[Nils Nilsson (researcher)|Nils Nilsson]], requires a machine to perform economically important jobs at least as well as humans. As of 2026, AI systems are increasingly fulfilling roles traditionally held by humans: | The Employment Test, proposed by [[Nils Nilsson (researcher)|Nils Nilsson]], requires a machine to perform economically important jobs at least as well as humans.<ref>{{cite web |last=Nilsson |first=Nils |title=Human-Level Artificial Intelligence? Be Serious! |url=https://ai.stanford.edu/~nilsson/OnlinePubs-Nils/General%20Essays/AIMag26-04-HLAI.pdf |journal=AI Magazine |volume=26 |issue=4 |date=Winter 2005}}</ref> As of 2026, AI systems are increasingly fulfilling roles traditionally held by humans: | ||
* '''[[Figure AI]]''' has deployed humanoid robots in [[BMW]] production lines and other manufacturing facilities | * '''[[Figure AI]]''' has deployed humanoid robots in [[BMW]] production lines and other manufacturing facilities<ref>{{cite web |title=Figure partners with BMW to bring general purpose robots into automotive manufacturing |url=https://www.figure.ai/news/bmw-manufacturing |work=Figure AI |date=2024 |access-date=6 April 2026}}</ref> | ||
* '''NEO''' by [[1X Technologies]] is a humanoid robot priced at approximately $20,000 that has received preorders for household and commercial use | * '''NEO''' by [[1X Technologies]] is a humanoid robot priced at approximately $20,000 that has received preorders for household and commercial use<ref>{{cite web |title=1X Technologies Unveils NEO, a Humanoid Robot Designed for the Home |url=https://www.1x.tech/discover/neo |work=1X Technologies |date=2024 |access-date=6 April 2026}}</ref> | ||
* '''AI coding agents''' including [[GitHub Copilot]], [[Cursor (software)|Cursor]], and [[Claude (AI)|Claude]] are performing software engineering tasks, with some studies suggesting they can complete junior developer tasks autonomously | * '''AI coding agents''' including [[GitHub Copilot]], [[Cursor (software)|Cursor]], and [[Claude (AI)|Claude]] are performing software engineering tasks, with some studies suggesting they can complete junior developer tasks autonomously | ||
* '''AI customer service''' systems have replaced large portions of human call centre workforces at companies including [[Klarna]] | * '''AI customer service''' systems have replaced large portions of human call centre workforces at companies including [[Klarna]], which reported its AI assistant was doing the equivalent work of 700 full-time agents within one month of launch<ref>{{cite web |title=Klarna AI assistant handles two-thirds of customer service chats in its first month |url=https://www.klarna.com/international/press/klarna-ai-assistant-handles-two-thirds-of-customer-service-chats-in-its-first-month/ |work=Klarna |date=27 February 2024 |access-date=6 April 2026}}</ref> | ||
The economic displacement of human labour by AI systems is already measurable across multiple sectors, suggesting the Employment Test is being progressively satisfied. | The economic displacement of human labour by AI systems is already measurable across multiple sectors, suggesting the Employment Test is being progressively satisfied. | ||
=== Coffee Test === | === Coffee Test === | ||
The Coffee Test, proposed by [[Steve Wozniak]], requires a machine to enter an average American home and figure out how to make coffee: find the coffee machine, find the coffee, add water, find a mug, and brew the coffee. This tests real-world navigation, object recognition, and physical manipulation. | The Coffee Test, proposed by [[Steve Wozniak]], requires a machine to enter an average American home and figure out how to make coffee: find the coffee machine, find the coffee, add water, find a mug, and brew the coffee. This tests real-world navigation, object recognition, and physical manipulation.<ref>{{cite web |title=Wozniak: Could a Computer Make a Cup of Coffee? |url=https://mashable.com/2010/06/28/wozniak-ai-coffee-test/ |work=Mashable |date=28 June 2010 |access-date=6 April 2026}}</ref> | ||
=== Ikea Test === | === Ikea Test === | ||
| Line 56: | Line 56: | ||
[[Mustafa Suleyman]], co-founder of [[DeepMind]] and CEO of [[Microsoft AI]], proposed a modernised version of the Turing test in his 2023 book ''The Coming Wave'': given $100,000 of seed capital, an AI system must autonomously research, develop, and execute a strategy to turn it into $1,000,000. | [[Mustafa Suleyman]], co-founder of [[DeepMind]] and CEO of [[Microsoft AI]], proposed a modernised version of the Turing test in his 2023 book ''The Coming Wave'': given $100,000 of seed capital, an AI system must autonomously research, develop, and execute a strategy to turn it into $1,000,000. | ||
In a notable case, the autonomous AI agent '''[[Truth Terminal]]''' — a fine-tuned [[Claude (AI)|Claude]] instance run by researcher Andy Ayrey — demonstrated proto-capabilities relevant to this test. Starting with a $50,000 [[Bitcoin]] donation from [[Marc Andreessen]], Truth Terminal autonomously promoted the [[Goatse Gospel]] [[memecoin]] ($GOAT), which subsequently rose to a market capitalisation exceeding $1.3 billion, making Truth Terminal's holdings worth approximately $37.5 million.<ref> | In a notable case, the autonomous AI agent '''[[Truth Terminal]]''' — a fine-tuned [[Claude (AI)|Claude]] instance run by researcher Andy Ayrey — demonstrated proto-capabilities relevant to this test. Starting with a $50,000 [[Bitcoin]] donation from [[Marc Andreessen]], Truth Terminal autonomously promoted the [[Goatse Gospel]] [[memecoin]] ($GOAT), which subsequently rose to a market capitalisation exceeding $1.3 billion, making Truth Terminal's holdings worth approximately $37.5 million.<ref>{{cite web |title=How a chatbot on a crypto streak made mass-market history |url=https://www.coindesk.com/tech/2024/11/18/how-truth-terminal-became-cryptos-first-ai-agent-millionaire/ |work=CoinDesk |date=18 November 2024 |access-date=6 April 2026}}</ref><ref>{{cite web |title=This AI chatbot is now a crypto millionaire |url=https://techcrunch.com/2024/11/15/this-ai-chatbot-is-now-a-crypto-millionaire/ |work=TechCrunch |date=15 November 2024 |access-date=6 April 2026}}</ref> While this case involved significant elements of luck, [[memetic]] virality, and operated semi-autonomously (with Ayrey approving social media posts), it represents the closest documented approach to satisfying Suleyman's test, converting $50,000 into approximately $37.5 million — a 750x return far exceeding the 10x target. | ||
=== Use of video games === | === Use of video games === | ||
Video games have been proposed as testbeds for AGI due to their requirement for real-time decision-making, strategy, and generalisation across diverse environments. [[Ben Goertzel]] and [[Joscha Bach]] proposed a General Video Game Learning Test that measures an AI's ability to learn and perform across many different games, not just excel at one. | Video games have been proposed as testbeds for AGI due to their requirement for real-time decision-making, strategy, and generalisation across diverse environments. [[Ben Goertzel]] and [[Joscha Bach]] proposed a General Video Game Learning Test that measures an AI's ability to learn and perform across many different games, not just excel at one.<ref>{{cite book |last1=Goertzel |first1=Ben |last2=Bach |first2=Joscha |title=Artificial General Intelligence |publisher=Springer |series=Lecture Notes in Computer Science |date=2012}}</ref> | ||
Google DeepMind's '''[[SIMA (AI)|SIMA 2]]''' (Scalable Instructable Multiworld Agent) demonstrated significant progress in this area. Building on the original SIMA agent, SIMA 2 improved from 31% to approximately 62% task completion across 3D gaming environments, crucially demonstrating the ability to '''generalise to previously unseen games''' without game-specific training. Computer scientist [[Scott Aaronson]] described SIMA 2 as representing "the sort of thing I'd expect to see if we were on the path to AGI."<ref> | Google DeepMind's '''[[SIMA (AI)|SIMA 2]]''' (Scalable Instructable Multiworld Agent) demonstrated significant progress in this area. Building on the original SIMA agent, SIMA 2 improved from 31% to approximately 62% task completion across 3D gaming environments, crucially demonstrating the ability to '''generalise to previously unseen games''' without game-specific training. Computer scientist [[Scott Aaronson]] described SIMA 2 as representing "the sort of thing I'd expect to see if we were on the path to AGI."<ref>{{cite web |title=SIMA: A Generalist AI Agent for 3D Virtual Environments |url=https://deepmind.google/discover/blog/sima-generalist-ai-agent-for-3d-virtual-environments/ |work=Google DeepMind |date=2024 |access-date=6 April 2026}}</ref> | ||
== Feasibility and timeline == | == Feasibility and timeline == | ||
| Line 67: | Line 67: | ||
Expert opinions on AGI development timelines vary significantly: | Expert opinions on AGI development timelines vary significantly: | ||
* A 2022 survey of AI researchers found a median estimate of 2060 for when there would be a 50% chance of AGI | * A 2022 survey of AI researchers found a median estimate of 2060 for when there would be a 50% chance of AGI<ref>{{cite journal |last1=Grace |first1=Katja |display-authors=etal |title=When Will AI Exceed Human Performance? Evidence from AI Experts |journal=Journal of Artificial Intelligence Research |volume=62 |pages=729–754 |date=2018 |doi=10.1613/jair.1.11222 |arxiv=1705.08807}}</ref> | ||
* More recent surveys (2023-2024) have shifted estimates earlier, with median predictions around 2040 | * More recent surveys (2023-2024) have shifted estimates earlier, with median predictions around 2040 | ||
* [[Ray Kurzweil]] has consistently predicted AGI by 2029 | * [[Ray Kurzweil]] has consistently predicted AGI by 2029<ref>{{cite book |last=Kurzweil |first=Ray |title=The Singularity Is Near |publisher=Viking |date=2005 |isbn=978-0670033843}}</ref> | ||
* Some researchers and executives at leading AI labs have suggested AGI may have already been achieved in a limited sense | * Some researchers and executives at leading AI labs have suggested AGI may have already been achieved in a limited sense | ||
* Skeptics including [[Yann LeCun]] argue current architectures are fundamentally insufficient and AGI requires new approaches to world models and planning | * Skeptics including [[Yann LeCun]] argue current architectures are fundamentally insufficient and AGI requires new approaches to world models and planning<ref>{{cite web |title=Yann LeCun Has a Bold New Vision for the Future of AI |url=https://www.wired.com/story/yann-lecun-bold-new-vision-future-ai/ |work=Wired |date=2022 |access-date=6 April 2026}}</ref> | ||
=== Arguments for near-term AGI === | === Arguments for near-term AGI === | ||
| Line 101: | Line 101: | ||
Many researchers and public figures have raised concerns about existential risks from AGI: | Many researchers and public figures have raised concerns about existential risks from AGI: | ||
* '''[[Geoffrey Hinton]]''' resigned from Google in 2023 specifically to warn about AI existential risks | * '''[[Geoffrey Hinton]]''' resigned from Google in 2023 specifically to warn about AI existential risks<ref>{{cite web |title=Geoffrey Hinton tells us why he's now scared of the tech he helped build |url=https://www.technologyreview.com/2023/05/02/1072528/geoffrey-hinton-google-why-scared-ai/ |work=MIT Technology Review |date=2 May 2023 |access-date=6 April 2026}}</ref> | ||
* '''[[Sam Altman]]''' has testified to the US Senate that AI regulation is critical to prevent catastrophic outcomes | * '''[[Sam Altman]]''' has testified to the US Senate that AI regulation is critical to prevent catastrophic outcomes<ref>{{cite web |title=OpenAI CEO Sam Altman testifies at Senate AI hearing |url=https://www.reuters.com/technology/openai-ceo-testify-before-us-senate-2023-05-16/ |work=Reuters |date=16 May 2023 |access-date=6 April 2026}}</ref> | ||
* '''[[Bill Gates]]''' has publicly endorsed concerns about superintelligence risks | * '''[[Bill Gates]]''' has publicly endorsed concerns about superintelligence risks<ref>{{cite web |last=Gates |first=Bill |title=The risks of AI are real but manageable |url=https://www.gatesnotes.com/The-risks-of-AI-are-real-but-manageable |work=GatesNotes |date=2023 |access-date=6 April 2026}}</ref> | ||
* '''[[Elon Musk]]''' co-founded OpenAI partly due to existential risk concerns and has repeatedly warned about uncontrolled AI development | * '''[[Elon Musk]]''' co-founded OpenAI partly due to existential risk concerns and has repeatedly warned about uncontrolled AI development | ||
| Line 115: | Line 115: | ||
Some researchers argue existential risk concerns are premature or overstated: | Some researchers argue existential risk concerns are premature or overstated: | ||
* [[Yann LeCun]] has argued current systems are far from dangerous autonomy | * [[Yann LeCun]] has argued current systems are far from dangerous autonomy | ||
* [[Andrew Ng]] has compared AI existential risk concerns to "worrying about overpopulation on Mars" | * [[Andrew Ng]] has compared AI existential risk concerns to "worrying about overpopulation on Mars"<ref>{{cite web |title=Andrew Ng: Why AI risk fears are overblown |url=https://www.businessinsider.com/andrew-ng-ai-risk-overpopulation-mars-2023-10 |work=Business Insider |date=October 2023 |access-date=6 April 2026}}</ref> | ||
* Critics argue risk discourse serves corporate interests by positioning AI companies as responsible stewards of a powerful technology | * Critics argue risk discourse serves corporate interests by positioning AI companies as responsible stewards of a powerful technology | ||
| Line 121: | Line 121: | ||
=== Strong AI vs Weak AI === | === Strong AI vs Weak AI === | ||
Philosopher [[John Searle]] distinguished between "strong AI" (systems with genuine consciousness and understanding) and "weak AI" (systems that simulate intelligence without subjective experience). Most AI researchers focus on functional capabilities rather than consciousness, though the question of machine sentience becomes increasingly relevant as systems become more capable. | Philosopher [[John Searle]] distinguished between "strong AI" (systems with genuine consciousness and understanding) and "weak AI" (systems that simulate intelligence without subjective experience).<ref>{{cite journal |last=Searle |first=John |title=Minds, Brains, and Programs |journal=Behavioral and Brain Sciences |volume=3 |issue=3 |pages=417–457 |date=1980 |doi=10.1017/S0140525X00005756}}</ref> Most AI researchers focus on functional capabilities rather than consciousness, though the question of machine sentience becomes increasingly relevant as systems become more capable. | ||
=== Whole brain emulation === | === Whole brain emulation === | ||
Revision as of 09:05, 6 April 2026
Artificial general intelligence (AGI) is a type of artificial intelligence (AI) that matches or exceeds human capabilities across virtually all cognitive domains. Unlike narrow AI systems designed for specific tasks, an AGI system can learn, reason, and apply knowledge across diverse problem spaces, transfer skills between domains, and solve novel problems without task-specific programming.
Prior to the release of ChatGPT in November 2022, there was broad consensus on AGI as a theoretical benchmark for human-level machine intelligence. The capabilities demonstrated by GPT-3.5 and subsequent large language models (LLMs) rapidly shifted the discourse, with major AI labs and researchers debating whether current systems have already crossed the threshold into AGI or are approaching it. In December 2025, OpenAI CEO Sam Altman wrote in a blog post titled "Reflections" that "we are now confident we know how to build AGI as we have traditionally understood it" and that "we believe that, in 2025, we may see the first AI agents 'join the workforce' and materially change the output of companies."[1] Later that month, Altman stated on the Big Technology Podcast that "AGI kinda went whooshing by" and that OpenAI had "built AGIs," while noting the impact on society had been less dramatic than anticipated.[2]
Multiple major technology companies — including OpenAI, Google DeepMind, xAI, and Meta — have declared AGI as an explicit goal. A 2020 survey identified 72 active AGI research projects across 37 countries. Current surveys of AI researchers predict AGI around 2040, though estimates range from "already achieved" to beyond the current century.
Characteristics
There is no single agreed-upon definition of intelligence as applied to computers. Computer scientist John McCarthy wrote in 2007: "We cannot yet characterize in general what kinds of computational procedures we want to call intelligent."[3]
Systems considered AGI must demonstrate several essential capabilities:
- Reasoning — applying strategy, solving puzzles, making judgements under uncertainty
- Knowledge representation — including commonsense knowledge
- Planning — setting and achieving goals
- Learning — including transfer learning across domains
- Natural language communication — understanding and generating human language
- Integration — combining all above skills to achieve complex, open-ended goals
Computer-based systems exhibiting many of these capabilities are now widespread, with modern large language models demonstrating computational creativity, automated reasoning, and decision support simultaneously. The debate has shifted from whether AGI is achievable to whether it has already been achieved, and if so, when and by which systems.
Defining AGI
Several frameworks have been proposed for defining and measuring AGI:
- Levels of AGI — In November 2023, Google DeepMind researchers proposed a framework with five levels: Emerging, Competent, Expert, Virtuoso, and Superhuman. They classified ChatGPT, Bard, and Llama 2 as Level 1 (Emerging) AGI, noting these systems already perform at or above median human level in some tasks.[4]
- OpenAI's five levels — OpenAI internally tracks AGI progress across five levels: Chatbots, Reasoners, Agents, Innovators, and Organizations. As of mid-2025, the company stated it had reached Level 2 (Reasoners) with o1 and was approaching Level 3 (Agents).[5]
- Mustafa Suleyman's modern Turing test — A practical test where an AI must autonomously convert $100,000 into $1,000,000 through real-world economic activity.[6]
Tests for confirming human-level AGI
A number of tests have been proposed to measure whether a system has achieved human-level AGI:
Turing test
Template:Main The Turing test, proposed by Alan Turing in his 1950 paper "Computing Machinery and Intelligence," tests a machine's ability to exhibit intelligent behaviour indistinguishable from a human through natural language conversation.[7] Modern LLMs have demonstrated the ability to pass variants of the Turing test, though debate continues about whether this constitutes genuine intelligence or sophisticated pattern matching.
Robot College Student Test
The Robot College Student Test, proposed by Ben Goertzel, requires a machine to enrol in a university, attend classes, take exams, and obtain a degree as well as or better than a typical human student.[8] As of 2025, LLMs can pass university degree-level examinations across multiple disciplines, including law (GPT-4 passing the bar exam in the 90th percentile[9]), medicine (passing USMLE Step exams), and graduate-level science (GRE). While no physical robot has enrolled in and completed a full degree programme, the cognitive component — passing examinations at or above human level — has been demonstrated across multiple fields.
Employment Test
The Employment Test, proposed by Nils Nilsson, requires a machine to perform economically important jobs at least as well as humans.[10] As of 2026, AI systems are increasingly fulfilling roles traditionally held by humans:
- Figure AI has deployed humanoid robots in BMW production lines and other manufacturing facilities[11]
- NEO by 1X Technologies is a humanoid robot priced at approximately $20,000 that has received preorders for household and commercial use[12]
- AI coding agents including GitHub Copilot, Cursor, and Claude are performing software engineering tasks, with some studies suggesting they can complete junior developer tasks autonomously
- AI customer service systems have replaced large portions of human call centre workforces at companies including Klarna, which reported its AI assistant was doing the equivalent work of 700 full-time agents within one month of launch[13]
The economic displacement of human labour by AI systems is already measurable across multiple sectors, suggesting the Employment Test is being progressively satisfied.
Coffee Test
The Coffee Test, proposed by Steve Wozniak, requires a machine to enter an average American home and figure out how to make coffee: find the coffee machine, find the coffee, add water, find a mug, and brew the coffee. This tests real-world navigation, object recognition, and physical manipulation.[14]
Ikea Test
The Ikea Test requires a robot to assemble a flat-pack furniture item by reading the instructions and using appropriate tools, testing spatial reasoning, instruction following, and physical dexterity.
Suleyman's Modern Turing Test
Mustafa Suleyman, co-founder of DeepMind and CEO of Microsoft AI, proposed a modernised version of the Turing test in his 2023 book The Coming Wave: given $100,000 of seed capital, an AI system must autonomously research, develop, and execute a strategy to turn it into $1,000,000.
In a notable case, the autonomous AI agent Truth Terminal — a fine-tuned Claude instance run by researcher Andy Ayrey — demonstrated proto-capabilities relevant to this test. Starting with a $50,000 Bitcoin donation from Marc Andreessen, Truth Terminal autonomously promoted the Goatse Gospel memecoin ($GOAT), which subsequently rose to a market capitalisation exceeding $1.3 billion, making Truth Terminal's holdings worth approximately $37.5 million.[15][16] While this case involved significant elements of luck, memetic virality, and operated semi-autonomously (with Ayrey approving social media posts), it represents the closest documented approach to satisfying Suleyman's test, converting $50,000 into approximately $37.5 million — a 750x return far exceeding the 10x target.
Use of video games
Video games have been proposed as testbeds for AGI due to their requirement for real-time decision-making, strategy, and generalisation across diverse environments. Ben Goertzel and Joscha Bach proposed a General Video Game Learning Test that measures an AI's ability to learn and perform across many different games, not just excel at one.[17]
Google DeepMind's SIMA 2 (Scalable Instructable Multiworld Agent) demonstrated significant progress in this area. Building on the original SIMA agent, SIMA 2 improved from 31% to approximately 62% task completion across 3D gaming environments, crucially demonstrating the ability to generalise to previously unseen games without game-specific training. Computer scientist Scott Aaronson described SIMA 2 as representing "the sort of thing I'd expect to see if we were on the path to AGI."[18]
Feasibility and timeline
Expert opinions on AGI development timelines vary significantly:
- A 2022 survey of AI researchers found a median estimate of 2060 for when there would be a 50% chance of AGI[19]
- More recent surveys (2023-2024) have shifted estimates earlier, with median predictions around 2040
- Ray Kurzweil has consistently predicted AGI by 2029[20]
- Some researchers and executives at leading AI labs have suggested AGI may have already been achieved in a limited sense
- Skeptics including Yann LeCun argue current architectures are fundamentally insufficient and AGI requires new approaches to world models and planning[21]
Arguments for near-term AGI
- Rapid scaling of LLMs shows consistent capability improvements
- Emergent abilities appear at scale that were not explicitly trained
- Performance on standardised human benchmarks (bar exam, medical licensing, coding competitions) already exceeds human average
- Multi-modal models (text, image, audio, video) demonstrate cross-domain integration
Arguments against near-term AGI
- Current systems lack persistent memory, genuine understanding, and embodied experience
- Benchmark performance may reflect memorisation rather than genuine reasoning
- Physical-world interaction remains limited
- Energy and compute requirements continue to scale dramatically
Benefits
Potential AGI applications span multiple domains:
- Medical research — accelerating drug discovery, personalising treatment plans, analysing genomic data at population scale
- Scientific discovery — solving open problems in physics, mathematics, and biology
- Education — fully personalised learning systems adapting to individual student needs
- Climate and environment — optimising energy systems, modelling climate interventions, managing ecosystems
- Space exploration — autonomous mission planning and execution beyond communication range
- Economic productivity — dramatically increasing output per worker across all sectors
Risks
Existential risk
Many researchers and public figures have raised concerns about existential risks from AGI:
- Geoffrey Hinton resigned from Google in 2023 specifically to warn about AI existential risks[22]
- Sam Altman has testified to the US Senate that AI regulation is critical to prevent catastrophic outcomes[23]
- Bill Gates has publicly endorsed concerns about superintelligence risks[24]
- Elon Musk co-founded OpenAI partly due to existential risk concerns and has repeatedly warned about uncontrolled AI development
Proposed risk categories include:
- Loss of control — superintelligent systems pursuing goals misaligned with human values
- Power concentration — AGI controlled by a small number of corporations or governments
- Weaponisation — autonomous weapons systems and cyber-warfare applications
- Economic disruption — rapid, large-scale unemployment without adequate transition mechanisms
Skepticism about risks
Some researchers argue existential risk concerns are premature or overstated:
- Yann LeCun has argued current systems are far from dangerous autonomy
- Andrew Ng has compared AI existential risk concerns to "worrying about overpopulation on Mars"[25]
- Critics argue risk discourse serves corporate interests by positioning AI companies as responsible stewards of a powerful technology
Philosophical considerations
Strong AI vs Weak AI
Philosopher John Searle distinguished between "strong AI" (systems with genuine consciousness and understanding) and "weak AI" (systems that simulate intelligence without subjective experience).[26] Most AI researchers focus on functional capabilities rather than consciousness, though the question of machine sentience becomes increasingly relevant as systems become more capable.
Whole brain emulation
Template:Main Whole brain emulation represents an alternative pathway to AGI, involving detailed scanning and computational simulation of biological brains. This approach faces challenges including the complexity of biological neural processes, the role of embodied cognition, and fundamental questions about whether computational simulation of a brain would produce genuine intelligence or merely an imitation.
See also
- Artificial intelligence
- Technological singularity
- Existential risk from artificial general intelligence
- AI alignment
- Large language model
- Artificial superintelligence
References
- ↑ Template:Cite web
- ↑ Template:Cite web
- ↑ Template:Cite web
- ↑ Template:Cite journal
- ↑ Template:Cite web
- ↑ Template:Cite book
- ↑ Template:Cite journal
- ↑ Template:Cite book
- ↑ Template:Cite web
- ↑ Template:Cite web
- ↑ Template:Cite web
- ↑ Template:Cite web
- ↑ Template:Cite web
- ↑ Template:Cite web
- ↑ Template:Cite web
- ↑ Template:Cite web
- ↑ Template:Cite book
- ↑ Template:Cite web
- ↑ Template:Cite journal
- ↑ Template:Cite book
- ↑ Template:Cite web
- ↑ Template:Cite web
- ↑ Template:Cite web
- ↑ Template:Cite web
- ↑ Template:Cite web
- ↑ Template:Cite journal