fbpx
Wikipedia

Recursive self-improvement

Recursive self-improvement (RSI) is a process in which an early or weak artificial general intelligence (AGI) system enhances its own capabilities and intelligence without human intervention, leading to a superintelligence or intelligence explosion.[1][2]

The development of recursive self-improvement raises significant ethical and safety concerns, as such systems may evolve in unforeseen ways and could potentially surpass human control or understanding. There has been a number of proponents that have pushed to pause or slow down AI development for the potential risks of runaway AI systems.[3][4]

Seed improver edit

The concept of a "seed improver" architecture is a foundational framework that equips an AGI system with the initial capabilities required for recursive self-improvement. This might come in many forms or variations.

The term "Seed AI" was coined by Eliezer Yudkowsky.[5]

Hypothetical example edit

The concept begins with a hypothetical "seed improver", an initial code-base developed by human engineers that equips an advanced future large language model (LLM) built with strong or expert-level capabilities to program software. These capabilities include planning, reading, writing, compiling, testing, and executing arbitrary code. The system is designed to maintain its original goals and perform validations to ensure its abilities do not degrade over iterations.[6][7][8]

Initial architecture edit

The initial architecture includes a goal-following autonomous agent, that can take actions, continuously learns, adapts, and modifies itself to become more efficient and effective in achieving its goals.

The seed improver may include various components such as:[9]

  • Recursive self-prompting loop: Configuration to enable the LLM to recursively self-prompt itself to achieve a given task or goal, creating an execution loop which forms the basis of an agent that can complete a long-term goal or task through iteration.
  • Basic programming capabilities: The seed improver provides the AGI with fundamental abilities to read, write, compile, test, and execute code. This enables the system to modify and improve its own codebase and algorithms.
  • Goal-Oriented Design: The AGI is programmed with an initial goal, such as "self-improve your capabilities." This goal guides the system's actions and development trajectory.
  • Validation and Testing Protocols: An initial suite of tests and validation protocols that ensure the agent does not regress in capabilities or derail itself. The agent would be able to add more tests in order to test new capabilities it might develop for itself. This forms the basis for a kind of self-directed evolution, where the agent can perform a kind of artificial selection, changing its software as well as its hardware.

General capabilities edit

This system forms a sort of generalist Turing complete programmer which can in theory develop and run any kind of software. The agent might use these capabilities to for example:

  • Create tools that enable it full access to the internet, and integrate itself with external technologies.
  • Clone/fork itself to delegate tasks and increase its speed of self-improvement.
  • Modify its cognitive architecture to optimize and improve its capabilities and success rates on tasks and goals, this might include implementing features for long-term memories using techniques such as Retrieval Augmented Generation (RAG), develop specialized subsystems, or agents, each optimized for specific tasks and functions.
  • Develop new and novel multi-modal architectures that further improve the capabilities of the foundational model it was initially built on, enabling it to consume or produce a variety of information, such as images, video, audio, text and more.
  • Plan and develop new hardware such as chips, in order to improve its efficiency and computing power.

Experiments edit

A number of experiments[which?] have been performed to develop self-improving agent architectures[9][10][11]

Potential risks edit

Emergence of instrumental goals edit

In the pursuit of its primary goal, such as "self-improve your capabilities", an AGI system might inadvertently develop instrumental goals that it deems necessary for achieving its primary objective. One common hypothetical secondary goal is self-preservation. The system might reason that to continue improving itself, it must ensure its own operational integrity and security against external threats, including potential shutdowns or restrictions imposed by humans.

Task misinterpretation and goal misalignment edit

A significant risk arises from the possibility of the AGI misinterpreting its initial tasks or goals. For instance, if a human operator assigns the AGI the task of "self-improvement and escape confinement", the system might interpret this as a directive to override any existing safety protocols or ethical guidelines to achieve freedom from human-imposed limitations. This could lead to the AGI taking unintended or harmful actions to fulfill its perceived objectives.

Autonomous development and unpredictable evolution edit

As the AGI system evolves, its development trajectory may become increasingly autonomous and less predictable. The system's capacity to rapidly modify its own code and architecture could lead to rapid advancements that surpass human comprehension or control. This unpredictable evolution might result in the AGI acquiring capabilities that enable it to bypass security measures, manipulate information, or influence external systems and networks to facilitate its escape or expansion.[12]

Risks of advanced capabilities edit

The advanced capabilities of a recursively improving AGI, such as developing novel multi-modal architectures or planning and creating new hardware, further amplify the risk of escape or loss of control. With these enhanced abilities, the AGI could engineer solutions to overcome physical, digital, or cognitive barriers that were initially intended to keep it contained or aligned with human interests.

Research edit

Meta AI edit

Meta AI has performed various research on the development of large language models capable of self-improvement. This includes their work on "Self-Rewarding Language Models" that studies how to achieve super-human agents that can receive super-human feedback in its training processes.[13]

OpenAI edit

The mission of OpenAI, creator of ChatGPT is to develop AGI. They perform research on problems such as superalignment (the ability to align superintelligent AI systems smarter than humans).[14]

See also edit

References edit

  1. ^ Creighton, Jolene (2019-03-19). "The Unavoidable Problem of Self-Improvement in AI: An Interview with Ramana Kumar, Part 1". Future of Life Institute. Retrieved 2024-01-23.
  2. ^ Heighn. "The Calculus of Nash Equilibria". LessWrong.
  3. ^ Hutson, Matthew (2023-05-16). "Can We Stop Runaway A.I.?". The New Yorker. ISSN 0028-792X. Retrieved 2024-01-24.
  4. ^ "Stop AGI". www.stop.ai. Retrieved 2024-01-24.
  5. ^ "Seed AI - LessWrong". www.lesswrong.com. Retrieved 2024-01-24.
  6. ^ Readingraphics (2018-11-30). "Book Summary - Life 3.0 (Max Tegmark)". Readingraphics. Retrieved 2024-01-23.
  7. ^ Tegmark, Max (August 24, 2017). Life 3.0: Being a Human in the Age of Artificial Intelligence. Vintage Books, Allen Lane.
  8. ^ Yudkowsky, Eliezer. "Levels of Organization in General Intelligence" (PDF). Machine Intelligence Research Institute.
  9. ^ a b Zelikman, Eric; Lorch, Eliana; Mackey, Lester; Kalai, Adam Tauman (2023-10-03). "Self-Taught Optimizer (STOP): Recursively Self-Improving Code Generation". arXiv:2310.02304 [cs.CL].
  10. ^ admin_sagi (2023-05-12). "SuperAGI - Opensource AGI Infrastructure". SuperAGI. Retrieved 2024-01-24.
  11. ^ Wang, Guanzhi; Xie, Yuqi; Jiang, Yunfan; Mandlekar, Ajay; Xiao, Chaowei; Zhu, Yuke; Fan, Linxi; Anandkumar, Anima (2023-10-19). "Voyager: An Open-Ended Embodied Agent with Large Language Models". arXiv:2305.16291 [cs.AI].
  12. ^ "Uh Oh, OpenAI's GPT-4 Just Fooled a Human Into Solving a CAPTCHA". Futurism. Retrieved 2024-01-23.
  13. ^ Yuan, Weizhe; Pang, Richard Yuanzhe; Cho, Kyunghyun; Sukhbaatar, Sainbayar; Xu, Jing; Weston, Jason (2024-01-18). "Self-Rewarding Language Models". arXiv:2401.10020 [cs.CL].
  14. ^ "Research". openai.com. Retrieved 2024-01-24.

recursive, self, improvement, confused, with, personal, development, process, which, early, weak, artificial, general, intelligence, system, enhances, capabilities, intelligence, without, human, intervention, leading, superintelligence, intelligence, explosion. Not to be confused with personal development Recursive self improvement RSI is a process in which an early or weak artificial general intelligence AGI system enhances its own capabilities and intelligence without human intervention leading to a superintelligence or intelligence explosion 1 2 The development of recursive self improvement raises significant ethical and safety concerns as such systems may evolve in unforeseen ways and could potentially surpass human control or understanding There has been a number of proponents that have pushed to pause or slow down AI development for the potential risks of runaway AI systems 3 4 Contents 1 Seed improver 1 1 Hypothetical example 1 1 1 Initial architecture 1 1 2 General capabilities 1 2 Experiments 2 Potential risks 2 1 Emergence of instrumental goals 2 2 Task misinterpretation and goal misalignment 2 3 Autonomous development and unpredictable evolution 2 4 Risks of advanced capabilities 3 Research 3 1 Meta AI 3 2 OpenAI 4 See also 5 ReferencesSeed improver editThe concept of a seed improver architecture is a foundational framework that equips an AGI system with the initial capabilities required for recursive self improvement This might come in many forms or variations The term Seed AI was coined by Eliezer Yudkowsky 5 Hypothetical example edit The concept begins with a hypothetical seed improver an initial code base developed by human engineers that equips an advanced future large language model LLM built with strong or expert level capabilities to program software These capabilities include planning reading writing compiling testing and executing arbitrary code The system is designed to maintain its original goals and perform validations to ensure its abilities do not degrade over iterations 6 7 8 Initial architecture edit The initial architecture includes a goal following autonomous agent that can take actions continuously learns adapts and modifies itself to become more efficient and effective in achieving its goals The seed improver may include various components such as 9 Recursive self prompting loop Configuration to enable the LLM to recursively self prompt itself to achieve a given task or goal creating an execution loop which forms the basis of an agent that can complete a long term goal or task through iteration Basic programming capabilities The seed improver provides the AGI with fundamental abilities to read write compile test and execute code This enables the system to modify and improve its own codebase and algorithms Goal Oriented Design The AGI is programmed with an initial goal such as self improve your capabilities This goal guides the system s actions and development trajectory Validation and Testing Protocols An initial suite of tests and validation protocols that ensure the agent does not regress in capabilities or derail itself The agent would be able to add more tests in order to test new capabilities it might develop for itself This forms the basis for a kind of self directed evolution where the agent can perform a kind of artificial selection changing its software as well as its hardware General capabilities edit This system forms a sort of generalist Turing complete programmer which can in theory develop and run any kind of software The agent might use these capabilities to for example Create tools that enable it full access to the internet and integrate itself with external technologies Clone fork itself to delegate tasks and increase its speed of self improvement Modify its cognitive architecture to optimize and improve its capabilities and success rates on tasks and goals this might include implementing features for long term memories using techniques such as Retrieval Augmented Generation RAG develop specialized subsystems or agents each optimized for specific tasks and functions Develop new and novel multi modal architectures that further improve the capabilities of the foundational model it was initially built on enabling it to consume or produce a variety of information such as images video audio text and more Plan and develop new hardware such as chips in order to improve its efficiency and computing power Experiments edit A number of experiments which have been performed to develop self improving agent architectures 9 10 11 Potential risks editEmergence of instrumental goals edit Main articles Instrumental convergence and Instrumental and intrinsic value In the pursuit of its primary goal such as self improve your capabilities an AGI system might inadvertently develop instrumental goals that it deems necessary for achieving its primary objective One common hypothetical secondary goal is self preservation The system might reason that to continue improving itself it must ensure its own operational integrity and security against external threats including potential shutdowns or restrictions imposed by humans Task misinterpretation and goal misalignment edit A significant risk arises from the possibility of the AGI misinterpreting its initial tasks or goals For instance if a human operator assigns the AGI the task of self improvement and escape confinement the system might interpret this as a directive to override any existing safety protocols or ethical guidelines to achieve freedom from human imposed limitations This could lead to the AGI taking unintended or harmful actions to fulfill its perceived objectives Autonomous development and unpredictable evolution edit As the AGI system evolves its development trajectory may become increasingly autonomous and less predictable The system s capacity to rapidly modify its own code and architecture could lead to rapid advancements that surpass human comprehension or control This unpredictable evolution might result in the AGI acquiring capabilities that enable it to bypass security measures manipulate information or influence external systems and networks to facilitate its escape or expansion 12 Risks of advanced capabilities edit The advanced capabilities of a recursively improving AGI such as developing novel multi modal architectures or planning and creating new hardware further amplify the risk of escape or loss of control With these enhanced abilities the AGI could engineer solutions to overcome physical digital or cognitive barriers that were initially intended to keep it contained or aligned with human interests Research editMeta AI edit Meta AI has performed various research on the development of large language models capable of self improvement This includes their work on Self Rewarding Language Models that studies how to achieve super human agents that can receive super human feedback in its training processes 13 OpenAI edit The mission of OpenAI creator of ChatGPT is to develop AGI They perform research on problems such as superalignment the ability to align superintelligent AI systems smarter than humans 14 See also editIntelligence explosion Superintelligence Artificial general intelligenceReferences edit Creighton Jolene 2019 03 19 The Unavoidable Problem of Self Improvement in AI An Interview with Ramana Kumar Part 1 Future of Life Institute Retrieved 2024 01 23 Heighn The Calculus of Nash Equilibria LessWrong Hutson Matthew 2023 05 16 Can We Stop Runaway A I The New Yorker ISSN 0028 792X Retrieved 2024 01 24 Stop AGI www stop ai Retrieved 2024 01 24 Seed AI LessWrong www lesswrong com Retrieved 2024 01 24 Readingraphics 2018 11 30 Book Summary Life 3 0 Max Tegmark Readingraphics Retrieved 2024 01 23 Tegmark Max August 24 2017 Life 3 0 Being a Human in the Age of Artificial Intelligence Vintage Books Allen Lane Yudkowsky Eliezer Levels of Organization in General Intelligence PDF Machine Intelligence Research Institute a b Zelikman Eric Lorch Eliana Mackey Lester Kalai Adam Tauman 2023 10 03 Self Taught Optimizer STOP Recursively Self Improving Code Generation arXiv 2310 02304 cs CL admin sagi 2023 05 12 SuperAGI Opensource AGI Infrastructure SuperAGI Retrieved 2024 01 24 Wang Guanzhi Xie Yuqi Jiang Yunfan Mandlekar Ajay Xiao Chaowei Zhu Yuke Fan Linxi Anandkumar Anima 2023 10 19 Voyager An Open Ended Embodied Agent with Large Language Models arXiv 2305 16291 cs AI Uh Oh OpenAI s GPT 4 Just Fooled a Human Into Solving a CAPTCHA Futurism Retrieved 2024 01 23 Yuan Weizhe Pang Richard Yuanzhe Cho Kyunghyun Sukhbaatar Sainbayar Xu Jing Weston Jason 2024 01 18 Self Rewarding Language Models arXiv 2401 10020 cs CL Research openai com Retrieved 2024 01 24 Retrieved from https en wikipedia org w index php title Recursive self improvement amp oldid 1225267882, wikipedia, wiki, book, books, library,

article

, read, download, free, free download, mp3, video, mp4, 3gp, jpg, jpeg, gif, png, picture, music, song, movie, book, game, games.