Difference between revisions of "AI predictions"

From GISAXS
Jump to: navigation, search
(Job Loss)
(Plans)
Line 97: Line 97:
 
* [https://cfg.eu/building-cern-for-ai/ Building CERN for AI: An institutional blueprint]
 
* [https://cfg.eu/building-cern-for-ai/ Building CERN for AI: An institutional blueprint]
 
* [https://arxiv.org/abs/2503.05710 AGI, Governments, and Free Societies]
 
* [https://arxiv.org/abs/2503.05710 AGI, Governments, and Free Societies]
 +
* [https://controlai.com/ Control AI]: [https://controlai.com/dip The Direct Institutional Plan]
  
 
==Philosophy==
 
==Philosophy==

Revision as of 14:38, 28 March 2025

Capability Scaling

GmZHL8xWQAAtFlF.jpeg


AGI Achievable

AGI Definition

Economic and Political

Job Loss

F-kVQuvWkAAemkr.png

Near-term

  • 2025-02: Sam Altman: Three Observations
    1. The intelligence of an AI model roughly equals the log of the resources used to train and run it.
    2. The cost to use a given level of AI falls about 10x every 12 months, and lower prices lead to much more use
    3. The socioeconomic value of linearly increasing intelligence is super-exponential in nature
  • 2025-03: Glimpses of AI Progress: Mental models for fast times

Overall

  • 2025-03: Kevin Roose (New York Times): Powerful A.I. Is Coming. We’re Not Ready. Three arguments for taking progress toward artificial general intelligence, or A.G.I., more seriously — whether you’re an optimist or a pessimist.
  • 2025-03: Nicholas Carlini: My Thoughts on the Future of "AI": "I have very wide error bars on the potential future of large language models, and I think you should too."

Surveys of Opinions/Predictions

Bad Outcomes

Intelligence Explosion

Psychology

Science & Technology Improvements

Plans

Philosophy

GlchEeObwAQ88NK.jpeg

Alignment

Strategic/Technical

Strategic/Policy

See Also