Difference between revisions of "AI safety"

From GISAXS
Jump to: navigation, search
(Light)
Line 1: Line 1:
 +
=Learning Resources=
 +
==Light==
 +
* [https://www.youtube.com/watch?v=xfMQ7hzyFW4 Writing Doom] (27m video): short film on Superintelligence (2024)
 +
* [https://orxl.org/ai-doom.html a casual intro to AI doom and alignment] (2022)
 +
* Anthony Aguirre: [https://keepthefuturehuman.ai/ Keep The Future Human]
 +
** [https://interactive.keepthefuturehuman.ai/ Interactive Explainer]
 +
** [https://keepthefuturehuman.ai/essay/ Essay: Keep the Future Human]
 +
** [https://www.youtube.com/watch?v=27KDl2uPiL8 We Can’t Stop AI – Here’s What To Do Instead] (4m video, 2025)
 +
** [https://www.youtube.com/watch?v=zeabrXV8zNE The 4 Rules That Could Stop AI Before It’s Too Late] (15m video, 2025)
 +
 +
==Deep==
 +
* [https://www.thecompendium.ai/ The Compendium: Humanity risks extinction from its very creations — AIs.] (2024)
 +
* [https://www.aisafetybook.com/ Introduction to AI Safety, Ethics, and Society] (Dan Hendrycks, [https://www.safe.ai/ Center for AI Safety])
 +
* [https://aisafety.info/ AI Safety FAQ]
 +
* [https://deepmindsafetyresearch.medium.com/introducing-our-short-course-on-agi-safety-1072adb7912c DeepMind short course on AGI safety]
  
 
=Description of Safety Concerns=
 
=Description of Safety Concerns=
Line 18: Line 33:
 
==Long-term  (x-risk)==
 
==Long-term  (x-risk)==
 
* [https://www.lesswrong.com/posts/uMQ3cqWDPHhjtiesc/agi-ruin-a-list-of-lethalities List AGI Ruin: A List of Lethalities] (Eliezer Yudkowsky)
 
* [https://www.lesswrong.com/posts/uMQ3cqWDPHhjtiesc/agi-ruin-a-list-of-lethalities List AGI Ruin: A List of Lethalities] (Eliezer Yudkowsky)
 
=Learning Resources=
 
==Light==
 
* [https://www.youtube.com/watch?v=xfMQ7hzyFW4 Writing Doom] (27m video): short film on Superintelligence (2024)
 
* [https://orxl.org/ai-doom.html a casual intro to AI doom and alignment] (2022)
 
* Anthony Aguirre: [https://keepthefuturehuman.ai/ Keep The Future Human]
 
** [https://interactive.keepthefuturehuman.ai/ Interactive Explainer]
 
** [https://keepthefuturehuman.ai/essay/ Essay: Keep the Future Human]
 
** [https://www.youtube.com/watch?v=27KDl2uPiL8 We Can’t Stop AI – Here’s What To Do Instead] (4m video, 2025)
 
** [https://www.youtube.com/watch?v=zeabrXV8zNE The 4 Rules That Could Stop AI Before It’s Too Late] (15m video, 2025)
 
 
==Deep==
 
* [https://www.thecompendium.ai/ The Compendium: Humanity risks extinction from its very creations — AIs.] (2024)
 
* [https://www.aisafetybook.com/ Introduction to AI Safety, Ethics, and Society] (Dan Hendrycks, [https://www.safe.ai/ Center for AI Safety])
 
* [https://aisafety.info/ AI Safety FAQ]
 
* [https://deepmindsafetyresearch.medium.com/introducing-our-short-course-on-agi-safety-1072adb7912c DeepMind short course on AGI safety]
 
  
 
=Status=
 
=Status=

Revision as of 09:45, 2 April 2025

Learning Resources

Light

Deep

Description of Safety Concerns

Key Concepts

Medium-term Risks

Long-term (x-risk)

Status

Policy

Proposals

Research

Demonstrations of Negative Use Capabilities

See Also