Holiday Special: AI Safety Update
Description
Welcome back to the show that keeps you informed on all things artificial intelligence and natural nonsense.
In our holiday episode, Mason opens a rather unique Christmas present from Perry, we invite a special guest to help explain the infamous "Paperclip Maximizer" thought experiment, and we discuss an interesting (and somewhat disturbing) new AI Safety paper from Apollo Research.
Want to leave us a voicemail? Here's the magic link to do just that: https://sayhi.chat/FAIK
You can also join our Discord server here: https://discord.gg/cThqEnMhJz
*** NOTES AND REFERENCES ***
An interesting cluster of new AI safety research papers:
- Apollo research: Frontier Models are Capable of In-context Scheming (Dec 5, 2024)
- YouTube Video: Apollo Research - AI Models Are Capable Of In Context Scheming Dec 2024
- YouTube Video: Cognitive Revolution - Emergency Pod: o1 Schemes Against Users, with Alexander Meinke from Apollo Research
- OpenAI o1 System Card (Dec 5, 2024)
- Anthropic: Alignment Faking in Large Language Models (Dec 18, 2024)
- Anthropic: Sycophancy to subterfuge: Investigating reward tampering in language models (June 17, 2024)
- Fudan University: Frontier AI systems have surpassed the self-replicating red line (Dec 9, 2024)
Other Interesting Bits:
- The Paperclip Maximizer thought experiment explanation
- Theory of Instrumental Convergence
- iPhone Game: Universal Paperclips
- VoxEU: AI and the paperclip problem
- Real Paperclips! 500 Pack Paper Clips (assorted sizes)
OpenAI Announces New o3 Reasoning Model:
- OpenAI's "12 Days of Ship-mas" announcement page
- YouTube video: OpenAI's announcement of their o3 Model
- TechCrunch: OpenAI announces new o3 models
- Wired: OpenAI Upgrades Its Smartest AI Model With Improved Reasoning Skills
- TechCrunch: OpenAI trained o1 and o3 to ‘think’ about its safety policy
- Matthew Berman YouTube video: OpenAI Unveils o3! AGI ACHIEVED!
- NewScientist: OpenAI's o3 model aced a test of AI reasoning – but it's still not AGI
- Yahoo Finance: OpenAI considers AGI clause removal for Microsoft investment
*** THE BOILERPLATE ***
About The FAIK Files:
The FAIK Files is an offshoot project from Perry Carpenter's most recent book, FAIK: A Practical Guide to Living in a World of Deepfakes, Disinformation, and AI-Generated Deceptions.
- Get the Book: FAIK: A Practical Guide to Living in a World of Deepfakes, Disinformation, and AI-Generated Deceptions (Amazon Associates link)
- Check out the website for more info: https://thisbookisfaik.com
Check out Perry & Mason's other show, the Digital Folklore Podcast:
- Apple Podcasts: https://podcasts.apple.com/us/podcast/digital-folklore/id1657374458
- Spotify: https://open.spotify.com/show/2v1BelkrbSRSkHEP4cYffj?si=u4XTTY4pR4qEqh5zMNSVQA
- Other: https://digitalfolklore.fm
Want to connect with us? Here's how:
Connect with Perry:
- Perry on LinkedIn: https://www.linkedin.com/in/perrycarpenter
- Perry on X: https://x.com/perrycarpenter
- Perry on BlueSky: https://bsky.app/profile/perrycarpenter.bsky.social
Connect with Mason:
- Mason on LinkedIn: https://www.linkedin.com/in/mason-amadeus-a853a7242/
- Mason on BlueSky: https://bsky.app/profile/pregnantsonic.com