DiscoverLex Fridman Podcast#431 – Roman Yampolskiy: Dangers of Superintelligent AI
#431 – Roman Yampolskiy: Dangers of Superintelligent AI

#431 – Roman Yampolskiy: Dangers of Superintelligent AI

Update: 2024-06-0213
Share

Description

Roman Yampolskiy is an AI safety researcher and author of a new book titled AI: Unexplainable, Unpredictable, Uncontrollable. Please support this podcast by checking out our sponsors:

Yahoo Finance: https://yahoofinance.com

MasterClass: https://masterclass.com/lexpod to get 15% off

NetSuite: http://netsuite.com/lex to get free product tour

LMNT: https://drinkLMNT.com/lex to get free sample pack

Eight Sleep: https://eightsleep.com/lex to get $350 off


Transcript: https://lexfridman.com/roman-yampolskiy-transcript


EPISODE LINKS:

Roman’s X: https://twitter.com/romanyam

Roman’s Website: http://cecs.louisville.edu/ry

Roman’s AI book: https://amzn.to/4aFZuPb


PODCAST INFO:

Podcast website: https://lexfridman.com/podcast

Apple Podcasts: https://apple.co/2lwqZIr

Spotify: https://spoti.fi/2nEwCF8

RSS: https://lexfridman.com/feed/podcast/

YouTube Full Episodes: https://youtube.com/lexfridman

YouTube Clips: https://youtube.com/lexclips


SUPPORT & CONNECT:

– Check out the sponsors above, it’s the best way to support this podcast

– Support on Patreon: https://www.patreon.com/lexfridman

– Twitter: https://twitter.com/lexfridman

– Instagram: https://www.instagram.com/lexfridman

– LinkedIn: https://www.linkedin.com/in/lexfridman

– Facebook: https://www.facebook.com/lexfridman

– Medium: https://medium.com/@lexfridman


OUTLINE:

Here’s the timestamps for the episode. On some podcast players you should be able to click the timestamp to jump to that time.

(00:00 ) – Introduction

(09:12 ) – Existential risk of AGI

(15:25 ) – Ikigai risk

(23:37 ) – Suffering risk

(27:12 ) – Timeline to AGI

(31:44 ) – AGI turing test

(37:06 ) – Yann LeCun and open source AI

(49:58 ) – AI control

(52:26 ) – Social engineering

(54:59 ) – Fearmongering

(1:04:49 ) – AI deception

(1:11:23 ) – Verification

(1:18:22 ) – Self-improving AI

(1:30:34 ) – Pausing AI development

(1:36:51 ) – AI Safety

(1:46:35 ) – Current AI

(1:51:58 ) – Simulation

(1:59:16 ) – Aliens

(2:00:50 ) – Human mind

(2:07:10 ) – Neuralink

(2:16:15 ) – Hope for the future

(2:20:11 ) – Meaning of life

Comments (2)

saul hernandez

8:57

Jun 4th
Reply (1)
In Channel
loading
00:00
00:00
x

0.5x

0.8x

1.0x

1.25x

1.5x

2.0x

3.0x

Sleep Timer

Off

End of Episode

5 Minutes

10 Minutes

15 Minutes

30 Minutes

45 Minutes

60 Minutes

120 Minutes

#431 – Roman Yampolskiy: Dangers of Superintelligent AI

#431 – Roman Yampolskiy: Dangers of Superintelligent AI

Lex Fridman