Home
cover of episode #431 – Roman Yampolskiy: Dangers of Superintelligent AI

#431 – Roman Yampolskiy: Dangers of Superintelligent AI

2024/6/2
logo of podcast Lex Fridman Podcast

Lex Fridman Podcast

Chapters

Shownotes Transcript

Roman Yampolskiy is an AI safety researcher and author of a new book titled AI: Unexplainable, Unpredictable, Uncontrollable. Please support this podcast by checking out our sponsors: – Yahoo Finance: https://yahoofinance.com) – MasterClass: https://masterclass.com/lexpod) to get 15% off – NetSuite: http://netsuite.com/lex) to get free product tour – LMNT: https://drinkLMNT.com/lex) to get free sample pack – Eight Sleep: https://eightsleep.com/lex) to get $350 off

Transcript: https://lexfridman.com/roman-yampolskiy-transcript)

EPISODE LINKS: Roman’s X: https://twitter.com/romanyam) Roman’s Website: http://cecs.louisville.edu/ry) Roman’s AI book: https://amzn.to/4aFZuPb)

PODCAST INFO: Podcast website: https://lexfridman.com/podcast) Apple Podcasts: https://apple.co/2lwqZIr) Spotify: https://spoti.fi/2nEwCF8) RSS: https://lexfridman.com/feed/podcast/) YouTube Full Episodes: https://youtube.com/lexfridman) YouTube Clips: https://youtube.com/lexclips)

SUPPORT & CONNECT: – Check out the sponsors above, it’s the best way to support this podcast – Support on Patreon: https://www.patreon.com/lexfridman) – Twitter: https://twitter.com/lexfridman) – Instagram: https://www.instagram.com/lexfridman) – LinkedIn: https://www.linkedin.com/in/lexfridman) – Facebook: https://www.facebook.com/lexfridman) – Medium: https://medium.com/@lexfridman)

OUTLINE: Here’s the timestamps for the episode. On some podcast players you should be able to click the timestamp to jump to that time. (00:00) – Introduction (09:12) – Existential risk of AGI (15:25) – Ikigai risk (23:37) – Suffering risk (27:12) – Timeline to AGI (31:44) – AGI turing test (37:06) – Yann LeCun and open source AI (49:58) – AI control (52:26) – Social engineering (54:59) – Fearmongering (1:04:49) – AI deception (1:11:23) – Verification (1:18:22) – Self-improving AI (1:30:34) – Pausing AI development (1:36:51) – AI Safety (1:46:35) – Current AI (1:51:58) – Simulation (1:59:16) – Aliens (2:00:50) – Human mind (2:07:10) – Neuralink (2:16:15) – Hope for the future (2:20:11) – Meaning of life