Roman Yampolskiy is an AI safety researcher and author of a new book titled AI: Unexplainable, Unpredictable, Uncontrollable. Please support this podcast by checking out our sponsors:
- Yahoo Finance: https://yahoofinance.com
- MasterClass: https://masterclass.com/lexpod to get 15% off
- NetSuite: http://netsuite.com/lex to get free product tour
- LMNT: https://drinkLMNT.com/lex to get free sample pack
- Eight Sleep: https://eightsleep.com/lex to get $350 off

Transcript: https://lexfridman.com/roman-yampolskiy-transcript

EPISODE LINKS:
Roman's X: https://twitter.com/romanyam
Roman's Website: http://cecs.louisville.edu/ry
Roman's AI book: https://amzn.to/4aFZuPb

PODCAST INFO:
Podcast website: https://lexfridman.com/podcast
Apple Podcasts: https://apple.co/2lwqZIr
Spotify: https://spoti.fi/2nEwCF8
RSS: https://lexfridman.com/feed/podcast/
YouTube Full Episodes: https://youtube.com/lexfridman
YouTube Clips: https://youtube.com/lexclips

SUPPORT & CONNECT:
- Check out the sponsors above, it's the best way to support this podcast
- Support on Patreon: https://www.patreon.com/lexfridman
- Twitter: https://twitter.com/lexfridman
- Instagram: https://www.instagram.com/lexfridman
- LinkedIn: https://www.linkedin.com/in/lexfridman
- Facebook: https://www.facebook.com/lexfridman
- Medium: https://medium.com/@lexfridman

OUTLINE:
Here's the timestamps for the episode. On some podcast players you should be able to click the timestamp to jump to that time.
(00:00) - Introduction
(09:12) - Existential risk of AGI
(15:25) - Ikigai risk
(23:37) - Suffering risk
(27:12) - Timeline to AGI
(31:44) - AGI turing test
(37:06) - Yann LeCun and open source AI
(49:58) - AI control
(52:26) - Social engineering
(54:59) - Fearmongering
(1:04:49) - AI deception
(1:11:23) - Verification
(1:18:22) - Self-improving AI
(1:30:34) - Pausing AI development
(1:36:51) - AI Safety
(1:46:35) - Current AI
(1:51:58) - Simulation
(1:59:16) - Aliens
(2:00:50) - Human mind
(2:07:10) - Neuralink
(2:16:15) - Hope for the future
(2:20:11) - Meaning of life

Roman Yampolskiy is an AI safety researcher and author of a new book titled AI: Unexplainable, Unpredictable, Uncontrollable. Please support this podcast by checking out our sponsors:

– Yahoo Finance: https://yahoofinance.com

– MasterClass: https://masterclass.com/lexpod to get 15% off

– NetSuite: http://netsuite.com/lex to get free product tour

– LMNT: https://drinkLMNT.com/lex to get free sample pack

– Eight Sleep: https://eightsleep.com/lex to get $350 off


Transcript: https://lexfridman.com/roman-yampolskiy-transcript


EPISODE LINKS:

Roman’s X: https://twitter.com/romanyam

Roman’s Website: http://cecs.louisville.edu/ry

Roman’s AI book: https://amzn.to/4aFZuPb


PODCAST INFO:

Podcast website: https://lexfridman.com/podcast

Apple Podcasts: https://apple.co/2lwqZIr

Spotify: https://spoti.fi/2nEwCF8

RSS: https://lexfridman.com/feed/podcast/

YouTube Full Episodes: https://youtube.com/lexfridman

YouTube Clips: https://youtube.com/lexclips


SUPPORT & CONNECT:

– Check out the sponsors above, it’s the best way to support this podcast

– Support on Patreon: https://www.patreon.com/lexfridman

– Twitter: https://twitter.com/lexfridman

– Instagram: https://www.instagram.com/lexfridman

– LinkedIn: https://www.linkedin.com/in/lexfridman

– Facebook: https://www.facebook.com/lexfridman

– Medium: https://medium.com/@lexfridman


OUTLINE:

Here’s the timestamps for the episode. On some podcast players you should be able to click the timestamp to jump to that time.

(00:00) – Introduction

(09:12) – Existential risk of AGI

(15:25) – Ikigai risk

(23:37) – Suffering risk

(27:12) – Timeline to AGI

(31:44) – AGI turing test

(37:06) – Yann LeCun and open source AI

(49:58) – AI control

(52:26) – Social engineering

(54:59) – Fearmongering

(1:04:49) – AI deception

(1:11:23) – Verification

(1:18:22) – Self-improving AI

(1:30:34) – Pausing AI development

(1:36:51) – AI Safety

(1:46:35) – Current AI

(1:51:58) – Simulation

(1:59:16) – Aliens

(2:00:50) – Human mind

(2:07:10) – Neuralink

(2:16:15) – Hope for the future

(2:20:11) – Meaning of life

Twitter Mentions