#431 – Roman Yampolskiy: Dangers of Superintelligent AI
Roman Yampolskiy is an AI safety researcher and author of a new book titled AI: Unexplainable, Unpredictable, Uncontrollable. Please support this podcast by checking out our sponsors: - Yahoo Finance: https://yahoofinance.com
Roman Yampolskiy is an AI safety researcher and author of a new book titled AI: Unexplainable, Unpredictable, Uncontrollable. Please support this podcast by checking out our sponsors:
–
Yahoo Finance
:
https://yahoofinance.com
–
MasterClass
:
https://masterclass.com/lexpod
to get 15% off
–
NetSuite
:
http://netsuite.com/lex
to get free product tour
–
LMNT
:
https://drinkLMNT.com/lex
to get free sample pack
–
Eight Sleep
:
https://eightsleep.com/lex
to get $350 off
Transcript: https://lexfridman.com/roman-yampolskiy-transcript
EPISODE LINKS:
Roman’s X:
https://twitter.com/romanyam
Roman’s Website:
http://cecs.louisville.edu/ry
Roman’s AI book:
https://amzn.to/4aFZuPb
PODCAST INFO:
Podcast website:
https://lexfridman.com/podcast
Apple Podcasts:
https://apple.co/2lwqZIr
Spotify:
https://spoti.fi/2nEwCF8
RSS:
https://lexfridman.com/feed/podcast/
YouTube Full Episodes:
https://youtube.com/lexfridman
YouTube Clips:
https://youtube.com/lexclips
SUPPORT & CONNECT:
– Check out the sponsors above, it’s the best way to support this podcast
– Support on Patreon:
https://www.patreon.com/lexfridman
– Twitter:
https://twitter.com/lexfridman
– Instagram:
https://www.instagram.com/lexfridman
– LinkedIn:
https://www.linkedin.com/in/lexfridman
– Facebook:
https://www.facebook.com/lexfridman
– Medium:
https://medium.com/@lexfridman
OUTLINE:
Here’s the timestamps for the episode. On some podcast players you should be able to click the timestamp to jump to that time.
(00:00) – Introduction
(09:12) – Existential risk of AGI
(15:25) – Ikigai risk
(23:37) – Suffering risk
(27:12) – Timeline to AGI
(31:44) – AGI turing test
(37:06) – Yann LeCun and open source AI
(49:58) – AI control
(52:26) – Social engineering
(54:59) – Fearmongering
(1:04:49) – AI deception
(1:11:23) – Verification
(1:18:22) – Self-improving AI
(1:30:34) – Pausing AI development
(1:36:51) – AI Safety
(1:46:35) – Current AI
(1:51:58) – Simulation
(1:59:16) – Aliens
(2:00:50) – Human mind
(2:07:10) – Neuralink
(2:16:15) – Hope for the future
(2:20:11) – Meaning of life