{"type":"rich","version":"1.0","provider_name":"Transistor","provider_url":"https://transistor.fm","author_name":"Future of Life Institute Podcast","title":"Connor Leahy on Why Humanity Risks Extinction from AGI","html":"<iframe width=\"100%\" height=\"180\" frameborder=\"no\" scrolling=\"no\" seamless src=\"https://share.transistor.fm/e/a8bde4d9\"></iframe>","width":"100%","height":180,"duration":7131,"description":"Connor Leahy joins the podcast to discuss the motivations of AGI corporations, how modern AI is \"grown\", the need for a science of intelligence, the effects of AI on work, the radical implications of superintelligence, open-source AI, and what you might be able to do about all of this.   \n\nHere's the document we discuss in the episode:   \n\nhttps://www.thecompendium.ai  \n\nTimestamps: \n\n00:00 The Compendium \n\n15:25 The motivations of AGI corps  \n\n31:17 AI is grown, not written  \n\n52:59 A science of intelligence \n\n01:07:50 Jobs, work, and AGI  \n\n01:23:19 Superintelligence  \n\n01:37:42 Open-source AI  \n\n01:45:07 What can we do?","thumbnail_url":"https://img.transistorcdn.com/fFhIC-s2qSlHXzmJI7qMGts2WuLwImi4tWmRLH9EdPg/rs:fill:0:0:1/w:400/h:400/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS81MmU5/MDZjZGQ5OTI0MDc5/YTk2ZTAxYTgwYTNk/M2VlOC5qcGc.webp","thumbnail_width":300,"thumbnail_height":300}