<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[Retort AI]]></title><description><![CDATA[Distilling the major events and challenges in the world of artificial intelligence and machine learning, from Thomas Krendl Gilbert and Nathan Lambert.

]]></description><link>https://www.retortai.com</link><image><url>https://substackcdn.com/image/fetch/$s_!AzON!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcbad298c-6074-441b-ad43-d5df6dbf101d_800x800.png</url><title>Retort AI</title><link>https://www.retortai.com</link></image><generator>Substack</generator><lastBuildDate>Fri, 03 Apr 2026 19:53:33 GMT</lastBuildDate><atom:link href="https://www.retortai.com/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Nathan Lambert]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[retortai@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[retortai@substack.com]]></itunes:email><itunes:name><![CDATA[Nathan Lambert]]></itunes:name></itunes:owner><itunes:author><![CDATA[Nathan Lambert]]></itunes:author><googleplay:owner><![CDATA[retortai@substack.com]]></googleplay:owner><googleplay:email><![CDATA[retortai@substack.com]]></googleplay:email><googleplay:author><![CDATA[Nathan Lambert]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[Tom leaves stealth: Hortus AI]]></title><description><![CDATA[An exciting week for Tom, who tells Nate about his company Hortus AI, whose mission is to make AI accountable to local communities.]]></description><link>https://www.retortai.com/p/tom-leaves-stealth-hortus-ai-5ad</link><guid isPermaLink="false">https://www.retortai.com/p/tom-leaves-stealth-hortus-ai-5ad</guid><dc:creator><![CDATA[Nathan Lambert]]></dc:creator><pubDate>Tue, 18 Feb 2025 21:26:52 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/162163031/b1aa3a40b326387d53f1ddbfcdf4c0b5.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>An exciting week for Tom, who tells Nate about his company Hortus AI, whose mission is to make AI accountable to local communities. We cover a lot of classic Retort themes as Tom makes a case for what's missing from AI development, and how models could be more healthily integrated into everyday people's lives.</p><p>Press release: https://hortus.ai/wp-content/uploads/2025/02/Hortus-AI-Press-Release_Feb25.pdf<br>Company website: https://hortus.ai/<br>Link to whitepaper: <a href="https://www.newamerica.org/rethinkai/policy-papers/a-sustainable-path-for-ai-development/">https://www.newamerica.org/rethinkai/policy-papers/a-sustainable-path-for-ai-development/</a></p>]]></content:encoded></item><item><title><![CDATA[We ask again: Is AI a science?]]></title><description><![CDATA[Tom and Nate sit down for a classic discussion of the role of AI in the modern philosophy of science.]]></description><link>https://www.retortai.com/p/we-ask-again-is-ai-a-science-3a7</link><guid isPermaLink="false">https://www.retortai.com/p/we-ask-again-is-ai-a-science-3a7</guid><dc:creator><![CDATA[Nathan Lambert]]></dc:creator><pubDate>Fri, 10 Jan 2025 12:00:00 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/162163032/bac65d7fc8d50a5ddcff9935402a609d.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>Tom and Nate sit down for a classic discussion of the role of AI in the modern philosophy of science. Much of this discussion is based on Thomas Samuel Kuhn's influential book&nbsp;<em>The Structure of Scientific Revolutions</em>. We ask -- is AI a science in the Kuhn'ian sense? Will the "paradigm" worldview apply to other sciences post AI? How will scientific institutions manage the addition of AI?</p><p>We promised an AI for science reading list, so here it is:</p><ul><li><p><a href="https://youtu.be/ugvHCXCOmm4?si=1hnlvue8M4pV2TqC">[Dario interview with Lex]</a> <a href="https://youtu.be/ugvHCXCOmm4?si=1hnlvue8M4pV2TqC">https://youtu.be/ugvHCXCOmm4?si=1hnlvue8M4pV2TqC</a></p></li><li><p><a href="https://open.substack.com/pub/cell/p/levers?r=68gy5&amp;utm_medium=ios">Levers for biological progress</a> <a href="https://open.substack.com/pub/cell/p/levers?r=68gy5&amp;utm_medium=ios">https://open.substack.com/pub/cell/p/levers?r=68gy5&amp;utm_medium=ios</a></p></li><li><p><a href="https://x.com/AdamMarblestone/status/1845158919523664019">X thread on theories of change in scienceshttps://x.com/AdamMarblestone/status/1845158919523664019</a></p></li><li><p>whitepaper linked by seb krier</p></li><li><p>Dwarkesh physics pod <a href="https://open.substack.com/pub/dwarkesh/p/adam-brown?r=68gy5&amp;utm_medium=ios">https://open.substack.com/pub/dwarkesh/p/adam-brown?r=68gy5&amp;utm_medium=ios</a> &#8212; Nobel in physics went to ai</p></li><li><p>Ai policy perspectives piece <em><a href="https://www.aipolicyperspectives.com/p/a-new-golden-age-of-discovery">A new golden age of discovery</a></em><a href="https://www.aipolicyperspectives.com/p/a-new-golden-age-of-discovery">https://www.aipolicyperspectives.com/p/a-new-golden-age-of-discovery</a></p></li><li><p>Owl posting checking recent NeurIPS papers <a href="https://www.owlposting.com/p/can-o1-preview-find-major-mistakes">https://www.owlposting.com/p/can-o1-preview-find-major-mistakes</a> based on idea from Ethan Mollick <a href="https://x.com/emollick/status/1868329599438037491">https://x.com/emollick/status/1868329599438037491</a></p></li><li><p>also another post on the subject <a href="https://open.substack.com/pub/amistrongeryet/p/the-black-spatula-project?r=68gy5&amp;utm_medium=ios">https://open.substack.com/pub/amistrongeryet/p/the-black-spatula-project?r=68gy5&amp;utm_medium=ios</a></p></li><li><p>Kuhn&#8217;s <em>The Structure of Scientific Revolutions</em></p></li><li><p>intrinsic perspective <a href="https://open.substack.com/pub/erikhoel/p/great-scientists-follow-intuition?r=68gy5&amp;utm_medium=ios">https://open.substack.com/pub/erikhoel/p/great-scientists-follow-intuition?r=68gy5&amp;utm_medium=ios</a></p></li></ul><p>Get The Retort (https://retortai.com/)&#8230;<br>&#8230; on YouTube: https://www.youtube.com/@TheRetortAIPodcast<br>&#8230; on Spotify: https://open.spotify.com/show/0FDjH8ujv7p8ELZGkBvrfv?si=fa17a4d408f245ee<br>&#8230; on Apple Podcasts: https://podcasts.apple.com/us/podcast/the-retort-ai-podcast/id1706223190<br>&#8230; Follow Interconnects: https://www.interconnects.ai/<br>&#8230; email us: mail@retortai.com</p>]]></content:encoded></item><item><title><![CDATA[The Retort's biggest AI stories of 2024]]></title><description><![CDATA[We're back!]]></description><link>https://www.retortai.com/p/the-retorts-biggest-ai-stories-of-c5a</link><guid isPermaLink="false">https://www.retortai.com/p/the-retorts-biggest-ai-stories-of-c5a</guid><dc:creator><![CDATA[Nathan Lambert]]></dc:creator><pubDate>Fri, 06 Dec 2024 12:00:00 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/162163033/f30e7c9d9cd03af043de3d06a9cb17a1.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>We're back! Tom and Nate catch up after the Thanksgiving holiday. Our main question was -- what were the biggest AI stories of the year? We touch on the core themes of the show: infrastructure, AI realities, and and antitrust. The power buildout to scale out AI is going to have very real long-term impacts.</p><p>Some links this week:<br>* Ben Thompson's, The End of the Beginning: https://stratechery.com/2020/the-end-of-the-beginning/<br>* Miles Brundage's Substack: https://milesbrundage.substack.com/p/why-im-leaving-openai-and-what-im<br>* Stochastic Parrots paper: https://dl.acm.org/doi/10.1145/3442188.3445922</p><p>Thanks for listening!&nbsp;</p><p>Get The Retort (https://retortai.com/)&#8230;<br>&#8230; on YouTube: https://www.youtube.com/@TheRetortAIPodcast<br>&#8230; on Spotify: https://open.spotify.com/show/0FDjH8ujv7p8ELZGkBvrfv?si=fa17a4d408f245ee<br>&#8230; on Apple Podcasts: https://podcasts.apple.com/us/podcast/the-retort-ai-podcast/id1706223190<br>&#8230; Follow Interconnects: https://www.interconnects.ai/<br>&#8230; email us: mail@retortai.com</p>]]></content:encoded></item><item><title><![CDATA[The Nobel Albatross]]></title><description><![CDATA[Tom and Nate catch up on the happenings in AI.]]></description><link>https://www.retortai.com/p/the-nobel-albatross-a02</link><guid isPermaLink="false">https://www.retortai.com/p/the-nobel-albatross-a02</guid><dc:creator><![CDATA[Nathan Lambert]]></dc:creator><pubDate>Fri, 11 Oct 2024 11:00:00 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/162163034/885421dc2eb62ddef9d93770cbec8b94.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>Tom and Nate catch up on the happenings in AI. Of course, we're focused on the biggest awards available to us as esteemed scientists (or something close enough) -- the Nobel Prizes! What does it mean in the trajectory of AI for Hinton and Hassabis to carry added scientific weight. Honestly, feels like a sinking ship.&nbsp;</p><p>Some links:<br>* Schmidhuber tweet: https://x.com/SchmidhuberAI/status/1844022724328394780<br>* Hinton "I'm proud my student fired Sam": https://x.com/Grady_Booch/status/1844145422824243290</p><p>00:00 Introduction<br>04:43 Criticism of AI-related Nobel Prize awards<br>09:06 Geoffrey Hinton's comments on winning the Nobel Prize<br>18:14 Debate on who should be credited for current AI advancements<br>25:53 Changes in the nature of scientific research and recognition<br>34:44 Changes in AI safety culture and company dynamics<br>37:27 Discussion on AI scaling and its impact on the industry<br>42:21 Reflection on the ongoing AI hype cycle</p><p>Retort on YouTube: https://www.youtube.com/@TheRetortAIPodcast<br>Retort on Twitter: https://x.com/retortai<br>Retort website: https://retortai.com/<br>Retort email: mail at retortai dot com</p>]]></content:encoded></item><item><title><![CDATA[Claude Needs a Constitutional Convention]]></title><description><![CDATA[Tom and Nate catch up on recent events (before the OpenAI o1 release) and opportunities in transparency/policy.]]></description><link>https://www.retortai.com/p/claude-needs-a-constitutional-convention-357</link><guid isPermaLink="false">https://www.retortai.com/p/claude-needs-a-constitutional-convention-357</guid><dc:creator><![CDATA[Nathan Lambert]]></dc:creator><pubDate>Fri, 13 Sep 2024 11:00:00 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/162163035/530de39d68da4a32926a554507766e75.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>Tom and Nate catch up on recent events (before the OpenAI o1 release) and opportunities in transparency/policy. We recap the legendary scam of Matt from IT department, why disclosing the outcomes of process is not enough, and more. This is a great episode on understanding why the process technology was birthed from is just as important as the outcome!</p><p>Some links:<br>* Nathan's post on Model Specs for regulation https://www.interconnects.ai/p/a-post-training-approach-to-ai-regulation<br>* Nathan's post on inference spend https://www.interconnects.ai/p/openai-strawberry-and-inference-scaling-laws</p><p>Send your questions to mail at retortai dot com</p>]]></content:encoded></item><item><title><![CDATA[Avoiding the AI burnout]]></title><description><![CDATA[Tom and Nate catch up on core themes of AI after a somewhat unintended summer break.]]></description><link>https://www.retortai.com/p/avoiding-the-ai-burnout-6e6</link><guid isPermaLink="false">https://www.retortai.com/p/avoiding-the-ai-burnout-6e6</guid><dc:creator><![CDATA[Nathan Lambert]]></dc:creator><pubDate>Mon, 26 Aug 2024 11:00:00 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/162163036/3a9370639bcdf02ad7ba8edd30cf2d34.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>Tom and Nate catch up on core themes of AI after a somewhat unintended summer break. We discuss the moral groundings and philosophy of what we're building, our travels, The Anxious Generation, AGI obsessions, an update on AI Ethics vs. AI Safety, and plenty more in between.</p><p>As always, contact us at mail@retortai.com</p><p>Some links we mention in the episode:<br>* The Emotional Dog and its Rational Tail https://motherjones.com/wp-content/uploads/emotional_dog_and_rational_tail.pdf<br>* The Anxious Generation https://www.amazon.com/Anxious-Generation-Rewiring-Childhood-Epidemic/dp/0593655036<br>* Shadow Lake Lodge https://shadowlakelodge.com/<br>* Recent Dwarkesh Podcast https://www.dwarkeshpatel.com/p/joe-carlsmith</p>]]></content:encoded></item><item><title><![CDATA[What we are getting wrong about AI regulation]]></title><description><![CDATA[Tom and Nate catch up on the rapidly evolving (and political) space of AI regulation.]]></description><link>https://www.retortai.com/p/what-we-are-getting-wrong-about-ai-63b</link><guid isPermaLink="false">https://www.retortai.com/p/what-we-are-getting-wrong-about-ai-63b</guid><dc:creator><![CDATA[Nathan Lambert]]></dc:creator><pubDate>Fri, 19 Jul 2024 11:00:00 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/162163037/d05ab7845328d96c4d8c51ae65d0d1f9.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>Tom and Nate catch up on the rapidly evolving (and political) space of AI regulation. We cover CA SB 1047, recent policing of data scraping, presidential appointees, antitrust intention vs. implementation, FLOP thresholds, and everything else touching the future of large ML models.</p><p>Nate's internet cut out, so this episode ends a little abruptly. Reach out with any questions to mail at retortai.com</p><p>Some links:<br>- night falls on the cumberlands https://en.wikipedia.org/wiki/Night_Comes_to_the_Cumberlands<br>- hillbilly elegy https://en.wikipedia.org/wiki/Hillbilly_Elegy<br>- wired piece on data https://www.wired.com/story/youtube-training-data-apple-nvidia-anthropic/<br>- nate's recent piece on AI regulation https://www.interconnects.ai/p/sb-1047-and-open-weights</p><p>00:00&nbsp; Intro&nbsp;<br>01:19 Training Data and the Media&nbsp;<br>03:43 Norms, Power, and the Limits of Regulation<br>08:52 OpenAI's Business Model<br>12:33&nbsp; Antitrust: The Essential Tool for Governing AI<br>17:11 Users as Afterthoughts<br>20:07 Depoliticizing AI&nbsp;<br>26:14&nbsp; "Breaking Bad" &amp; the AI Parallel<br>28:11&nbsp; The "Little Tech" Agenda<br>31:03&nbsp; Reframing the Narrative of Big Tech &nbsp;<br>32:20&nbsp; "The Lean Startup" &amp; AI's Uncertainty</p>]]></content:encoded></item><item><title><![CDATA[AI, feedback, and population public health]]></title><description><![CDATA[Tom and Nate revisit one of their old ideas -- AI through the lens of public health infrastructure, and especially alignment.]]></description><link>https://www.retortai.com/p/ai-feedback-and-population-public-5e0</link><guid isPermaLink="false">https://www.retortai.com/p/ai-feedback-and-population-public-5e0</guid><dc:creator><![CDATA[Nathan Lambert]]></dc:creator><pubDate>Fri, 28 Jun 2024 11:00:00 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/162163038/9a494aace54b658884d4ffb21f53cdfd.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>Tom and Nate revisit one of their old ideas -- AI through the lens of public health infrastructure, and especially alignment. Sorry about Tom's glitchy audio, I figured it out after the fact that he was talking into the microphone at the wrong angle. Regardless, here are some links for this week. Links:<br>- Data foundry for AI https://scale.com/blog/scale-ai-series-f&nbsp;<br>- Information piece on Scale AI ($) https://www.theinformation.com/articles/why-a-14-billion-startup-is-now-hiring-phds-to-train-ai-from-their-living-rooms?shared=168f685a864ca709&nbsp;<br>- ChatGPT compounding math: https://chatgpt.com/share/2c19a357-acb2-441d-8203-946b74ce785c&nbsp;</p><p>contact us at mail at retortai dot com</p><p>00:00 Intro<br>00:39 Chicago's Tech Scene and the "The Bear"<br>01:22 AI and Public Health: A New Framework<br>08:17 Lessons for AI from Sanitation Infrastructure<br>12:58 The Mental Health Impact of Generative AI<br>23:28 Aligning AI with Diverse Societal Values<br>27:06 Power Dynamics in AI's Development<br>33:02 The Need for a Neutral AI Research Body (NAIRR)<br>36:57 New Regulations for a New Era of AI<br>41:05 Outro: Join the Conversation</p>]]></content:encoded></item><item><title><![CDATA[Apple sends a memo to the AGI faithful]]></title><description><![CDATA[Tom and Nate caught up last week (sorry for the editing delay) on the big two views of the AI future: Apple Intelligence and Situational Awareness (Nationalistic AI doom prevention).]]></description><link>https://www.retortai.com/p/apple-sends-a-memo-to-the-agi-faithful-700</link><guid isPermaLink="false">https://www.retortai.com/p/apple-sends-a-memo-to-the-agi-faithful-700</guid><dc:creator><![CDATA[Nathan Lambert]]></dc:creator><pubDate>Fri, 21 Jun 2024 14:48:18 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/162163039/c9576e689172d65e59378b4eeaff1b7e.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>Tom and Nate caught up last week (sorry for the editing delay) on the big two views of the AI future: Apple Intelligence and Situational Awareness (Nationalistic AI doom prevention). One of our best episodes, here are the links:<br>* The Kekul&#233; Problem https://en.wikipedia.org/wiki/The_Kekul%C3%A9_Problem<br>* Truth and Method https://en.wikipedia.org/wiki/Truth_and_Method<br>* Situational Awareness https://situational-awareness.ai/</p><p>00:00 A Hypothetical Life: From Germany to AGI<br>01:20 Leopold Aschenbrenner: Situational Awareness and Extrapolation<br>02:01 The Retort: Apple vs. Doomsday AI<br>03:40 Credentials and Social Choice Theory<br>05:14 Dissecting "Situational Awareness": Hype vs. Reality<br>07:16 The Limits of Language Models: Are They Really Intelligent?<br>11:04 Apple's Vision: AI for Consumers, Not Conquerors<br>13:53 Silicon Valley Myopia and the Geopolitics of AI<br>18:25 Beyond Benchmarks: The Scientist vs. The Engineer<br>22:04 What is Intelligence? The Narrowness of Human Fixation<br>24:32 A Growing Disrespect for Language?<br>27:40 The Power of Talking to Language Models<br>32:50 Language: Representation or Revelation?<br>38:54 The Future of Meaning: Will AI Obliterate Art?<br>45:32 A Vision for AI as Public Infrastructure</p>]]></content:encoded></item><item><title><![CDATA[Murky waters in AI policy]]></title><description><![CDATA[Tom and Nate catch up on many AI policy happenings recently.]]></description><link>https://www.retortai.com/p/murky-waters-in-ai-policy-b5d</link><guid isPermaLink="false">https://www.retortai.com/p/murky-waters-in-ai-policy-b5d</guid><dc:creator><![CDATA[Nathan Lambert]]></dc:creator><pubDate>Fri, 31 May 2024 11:00:00 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/162163040/fab92d3b9341703bc70708dfa18a7425.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>Tom and Nate catch up on many AI policy happenings recently. California's "anti open source" 1047 bill, the senate AI roadmap, Google's search snaifu, OpenAI's normal nonsense, and reader feedback! A bit of a mailbag. Enjoy.</p><p>00:00 Murky waters in AI policy<br>00:33 The Senate AI Roadmap<br>05:14 The Executive Branch Takes the Lead<br>08:33 California's Senate AI Bill<br>22:22 OpenAI's Two Audiences<br>28:53 The Problem with OpenAI Model Spec<br>39:50 A New World of AI Regulation</p><p>A bunch of links...<br>Data and society whitepaper: https://static1.squarespace.com/static/66465fcd83d1881b974fe099/t/664b866c9524f174acd7931c/1716225644575/24.05.18+-+AI+Shadow+Report+V4.pdf<br>https://senateshadowreport.com/&nbsp;</p><p>California bill<br>https://www.hyperdimensional.co/p/california-senate-passes-sb-1047&nbsp;<br>https://legiscan.com/CA/text/SB1047/id/2999979&nbsp;</p><p>Data walls<br>https://www.interconnects.ai/p/the-data-wall&nbsp;</p><p>Interconnects Merch<br>https://interconnects.myshopify.com/</p>]]></content:encoded></item><item><title><![CDATA[ChatGPT talks: diamond of the season or quite the scandal?]]></title><description><![CDATA[Tom and Nate discuss two major OpenAI happenings in the last week.]]></description><link>https://www.retortai.com/p/chatgpt-talks-diamond-of-the-season-850</link><guid isPermaLink="false">https://www.retortai.com/p/chatgpt-talks-diamond-of-the-season-850</guid><dc:creator><![CDATA[Nathan Lambert]]></dc:creator><pubDate>Fri, 17 May 2024 11:00:00 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/162163041/de04cb942fe7e22790670bdcb75ac8b6.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>Tom and Nate discuss two major OpenAI happenings in the last week. The popular one, the chat assistant, and what it reveals about OpenAI's worldview. We pair this with discussion of OpenAI's new Model Spec, which details their RLHF goals: https://cdn.openai.com/spec/model-spec-2024-05-08.html</p><p>This is a monumental week for AI. The product transition is completed, we can't just be researchers anymore.</p><p>00:00 Guess the Donkey Kong Character<br>00:50 OpenAI's New AI Girlfriend<br>07:08 OpenAI's Business Model and Responsible AI<br>08:45 GPT-2 Chatbot Thing and OpenAI's Weirdness<br>12:48 OpenAI and the Mystery Box<br>19:10 The Blurring Boundaries of Intimacy and Technology<br>22:05 Rousseau's Discourse on Inequality and the Impact of Technology<br>26:16 OpenAI's Model Spec and Its Objectives<br>30:10 The Unintelligibility of "Benefiting Humanity"<br>37:01 The Chain of Command and the Paradox of AI Love<br>45:46 The Form and Content of OpenAI's Model Spec<br>48:51 The Future of AI and Societal Disruptions</p>]]></content:encoded></item><item><title><![CDATA[Three pillars of AI power]]></title><description><![CDATA[Tom and Nate discuss the shifting power landscape in AI.]]></description><link>https://www.retortai.com/p/three-pillars-of-ai-power-f08</link><guid isPermaLink="false">https://www.retortai.com/p/three-pillars-of-ai-power-f08</guid><dc:creator><![CDATA[Nathan Lambert]]></dc:creator><pubDate>Fri, 10 May 2024 11:00:00 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/162163042/129bf3f2a18a373de231b9439d680437.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>Tom and Nate discuss the shifting power landscape in AI. They try to discern what is special about Silicon Valley's grasp on the ecosystem and what other types of power (e.g. those in New York and Washington DC) will do to mobilize their influence.&nbsp;</p><p>Here's the one Tweet we referenced on the FAccT community: https://twitter.com/KLdivergence/status/1653843497932267520</p><p>00:00: Introduction and Cryptozoologists<br>02:00: DC and the National AI Research Resource (NAIR)<br>05:34: The Three Legs of the AI World: Silicon Valley, New York, and DC<br>11:00: The AI Safety vs. Ethics Debate<br>13:42: The Rise of the Third Entity: The Government's Role in AI<br>19:42: New York's Influence and the Power of Narrative<br>29:36: Silicon Valley's Insularity and the Need for Regulation<br>36:50: The Amazon Antitrust Paradox and the Shifting Landscape<br>48:20: The Energy Conundrum and the Need for Policy Solutions<br>56:34: Conclusion: Finding Common Ground and Building a Better Future for AI</p>]]></content:encoded></item><item><title><![CDATA[Llama 3: Can't Compete with a Capuchin]]></title><description><![CDATA[Tom and Nate cover the state of the industry after Llama 3.]]></description><link>https://www.retortai.com/p/llama-3-cant-compete-with-a-capuchin-b38</link><guid isPermaLink="false">https://www.retortai.com/p/llama-3-cant-compete-with-a-capuchin-b38</guid><dc:creator><![CDATA[Nathan Lambert]]></dc:creator><pubDate>Fri, 26 Apr 2024 11:00:00 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/162163043/3f57b662f44f548d9e7c23bc1bea08e2.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>Tom and Nate cover the state of the industry after Llama 3. Is Zuck the best storyteller in AI? Is he the best CEO? Are CEOs doing anything other than buying compute? We cover what it means to be successful at the highest level this week.&nbsp;</p><p>Links:<br>Dwarkesh interview with Zuck https://www.dwarkeshpatel.com/p/mark-zuckerberg&nbsp;<br>Capuchin monkey https://en.wikipedia.org/wiki/Capuchin_monkey&nbsp;</p><p>00:00 Introductions &amp; advice from a wolf<br>00:45 Llama 3<br>07:15 Resources and investment required for large language models<br>14:10 What it means to be a leader in the rapidly evolving AI landscape<br>22:07 How much of AI progress is driven by stories vs resources<br>29:41 Critiquing the concept of Artificial General Intelligence (AGI)<br>38:10 Misappropriation of the term AGI by tech leaders<br>42:09 The future of open models and AI development</p>]]></content:encoded></item><item><title><![CDATA[Into the AI Trough of Disillusionment]]></title><description><![CDATA[Tom and Nate catch up after a few weeks off the pod.]]></description><link>https://www.retortai.com/p/into-the-ai-trough-of-disillusionment-900</link><guid isPermaLink="false">https://www.retortai.com/p/into-the-ai-trough-of-disillusionment-900</guid><dc:creator><![CDATA[Nathan Lambert]]></dc:creator><pubDate>Fri, 12 Apr 2024 11:00:00 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/162163044/79ee8d9218c08266f953e1a338f456e4.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>Tom and Nate catch up after a few weeks off the pod. We discuss what it means for the pace and size of open models to get bigger and bigger. In some ways, this disillusionment is a great way to zoom our into the big picture. These models are coming. These models are getting cheaper. We need to think about risks and infrastructure more than open vs. closed.</p><p>00:00 Introduction&nbsp;<br>01:16 Recent developments in open model releases&nbsp;<br>04:21 Tom's experience viewing the total solar eclipse<br>09:38 The Three-Body Problem book and Netflix<br>14:06 The Gartner Hype Cycle<br>22:51 Infrastructure constraints on scaling AI<br>28:47 Metaphors and narratives around AI risk<br>34:43 Rethinking AI risk as public health problems<br>37:37 The "one-way door" nature of releasing open model weights<br>44:04 The relationship between the AI ecosystem and the models<br>48:24 Wrapping up the discussion in the "trough of disillusionment"</p><p>We've got some links for you again:<br>- Gartner hype cycle https://en.wikipedia.org/wiki/Gartner_hype_cycle&nbsp;<br>- MSFT Supercomputer https://www.theinformation.com/articles/microsoft-and-openai-plot-100-billion-stargate-ai-supercomputer&nbsp;<br>- Safety is about systems https://www.aisnakeoil.com/p/ai-safety-is-not-a-model-property&nbsp;<br>- Earth day history https://www.earthday.org/history/&nbsp;<br>- For our loyal listeners http://tudorsbiscuitworld.com/</p>]]></content:encoded></item><item><title><![CDATA[AI's Eras Tour: Performance, Trust, and Legitimacy]]></title><description><![CDATA[Tom and Nate catch up on the ridiculous of Nvidia GTC, the lack of trust in AI, and some important taxonomies and politics around governing AI.]]></description><link>https://www.retortai.com/p/ais-eras-tour-performance-trust-and-363</link><guid isPermaLink="false">https://www.retortai.com/p/ais-eras-tour-performance-trust-and-363</guid><dc:creator><![CDATA[Nathan Lambert]]></dc:creator><pubDate>Fri, 22 Mar 2024 11:00:00 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/162163045/374046d4a02a6b912f588b26322878d4.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>Tom and Nate catch up on the ridiculous of Nvidia GTC, the lack of trust in AI, and some important taxonomies and politics around governing AI. Safety institutes, reward model benchmarks, Nathan's bad joke delivery, and all the normal good stuff in this episode! Yes, we're also sick of the Taylor Swift jokes, but they get the clicks.</p><p>The Taylor moment: https://twitter.com/DrJimFan/status/1769817948930072930</p><p>00:00 Intros and discussion on NVIDIA's influence in AI and the Bay Area<br>09:08 Mustafa Suleyman's new role and discussion on AI safety<br>11:31 The shift from performance to trust in AI evaluation<br>17:31 The role of government agencies in AI policy and regulation<br>24:07 The role of accreditation in establishing legitimacy and trust<br>32:11 Grok's open source release and its impact on the AI community<br>39:34 Responsibility and accountability in AI and social media platforms</p>]]></content:encoded></item><item><title><![CDATA[Claude 3: Is Nathan too bought into the hype?]]></title><description><![CDATA[Tom and Nate sit down to discuss Claude 3 and some updates on what it means to be open.]]></description><link>https://www.retortai.com/p/claude-3-is-nathan-too-bought-into-67e</link><guid isPermaLink="false">https://www.retortai.com/p/claude-3-is-nathan-too-bought-into-67e</guid><dc:creator><![CDATA[Nathan Lambert]]></dc:creator><pubDate>Fri, 08 Mar 2024 15:37:24 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/162163046/0696c98a25b32e5d41048738516d9fa3.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>Tom and Nate sit down to discuss Claude 3 and some updates on what it means to be open. Not surprisingly, we get into debating some different views. We cover Dune 2's impact on AI and have a brief giveaway at the end. Cheers!</p><p>More at retortai.com. Contact us at mail at domain.</p><p>Some topics:<br>- The pace of progress in AI and whether it feels meaningful or like "progress fatigue" to different groups</p><p>- The role of hype and "vibes" in driving interest and investment in new AI models&nbsp;</p><p>- Whether the value being created by large language models is actually just being concentrated in a few big tech companies</p><p>- The debate around whether open source AI is feasible given the massive compute requirements</p><p>- The limitations of "open letters" and events with Chatham House rules as forms of politics and accountability around AI</p><p>- The analogy between the AI arms race and historical arms races like the dreadnought naval arms race</p><p>- The role of narratives, pop culture, and "priesthoods" in shaping public understanding of AI</p><p>Chapters &amp; transcript partially created with https://github.com/FanaHOVA/smol-podcaster.</p><p>00:00 Introduction and the spirit of open source<br>04:32 Historical parallels of technology arms races<br>10:26 The practical use of language models and their impact on society<br>22:21 The role and potential of open source in AI development<br>28:05 The challenges of achieving coordination and scale in open AI development<br>34:18 Pop culture's influence on the AI conversation, specifically through "Dune"</p>]]></content:encoded></item><item><title><![CDATA[Model release therapy session #1]]></title><description><![CDATA[This week Tom and Nate cover all the big topics from the big picture lens.]]></description><link>https://www.retortai.com/p/model-release-therapy-session-1-928</link><guid isPermaLink="false">https://www.retortai.com/p/model-release-therapy-session-1-928</guid><dc:creator><![CDATA[Nathan Lambert]]></dc:creator><pubDate>Fri, 23 Feb 2024 12:00:00 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/162163047/f9e0c5bc9b536ba21e1d5042dfa04000.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>This week Tom and Nate cover all the big topics from the big picture lens. Sora, Gemini 1.5's context length, Gemini's bias backlash, Gemma open models, it was a busy week in AI. We come to the conclusion that we can no longer trust a lot of these big companies to do much. We are the gladiators playing the crowd of AI. This was a great one, I'm proud on one of Tom's all time best jokes.&nbsp;<br>Thanks for listening, and reach out with any questions.</p>]]></content:encoded></item><item><title><![CDATA[Waymo vs. the time honored human experiences, vandalism and defacement]]></title><description><![CDATA[A metaphor episode!]]></description><link>https://www.retortai.com/p/waymo-vs-the-time-honored-human-experiences-90a</link><guid isPermaLink="false">https://www.retortai.com/p/waymo-vs-the-time-honored-human-experiences-90a</guid><dc:creator><![CDATA[Nathan Lambert]]></dc:creator><pubDate>Fri, 16 Feb 2024 12:00:00 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/162163048/d8fa98bea53630897e3439ad99405975.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>A metaphor episode! We are trying to figure how much the Waymo incident is or is not about AI. We bring back our Berkeley roots and talk about traditions in the Bay around distributed technology. Scooters and robots are not safe in this episode, sadly. Here's the link to the Verge piece Tom read from: <a href="https://www.theverge.com/2024/2/11/24069251/waymo-driverless-taxi-fire-vandalized-video-san-francisco-china-town">https://www.theverge.com/2024/2/11/24069251/waymo-driverless-taxi-fire-vandalized-video-san-francisco-china-town</a></p>]]></content:encoded></item><item><title><![CDATA[We believe in the metaverse]]></title><description><![CDATA[...]]></description><link>https://www.retortai.com/p/we-believe-in-the-metaverse-fd2</link><guid isPermaLink="false">https://www.retortai.com/p/we-believe-in-the-metaverse-fd2</guid><dc:creator><![CDATA[Nathan Lambert]]></dc:creator><pubDate>Fri, 09 Feb 2024 12:00:00 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/162163049/66cb838f9b6be5c639b0ea63f279f935.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>... and you should too. We catch up this week on all things Apple Vision Pro and how these devices will intersect with AI. It really turned more into a commentary on the future of society, and how various technologies may or may not tap into our subconscious.&nbsp;<br>The only link we've got for you is DeepDream:&nbsp;<strong><a href="https://en.wikipedia.org/wiki/DeepDream">https://en.wikipedia.org/wiki/DeepDream</a>&nbsp;</strong></p>]]></content:encoded></item><item><title><![CDATA[How to OLMost find a Waifu]]></title><description><![CDATA[Wow, one of our favorites.]]></description><link>https://www.retortai.com/p/how-to-olmost-find-a-waifu-fda</link><guid isPermaLink="false">https://www.retortai.com/p/how-to-olmost-find-a-waifu-fda</guid><dc:creator><![CDATA[Nathan Lambert]]></dc:creator><pubDate>Fri, 02 Feb 2024 12:00:00 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/162163050/1bef4e25559cf6d5aa2e41d4520bd6a3.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>Wow, one of our favorites. This week Tom and Nate have a lot to cover. We cover AI2's new OPEN large language models (OLMo) and all that means, the alchemical model merging craze powering waifu factories, model weight leaks from Mistral, the calling card for our loyal fans, and more topics.</p><p>We have a lot of links you'll enjoy as you'll go through it:</p><ul><li><p>The Mistral leak: <a href="https://huggingface.co/miqudev/miqu-1-70b/discussions/10">https://huggingface.co/miqudev/miqu-1-70b/discussions/10</a>&nbsp;</p></li><li><p>Writing on model merging: <a href="https://www.interconnects.ai/p/model-merging">https://www.interconnects.ai/p/model-merging</a>&nbsp;</p></li><li><p>Writing on open LLMs: <a href="https://www.interconnects.ai/p/olmo">https://www.interconnects.ai/p/olmo</a>&nbsp;</p></li><li><p>The original mechanical turk: <a href="https://en.wikipedia.org/wiki/Mechanical_Turk">https://en.wikipedia.org/wiki/Mechanical_Turk</a>&nbsp;</p></li><li><p>This Waifu does not exist: <a href="https://thisanimedoesnotexist.ai/">https://thisanimedoesnotexist.ai/</a>&nbsp;</p></li><li><p>The warriors film <a href="https://www.youtube.com/watch?v=--gdB-nnQkU">https://www.youtube.com/watch?v=--gdB-nnQkU</a>&nbsp;</p></li><li><p>The Waifu Research Department: <a href="https://huggingface.co/waifu-research-department">https://huggingface.co/waifu-research-department</a></p></li></ul>]]></content:encoded></item></channel></rss>