<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:podcast="https://podcastindex.org/namespace/1.0" xmlns:jellypod="https://jellypod.ai/namespace/1.0" xmlns:psc="http://podlove.org/simple-chapters"><channel><title><![CDATA[Inside the Black Box: Cracking AI and Deep Learning]]></title><description><![CDATA[How do Large Language Models like ChatGPT work, anyway? (Powered by Jellypod)]]></description><link>https://arshavir.jellypod.com</link><generator>Powered by Jellypod (https://www.jellypod.com)</generator><lastBuildDate>Fri, 15 May 2026 00:54:10 GMT</lastBuildDate><atom:link href="https://arshavir.jellypod.com/rss" rel="self" type="application/rss+xml"/><pubDate>Fri, 31 Oct 2025 17:35:30 GMT</pubDate><copyright><![CDATA[Copyright 2026 Inside the Black Box: Cracking AI and Deep Learning]]></copyright><language><![CDATA[en]]></language><podcast:locked owner="feed+f393967d@podcasts.jellypod.com">yes</podcast:locked><podcast:guid>9330ccd8-4e44-44e7-bbb5-346de7829edc</podcast:guid><itunes:author>Jellypod</itunes:author><itunes:subtitle>How do Large Language Models like ChatGPT work, anyway? (Powered by Jellypod)</itunes:subtitle><itunes:summary>How do Large Language Models like ChatGPT work, anyway? (Powered by Jellypod)</itunes:summary><itunes:type>episodic</itunes:type><itunes:owner><itunes:name>Jellypod</itunes:name><itunes:email>feed+f393967d@podcasts.jellypod.com</itunes:email></itunes:owner><itunes:explicit>false</itunes:explicit><itunes:category text="Technology"/><itunes:category text="Education"/><itunes:image href="https://auth.jellypod.ai/storage/v1/object/public/CoverImages/org_01K8XHYJK76KCD363NADM5CKC2/users/user_01K8XHWVRMZ2M2PFTJPW4582K9/image_1761932049514.jpg"/><item><title><![CDATA[When Fluent Answers Start Sounding True]]></title><description><![CDATA[This episode explores why smooth, coherent language can feel more credible than it is, and how processing fluency, familiarity, and authority cues shape what we believe. It also digs into why conversational AI is especially persuasive, from polished explanations to confident-sounding confabulations.]]></description><link>https://arshavir.jellypod.com/episodes/20c90d86-1e72-49cd-a201-6be6023bb204</link><guid isPermaLink="false">20c90d86-1e72-49cd-a201-6be6023bb204</guid><pubDate>Sat, 02 May 2026 17:41:57 GMT</pubDate><enclosure url="https://op3.dev/e,pg=9330ccd8-4e44-44e7-bbb5-346de7829edc/auth.jellypod.ai/storage/v1/object/public/Podcasts/org_01K8XHYJK76KCD363NADM5CKC2/20c90d86-1e72-49cd-a201-6be6023bb204/audio.mp3" length="0" type="audio/mpeg"/><podcast:generator uri="https://www.jellypod.com"></podcast:generator><podcast:episode>28</podcast:episode><podcast:transcript url="https://auth.jellypod.ai/storage/v1/object/public/Podcasts/20c90d86-1e72-49cd-a201-6be6023bb204/captions_1777743704.srt" type="application/x-subrip" language="en" rel="captions"></podcast:transcript><itunes:author>Jellypod</itunes:author><itunes:subtitle>This episode explores why smooth, coherent language can feel more credible than it is, and how processing fluency, familiarity, and authority cues shape what we believe. It also digs into why conversational AI is especially persuasive, from polished expla</itunes:subtitle><itunes:summary>This episode explores why smooth, coherent language can feel more credible than it is, and how processing fluency, familiarity, and authority cues shape what we believe. It also digs into why conversational AI is especially persuasive, from polished explanations to confident-sounding confabulations.</itunes:summary><itunes:explicit>false</itunes:explicit><itunes:duration>00:15:20</itunes:duration><itunes:image href="https://auth.jellypod.ai/storage/v1/object/public/CoverImages/org_01K8XHYJK76KCD363NADM5CKC2/resized_417F9093-BD99-4E97-8975-BD271E7A6A8A.jpg"/><itunes:episodeType>full</itunes:episodeType></item><item><title><![CDATA[Why Your Brain Believes the Model]]></title><description><![CDATA[The Heuristic Loop You Can't Break from Inside]]></description><link>https://arshavir.jellypod.com/episodes/002e12f5-14a0-43bd-a3e2-c079d12f0434</link><guid isPermaLink="false">002e12f5-14a0-43bd-a3e2-c079d12f0434</guid><pubDate>Mon, 27 Apr 2026 01:12:49 GMT</pubDate><enclosure url="https://op3.dev/e,pg=9330ccd8-4e44-44e7-bbb5-346de7829edc/auth.jellypod.ai/storage/v1/object/public/Podcasts/org_01K8XHYJK76KCD363NADM5CKC2/002e12f5-14a0-43bd-a3e2-c079d12f0434/audio.mp3" length="0" type="audio/mpeg"/><podcast:generator uri="https://www.jellypod.com"></podcast:generator><podcast:episode>27</podcast:episode><podcast:transcript url="https://auth.jellypod.ai/storage/v1/object/public/Podcasts/002e12f5-14a0-43bd-a3e2-c079d12f0434/captions_1777252356.srt" type="application/x-subrip" language="en" rel="captions"></podcast:transcript><itunes:author>Jellypod</itunes:author><itunes:subtitle>The Heuristic Loop You Can&apos;t Break from Inside</itunes:subtitle><itunes:summary>The Heuristic Loop You Can&apos;t Break from Inside</itunes:summary><itunes:explicit>false</itunes:explicit><itunes:duration>00:24:55</itunes:duration><itunes:image href="https://auth.jellypod.ai/storage/v1/object/public/CoverImages/org_01K8XHYJK76KCD363NADM5CKC2/resized_7D1CCBAA-2F3D-4101-B821-9BB15D107A21.jpg"/><itunes:episodeType>full</itunes:episodeType></item><item><title><![CDATA[When Polished Answers Feel Finished]]></title><description><![CDATA[This episode explores fluency-as-validity: the way polished AI responses can make us feel like the work of judgment is already done. It also looks at why large language models are so effective at creating the sensation of clarity, and why mechanistic interpretability may be a way to push back against that enchantment.]]></description><link>https://arshavir.jellypod.com/episodes/38b5a072-f1fc-4d41-b249-1e2166aea1d4</link><guid isPermaLink="false">38b5a072-f1fc-4d41-b249-1e2166aea1d4</guid><pubDate>Mon, 20 Apr 2026 18:11:03 GMT</pubDate><enclosure url="https://op3.dev/e,pg=9330ccd8-4e44-44e7-bbb5-346de7829edc/auth.jellypod.ai/storage/v1/object/public/Podcasts/org_01K8XHYJK76KCD363NADM5CKC2/38b5a072-f1fc-4d41-b249-1e2166aea1d4/audio.mp3" length="0" type="audio/mpeg"/><podcast:generator uri="https://www.jellypod.com"></podcast:generator><podcast:episode>26</podcast:episode><podcast:transcript url="https://auth.jellypod.ai/storage/v1/object/public/Podcasts/38b5a072-f1fc-4d41-b249-1e2166aea1d4/captions_1776708647.srt" type="application/x-subrip" language="en" rel="captions"></podcast:transcript><itunes:author>Jellypod</itunes:author><itunes:subtitle>This episode explores fluency-as-validity: the way polished AI responses can make us feel like the work of judgment is already done. It also looks at why large language models are so effective at creating the sensation of clarity, and why mechanistic inte</itunes:subtitle><itunes:summary>This episode explores fluency-as-validity: the way polished AI responses can make us feel like the work of judgment is already done. It also looks at why large language models are so effective at creating the sensation of clarity, and why mechanistic interpretability may be a way to push back against that enchantment.</itunes:summary><itunes:explicit>false</itunes:explicit><itunes:duration>00:27:53</itunes:duration><itunes:image href="https://auth.jellypod.ai/storage/v1/object/public/CoverImages/org_01K8XHYJK76KCD363NADM5CKC2/resized_69E64DD4-395A-43DF-B3D4-252DAE2E60B6.jpg"/><itunes:episodeType>full</itunes:episodeType></item><item><title><![CDATA[What Seneca Teaches Us that Marcus Couldn't]]></title><description><![CDATA[716 features fire on both Seneca and Marcus Aurelius but stay dark for ad copy. The model learned Stoic philosophy, not just an author's style. Plus: why 'inert' features aren't all the same thing.]]></description><link>https://arshavir.jellypod.com/episodes/bf3cf380-31ac-49a5-a779-9353db2eb50c</link><guid isPermaLink="false">bf3cf380-31ac-49a5-a779-9353db2eb50c</guid><pubDate>Sun, 12 Apr 2026 04:10:39 GMT</pubDate><enclosure url="https://op3.dev/e,pg=9330ccd8-4e44-44e7-bbb5-346de7829edc/auth.jellypod.ai/storage/v1/object/public/Podcasts/org_01K8XHYJK76KCD363NADM5CKC2/bf3cf380-31ac-49a5-a779-9353db2eb50c/audio.mp3" length="0" type="audio/mpeg"/><podcast:generator uri="https://www.jellypod.com"></podcast:generator><podcast:episode>25</podcast:episode><podcast:transcript url="https://auth.jellypod.ai/storage/v1/object/public/Podcasts/bf3cf380-31ac-49a5-a779-9353db2eb50c/captions_1775967026.srt" type="application/x-subrip" language="en" rel="captions"></podcast:transcript><itunes:author>Jellypod</itunes:author><itunes:subtitle>716 features fire on both Seneca and Marcus Aurelius but stay dark for ad copy. The model learned Stoic philosophy, not just an author&apos;s style. Plus: why &apos;inert&apos; features aren&apos;t all the same thing.</itunes:subtitle><itunes:summary>716 features fire on both Seneca and Marcus Aurelius but stay dark for ad copy. The model learned Stoic philosophy, not just an author&apos;s style. Plus: why &apos;inert&apos; features aren&apos;t all the same thing.</itunes:summary><itunes:explicit>false</itunes:explicit><itunes:duration>00:17:26</itunes:duration><itunes:image href="https://auth.jellypod.ai/storage/v1/object/public/CoverImages/org_01K8XHYJK76KCD363NADM5CKC2/resized_D1977719-5C78-444D-97EA-51B5C208062D.jpg"/><itunes:episodeType>full</itunes:episodeType></item><item><title><![CDATA[The Pattern Holds for Another Author]]></title><description><![CDATA[We trained a fresh LoRA on the letters of Seneca and ran the same analysis pipeline we used on Marcus Aurelius and advertising copy. Every structural finding replicated. The model organizes its adaptation into five clusters: one tight (features moving in lockstep) and four loose (features cooperating more independently). Seneca produced the cleanest clustering we've measured and the strongest workhorse cluster, a group of 141 features encoding philosophical argumentation with a causal effect more than three times stronger than anything in Marcus. Done in collaboration with John Holman.]]></description><link>https://arshavir.jellypod.com/episodes/3efa98e6-ab9e-4b00-afd1-0bc7a196012c</link><guid isPermaLink="false">3efa98e6-ab9e-4b00-afd1-0bc7a196012c</guid><pubDate>Sat, 04 Apr 2026 01:19:28 GMT</pubDate><enclosure url="https://op3.dev/e,pg=9330ccd8-4e44-44e7-bbb5-346de7829edc/auth.jellypod.ai/storage/v1/object/public/Podcasts/org_01K8XHYJK76KCD363NADM5CKC2/3efa98e6-ab9e-4b00-afd1-0bc7a196012c/audio.mp3" length="0" type="audio/mpeg"/><podcast:generator uri="https://www.jellypod.com"></podcast:generator><podcast:episode>24</podcast:episode><podcast:transcript url="https://auth.jellypod.ai/storage/v1/object/public/Podcasts/3efa98e6-ab9e-4b00-afd1-0bc7a196012c/captions_1775265555.srt" type="application/x-subrip" language="en" rel="captions"></podcast:transcript><itunes:author>Jellypod</itunes:author><itunes:subtitle>We trained a fresh LoRA on the letters of Seneca and ran the same analysis pipeline we used on Marcus Aurelius and advertising copy. Every structural finding replicated. The model organizes its adaptation into five clusters: one tight (features moving in </itunes:subtitle><itunes:summary>We trained a fresh LoRA on the letters of Seneca and ran the same analysis pipeline we used on Marcus Aurelius and advertising copy. Every structural finding replicated. The model organizes its adaptation into five clusters: one tight (features moving in lockstep) and four loose (features cooperating more independently). Seneca produced the cleanest clustering we&apos;ve measured and the strongest workhorse cluster, a group of 141 features encoding philosophical argumentation with a causal effect more than three times stronger than anything in Marcus. Done in collaboration with John Holman.</itunes:summary><itunes:explicit>false</itunes:explicit><itunes:duration>00:15:31</itunes:duration><itunes:image href="https://auth.jellypod.ai/storage/v1/object/public/CoverImages/org_01K8XHYJK76KCD363NADM5CKC2/resized_a.jpg"/><itunes:episodeType>full</itunes:episodeType></item><item><title><![CDATA[The Pattern Holds]]></title><description><![CDATA[We replicated our Marcus Aurelius findings at a new layer, then threw the whole method at 12 commercial ad copy styles trained into a single LoRA. The patterns held, and the new domain revealed something we couldn't have seen before: the model organizes its adaptations by register family, not by individual style.]]></description><link>https://arshavir.jellypod.com/episodes/7c312241-afe5-48fe-8186-3956d7004d8c</link><guid isPermaLink="false">7c312241-afe5-48fe-8186-3956d7004d8c</guid><pubDate>Mon, 30 Mar 2026 00:13:15 GMT</pubDate><enclosure url="https://op3.dev/e,pg=9330ccd8-4e44-44e7-bbb5-346de7829edc/auth.jellypod.ai/storage/v1/object/public/Podcasts/org_01K8XHYJK76KCD363NADM5CKC2/7c312241-afe5-48fe-8186-3956d7004d8c/audio.mp3" length="0" type="audio/mpeg"/><podcast:generator uri="https://www.jellypod.com"></podcast:generator><podcast:episode>23</podcast:episode><podcast:transcript url="https://auth.jellypod.ai/storage/v1/object/public/Podcasts/7c312241-afe5-48fe-8186-3956d7004d8c/captions_1774829582.srt" type="application/x-subrip" language="en" rel="captions"></podcast:transcript><itunes:author>Jellypod</itunes:author><itunes:subtitle>We replicated our Marcus Aurelius findings at a new layer, then threw the whole method at 12 commercial ad copy styles trained into a single LoRA. The patterns held, and the new domain revealed something we couldn&apos;t have seen before: the model organizes i</itunes:subtitle><itunes:summary>We replicated our Marcus Aurelius findings at a new layer, then threw the whole method at 12 commercial ad copy styles trained into a single LoRA. The patterns held, and the new domain revealed something we couldn&apos;t have seen before: the model organizes its adaptations by register family, not by individual style.</itunes:summary><itunes:explicit>false</itunes:explicit><itunes:duration>00:18:27</itunes:duration><itunes:image href="https://auth.jellypod.ai/storage/v1/object/public/CoverImages/org_01K8XHYJK76KCD363NADM5CKC2/resized_B0CD0813-0213-4E73-8559-26B17C999DE6.jpg"/><itunes:episodeType>full</itunes:episodeType></item><item><title><![CDATA[Cracking Open the Black Box]]></title><description><![CDATA[We opened the 65%. The features that resisted interpretation one at a time turned out to organize into five co-activation clusters with clear thematic identities and causal effects nearly ten times stronger than any individual feature. Second in a series with John Holman.]]></description><link>https://arshavir.jellypod.com/episodes/49d3d9a6-3fea-4475-bc3c-4aceb2a5e536</link><guid isPermaLink="false">49d3d9a6-3fea-4475-bc3c-4aceb2a5e536</guid><pubDate>Sun, 22 Mar 2026 22:26:18 GMT</pubDate><enclosure url="https://op3.dev/e,pg=9330ccd8-4e44-44e7-bbb5-346de7829edc/auth.jellypod.ai/storage/v1/object/public/Podcasts/org_01K8XHYJK76KCD363NADM5CKC2/49d3d9a6-3fea-4475-bc3c-4aceb2a5e536/audio.mp3" length="0" type="audio/mpeg"/><podcast:generator uri="https://www.jellypod.com"></podcast:generator><podcast:episode>22</podcast:episode><podcast:transcript url="https://auth.jellypod.ai/storage/v1/object/public/Podcasts/49d3d9a6-3fea-4475-bc3c-4aceb2a5e536/captions_1774218366.srt" type="application/x-subrip" language="en" rel="captions"></podcast:transcript><itunes:author>Jellypod</itunes:author><itunes:subtitle>We opened the 65%. The features that resisted interpretation one at a time turned out to organize into five co-activation clusters with clear thematic identities and causal effects nearly ten times stronger than any individual feature. Second in a series </itunes:subtitle><itunes:summary>We opened the 65%. The features that resisted interpretation one at a time turned out to organize into five co-activation clusters with clear thematic identities and causal effects nearly ten times stronger than any individual feature. Second in a series with John Holman.</itunes:summary><itunes:explicit>false</itunes:explicit><itunes:duration>00:11:21</itunes:duration><itunes:image href="https://auth.jellypod.ai/storage/v1/object/public/CoverImages/org_01K8XHYJK76KCD363NADM5CKC2/users/user_01K8XHWVRMZ2M2PFTJPW4582K9/resized_84D666C5-289A-43E2-B2D9-268AD4836CBC.jpg"/><itunes:episodeType>full</itunes:episodeType></item><item><title><![CDATA[Inside a Fine-Tuned Language Model]]></title><description><![CDATA[A concise, single-segment episode of Inside the Black Box: Cracking AI and Deep Learning where Arshavir Blackwell explains, in one continuous narrative, what neural networks are, how their simple units combine into powerful systems, and how learning by backpropagation sculpts their behavior. This short episode is designed as an elegant, one-paragraph-style monologue that introduces listeners to neural nets without equations or jargon.]]></description><link>https://arshavir.jellypod.com/episodes/558f6836-eccc-4e92-80e6-771bc2942b26</link><guid isPermaLink="false">558f6836-eccc-4e92-80e6-771bc2942b26</guid><pubDate>Thu, 12 Mar 2026 05:06:57 GMT</pubDate><enclosure url="https://op3.dev/e,pg=9330ccd8-4e44-44e7-bbb5-346de7829edc/auth.jellypod.ai/storage/v1/object/public/Podcasts/org_01K8XHYJK76KCD363NADM5CKC2/users/user_01K8XHWVRMZ2M2PFTJPW4582K9/558f6836-eccc-4e92-80e6-771bc2942b26/audio.mp3" length="0" type="audio/mpeg"/><podcast:generator uri="https://www.jellypod.com"></podcast:generator><podcast:episode>21</podcast:episode><podcast:transcript url="https://auth.jellypod.ai/storage/v1/object/public/Podcasts/558f6836-eccc-4e92-80e6-771bc2942b26/captions_1773292007.srt" type="application/x-subrip" language="en" rel="captions"></podcast:transcript><itunes:author>Jellypod</itunes:author><itunes:subtitle>A concise, single-segment episode of Inside the Black Box: Cracking AI and Deep Learning where Arshavir Blackwell explains, in one continuous narrative, what neural networks are, how their simple units combine into powerful systems, and how learning by ba</itunes:subtitle><itunes:summary>A concise, single-segment episode of Inside the Black Box: Cracking AI and Deep Learning where Arshavir Blackwell explains, in one continuous narrative, what neural networks are, how their simple units combine into powerful systems, and how learning by backpropagation sculpts their behavior. This short episode is designed as an elegant, one-paragraph-style monologue that introduces listeners to neural nets without equations or jargon.</itunes:summary><itunes:explicit>false</itunes:explicit><itunes:duration>00:18:50</itunes:duration><itunes:image href="https://auth.jellypod.ai/storage/v1/object/public/CoverImages/org_01K8XHYJK76KCD363NADM5CKC2/users/user_01K8XHWVRMZ2M2PFTJPW4582K9/cover-art-1772895617464.jpeg"/><itunes:episodeType>full</itunes:episodeType></item><item><title><![CDATA[What Counts as Structure? From Harris and Elman to Today’s Neural Nets]]></title><description><![CDATA[This episode of Inside the Black Box: Cracking AI and Deep Learning tells the story of an unexpected convergence in the history of language and AI. In 1995, Peter Bensch noticed that Zelig Harris, a mid‑century structural linguist, and Jeff Elman, a pioneer of simple recurrent networks, had independently uncovered the same deep insight about language: structure lives in patterns of use.Arshavir Blackwell, PhD, guides listeners through Harris’s world of distributional linguistics and operator grammar—where you infer structure from where words can substitute for one another—and contrasts it with Elman’s tiny recurrent neural networks that learn to predict the next word. Along the way, we see how these very different traditions arrive at the same place: hidden geometric structure in how language is used.From there, the episode bridges to today’s large language models and mechanistic interpretability, asking a deceptively simple question: what counts as "structure" inside a model? We explore how patterns, clusters, and features relate to genuine internal organization, and why Harris and Elman’s convergence still shapes how we think about circuits, features, and the geometry of meaning in modern AI.]]></description><link>https://arshavir.jellypod.com/episodes/8f6e56c6-19fc-4d98-81fa-81a889a2945c</link><guid isPermaLink="false">8f6e56c6-19fc-4d98-81fa-81a889a2945c</guid><pubDate>Fri, 06 Mar 2026 17:20:04 GMT</pubDate><enclosure url="https://op3.dev/e,pg=9330ccd8-4e44-44e7-bbb5-346de7829edc/auth.jellypod.ai/storage/v1/object/public/Podcasts/org_01K8XHYJK76KCD363NADM5CKC2/users/user_01K8XHWVRMZ2M2PFTJPW4582K9/8f6e56c6-19fc-4d98-81fa-81a889a2945c/audio.mp3" length="0" type="audio/mpeg"/><podcast:generator uri="https://www.jellypod.com"></podcast:generator><podcast:episode>20</podcast:episode><podcast:transcript url="https://auth.jellypod.ai/storage/v1/object/public/Podcasts/8f6e56c6-19fc-4d98-81fa-81a889a2945c/captions_1772817592.srt" type="application/x-subrip" language="en" rel="captions"></podcast:transcript><itunes:author>Jellypod</itunes:author><itunes:subtitle>This episode of Inside the Black Box: Cracking AI and Deep Learning tells the story of an unexpected convergence in the history of language and AI. In 1995, Peter Bensch noticed that Zelig Harris, a mid‑century structural linguist, and Jeff Elman, a pione</itunes:subtitle><itunes:summary>This episode of Inside the Black Box: Cracking AI and Deep Learning tells the story of an unexpected convergence in the history of language and AI. In 1995, Peter Bensch noticed that Zelig Harris, a mid‑century structural linguist, and Jeff Elman, a pioneer of simple recurrent networks, had independently uncovered the same deep insight about language: structure lives in patterns of use.Arshavir Blackwell, PhD, guides listeners through Harris’s world of distributional linguistics and operator grammar—where you infer structure from where words can substitute for one another—and contrasts it with Elman’s tiny recurrent neural networks that learn to predict the next word. Along the way, we see how these very different traditions arrive at the same place: hidden geometric structure in how language is used.From there, the episode bridges to today’s large language models and mechanistic interpretability, asking a deceptively simple question: what counts as &quot;structure&quot; inside a model? We explore how patterns, clusters, and features relate to genuine internal organization, and why Harris and Elman’s convergence still shapes how we think about circuits, features, and the geometry of meaning in modern AI.</itunes:summary><itunes:explicit>false</itunes:explicit><itunes:duration>00:13:34</itunes:duration><itunes:image href="https://auth.jellypod.ai/storage/v1/object/public/CoverImages/org_01K8XHYJK76KCD363NADM5CKC2/users/user_01K8XHWVRMZ2M2PFTJPW4582K9/resized_F901E45A-1C5A-4D31-84B0-E7F4E68B9AA6.jpg"/><itunes:episodeType>full</itunes:episodeType></item><item><title><![CDATA[Building a House Without Blueprints: When Interpretability Tools Work — and When They Don’t]]></title><description><![CDATA[This episode of Inside the Black Box: Cracking AI and Deep Learning explores a new theoretical framework that unifies sparse autoencoders (SAEs), transcoders, and crosscoders — and what it tells us about when mechanistic interpretability actually works.

We start by demystifying these tools and how they use sparse features to uncover internal concepts and computations in large language models, from DNA detectors to deception circuits in Claude 3 Sonnet. Then we introduce the linear representation hypothesis and the geometry of concepts as directions in activation space, along with the challenge of superposition when thousands of concepts must fit into limited dimensions.

Finally, we dive into Tang et al.’s recovery theorems, the compressed sensing roots of their approach, and why these results matter for using SAEs as a reliable “microscope” on model internals, especially in the context of fine-tuning and LoRA experiments. Along the way, we confront the uncomfortable possibility that the linear picture may break down at frontier scales — and what that would mean for the future of interpretability as a safety strategy.]]></description><link>https://arshavir.jellypod.com/episodes/2647048c-ab9f-48ce-936c-d1069de7b443</link><guid isPermaLink="false">2647048c-ab9f-48ce-936c-d1069de7b443</guid><pubDate>Fri, 27 Feb 2026 01:14:22 GMT</pubDate><enclosure url="https://op3.dev/e,pg=9330ccd8-4e44-44e7-bbb5-346de7829edc/auth.jellypod.ai/storage/v1/object/public/Podcasts/org_01K8XHYJK76KCD363NADM5CKC2/users/user_01K8XHWVRMZ2M2PFTJPW4582K9/2647048c-ab9f-48ce-936c-d1069de7b443/audio.mp3" length="0" type="audio/mpeg"/><podcast:generator uri="https://www.jellypod.com"></podcast:generator><podcast:episode>19</podcast:episode><podcast:transcript url="https://auth.jellypod.ai/storage/v1/object/public/Podcasts/2647048c-ab9f-48ce-936c-d1069de7b443/captions_1772154851.srt" type="application/x-subrip" language="en" rel="captions"></podcast:transcript><itunes:author>Jellypod</itunes:author><itunes:subtitle>This episode of Inside the Black Box: Cracking AI and Deep Learning explores a new theoretical framework that unifies sparse autoencoders (SAEs), transcoders, and crosscoders — and what it tells us about when mechanistic interpretability actually works.

</itunes:subtitle><itunes:summary>This episode of Inside the Black Box: Cracking AI and Deep Learning explores a new theoretical framework that unifies sparse autoencoders (SAEs), transcoders, and crosscoders — and what it tells us about when mechanistic interpretability actually works.

We start by demystifying these tools and how they use sparse features to uncover internal concepts and computations in large language models, from DNA detectors to deception circuits in Claude 3 Sonnet. Then we introduce the linear representation hypothesis and the geometry of concepts as directions in activation space, along with the challenge of superposition when thousands of concepts must fit into limited dimensions.

Finally, we dive into Tang et al.’s recovery theorems, the compressed sensing roots of their approach, and why these results matter for using SAEs as a reliable “microscope” on model internals, especially in the context of fine-tuning and LoRA experiments. Along the way, we confront the uncomfortable possibility that the linear picture may break down at frontier scales — and what that would mean for the future of interpretability as a safety strategy.</itunes:summary><itunes:explicit>false</itunes:explicit><itunes:duration>00:18:04</itunes:duration><itunes:image href="https://auth.jellypod.ai/storage/v1/object/public/CoverImages/org_01K8XHYJK76KCD363NADM5CKC2/users/user_01K8XHWVRMZ2M2PFTJPW4582K9/cover-art-1772154733733.jpeg"/><itunes:episodeType>full</itunes:episodeType></item><item><title><![CDATA[I Told My LLM Not to Say "Empower"]]></title><description><![CDATA[In this episode of Inside the Black Box: Cracking AI and Deep Learning, Arshavir Blackwell, PhD, takes engineers and researchers inside the practical mechanics of LoRA, low‑rank adaptation methods that make it possible to fine‑tune multi‑billion‑parameter language models on a single GPU.]]></description><link>https://arshavir.jellypod.com/episodes/1a6e0162-1abe-4a56-9c98-c09c55b7200d</link><guid isPermaLink="false">1a6e0162-1abe-4a56-9c98-c09c55b7200d</guid><pubDate>Thu, 19 Feb 2026 21:57:46 GMT</pubDate><enclosure url="https://op3.dev/e,pg=9330ccd8-4e44-44e7-bbb5-346de7829edc/auth.jellypod.ai/storage/v1/object/public/Podcasts/org_01K8XHYJK76KCD363NADM5CKC2/users/user_01K8XHWVRMZ2M2PFTJPW4582K9/1a6e0162-1abe-4a56-9c98-c09c55b7200d/audio.mp3" length="0" type="audio/mpeg"/><podcast:generator uri="https://www.jellypod.com"></podcast:generator><podcast:episode>18</podcast:episode><podcast:transcript url="https://auth.jellypod.ai/storage/v1/object/public/Podcasts/1a6e0162-1abe-4a56-9c98-c09c55b7200d/captions_1771538259.srt" type="application/x-subrip" language="en" rel="captions"></podcast:transcript><itunes:author>Jellypod</itunes:author><itunes:subtitle>In this episode of Inside the Black Box: Cracking AI and Deep Learning, Arshavir Blackwell, PhD, takes engineers and researchers inside the practical mechanics of LoRA, low‑rank adaptation methods that make it possible to fine‑tune multi‑billion‑parameter</itunes:subtitle><itunes:summary>In this episode of Inside the Black Box: Cracking AI and Deep Learning, Arshavir Blackwell, PhD, takes engineers and researchers inside the practical mechanics of LoRA, low‑rank adaptation methods that make it possible to fine‑tune multi‑billion‑parameter language models on a single GPU.</itunes:summary><itunes:explicit>false</itunes:explicit><itunes:duration>00:06:34</itunes:duration><itunes:image href="https://auth.jellypod.ai/storage/v1/object/public/CoverImages/org_01K8XHYJK76KCD363NADM5CKC2/users/user_01K8XHWVRMZ2M2PFTJPW4582K9/resized_F5530319-503F-4377-B656-0C3CA1E52E48.jpg"/><itunes:episodeType>full</itunes:episodeType></item><item><title><![CDATA[Beyond the Surface of AI Intelligence]]></title><description><![CDATA[This episode dives into why judging AI by behavior alone falls short of proving true intelligence. We explore how insights from mechanistic interpretability and cognitive science reveal what’s really happening inside AI models. Join us as we challenge the limits of behavioral tests and rethink what intelligence means for future AI.]]></description><link>https://arshavir.jellypod.com/episodes/1321425c-360a-4ed8-bb85-0a690245aba3</link><guid isPermaLink="false">1321425c-360a-4ed8-bb85-0a690245aba3</guid><pubDate>Mon, 09 Feb 2026 14:00:03 GMT</pubDate><enclosure url="https://op3.dev/e,pg=9330ccd8-4e44-44e7-bbb5-346de7829edc/auth.jellypod.ai/storage/v1/object/public/Podcasts/org_01K8XHYJK76KCD363NADM5CKC2/users/user_01K8XHWVRMZ2M2PFTJPW4582K9/1321425c-360a-4ed8-bb85-0a690245aba3/audio.mp3" length="0" type="audio/mpeg"/><podcast:generator uri="https://www.jellypod.com"></podcast:generator><podcast:episode>17</podcast:episode><podcast:transcript url="https://auth.jellypod.ai/storage/v1/object/public/Podcasts/1321425c-360a-4ed8-bb85-0a690245aba3/captions_1770482418.srt" type="application/x-subrip" language="en" rel="captions"></podcast:transcript><itunes:author>Jellypod</itunes:author><itunes:subtitle>This episode dives into why judging AI by behavior alone falls short of proving true intelligence. We explore how insights from mechanistic interpretability and cognitive science reveal what’s really happening inside AI models. Join us as we challenge the</itunes:subtitle><itunes:summary>This episode dives into why judging AI by behavior alone falls short of proving true intelligence. We explore how insights from mechanistic interpretability and cognitive science reveal what’s really happening inside AI models. Join us as we challenge the limits of behavioral tests and rethink what intelligence means for future AI.</itunes:summary><itunes:explicit>false</itunes:explicit><itunes:duration>00:14:43</itunes:duration><itunes:image href="https://auth.jellypod.ai/storage/v1/object/public/CoverImages/org_01K8XHYJK76KCD363NADM5CKC2/users/user_01K8XHWVRMZ2M2PFTJPW4582K9/image_1761932049514.jpg"/><itunes:episodeType>full</itunes:episodeType></item><item><title><![CDATA[Unlocking BERTs Hidden Grammar]]></title><description><![CDATA[Explore how BERT’s attention heads reveal an emergent understanding of language structure without explicit supervision. Discover the role of attention as a form of memory and what it means for the future of AI language models.]]></description><link>https://arshavir.jellypod.com/episodes/4d8ebd67-60c2-475f-a425-150ed825ec0e</link><guid isPermaLink="false">4d8ebd67-60c2-475f-a425-150ed825ec0e</guid><pubDate>Tue, 03 Feb 2026 17:44:29 GMT</pubDate><enclosure url="https://op3.dev/e,pg=9330ccd8-4e44-44e7-bbb5-346de7829edc/auth.jellypod.ai/storage/v1/object/public/Podcasts/org_01K8XHYJK76KCD363NADM5CKC2/users/user_01K8XHWVRMZ2M2PFTJPW4582K9/4d8ebd67-60c2-475f-a425-150ed825ec0e/audio.mp3" length="0" type="audio/mpeg"/><podcast:generator uri="https://www.jellypod.com"></podcast:generator><podcast:episode>16</podcast:episode><podcast:transcript url="https://auth.jellypod.ai/storage/v1/object/public/Podcasts/4d8ebd67-60c2-475f-a425-150ed825ec0e/captions_1770140659.srt" type="application/x-subrip" language="en" rel="captions"></podcast:transcript><itunes:author>Jellypod</itunes:author><itunes:subtitle>Explore how BERT’s attention heads reveal an emergent understanding of language structure without explicit supervision. Discover the role of attention as a form of memory and what it means for the future of AI language models.</itunes:subtitle><itunes:summary>Explore how BERT’s attention heads reveal an emergent understanding of language structure without explicit supervision. Discover the role of attention as a form of memory and what it means for the future of AI language models.</itunes:summary><itunes:explicit>false</itunes:explicit><itunes:duration>00:09:06</itunes:duration><itunes:image href="https://auth.jellypod.ai/storage/v1/object/public/CoverImages/org_01K8XHYJK76KCD363NADM5CKC2/users/user_01K8XHWVRMZ2M2PFTJPW4582K9/cover-art-1770140746186.jpeg"/><itunes:episodeType>full</itunes:episodeType></item><item><title><![CDATA[Cracking the Code of AI Interpretation]]></title><description><![CDATA[Dive into how we naturally explain neural networks with folk interpretability and why these simple stories fall short. Discover the journey toward mechanistic understandability in AI and what that means for how we talk about and trust large language models.]]></description><link>https://arshavir.jellypod.com/episodes/5a3647a6-81ef-433b-8389-4feabfbfa937</link><guid isPermaLink="false">5a3647a6-81ef-433b-8389-4feabfbfa937</guid><pubDate>Wed, 28 Jan 2026 18:55:51 GMT</pubDate><enclosure url="https://op3.dev/e,pg=9330ccd8-4e44-44e7-bbb5-346de7829edc/auth.jellypod.ai/storage/v1/object/public/Podcasts/org_01K8XHYJK76KCD363NADM5CKC2/users/user_01K8XHWVRMZ2M2PFTJPW4582K9/5a3647a6-81ef-433b-8389-4feabfbfa937/audio.mp3" length="0" type="audio/mpeg"/><podcast:generator uri="https://www.jellypod.com"></podcast:generator><podcast:episode>15</podcast:episode><podcast:transcript url="https://auth.jellypod.ai/storage/v1/object/public/Podcasts/org_01K8XHYJK76KCD363NADM5CKC2/users/user_01K8XHWVRMZ2M2PFTJPW4582K9/5a3647a6-81ef-433b-8389-4feabfbfa937/captions_1769626499.srt" type="application/x-subrip" language="en" rel="captions"></podcast:transcript><itunes:author>Jellypod</itunes:author><itunes:subtitle>Dive into how we naturally explain neural networks with folk interpretability and why these simple stories fall short. Discover the journey toward mechanistic understandability in AI and what that means for how we talk about and trust large language model</itunes:subtitle><itunes:summary>Dive into how we naturally explain neural networks with folk interpretability and why these simple stories fall short. Discover the journey toward mechanistic understandability in AI and what that means for how we talk about and trust large language models.</itunes:summary><itunes:explicit>false</itunes:explicit><itunes:duration>00:10:11</itunes:duration><itunes:image href="https://auth.jellypod.ai/storage/v1/object/public/CoverImages/org_01K8XHYJK76KCD363NADM5CKC2/users/user_01K8XHWVRMZ2M2PFTJPW4582K9/resized_A629E3F5-ED5D-421C-A3CA-0636CB5E7911.jpg"/><itunes:episodeType>full</itunes:episodeType></item><item><title><![CDATA[Decoding GPTs Hidden Circuits]]></title><description><![CDATA[Explore how sparse autoencoders and transcoders unveil the inner workings of GPT-2 by revealing functional features and computational circuits. Discover breakthrough methods that shift from observing raw network activations to mapping the model's actual computation, making AI behavior more interpretable than ever.]]></description><link>https://arshavir.jellypod.com/episodes/ee48b068-7fb5-4115-a86e-f23d1ad97a18</link><guid isPermaLink="false">ee48b068-7fb5-4115-a86e-f23d1ad97a18</guid><pubDate>Mon, 26 Jan 2026 00:00:51 GMT</pubDate><enclosure url="https://op3.dev/e,pg=9330ccd8-4e44-44e7-bbb5-346de7829edc/auth.jellypod.ai/storage/v1/object/public/Podcasts/org_01K8XHYJK76KCD363NADM5CKC2/users/user_01K8XHWVRMZ2M2PFTJPW4582K9/ee48b068-7fb5-4115-a86e-f23d1ad97a18/audio.mp3" length="0" type="audio/mpeg"/><podcast:generator uri="https://www.jellypod.com"></podcast:generator><podcast:episode>14</podcast:episode><podcast:transcript url="https://auth.jellypod.ai/storage/v1/object/public/Podcasts/org_01K8XHYJK76KCD363NADM5CKC2/users/user_01K8XHWVRMZ2M2PFTJPW4582K9/ee48b068-7fb5-4115-a86e-f23d1ad97a18/captions_1769385593.srt" type="application/x-subrip" language="en" rel="captions"></podcast:transcript><itunes:author>Jellypod</itunes:author><itunes:subtitle>Explore how sparse autoencoders and transcoders unveil the inner workings of GPT-2 by revealing functional features and computational circuits. Discover breakthrough methods that shift from observing raw network activations to mapping the model&apos;s actual c</itunes:subtitle><itunes:summary>Explore how sparse autoencoders and transcoders unveil the inner workings of GPT-2 by revealing functional features and computational circuits. Discover breakthrough methods that shift from observing raw network activations to mapping the model&apos;s actual computation, making AI behavior more interpretable than ever.</itunes:summary><itunes:explicit>false</itunes:explicit><itunes:duration>00:11:27</itunes:duration><itunes:image href="https://auth.jellypod.ai/storage/v1/object/public/CoverImages/org_01K8XHYJK76KCD363NADM5CKC2/users/user_01K8XHWVRMZ2M2PFTJPW4582K9/resized_8BC983FC-F55E-4B77-BD8C-CB01F72FDBEB.jpg"/><itunes:episodeType>full</itunes:episodeType></item><item><title><![CDATA[Decoding Attention and Emergence in AI]]></title><description><![CDATA[Explore how attention heads uncover patterns through learned queries and keys, revealing emergent behaviors shaped by optimization. Dive into parallels with natural selection and psycholinguistics to understand how meaning arises not by design but through experience in both machines and brains.]]></description><link>https://arshavir.jellypod.com/episodes/6d50e174-a0b4-4cc5-acb1-b728fcdb7268</link><guid isPermaLink="false">6d50e174-a0b4-4cc5-acb1-b728fcdb7268</guid><pubDate>Wed, 14 Jan 2026 23:47:15 GMT</pubDate><enclosure url="https://op3.dev/e,pg=9330ccd8-4e44-44e7-bbb5-346de7829edc/auth.jellypod.ai/storage/v1/object/public/Podcasts/org_01K8XHYJK76KCD363NADM5CKC2/users/user_01K8XHWVRMZ2M2PFTJPW4582K9/6d50e174-a0b4-4cc5-acb1-b728fcdb7268/audio.mp3" length="0" type="audio/mpeg"/><podcast:generator uri="https://www.jellypod.com"></podcast:generator><podcast:episode>13</podcast:episode><podcast:transcript url="https://auth.jellypod.ai/storage/v1/object/public/Podcasts/org_01K8XHYJK76KCD363NADM5CKC2/users/user_01K8XHWVRMZ2M2PFTJPW4582K9/6d50e174-a0b4-4cc5-acb1-b728fcdb7268/captions_1768434383.srt" type="application/x-subrip" language="en" rel="captions"></podcast:transcript><itunes:author>Jellypod</itunes:author><itunes:subtitle>Explore how attention heads uncover patterns through learned queries and keys, revealing emergent behaviors shaped by optimization. Dive into parallels with natural selection and psycholinguistics to understand how meaning arises not by design but through</itunes:subtitle><itunes:summary>Explore how attention heads uncover patterns through learned queries and keys, revealing emergent behaviors shaped by optimization. Dive into parallels with natural selection and psycholinguistics to understand how meaning arises not by design but through experience in both machines and brains.</itunes:summary><itunes:explicit>false</itunes:explicit><itunes:duration>00:06:06</itunes:duration><itunes:image href="https://auth.jellypod.ai/storage/v1/object/public/CoverImages/org_01K8XHYJK76KCD363NADM5CKC2/users/user_01K8XHWVRMZ2M2PFTJPW4582K9/cover-art-1768435567918.jpeg"/><itunes:episodeType>full</itunes:episodeType></item><item><title><![CDATA[When Knowledge Battles Noise in GPT Models]]></title><description><![CDATA[Explore how GPT-2 balances fleeting factual recall with generic responses through internal competition among candidate answers. Discover parallels with human cognition and how larger models navigate indirect recall to reveal hidden knowledge beneath suppression.]]></description><link>https://arshavir.jellypod.com/episodes/8fe6e7de-512e-48af-8b16-38844953fd32</link><guid isPermaLink="false">8fe6e7de-512e-48af-8b16-38844953fd32</guid><pubDate>Wed, 07 Jan 2026 00:13:17 GMT</pubDate><enclosure url="https://op3.dev/e,pg=9330ccd8-4e44-44e7-bbb5-346de7829edc/auth.jellypod.ai/storage/v1/object/public/Podcasts/org_01K8XHYJK76KCD363NADM5CKC2/users/user_01K8XHWVRMZ2M2PFTJPW4582K9/8fe6e7de-512e-48af-8b16-38844953fd32/audio.mp3" length="0" type="audio/mpeg"/><podcast:generator uri="https://www.jellypod.com"></podcast:generator><podcast:episode>12</podcast:episode><podcast:transcript url="https://auth.jellypod.ai/storage/v1/object/public/Podcasts/org_01K8XHYJK76KCD363NADM5CKC2/users/user_01K8XHWVRMZ2M2PFTJPW4582K9/8fe6e7de-512e-48af-8b16-38844953fd32/captions_1767744759.srt" type="application/x-subrip" language="en" rel="captions"></podcast:transcript><itunes:author>Jellypod</itunes:author><itunes:subtitle>Explore how GPT-2 balances fleeting factual recall with generic responses through internal competition among candidate answers. Discover parallels with human cognition and how larger models navigate indirect recall to reveal hidden knowledge beneath suppr</itunes:subtitle><itunes:summary>Explore how GPT-2 balances fleeting factual recall with generic responses through internal competition among candidate answers. Discover parallels with human cognition and how larger models navigate indirect recall to reveal hidden knowledge beneath suppression.</itunes:summary><itunes:explicit>false</itunes:explicit><itunes:duration>00:07:24</itunes:duration><itunes:image href="https://auth.jellypod.ai/storage/v1/object/public/CoverImages/org_01K8XHYJK76KCD363NADM5CKC2/users/user_01K8XHWVRMZ2M2PFTJPW4582K9/resized_40462D08-F6FF-4189-B987-893F62BFD294.jpg"/><itunes:episodeType>full</itunes:episodeType></item><item><title><![CDATA[Inside Circuits: How Large Language Models Understand]]></title><description><![CDATA[Dive into the world of neural circuits within large language models. In this episode, Arshavir Blackwell unpacks how transformer circuits, attention mechanisms, and high-dimensional geometry combine to create the magic—and limits—of modern AI language systems.]]></description><link>https://arshavir.jellypod.com/episodes/61919a54-cdec-4328-bcd8-fc69616baae0</link><guid isPermaLink="false">61919a54-cdec-4328-bcd8-fc69616baae0</guid><pubDate>Thu, 01 Jan 2026 23:13:33 GMT</pubDate><enclosure url="https://op3.dev/e,pg=9330ccd8-4e44-44e7-bbb5-346de7829edc/auth.jellypod.ai/storage/v1/object/public/Podcasts/org_01K8XHYJK76KCD363NADM5CKC2/users/user_01K8XHWVRMZ2M2PFTJPW4582K9/61919a54-cdec-4328-bcd8-fc69616baae0/audio.mp3" length="0" type="audio/mpeg"/><podcast:generator uri="https://www.jellypod.com"></podcast:generator><podcast:episode>11</podcast:episode><podcast:transcript url="https://auth.jellypod.ai/storage/v1/object/public/Podcasts/org_01K8XHYJK76KCD363NADM5CKC2/users/user_01K8XHWVRMZ2M2PFTJPW4582K9/61919a54-cdec-4328-bcd8-fc69616baae0/captions_1767309159.srt" type="application/x-subrip" language="en" rel="captions"></podcast:transcript><itunes:author>Jellypod</itunes:author><itunes:subtitle>Dive into the world of neural circuits within large language models. In this episode, Arshavir Blackwell unpacks how transformer circuits, attention mechanisms, and high-dimensional geometry combine to create the magic—and limits—of modern AI language sys</itunes:subtitle><itunes:summary>Dive into the world of neural circuits within large language models. In this episode, Arshavir Blackwell unpacks how transformer circuits, attention mechanisms, and high-dimensional geometry combine to create the magic—and limits—of modern AI language systems.</itunes:summary><itunes:explicit>false</itunes:explicit><itunes:duration>00:07:41</itunes:duration><itunes:image href="https://auth.jellypod.ai/storage/v1/object/public/CoverImages/org_01K8XHYJK76KCD363NADM5CKC2/users/user_01K8XHWVRMZ2M2PFTJPW4582K9/resized_image.jpg"/><itunes:episodeType>full</itunes:episodeType></item><item><title><![CDATA[Hallucinations, Interpretability, and the Seahorse Mirage]]></title><description><![CDATA[This episode dives into why advanced language models still generate hallucinations, how interpretability tools help us uncover their hidden workings, and what the seahorse emoji teaches us about model and human reasoning. Arshavir connects groundbreaking research, practical business importance, and the statistical quirks that shape AI's version of 'truth.']]></description><link>https://arshavir.jellypod.com/episodes/7cbe7f65-4454-4d61-9a92-19a5b356dd88</link><guid isPermaLink="false">7cbe7f65-4454-4d61-9a92-19a5b356dd88</guid><pubDate>Mon, 29 Dec 2025 20:38:45 GMT</pubDate><enclosure url="https://op3.dev/e,pg=9330ccd8-4e44-44e7-bbb5-346de7829edc/auth.jellypod.ai/storage/v1/object/public/Podcasts/org_01K8XHYJK76KCD363NADM5CKC2/users/user_01K8XHWVRMZ2M2PFTJPW4582K9/7cbe7f65-4454-4d61-9a92-19a5b356dd88/audio.mp3" length="0" type="audio/mpeg"/><podcast:generator uri="https://www.jellypod.com"></podcast:generator><podcast:episode>10</podcast:episode><podcast:transcript url="https://auth.jellypod.ai/storage/v1/object/public/Podcasts/org_01K8XHYJK76KCD363NADM5CKC2/users/user_01K8XHWVRMZ2M2PFTJPW4582K9/7cbe7f65-4454-4d61-9a92-19a5b356dd88/captions_1767040684.srt" type="application/x-subrip" language="en" rel="captions"></podcast:transcript><itunes:author>Jellypod</itunes:author><itunes:subtitle>This episode dives into why advanced language models still generate hallucinations, how interpretability tools help us uncover their hidden workings, and what the seahorse emoji teaches us about model and human reasoning. Arshavir connects groundbreaking </itunes:subtitle><itunes:summary>This episode dives into why advanced language models still generate hallucinations, how interpretability tools help us uncover their hidden workings, and what the seahorse emoji teaches us about model and human reasoning. Arshavir connects groundbreaking research, practical business importance, and the statistical quirks that shape AI&apos;s version of &apos;truth.&apos;</itunes:summary><itunes:explicit>false</itunes:explicit><itunes:duration>00:09:37</itunes:duration><itunes:image href="https://auth.jellypod.ai/storage/v1/object/public/CoverImages/org_01K8XHYJK76KCD363NADM5CKC2/users/user_01K8XHWVRMZ2M2PFTJPW4582K9/resized_pic.jpg"/><itunes:episodeType>full</itunes:episodeType></item><item><title><![CDATA[How Transformers Stack Meaning Like Finnish Words]]></title><description><![CDATA[Explore how large language models build up meaning in ways strikingly similar to the layered grammar of Finnish. Arshavir Blackwell reveals why understanding Finnish morphology offers a powerful analogy for interpreting the compositional logic inside modern AI systems.]]></description><link>https://arshavir.jellypod.com/episodes/1a01316f-523f-411f-9ffd-35ddcee4f868</link><guid isPermaLink="false">1a01316f-523f-411f-9ffd-35ddcee4f868</guid><pubDate>Fri, 19 Dec 2025 23:10:46 GMT</pubDate><enclosure url="https://op3.dev/e,pg=9330ccd8-4e44-44e7-bbb5-346de7829edc/auth.jellypod.ai/storage/v1/object/public/Podcasts/org_01K8XHYJK76KCD363NADM5CKC2/users/user_01K8XHWVRMZ2M2PFTJPW4582K9/1a01316f-523f-411f-9ffd-35ddcee4f868/audio.mp3" length="0" type="audio/mpeg"/><podcast:generator uri="https://www.jellypod.com"></podcast:generator><podcast:episode>9</podcast:episode><podcast:transcript url="https://auth.jellypod.ai/storage/v1/object/public/Podcasts/org_01K8XHYJK76KCD363NADM5CKC2/users/user_01K8XHWVRMZ2M2PFTJPW4582K9/1a01316f-523f-411f-9ffd-35ddcee4f868/captions_1766185490.srt" type="application/x-subrip" language="en" rel="captions"></podcast:transcript><itunes:author>Jellypod</itunes:author><itunes:subtitle>Explore how large language models build up meaning in ways strikingly similar to the layered grammar of Finnish. Arshavir Blackwell reveals why understanding Finnish morphology offers a powerful analogy for interpreting the compositional logic inside mode</itunes:subtitle><itunes:summary>Explore how large language models build up meaning in ways strikingly similar to the layered grammar of Finnish. Arshavir Blackwell reveals why understanding Finnish morphology offers a powerful analogy for interpreting the compositional logic inside modern AI systems.</itunes:summary><itunes:explicit>false</itunes:explicit><itunes:duration>00:13:26</itunes:duration><itunes:image href="https://auth.jellypod.ai/storage/v1/object/public/CoverImages/org_01K8XHYJK76KCD363NADM5CKC2/users/user_01K8XHWVRMZ2M2PFTJPW4582K9/resized_fin.jpg"/><itunes:episodeType>full</itunes:episodeType></item><item><title><![CDATA[The Mandela Effect in AI: Why Language Models Misremember]]></title><description><![CDATA[Dive into how and why large language models like ChatGPT mirror the human Mandela Effect, reproducing our collective false memories and misquotations. Arshavir Blackwell examines the science behind errors in models and minds, and explores how new techniques can counteract these uncanny AI confabulations.]]></description><link>https://arshavir.jellypod.com/episodes/7b1e371e-3e28-4ffe-bec5-d40ef1653eed</link><guid isPermaLink="false">7b1e371e-3e28-4ffe-bec5-d40ef1653eed</guid><pubDate>Sun, 14 Dec 2025 00:09:07 GMT</pubDate><enclosure url="https://op3.dev/e,pg=9330ccd8-4e44-44e7-bbb5-346de7829edc/auth.jellypod.ai/storage/v1/object/public/Podcasts/org_01K8XHYJK76KCD363NADM5CKC2/users/user_01K8XHWVRMZ2M2PFTJPW4582K9/7b1e371e-3e28-4ffe-bec5-d40ef1653eed/audio.mp3" length="0" type="audio/mpeg"/><podcast:generator uri="https://www.jellypod.com"></podcast:generator><podcast:episode>8</podcast:episode><podcast:transcript url="https://auth.jellypod.ai/storage/v1/object/public/Podcasts/org_01K8XHYJK76KCD363NADM5CKC2/users/user_01K8XHWVRMZ2M2PFTJPW4582K9/7b1e371e-3e28-4ffe-bec5-d40ef1653eed/captions_1765670896.srt" type="application/x-subrip" language="en" rel="captions"></podcast:transcript><itunes:author>Jellypod</itunes:author><itunes:subtitle>Dive into how and why large language models like ChatGPT mirror the human Mandela Effect, reproducing our collective false memories and misquotations. Arshavir Blackwell examines the science behind errors in models and minds, and explores how new techniqu</itunes:subtitle><itunes:summary>Dive into how and why large language models like ChatGPT mirror the human Mandela Effect, reproducing our collective false memories and misquotations. Arshavir Blackwell examines the science behind errors in models and minds, and explores how new techniques can counteract these uncanny AI confabulations.</itunes:summary><itunes:explicit>false</itunes:explicit><itunes:duration>00:11:24</itunes:duration><itunes:image href="https://auth.jellypod.ai/storage/v1/object/public/CoverImages/org_01K8XHYJK76KCD363NADM5CKC2/users/user_01K8XHWVRMZ2M2PFTJPW4582K9/resized_Screenshot-2025-12-13-at-16-22-14.jpg"/><itunes:episodeType>full</itunes:episodeType></item><item><title><![CDATA[Bridging Circuits and Concepts in Large Language Models]]></title><description><![CDATA[How do millions of computations inside large language models add up to something like understanding? This episode explores the latest breakthroughs in mechanistic interpretability, showing how tools like representational geometry, circuit decomposition, and compression theory illuminate the missing middle between circuits and meaning. Join Arshavir Blackwell as he opens the black box and challenges what we really mean by 'understanding' in machines.]]></description><link>https://arshavir.jellypod.com/episodes/93ac77dc-6732-4bf1-890b-faaa93c4749e</link><guid isPermaLink="false">93ac77dc-6732-4bf1-890b-faaa93c4749e</guid><pubDate>Fri, 05 Dec 2025 19:17:25 GMT</pubDate><enclosure url="https://op3.dev/e,pg=9330ccd8-4e44-44e7-bbb5-346de7829edc/auth.jellypod.ai/storage/v1/object/public/Podcasts/org_01K8XHYJK76KCD363NADM5CKC2/users/user_01K8XHWVRMZ2M2PFTJPW4582K9/93ac77dc-6732-4bf1-890b-faaa93c4749e/audio.mp3" length="0" type="audio/mpeg"/><podcast:generator uri="https://www.jellypod.com"></podcast:generator><podcast:episode>7</podcast:episode><podcast:transcript url="https://auth.jellypod.ai/storage/v1/object/public/Podcasts/org_01K8XHYJK76KCD363NADM5CKC2/users/user_01K8XHWVRMZ2M2PFTJPW4582K9/93ac77dc-6732-4bf1-890b-faaa93c4749e/captions_1764962195.srt" type="application/x-subrip" language="en" rel="captions"></podcast:transcript><itunes:author>Jellypod</itunes:author><itunes:subtitle>How do millions of computations inside large language models add up to something like understanding? This episode explores the latest breakthroughs in mechanistic interpretability, showing how tools like representational geometry, circuit decomposition, a</itunes:subtitle><itunes:summary>How do millions of computations inside large language models add up to something like understanding? This episode explores the latest breakthroughs in mechanistic interpretability, showing how tools like representational geometry, circuit decomposition, and compression theory illuminate the missing middle between circuits and meaning. Join Arshavir Blackwell as he opens the black box and challenges what we really mean by &apos;understanding&apos; in machines.</itunes:summary><itunes:explicit>false</itunes:explicit><itunes:duration>00:14:51</itunes:duration><itunes:image href="https://auth.jellypod.ai/storage/v1/object/public/CoverImages/org_01K8XHYJK76KCD363NADM5CKC2/users/user_01K8XHWVRMZ2M2PFTJPW4582K9/resized_image-2.jpg"/><itunes:episodeType>full</itunes:episodeType></item><item><title><![CDATA[How Transformers Turn Words Into Meaning]]></title><description><![CDATA[Embark on a step-by-step journey through the inner workings of transformer models like those powering ChatGPT. Arshavir Blackwell breaks down how context, attention, and high-dimensional geometry turn isolated tokens into fluent, meaningful language—revealing the mathematics of understanding inside the black box.]]></description><link>https://arshavir.jellypod.com/episodes/91e214c9-25a3-4eda-832f-703738a7c893</link><guid isPermaLink="false">91e214c9-25a3-4eda-832f-703738a7c893</guid><pubDate>Fri, 28 Nov 2025 17:37:33 GMT</pubDate><enclosure url="https://op3.dev/e,pg=9330ccd8-4e44-44e7-bbb5-346de7829edc/auth.jellypod.ai/storage/v1/object/public/Podcasts/org_01K8XHYJK76KCD363NADM5CKC2/users/user_01K8XHWVRMZ2M2PFTJPW4582K9/91e214c9-25a3-4eda-832f-703738a7c893/audio.mp3" length="0" type="audio/mpeg"/><podcast:generator uri="https://www.jellypod.com"></podcast:generator><podcast:episode>6</podcast:episode><podcast:transcript url="https://auth.jellypod.ai/storage/v1/object/public/Podcasts/org_01K8XHYJK76KCD363NADM5CKC2/users/user_01K8XHWVRMZ2M2PFTJPW4582K9/91e214c9-25a3-4eda-832f-703738a7c893/captions_1764351423.srt" type="application/x-subrip" language="en" rel="captions"></podcast:transcript><itunes:author>Jellypod</itunes:author><itunes:subtitle>Embark on a step-by-step journey through the inner workings of transformer models like those powering ChatGPT. Arshavir Blackwell breaks down how context, attention, and high-dimensional geometry turn isolated tokens into fluent, meaningful language—revea</itunes:subtitle><itunes:summary>Embark on a step-by-step journey through the inner workings of transformer models like those powering ChatGPT. Arshavir Blackwell breaks down how context, attention, and high-dimensional geometry turn isolated tokens into fluent, meaningful language—revealing the mathematics of understanding inside the black box.</itunes:summary><itunes:explicit>false</itunes:explicit><itunes:duration>00:06:54</itunes:duration><itunes:image href="https://auth.jellypod.ai/storage/v1/object/public/CoverImages/org_01K8XHYJK76KCD363NADM5CKC2/users/user_01K8XHWVRMZ2M2PFTJPW4582K9/resized_image.jpg"/><itunes:episodeType>full</itunes:episodeType></item><item><title><![CDATA[Can Smaller Language Models Be Smarter?]]></title><description><![CDATA[Today we explore whether mechanistic interpretability could hold the key to building leaner, more transparent—and perhaps even smarter—large language models. From knowledge distillation and pruning to low-rank adaptation, we examine cutting-edge strategies to make AI models both smaller and more explainable. Join Arshavir as he breaks down the surprising challenges of making models efficient without sacrificing understanding.]]></description><link>https://arshavir.jellypod.com/episodes/876011f4-cea2-44ed-a573-db4db24bcacd</link><guid isPermaLink="false">876011f4-cea2-44ed-a573-db4db24bcacd</guid><pubDate>Wed, 19 Nov 2025 03:53:28 GMT</pubDate><enclosure url="https://op3.dev/e,pg=9330ccd8-4e44-44e7-bbb5-346de7829edc/auth.jellypod.ai/storage/v1/object/public/Podcasts/org_01K8XHYJK76KCD363NADM5CKC2/users/user_01K8XHWVRMZ2M2PFTJPW4582K9/876011f4-cea2-44ed-a573-db4db24bcacd/audio.mp3" length="0" type="audio/mpeg"/><podcast:generator uri="https://www.jellypod.com"></podcast:generator><podcast:episode>5</podcast:episode><podcast:transcript url="https://auth.jellypod.ai/storage/v1/object/public/Podcasts/org_01K8XHYJK76KCD363NADM5CKC2/users/user_01K8XHWVRMZ2M2PFTJPW4582K9/876011f4-cea2-44ed-a573-db4db24bcacd/captions_1763524359.srt" type="application/x-subrip" language="en" rel="captions"></podcast:transcript><itunes:author>Jellypod</itunes:author><itunes:subtitle>Today we explore whether mechanistic interpretability could hold the key to building leaner, more transparent—and perhaps even smarter—large language models. From knowledge distillation and pruning to low-rank adaptation, we examine cutting-edge strategie</itunes:subtitle><itunes:summary>Today we explore whether mechanistic interpretability could hold the key to building leaner, more transparent—and perhaps even smarter—large language models. From knowledge distillation and pruning to low-rank adaptation, we examine cutting-edge strategies to make AI models both smaller and more explainable. Join Arshavir as he breaks down the surprising challenges of making models efficient without sacrificing understanding.</itunes:summary><itunes:explicit>false</itunes:explicit><itunes:duration>00:06:11</itunes:duration><itunes:image href="https://auth.jellypod.ai/storage/v1/object/public/CoverImages/org_01K8XHYJK76KCD363NADM5CKC2/users/user_01K8XHWVRMZ2M2PFTJPW4582K9/resized_one.jpg"/><itunes:episodeType>full</itunes:episodeType></item><item><title><![CDATA[The Weird Geometry That Makes AI Think]]></title><description><![CDATA[Explore how large language models use high-dimensional geometry to produce intelligent behavior. We peer into the mathematical wilderness inside transformers, revealing how intuition fails, and meaning emerges.]]></description><link>https://arshavir.jellypod.com/episodes/c7771dee-a476-4e8c-8018-1d6fffc6e18a</link><guid isPermaLink="false">c7771dee-a476-4e8c-8018-1d6fffc6e18a</guid><pubDate>Thu, 13 Nov 2025 09:30:01 GMT</pubDate><enclosure url="https://op3.dev/e,pg=9330ccd8-4e44-44e7-bbb5-346de7829edc/auth.jellypod.ai/storage/v1/object/public/Podcasts/org_01K8XHYJK76KCD363NADM5CKC2/users/user_01K8XHWVRMZ2M2PFTJPW4582K9/audio.mp3" length="0" type="audio/mpeg"/><podcast:generator uri="https://www.jellypod.com"></podcast:generator><podcast:episode>4</podcast:episode><podcast:transcript url="https://auth.jellypod.ai/storage/v1/object/public/Podcasts/org_01K8XHYJK76KCD363NADM5CKC2/users/user_01K8XHWVRMZ2M2PFTJPW4582K9/c7771dee-a476-4e8c-8018-1d6fffc6e18a/captions_1762892995.srt" type="application/x-subrip" language="en" rel="captions"></podcast:transcript><itunes:author>Jellypod</itunes:author><itunes:subtitle>Explore how large language models use high-dimensional geometry to produce intelligent behavior. We peer into the mathematical wilderness inside transformers, revealing how intuition fails, and meaning emerges.</itunes:subtitle><itunes:summary>Explore how large language models use high-dimensional geometry to produce intelligent behavior. We peer into the mathematical wilderness inside transformers, revealing how intuition fails, and meaning emerges.</itunes:summary><itunes:explicit>false</itunes:explicit><itunes:image href="https://auth.jellypod.ai/storage/v1/object/public/CoverImages/org_01K8XHYJK76KCD363NADM5CKC2/users/user_01K8XHWVRMZ2M2PFTJPW4582K9/image_1762837557524.jpg"/><itunes:episodeType>full</itunes:episodeType></item><item><title><![CDATA[Can We Fix It?]]></title><description><![CDATA[Arshavir Blackwell takes you on a journey inside the black box of large language models, showing how cutting-edge methods help researchers identify, understand, and even fix the inner quirks of AI. Through concrete case studies, he demonstrates how interpretability is evolving from an arcane art to a collaborative science—while revealing the daunting puzzles that remain. This episode unpacks the step-by-step workflow and surprising realities of mechanistically mapping model cognition.]]></description><link>https://arshavir.jellypod.com/episodes/f5282a9f-749f-40d1-892c-96f9dfc05865</link><guid isPermaLink="false">f5282a9f-749f-40d1-892c-96f9dfc05865</guid><pubDate>Wed, 05 Nov 2025 22:06:00 GMT</pubDate><enclosure url="https://op3.dev/e,pg=9330ccd8-4e44-44e7-bbb5-346de7829edc/auth.jellypod.ai/storage/v1/object/public/Podcasts/org_01K8XHYJK76KCD363NADM5CKC2/users/user_01K8XHWVRMZ2M2PFTJPW4582K9/f5282a9f-749f-40d1-892c-96f9dfc05865/fa1a1fe8.mp3?" length="0" type="audio/mpeg"/><podcast:generator uri="https://www.jellypod.com"></podcast:generator><podcast:episode>3</podcast:episode><podcast:transcript url="https://auth.jellypod.ai/storage/v1/object/public/Podcasts/org_01K8XHYJK76KCD363NADM5CKC2/users/user_01K8XHWVRMZ2M2PFTJPW4582K9/f5282a9f-749f-40d1-892c-96f9dfc05865/captions_1762207784.srt" type="application/x-subrip" language="en" rel="captions"></podcast:transcript><itunes:author>Jellypod</itunes:author><itunes:subtitle>Arshavir Blackwell takes you on a journey inside the black box of large language models, showing how cutting-edge methods help researchers identify, understand, and even fix the inner quirks of AI. Through concrete case studies, he demonstrates how interp</itunes:subtitle><itunes:summary>Arshavir Blackwell takes you on a journey inside the black box of large language models, showing how cutting-edge methods help researchers identify, understand, and even fix the inner quirks of AI. Through concrete case studies, he demonstrates how interpretability is evolving from an arcane art to a collaborative science—while revealing the daunting puzzles that remain. This episode unpacks the step-by-step workflow and surprising realities of mechanistically mapping model cognition.</itunes:summary><itunes:explicit>false</itunes:explicit><itunes:image href="https://auth.jellypod.ai/storage/v1/object/public/CoverImages/org_01K8XHYJK76KCD363NADM5CKC2/users/user_01K8XHWVRMZ2M2PFTJPW4582K9/image_1762837691925.jpg"/><itunes:episodeType>full</itunes:episodeType></item><item><title><![CDATA[Using Symbolic AI to Explain LLMs]]></title><description><![CDATA[Delve into the mysterious world of neural circuits within large language models. We’ll dismantle the jargon, connect these abstract ideas to real examples, and discuss how circuits help bridge the gap between machine learning and human cognition.]]></description><link>https://arshavir.jellypod.com/episodes/c58493e3-e202-4665-b005-656415c3a79e</link><guid isPermaLink="false">c58493e3-e202-4665-b005-656415c3a79e</guid><pubDate>Sun, 02 Nov 2025 22:02:58 GMT</pubDate><enclosure url="https://op3.dev/e,pg=9330ccd8-4e44-44e7-bbb5-346de7829edc/auth.jellypod.ai/storage/v1/object/public/Podcasts/org_01K8XHYJK76KCD363NADM5CKC2/users/user_01K8XHWVRMZ2M2PFTJPW4582K9/c58493e3-e202-4665-b005-656415c3a79e/8ccbf42e.mp3?" length="0" type="audio/mpeg"/><podcast:generator uri="https://www.jellypod.com"></podcast:generator><podcast:episode>2</podcast:episode><podcast:transcript url="https://auth.jellypod.ai/storage/v1/object/public/Podcasts/org_01K8XHYJK76KCD363NADM5CKC2/users/user_01K8XHWVRMZ2M2PFTJPW4582K9/c58493e3-e202-4665-b005-656415c3a79e/captions_1762120949.srt" type="application/x-subrip" language="en" rel="captions"></podcast:transcript><itunes:author>Jellypod</itunes:author><itunes:subtitle>Delve into the mysterious world of neural circuits within large language models. We’ll dismantle the jargon, connect these abstract ideas to real examples, and discuss how circuits help bridge the gap between machine learning and human cognition.</itunes:subtitle><itunes:summary>Delve into the mysterious world of neural circuits within large language models. We’ll dismantle the jargon, connect these abstract ideas to real examples, and discuss how circuits help bridge the gap between machine learning and human cognition.</itunes:summary><itunes:explicit>false</itunes:explicit><itunes:image href="https://auth.jellypod.ai/storage/v1/object/public/CoverImages/org_01K8XHYJK76KCD363NADM5CKC2/users/user_01K8XHWVRMZ2M2PFTJPW4582K9/image_1762837782526.jpg"/><itunes:episodeType>full</itunes:episodeType></item><item><title><![CDATA[Peering Inside the Black Box]]></title><description><![CDATA[Mechanistic interpretability and artificial psycholinguistics are transforming our understanding of large language models. In this episode, Arshavir Blackwell explores how probing neural circuits, behavioral tests, and new tools are unraveling the mysteries of AI reasoning.]]></description><link>https://arshavir.jellypod.com/episodes/caa3a9a7-a2fb-4992-9c13-55ab87827c27</link><guid isPermaLink="false">caa3a9a7-a2fb-4992-9c13-55ab87827c27</guid><pubDate>Sun, 02 Nov 2025 17:15:23 GMT</pubDate><enclosure url="https://op3.dev/e,pg=9330ccd8-4e44-44e7-bbb5-346de7829edc/auth.jellypod.ai/storage/v1/object/public/Podcasts/org_01K8XHYJK76KCD363NADM5CKC2/users/user_01K8XHWVRMZ2M2PFTJPW4582K9/caa3a9a7-a2fb-4992-9c13-55ab87827c27/35f935c2.mp3?" length="0" type="audio/mpeg"/><podcast:generator uri="https://www.jellypod.com"></podcast:generator><podcast:episode>1</podcast:episode><podcast:transcript url="https://auth.jellypod.ai/storage/v1/object/public/Podcasts/org_01K8XHYJK76KCD363NADM5CKC2/users/user_01K8XHWVRMZ2M2PFTJPW4582K9/caa3a9a7-a2fb-4992-9c13-55ab87827c27/captions_1762103676.srt" type="application/x-subrip" language="en" rel="captions"></podcast:transcript><itunes:author>Jellypod</itunes:author><itunes:subtitle>Mechanistic interpretability and artificial psycholinguistics are transforming our understanding of large language models. In this episode, Arshavir Blackwell explores how probing neural circuits, behavioral tests, and new tools are unraveling the mysteri</itunes:subtitle><itunes:summary>Mechanistic interpretability and artificial psycholinguistics are transforming our understanding of large language models. In this episode, Arshavir Blackwell explores how probing neural circuits, behavioral tests, and new tools are unraveling the mysteries of AI reasoning.</itunes:summary><itunes:explicit>false</itunes:explicit><itunes:image href="https://auth.jellypod.ai/storage/v1/object/public/CoverImages/org_01K8XHYJK76KCD363NADM5CKC2/users/user_01K8XHWVRMZ2M2PFTJPW4582K9/image_1762837846087.jpg"/><itunes:episodeType>full</itunes:episodeType></item></channel></rss>