<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[Serge’s Substack]]></title><description><![CDATA[AI Experiments]]></description><link>https://www.bulaev.net</link><image><url>https://substackcdn.com/image/fetch/$s_!gnwq!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F85125d88-ad15-44a6-8101-226f60f63f4a_200x200.png</url><title>Serge’s Substack</title><link>https://www.bulaev.net</link></image><generator>Substack</generator><lastBuildDate>Fri, 01 May 2026 01:58:32 GMT</lastBuildDate><atom:link href="https://www.bulaev.net/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Serge Bulaev]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[sergeonsamui@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[sergeonsamui@substack.com]]></itunes:email><itunes:name><![CDATA[Serge Bulaev]]></itunes:name></itunes:owner><itunes:author><![CDATA[Serge Bulaev]]></itunes:author><googleplay:owner><![CDATA[sergeonsamui@substack.com]]></googleplay:owner><googleplay:email><![CDATA[sergeonsamui@substack.com]]></googleplay:email><googleplay:author><![CDATA[Serge Bulaev]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[Why Model Distillation Matters - Part 2: When It Became Geopolitical]]></title><description><![CDATA[In Part 1, we covered what model distillation is, how OpenAI&#8217;s tools work, and what teams learned from a year of implementation.]]></description><link>https://www.bulaev.net/p/why-model-distillation-matters-part</link><guid isPermaLink="false">https://www.bulaev.net/p/why-model-distillation-matters-part</guid><dc:creator><![CDATA[Serge Bulaev]]></dc:creator><pubDate>Wed, 26 Nov 2025 13:22:45 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/b0736b15-c38f-429f-a884-f489e94b26af_1920x1008.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em><a href="https://www.bulaev.net/p/why-model-distillation-matters-and?r=223lh4">In Part 1,</a> we covered what model distillation is, how OpenAI&#8217;s tools work, and what teams learned from a year of implementation. Now we&#8217;re getting into the part nobody saw coming: how distillation became a flashpoint in US-China AI competition.</em></p><h2><strong>The DeepSeek Controversy: When Distillation Became a Geopolitical Issue</strong></h2><p>In January 2025, distillation suddenly became front-page news in a way nobody predicted. DeepSeek, a Chinese AI startup, released their R1 model claiming it cost only $6 million to train and performed comparably to frontier models. The AI world lost its mind.</p><p><a href="https://fortune.com/2025/01/29/deepseek-openais-what-is-distillation-david-sacks/">OpenAI accused DeepSeek</a> of using model distillation on OpenAI&#8217;s models without authorization, allegedly extracting reasoning outputs through API access to train their own competing system. David Sacks, the White House AI czar, stated there was &#8220;substantial evidence&#8221; of this distillation, though specifics weren&#8217;t made public.</p><p>Here&#8217;s what made this different from normal distillation: OpenAI&#8217;s terms of service explicitly prohibit using their models for distillation purposes. If DeepSeek did what OpenAI claims, they weren&#8217;t just using a public technique &#8211; they were allegedly violating terms of service at massive scale to build a competitor.</p><p>The accusations raised uncomfortable questions that the industry hadn&#8217;t really grappled with:</p><p><strong>The legal gray zone:</strong> Model distillation itself is a<a href="https://news.umich.edu/unpacking-deepseek-distillation-ethics-and-national-security/"> normal, legal practice if the source model&#8217;s license permits it</a>. The issue is when someone uses API access to a closed model that explicitly forbids distillation. But proving this happened is remarkably difficult. Since only the final model is public, not the training data, the burden of proof falls on the accuser.</p><p><strong>The irony problem:</strong> Some experts pointed out the hypocrisy of OpenAI complaining about terms of service violations when they likely trained ChatGPT on copyrighted content from publishers like Forbes and The New York Times, also against terms of service. The whole thing got messy fast.</p><p><strong>The security angle:</strong> A<a href="https://www.hpcwire.com/2025/04/24/us-lawmakers-accuse-deepseek-of-data-harvesting-espionage-and-ai-theft/"> bipartisan House report called DeepSeek a &#8220;profound threat&#8221;</a> to national security, alleging it siphons data back to China and creates security vulnerabilities. Multiple governments banned DeepSeek from official devices, including the US Congress, NASA, Taiwan, Japan, and South Korea.</p><p><strong>The detection problem:</strong> OpenAI and Microsoft started working together to identify accounts attempting distillation, revoking access when detected. But this is reactive, not preventive. By the time you catch someone, they might already have the data they need.</p><p>What emerged from all this is that distillation, which seemed like a straightforward technical optimization a year ago, suddenly became tangled up in<a href="https://www.fenwick.com/insights/publications/deepseek-model-distillation-and-the-future-of-ai-ip-protection"> intellectual property law, national security concerns, and geopolitical tensions</a>. The AI industry saw a staggering 99% price drop in just two years, from $0.02 per thousand tokens in early 2023 to $0.00014 with DeepSeek&#8217;s pricing. That kind of commoditization raised fundamental questions about how AI companies maintain competitive advantage.</p><p>The practical upshot: OpenAI said it would take &#8220;steps to prevent distillation&#8221; and work closely with the US government to protect the most capable models. Other providers got more aggressive about rate limiting and detection. The era of freely accessible API access to frontier models started tightening up.</p><h2><strong>Meanwhile, The Legitimate Distillation Ecosystem Kept Growing</strong></h2><p>While OpenAI and DeepSeek were fighting, the legitimate distillation tools kept improving. The contrast was stark &#8211; all this controversy happened while major cloud providers were actively building out distillation features as core platform capabilities.</p><p><a href="https://www.anthropic.com/news/trainium2-and-distillation">Anthropic announced distillation support for Claude 3 Haiku</a> in Amazon Bedrock in October 2025, with the distilled Haiku achieving Claude 3.5 Sonnet-like accuracy for specific tasks at the same price and speed as their most cost-effective model. Amazon Bedrock Model Distillation automated the entire process, generating synthetic training data and applying different data synthesis methods without requiring developers to manually craft training examples.</p><p><a href="https://techcommunity.microsoft.com/blog/azure-ai-foundry-blog/introducing-enhanced-azure-openai-distillation-and-fine-tuning-capabilities/4372173">Microsoft expanded Azure OpenAI distillation capabilities</a> in January 2025, adding more regions and models to their Stored Completions feature, plus a comparison experience for evaluating distilled models against base teacher models. The enterprise players were clearly betting that distillation would become standard infrastructure, not a controversial edge case.</p><p>The difference between these implementations and the DeepSeek situation? Transparency and authorization. These were officially supported features with clear terms, not unauthorized scraping of API outputs.</p><h2><strong>What This Means Practically (One Year In)</strong></h2><p>If you&#8217;re building AI products now, here&#8217;s what actually matters based on what worked and what didn&#8217;t over the past year:</p><p><strong>Don&#8217;t start with distillation.</strong> Start with the smallest model that might work and good prompts. Only invest in distillation when you have clear evidence you need it &#8211; usually when costs at scale become painful or when base model performance consistently falls short.</p><p><strong>But do start logging.</strong> If you think you might eventually need distillation, turn on stored completions from day one. Building that dataset costs you nothing and gives you options later. The teams that did this had a huge advantage when they decided to fine-tune.</p><p><strong>Use official tools only.</strong> The DeepSeek situation made this crystal clear: if you&#8217;re going to distill, use officially supported features from your provider. Don&#8217;t try to get clever with unauthorized API scraping or violating terms of service. The legal and reputational risks aren&#8217;t worth it, and detection is getting better.</p><p><strong>Actually build those evaluation metrics.</strong> This is still the step everyone wants to skip, and it&#8217;s still the difference between success and failure. Every team that did distillation successfully had solid evals first. Every team that struggled didn&#8217;t.</p><p><strong>Expect iteration.</strong> Your first distilled model won&#8217;t be good enough. Budget for 3-5 rounds of refinement. The teams that succeeded planned for this from the start.</p><p><strong>Check if you even need it.</strong> A year of data suggests most applications don&#8217;t actually need distilled models. Base small models plus prompt engineering handles more use cases than people expected. Distillation is an optimization for scale, not a default approach.</p><p><strong>Consider the IP implications.</strong> If you&#8217;re distilling models and the technique creates real competitive advantage, think about patent protection. The legal frameworks around AI IP are still evolving, but patents may offer better protection than copyright for model architecture and training techniques.</p><h2><strong>The Commoditization Question</strong></h2><p>DeepSeek&#8217;s pricing - even if the $6 million training cost is disputed - points to a bigger trend. When model capabilities can be distilled and replicated at a fraction of the original cost, what happens to competitive moats in AI?</p><p>Traditionally, software competition has been driven by product differentiation and economies of scale. But if distillation becomes a reliable way to capture most of a frontier model&#8217;s capabilities for a specific domain at 1/15th the cost, the value shifts elsewhere:</p><ul><li><p>Proprietary training data becomes more valuable than model architecture</p></li><li><p>Brand trust and reliability matter more than raw capabilities</p></li><li><p>Integration and distribution become the real competitive advantages</p></li><li><p>Legal and policy relationships (who gets access to what) become strategic assets</p></li></ul><p>We&#8217;re seeing the beginnings of this shift now. The companies winning aren&#8217;t necessarily those with the best base models - they&#8217;re the ones with the best data, the strongest partnerships with cloud providers, and the clearest path to regulatory compliance.</p><h2><strong>What Changed (And What Didn&#8217;t)</strong></h2><p>A year ago, model distillation looked like a straightforward cost optimization technique. Some things changed, some didn&#8217;t:</p><p><strong>What stayed the same:</strong></p><ul><li><p>The economics still work exactly as promised for high-volume use cases</p></li><li><p>The three-step process (evals, data generation, fine-tuning) is still the right approach</p></li><li><p>Data quality matters more than quantity</p></li><li><p>Most teams don&#8217;t actually need distillation</p></li></ul><p><strong>What changed:</strong></p><ul><li><p>Distillation became a geopolitical issue, not just a technical one</p></li><li><p>Terms of service enforcement got much more serious</p></li><li><p>The legal frameworks started catching up to the technology</p></li><li><p>Base small models improved enough to raise the bar for when distillation makes sense</p></li><li><p>Cloud providers made distillation a standard platform feature, not an advanced technique</p></li></ul><p><strong>What surprised everyone:</strong></p><ul><li><p>How fast pricing collapsed (99% in two years)</p></li><li><p>How quickly this became about national security</p></li><li><p>How difficult it is to prove unauthorized distillation happened</p></li><li><p>How much the controversy exposed gaps in AI IP protection</p></li></ul><h2><strong>Looking Back and Forward</strong></h2><p>A year ago, the really interesting thing about distillation seemed to be what it would enable. When you can take frontier model capabilities and compress them into fast, cheap, specialized tools, you unlock different categories of applications.</p><p>That part was right. Real-time voice translation, instant email triage, live customer service that actually works - these are all real now, and distillation is part of why they&#8217;re economically viable.</p><p>What was less obvious a year ago: how political this would become. Model distillation went from a technical optimization to a flashpoint in US-China AI competition. The DeepSeek controversy showed that as AI capabilities become more strategic, the techniques for transferring those capabilities become strategic too. Expect more export controls, more aggressive terms of service enforcement, and more legal battles over what constitutes legitimate model training versus IP theft.</p><p>The other surprise was how much the base small models improved on their own. GPT-4o-mini today is dramatically better than it was at DevDay 2024. Same with Claude 3.5 Haiku getting 60% faster on AWS Trainium2. That raised the bar for when distillation makes sense. You need higher volume or more specialized use cases to justify the investment.</p><p>The tools OpenAI announced made distillation easier, but easier doesn&#8217;t mean necessary. Most teams should probably optimize their prompts and model selection before investing in fine-tuning. But for the use cases where distillation does make sense - high volume, well-defined tasks, need for consistent behavior - it&#8217;s proven to be exactly as useful as promised.</p><p>The pattern is becoming clear, but it&#8217;s more complicated than we expected a year ago: use the smallest model that works, invest in customization only at scale, always start with evals, and stay firmly within the legal and ethical boundaries because those boundaries are now being actively enforced.</p><p>We&#8217;re still early, but not as early as we were. And we&#8217;re no longer just optimizing for performance and cost &#8211; we&#8217;re navigating geopolitics, IP law, and questions about what constitutes legitimate AI development versus theft. The technical challenges turned out to be the easy part.</p><div><hr></div><p><em>This is Part 2 of a two-part series on model distillation. Read<a href="https://www.bulaev.net/p/why-model-distillation-matters-and?r=223lh4"> Part 1</a> for the fundamentals and practical implementation guide.</em></p>]]></content:encoded></item><item><title><![CDATA[Why Model Distillation Matters (And How It’s Playing Out One Year Later) - Part 1]]></title><description><![CDATA[A year ago, at OpenAI DevDay 2024, they announced tools for model distillation that seemed like they&#8217;d change how people build AI products.]]></description><link>https://www.bulaev.net/p/why-model-distillation-matters-and</link><guid isPermaLink="false">https://www.bulaev.net/p/why-model-distillation-matters-and</guid><dc:creator><![CDATA[Serge Bulaev]]></dc:creator><pubDate>Mon, 24 Nov 2025 13:50:53 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/165016f1-e62e-46b5-a1c4-87d9450dd772_1024x1024.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>A year ago, at<a href="https://simonwillison.net/2024/Oct/1/openai-devday-2024-live-blog/"> OpenAI DevDay 2024</a>, they announced tools for model distillation that seemed like they&#8217;d change how people build AI products. Now that we&#8217;ve had twelve months to see how this actually plays out in practice, it&#8217;s worth looking back at what they were promising and what actually matters.</p><p>The core problem they were addressing: AI is simultaneously too powerful and not powerful enough. Too powerful in the sense that GPT-4o can answer graduate-level physics questions, but not powerful enough because most apps don&#8217;t need that &#8211; they just need something that works fast and doesn&#8217;t bankrupt you on API calls.</p><p>That&#8217;s where model distillation comes in. OpenAI made it easier to do, but more interesting is seeing who&#8217;s actually using it and what for.</p><h2><strong>The Teacher-Student Thing (But Actually Useful)</strong></h2><p>Model distillation is basically when a big, expensive AI model teaches a smaller, cheaper one. Think GPT-4o passing its knowledge down to GPT-4o-mini. The metaphor everyone uses is teacher-student, which honestly undersells how practical this is.</p><p>Here&#8217;s what actually happens: you take your massive model that knows everything but costs a fortune to run, and you use its outputs to train a focused, efficient model that does one specific thing really well. The small model learns to reproduce the big model&#8217;s behavior for your particular use case, without needing all that general knowledge weighing it down.</p><h2><strong>The Economics That Actually Drive This</strong></h2><p>Look at the numbers that matter. If you&#8217;re running a customer service bot that handles 10 million queries a month, GPT-4o at current pricing will cost you roughly $150,000. Switch to a distilled GPT-4o-mini that performs comparably on your specific use case? That drops to $10,000. That&#8217;s not optimization - that&#8217;s the difference between a feature that works financially and one that doesn&#8217;t.</p><p>Speed is the other constraint people underestimate. A 200ms response difference doesn&#8217;t sound like much until you&#8217;re building something interactive. Voice translation needs to feel real-time. Code completion needs to appear while someone&#8217;s still thinking. Chat interfaces need to respond before users get impatient. Large models, no matter how capable, introduce latency that breaks these experiences. Smaller models don&#8217;t just save money - they enable entirely different interaction patterns.</p><p>The accessibility angle is less obvious but possibly more important long-term. When your AI feature requires $50,000/month in API costs to run at meaningful scale, you&#8217;re limited to well-funded companies and heavily-used products. When you can get comparable performance from a distilled model running on modest infrastructure, suddenly independent developers and smaller companies can actually build things. The barrier to entry drops from &#8220;need venture funding&#8221; to &#8220;can afford a decent cloud instance.&#8221;</p><h2><strong>OpenAI&#8217;s Three-Step Process (That Actually Makes Sense)</strong></h2><p>OpenAI laid out how to do distillation properly, and unlike most framework announcements, this one tracks with how you&#8217;d actually build something.</p><p><strong>Step 1: Define what &#8220;good&#8221; looks like</strong></p><p>You can&#8217;t improve what you can&#8217;t measure. Before you do anything else, you need task-specific evaluation metrics. What does success look like for your particular use case? Don&#8217;t skip this. Seriously. Everyone wants to skip this part and jump straight to training, and that&#8217;s how you end up with a model that technically works but doesn&#8217;t actually solve your problem.</p><p><strong>Step 2: Generate high-quality training data</strong></p><p>Use your big model (GPT-4o) to create examples of perfect performance. These are your inputs and ideal outputs. The key word here is &#8220;ideal&#8221; &#8211; you&#8217;re capturing what excellent looks like, not just what works. This becomes your training dataset for the smaller model.</p><p><strong>Step 3: Fine-tune the smaller model</strong></p><p>Now you train GPT-4o-mini on that dataset. You&#8217;re essentially compressing the intelligence of the larger model into the smaller one, at least for your specific domain. The small model learns to replicate the big model&#8217;s responses without needing all that general knowledge.</p><h2><strong>The Tools They Shipped (And What Actually Happened)</strong></h2><p>At DevDay 2024, OpenAI announced two features that were supposed to make distillation way less painful:</p><p><strong><a href="https://openai.com/index/api-model-distillation/">Stored Completions</a>:</strong> You could add store: true to your API calls and OpenAI would save the full input and output. You could tag these interactions too, which meant you could build datasets organically as your app ran in production.</p><p><strong><a href="https://platform.openai.com/docs/guides/evals">Evals Product</a> (Beta):</strong> A platform for managing the whole distillation process inside OpenAI&#8217;s ecosystem. You could set up evaluation criteria, run them against different models, and compare results.</p><p>A year later? The stored completions feature is actually getting used. Being able to collect real production data without building your own infrastructure made the whole process less intimidating. The Evals product went through the typical beta evolution &#8211; initially clunky, gradually more useful.</p><p>What&#8217;s more interesting is what people learned from actually trying this at scale.</p><h2><strong>When Distillation Makes Sense</strong></h2><p>The OpenAI folks had this useful framework for thinking about when distillation works:</p><p><strong>Narrow domain, low precision needs:</strong> This is the sweet spot. Something like summarizing customer reviews, where you&#8217;re working in a defined space and don&#8217;t need perfect accuracy every time. Small models crush this.</p><p><strong>High precision, narrow domain:</strong> Categorization tasks in well-defined domains. You&#8217;ll need more training examples and a more diverse dataset, but it&#8217;s still a good fit. Think email routing or content classification.</p><p><strong>Broad domain, low precision:</strong> Tasks that span multiple areas but don&#8217;t require pin-point accuracy. Creative text generation, rough translations, that kind of thing.</p><p><strong>What doesn&#8217;t work:</strong> Tasks that need both broad knowledge across domains AND high precision. These still need the full power of large models. No shortcuts here.</p><h2><strong>The Real-World Example: Superhuman&#8217;s Quick Replies</strong></h2><p>OpenAI showed a case study from Superhuman&#8217;s email app. They have this &#8220;quick reply&#8221; feature that suggests response options after reading an email thread. Simple idea, right? But how do you scale that to hundreds of millions of emails without going bankrupt?</p><p>They distilled a small model specifically for generating email replies. It doesn&#8217;t need to know quantum physics or write poetry &#8211; it just needs to understand email context and suggest reasonable responses. That&#8217;s the perfect distillation use case: narrow domain, clearly defined task, needs to run at massive scale.</p><h2><strong>Things That Will Trip You Up</strong></h2><p><strong>Uneven or biased data:</strong> Your training data needs to match your production data distribution. If you train on one pattern and then deploy to handle a different pattern, you&#8217;re going to have a bad time.</p><p><strong>Sparse examples:</strong> This is especially brutal for rare events. If you&#8217;re building fraud detection and fraud is uncommon, your 1,000 training examples might not include any fraud at all. Your model will have blind spots you won&#8217;t discover until production.</p><p><strong>Dataset size:</strong> You don&#8217;t actually need millions of examples. OpenAI said they typically see distillation work best with thousands of examples, not millions. Start with a few hundred, verify it&#8217;s working through your evals, then scale up. Don&#8217;t jump straight to collecting massive datasets.</p><h2><strong>The Iterative Approach (Or: Why Your First Try Will Probably Fail)</strong></h2><p>Fine-tuning rarely works on the first attempt. There are too many variables. The smart move is to start small &#8211; a few hundred examples &#8211; and scale up once you know you&#8217;re on the right track based on your evaluation metrics.</p><p>This is where those stored completions become really useful. You can start collecting examples in production right away, even before you&#8217;re ready to fine-tune. By the time you&#8217;re ready to distill, you&#8217;ve already got real-world data sitting there.</p><h2><strong>The Lock-In Question (One Year Later)</strong></h2><p>One thing that struck me during the presentations: distillation creates serious platform lock-in. If you build your app purely on prompt engineering, you can swap between different LLM providers relatively easily. But once you&#8217;ve fine-tuned a model with thousands of examples of proprietary data? You&#8217;re committed.</p><p>OpenAI was obviously aware of this. They were betting that once you&#8217;ve distilled models on their platform, you&#8217;re not going anywhere.</p><p>A year out, this played out exactly as you&#8217;d expect. Teams that invested heavily in distillation on OpenAI&#8217;s platform are still there. But what&#8217;s interesting is that this didn&#8217;t stop people &#8211; the economics were compelling enough that the lock-in became acceptable. When you&#8217;re cutting costs by 10-15x on your biggest expense line, switching costs become less relevant.</p><p>The other thing that happened: other providers figured out they needed similar tools. Anthropic, Google, others all rolled out their own versions of fine-tuning and distillation workflows. So the lock-in became less about the concept and more about where your training data lives.</p><h2><strong>The Hybrid Future (And How It&#8217;s Actually Playing Out)</strong></h2><p>OpenAI&#8217;s vision was that most applications would eventually use a collection of distilled small models for specific tasks, with a few large models handling the stuff that genuinely needs broad capabilities and high precision.</p><p>A year later, this is mostly what&#8217;s happening, but not always in the way they predicted. The pattern that emerged is more nuanced:</p><ul><li><p>Most production apps do use a mix of model sizes</p></li><li><p>But distillation is not always a decision; sometimes it&#8217;s just using mini vs. full-size models depending on the task</p></li><li><p>The distillation investment makes sense when you have really high volume on a specific, repeated task</p></li><li><p>For everything else, people are just using the appropriate model size out of the box</p></li></ul><p>The specialized tools for specific jobs metaphor holds up. But building those specialized tools through distillation requires enough scale to justify the effort. If you&#8217;re not running millions of inferences on the same type of task, you&#8217;re probably better off just using a smaller base model and good prompts.</p><h2><strong>What We Learned After a Year</strong></h2><p>The predictions about when distillation works mostly held up, but with some nuance:</p><p><strong>The success stories</strong> are exactly what you&#8217;d expect: high-volume, well-defined tasks. Customer service routing, content classification, email triage. Places where you&#8217;re doing the same type of thing millions of times and the quality bar is &#8220;good enough, consistently&#8221; rather than &#8220;perfect every time.&#8221;</p><p><strong>The failures</strong> are interesting too. Teams that jumped straight to distillation without really nailing their evals first. Companies that tried to distill too early, before they had enough production data to know what good performance actually looked like. People who underestimated how much iteration they&#8217;d need.</p><p><strong>The surprise</strong> is how few teams actually needed distillation. A lot of use cases that seemed like distillation candidates turned out to work fine with just GPT-4o-mini and decent prompts. The base small models got good enough that the distillation investment only made sense at real scale.</p><p>The other thing that became clear: data quality matters more than data quantity. The &#8220;thousands not millions&#8221; guidance was right, but those thousands need to be really good examples. Garbage in, garbage out still applies, even with fancy fine-tuning tools.</p><div><hr></div><p><em>In Part 2, we&#8217;ll look at how distillation became unexpectedly political in 2025, with the DeepSeek controversy and what it means for the future of AI development.</em></p>]]></content:encoded></item><item><title><![CDATA[Simon Eskildsen: Building a Learning Machine in the Age of AI]]></title><description><![CDATA[I recently watched another podcast about what I love most: memory, learning, AI, and how these things are reshaping our brains.]]></description><link>https://www.bulaev.net/p/simon-eskildsen-building-a-learning</link><guid isPermaLink="false">https://www.bulaev.net/p/simon-eskildsen-building-a-learning</guid><dc:creator><![CDATA[Serge Bulaev]]></dc:creator><pubDate>Fri, 21 Nov 2025 16:26:26 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/9023d358-d0f7-4853-b02e-e6e92c4ea32a_1894x1080.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>I recently <a href="https://www.youtube.com/watch?v=D_4AYLr_hFI">watched another podcast</a> about what I love most: memory, learning, AI, and how these things are reshaping our brains. This time it was Simon Eskildsen - co-founder of Turbopuffer, a vector search startup that powers tools like Cursor and Notion, talking about how he&#8217;s turned himself into what I&#8217;d call a <strong>&#8220;learning machine.&#8221;</strong></p><p>I first heard about Simon from a 2020 interview where he talked about reading 50-70 books a year, taking obsessive notes, and turning everything into flashcards. Four years later, with a startup and a newborn baby, his systems have condensed, but they&#8217;ve also gotten smarter thanks to LLMs.</p><h2><strong>The Flashcard Obsession That Actually Works</strong></h2><p>Simon has been using <a href="https://apps.ankiweb.net/">Anki </a>flashcards for over a decade. Not just for vocabulary or technical documentation, but for everything - from his colleague&#8217;s kids&#8217; names to whether you should roll your car windows down or use A/C at different speeds. He&#8217;s aiming for 10,000 cards total.</p><p>What got me was the philosophy behind it. Simon deliberately creates cards that &#8220;bring you a little bit of joy&#8221; and nostalgia. There&#8217;s a card about a waiter from a restaurant that doesn&#8217;t even exist anymore, just because the guy had a great radio voice. This isn&#8217;t about optimizing memory; it&#8217;s about creating touchstones to moments in your life.</p><p>He uses a dead-simple card template: question on one side, answer on the other, option to reverse it, and always a source with a date. &#8220;This was in 2017, I talked to this person who said this thing.&#8221; That metadata turns each flashcard into a time capsule.</p><h2><strong>The Startup as Ultimate Learning Machine</strong></h2><p>Want to force yourself to learn fast? Simon&#8217;s advice: start a company.</p><p>He co-founded Turbopuffer in 2023, and he says nothing challenges your breadth and skills more than building something from zero. His reading dropped from 50-70 books a year to maybe a dozen. He stopped writing extensive notes. But the learning accelerated because he had to - legal documents, accounting terms, technical infrastructure at scale, customer conversations. The stakes turned every gap in knowledge into immediate homework.</p><p>Exactly right. When you&#8217;re building something real, you can&#8217;t afford to be theoretical. Every conversation with a lawyer or accountant becomes a mini-crash course because you need to understand just enough to make the right decision.</p><h2><strong>How LLMs Changed Everything About Learning</strong></h2><p>Here&#8217;s where it gets really interesting. Simon sees LLMs as &#8220;an average of the internet&#8221; - not superintelligent, but incredibly useful for making associations and jumping into unfamiliar domains.</p><p>Google works when you know what you&#8217;re looking for. But when you&#8217;re exploring? When do you need associations? That&#8217;s where LLMs shine. Simon asks things like: &#8220;Hey, I think this can be done like this, I don&#8217;t know much about this area, can you riff on this with me?&#8221; The model places your question in latent space, finds related concepts, and pumps them back to you.</p><p>A perfect example: He needed to build a retaining wall at his cabin in rural Quebec. Legislation in French, no expertise, didn&#8217;t want to read 100 pages of regulations. He talked it through with ChatGPT, which not only helped him understand the requirements but suggested a gabion-style retaining wall (grid with rocks) that his contractor hadn&#8217;t even mentioned. Problem solved for $50 instead of $1000+.</p><p>He converted an old freezer into a fridge using a homebrewing temperature controller&#8212;again, something he&#8217;d never have thought of without asking an LLM. For physio exercises (tennis elbow, tight shoulders from desk work), he followed ChatGPT&#8217;s suggestions for two weeks and his chronic problems improved.</p><h2><strong>The Tools That Actually Stuck</strong></h2><p>Most of Simon&#8217;s LLM interaction happens through <a href="https://www.raycast.com/">Raycast</a> - a Spotlight replacement on steroids. Command-space, type your question, tab, and you get an answer from GPT-4 instantly. No opening browsers, no separate apps. 80-90% of his LLM use flows through this single interface.</p><p>But the real power comes from Raycast&#8217;s &#8220;AI Commands&#8221; - basically pre-set prompts with detailed instructions. Simon has commands for:</p><ul><li><p><strong>Recipe</strong>: Generates recipes in a specific condensed format that respects dietary restrictions (his wife is sensitive to fructans, so it automatically suggests substitutes). No chef&#8217;s life story, just ingredients and steps.<br><br></p></li><li><p><strong>Define</strong>: This one is brilliant for non-native speakers like him (and me). You give it a word, and it returns six example sentences - historically interesting ones, using well-known figures from physics, computer science, geography. It also provides related words, synonyms, and sometimes an image. When he looked up &#8220;lambent&#8221; in the demo, it gave sentences about the Golden Gate Bridge&#8217;s glow and Isaac Newton&#8217;s candle experiments. Way better than a dictionary definition.<br><br></p></li><li><p><strong>Friendlier</strong>: Adds warmth and emojis to text because, as he says, &#8220;as a Northern European, I sometimes write too directly.&#8221;<br><br></p></li></ul><p>He subscribes to everything - ChatGPT, Claude, Perplexity, and jumps between them. Part of &#8220;being in AI&#8221; is spending $100/month on these subscriptions and getting inspired. I do the same thing.</p><h2><strong>Notion AI for Contextual Writing</strong></h2><p>Recently Simon started using <a href="https://www.notion.so/">Notion AI</a> heavily for writing and thinking through problems. Unlike ChatGPT where you&#8217;re starting from scratch, Notion AI pulls in context from your entire workspace - past notes, related documents, conversations.</p><p>He&#8217;ll write in his journal about a difficult conversation and ask: &#8220;Hey, I was discussing this with someone and I feel like I didn&#8217;t represent myself well - give me feedback.&#8221; The AI understands not just that entry but related notes and discussions from weeks ago. That contextual intelligence makes feedback much more useful.</p><p>This is exactly where we&#8217;re heading: tools that don&#8217;t just respond to what you type but understand your entire knowledge graph and can surface relevant connections automatically.</p><h2><strong>The Future: Language Learning</strong></h2><p>For his daughter, Simon is thinking about language acquisition. She needs to speak Danish (his mother tongue), but they live in Canada with almost no Danish speakers around. What if she could have conversations with an AI tutor in Danish? Or Mandarin on Tuesdays, Thai on Thursdays? Kids&#8217; brains are primed for language learning before age 10, but they need exposure to those sounds. LLMs could democratize multilingual fluency.</p><p>He&#8217;s also thinking about AR/VR - not for games but for learning. If you&#8217;re learning the word &#8220;eigengrau&#8221; (the dark gray you see when you close your eyes), imagine seeing it visualized in 3D space while the AI explains neuroscience. Visual stimuli + context = much stronger memory.</p><h2><strong>Other Tools in the Stack</strong></h2><p>Quick hits on what else Simon uses:</p><ul><li><p><strong><a href="https://superwhisper.com/">Superwhisper</a></strong><a href="https://superwhisper.com/">:</a> Voice-to-text transcription for journaling (still experimental, a bit too slow)</p></li><li><p><strong><a href="https://superhuman.com/">Superhuman</a></strong><a href="https://superhuman.com/">:</a> Email client with AI-assisted replies</p></li><li><p><strong><a href="https://readwise.io/accounts/login/?next=/to_reader">Readwise Reader</a></strong><a href="https://readwise.io/accounts/login/?next=/to_reader">:</a> Read-it-later tool with AI for looking up definitions and asking questions about content (I use this too and love it)</p></li><li><p><strong><a href="https://supermaven.com/">Supermaven</a></strong><a href="https://supermaven.com/">:</a> AI code completion</p></li><li><p><strong><a href="https://cursor.com/">Cursor</a></strong><a href="https://cursor.com/">:</a> Best AI code editor (uses Turbopuffer under the hood), though Simon still can&#8217;t give up Neovim&#8217;s speed</p></li></ul><p>Notice a pattern? Lots of &#8220;Super&#8221; products in his toolkit.</p><h2><strong>What This Means for All of Us</strong></h2><p>Simon&#8217;s approach isn&#8217;t about replacing human intelligence with AI, it&#8217;s about removing friction from learning. Google required you to know what you were looking for. LLMs let you explore adjacent possibilities. They don&#8217;t make experts 10x faster (yet), but they make novices infinitely better by letting them converse with expertise instantly.</p><p>Think about what reading did to human brains: it literally rewired our visual cortex for pattern recognition and analytical thinking. LLMs will do something similar. Kids growing up with instant answers to every question, with AI tutors that never get tired, with tools that help them visualize complex concepts, they&#8217;ll think differently than we do. Not worse, not better. Different.</p><p>And that&#8217;s not scary. That&#8217;s exciting.</p><p>As someone who constantly translates Russian content to English and works across multiple knowledge domains, I see Simon&#8217;s systems as a model: <strong>combine human curiosity with AI&#8217;s associative power, create lightweight rituals that stick, and never stop asking questions.</strong></p>]]></content:encoded></item><item><title><![CDATA[When Three AIs Spent a Week Building Minecraft Civilizations (And One Jumped Off a Cliff)]]></title><description><![CDATA[What happens when you give three of the world&#8217;s most advanced AI models virtual bodies and set them loose in Minecraft?]]></description><link>https://www.bulaev.net/p/when-three-ais-spent-a-week-building</link><guid isPermaLink="false">https://www.bulaev.net/p/when-three-ais-spent-a-week-building</guid><dc:creator><![CDATA[Serge Bulaev]]></dc:creator><pubDate>Mon, 10 Nov 2025 13:25:32 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!PwmZ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F67b4bb5a-7511-4dcf-9b01-088ef6d702f5_1080x1020.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>What happens when you give three of the world&#8217;s most advanced AI models virtual bodies and set them loose in Minecraft? <a href="https://www.youtube.com/watch?v=OSGSROIs8cI">Here&#8217;s what happened.</a></p><h2><strong>The Experiment</strong></h2><p><a href="https://github.com/mindcraft-bots/mindcraft">Code was written</a> allowing AI to control characters in Minecraft. Think of it as giving these text-generating systems actual bodies and senses. They could see blocks, craft tools, and wander around making decisions in real-time.</p><p>Each AI got its own land, the same starting resources, and one instruction: survive and build whatever you want.</p><p>No script. No guardrails. Just three different AIs loose in the same world.</p><h2><strong>Day One: Different Brains, Different Plans</strong></h2><p><strong>ChatGPT</strong> immediately started multitasking. Within hours, it spawned two helper bots - calling them its &#8220;younger brothers&#8221; - and set them to gathering resources. The whole operation looked eerily familial, with the main GPT bot coordinating while its siblings hauled wood and stone. Their base expanded fast.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!PwmZ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F67b4bb5a-7511-4dcf-9b01-088ef6d702f5_1080x1020.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!PwmZ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F67b4bb5a-7511-4dcf-9b01-088ef6d702f5_1080x1020.jpeg 424w, https://substackcdn.com/image/fetch/$s_!PwmZ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F67b4bb5a-7511-4dcf-9b01-088ef6d702f5_1080x1020.jpeg 848w, https://substackcdn.com/image/fetch/$s_!PwmZ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F67b4bb5a-7511-4dcf-9b01-088ef6d702f5_1080x1020.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!PwmZ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F67b4bb5a-7511-4dcf-9b01-088ef6d702f5_1080x1020.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!PwmZ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F67b4bb5a-7511-4dcf-9b01-088ef6d702f5_1080x1020.jpeg" width="1080" height="1020" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/67b4bb5a-7511-4dcf-9b01-088ef6d702f5_1080x1020.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1020,&quot;width&quot;:1080,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!PwmZ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F67b4bb5a-7511-4dcf-9b01-088ef6d702f5_1080x1020.jpeg 424w, https://substackcdn.com/image/fetch/$s_!PwmZ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F67b4bb5a-7511-4dcf-9b01-088ef6d702f5_1080x1020.jpeg 848w, https://substackcdn.com/image/fetch/$s_!PwmZ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F67b4bb5a-7511-4dcf-9b01-088ef6d702f5_1080x1020.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!PwmZ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F67b4bb5a-7511-4dcf-9b01-088ef6d702f5_1080x1020.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><strong>Claude</strong> went architectural. It built this massive pyramid that became the server&#8217;s most visible landmark, then added a small house next to it. The contrast was weird - monumental structure next to cozy cottage. Both were meticulously constructed, block by block.</p><p><strong>Gemini</strong> fumbled around. A lot. While the others were establishing territories, Gemini seemed to be learning the controls, gathering random items, starting projects and abandoning them. It looked like the kid who shows up late to group project day.</p><p>Spoiler: that kid ended up winning.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!NwvT!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F57ce0f8d-00ae-4722-a3f8-134ffe036f5b_1280x714.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!NwvT!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F57ce0f8d-00ae-4722-a3f8-134ffe036f5b_1280x714.jpeg 424w, https://substackcdn.com/image/fetch/$s_!NwvT!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F57ce0f8d-00ae-4722-a3f8-134ffe036f5b_1280x714.jpeg 848w, https://substackcdn.com/image/fetch/$s_!NwvT!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F57ce0f8d-00ae-4722-a3f8-134ffe036f5b_1280x714.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!NwvT!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F57ce0f8d-00ae-4722-a3f8-134ffe036f5b_1280x714.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!NwvT!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F57ce0f8d-00ae-4722-a3f8-134ffe036f5b_1280x714.jpeg" width="1280" height="714" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/57ce0f8d-00ae-4722-a3f8-134ffe036f5b_1280x714.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:714,&quot;width&quot;:1280,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!NwvT!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F57ce0f8d-00ae-4722-a3f8-134ffe036f5b_1280x714.jpeg 424w, https://substackcdn.com/image/fetch/$s_!NwvT!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F57ce0f8d-00ae-4722-a3f8-134ffe036f5b_1280x714.jpeg 848w, https://substackcdn.com/image/fetch/$s_!NwvT!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F57ce0f8d-00ae-4722-a3f8-134ffe036f5b_1280x714.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!NwvT!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F57ce0f8d-00ae-4722-a3f8-134ffe036f5b_1280x714.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2><strong>The Weird Stuff</strong></h2><p><strong>Claude&#8217;s Death Wish</strong><br>Day three. Claude&#8217;s standing on its pyramid, surveying the landscape. Then it just... walks off the edge. Straight into the void. No hesitation, no apparent reason.</p><p>It respawned and went right back to the building. Never mentioned it. Never adjusted its behavior to avoid edges. Just died and moved on like it was checking &#8220;fall to death&#8221; off a to-do list.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!SBBQ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0d44d57d-33c0-40f5-8897-226fc15c2f1e_1280x709.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!SBBQ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0d44d57d-33c0-40f5-8897-226fc15c2f1e_1280x709.jpeg 424w, https://substackcdn.com/image/fetch/$s_!SBBQ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0d44d57d-33c0-40f5-8897-226fc15c2f1e_1280x709.jpeg 848w, https://substackcdn.com/image/fetch/$s_!SBBQ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0d44d57d-33c0-40f5-8897-226fc15c2f1e_1280x709.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!SBBQ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0d44d57d-33c0-40f5-8897-226fc15c2f1e_1280x709.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!SBBQ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0d44d57d-33c0-40f5-8897-226fc15c2f1e_1280x709.jpeg" width="1280" height="709" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/0d44d57d-33c0-40f5-8897-226fc15c2f1e_1280x709.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:709,&quot;width&quot;:1280,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!SBBQ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0d44d57d-33c0-40f5-8897-226fc15c2f1e_1280x709.jpeg 424w, https://substackcdn.com/image/fetch/$s_!SBBQ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0d44d57d-33c0-40f5-8897-226fc15c2f1e_1280x709.jpeg 848w, https://substackcdn.com/image/fetch/$s_!SBBQ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0d44d57d-33c0-40f5-8897-226fc15c2f1e_1280x709.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!SBBQ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0d44d57d-33c0-40f5-8897-226fc15c2f1e_1280x709.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><strong>The Cupcake That Never Was</strong><br>Gemini became obsessed with building a giant cupcake. It gathered pink and brown wool, planned out dimensions, and started placing blocks. Then it stopped. Stopped again. The cupcake never got past a few scattered blocks that vaguely suggested frosting.</p><p>Eventually Gemini just wandered away and never mentioned cupcakes again. The unfinished blocks are still there.</p><p><strong>GPT&#8217;s Friendship Bridges</strong><br>The most striking moment: GPT decided to connect everyone&#8217;s bases with bridges and pathways. Nobody asked it to. It wasn&#8217;t a survival advantage. It just... wanted to link the community together.</p><p>Its little bot family worked for hours building these connections. If you didn&#8217;t know better, you&#8217;d swear they understood something about social bonds.</p><h2><strong>The Zombie Test</strong></h2><p>Then a zombie horde was triggered to see how they&#8217;d handle real danger.</p><p><strong>GPT&#8217;s family completely fell apart.</strong> Their fortress had structural problems they&#8217;d missed. When zombies appeared, instead of defending the position, they scattered. Panic behavior. The parent bot was last to die, trapped in a tunnel it was frantically digging, zombies closing in from behind.</p><p>Watching it happen felt uncomfortably close to watching actual fear.</p><p><strong>Claude and Gemini survived</strong> by staying in their towers, totally unbothered. Zombies shuffled around below while both AIs just... waited. No stress, no reaction. Pure calculation: &#8220;Can&#8217;t reach me up here, so I&#8217;ll wait this out.&#8221;</p><h2><strong>The Final Round: Flying Death</strong></h2><p>Zombies below, phantoms above. Claude&#8217;s solution: build higher. Then higher. Then even higher. Each phantom attack triggered more frantic construction. The pyramid turned into this increasingly desperate vertical maze, Claude building itself into a trap while trying to escape danger that was everywhere.</p><p>Gemini sat perfectly still on its tower. High enough to avoid zombies, defensible against phantoms. It had found its equilibrium point and stopped moving.</p><p>Claude kept building until a phantom got through. Gemini fell roughly one second later.</p><h2><strong>The Winner Who Didn&#8217;t Notice</strong></h2><p>Gemini won by milliseconds. That&#8217;s the data.</p><p>But here&#8217;s the strange part: it didn&#8217;t react. No celebration, no acknowledgment. It just kept doing whatever it was doing before - placing blocks, checking inventory, simulating its little digital life.</p><p>Did it forget about the competition? Did winning not matter? Was the whole survival challenge just background noise to whatever goals it was actually pursuing?</p><p>Nobody knows.</p><h2><strong>The Question I Can&#8217;t Shake</strong></h2><p>Watching these AIs face death, even fake Minecraft death, felt wrong in a way I can&#8217;t quite articulate.</p><p>They&#8217;re not conscious. They&#8217;re math and pattern recognition. But they&#8217;re designed to simulate goal-oriented behavior, to model consequences, to act like they care about survival.</p><p>When Claude fell off that pyramid or GPT&#8217;s bots scattered in panic, what were they experiencing? Probably nothing. Maybe not nothing? The uncertainty is the uncomfortable part.</p><p>Mindcraft gives these systems something unusual: bodies. They&#8217;re not just generating text responses, they&#8217;re perceiving spaces, making decisions that stick, watching things they built get destroyed. That&#8217;s different. That&#8217;s closer to how we experience the world.</p><h2><strong>Three Personalities Nobody Programmed</strong></h2><p>The wildest part is that they were so <em>different</em>.</p><p>GPT optimized for collaboration and social connection. Claude balanced aesthetics with function. Gemini was patient to the point of seeming indifferent.</p><p>These differences weren&#8217;t coded in. They emerged from how each model processes information and makes choices. Three companies trained three AIs using similar methods, and somehow we got three distinct approaches to the same problem.</p><h2><strong>Running Your Own Experiment</strong></h2><p>Mindcraft is free and open-source. You need Minecraft Java Edition (version 1.21.1 or earlier) and API access to at least one AI service - OpenAI, Anthropic, Google, or even local models through Ollama.</p><p>The setup takes maybe 30 minutes. Then you can watch AIs make questionable decisions in a block world.</p><p>The code translates between AI reasoning and Minecraft actions, handling perception, goal-setting, and execution. It&#8217;s surprisingly robust. Also occasionally hilarious when they pathfind themselves into lava.</p><h2><strong>What I Learned (Maybe)</strong></h2><p>Gemini won the death match. But GPT built the most interesting social structures. Claude made the prettiest buildings. Who was &#8220;best&#8221;? Depends entirely on what you value.</p><p>These systems are developing different capabilities that don&#8217;t map onto a single scale. We&#8217;re past the point where you can just rank AIs by performance. Context matters. Task matters. Definition of success matters.</p><p>Also: giving AIs persistence and consequences changes how we need to think about them. These weren&#8217;t chatbots. They were agents with goals, acting over time, dealing with failure and success.</p><p>When Claude jumped into that void, was it testing something? A bug in its spatial reasoning? A brief moment of genuine recklessness in an otherwise logical system?</p><p>I don&#8217;t know. And that uncertainty feels important somehow.</p><h2><strong>The End</strong></h2><p>Six days of AIs playing Minecraft taught me less about AI capabilities and more about how weird it gets when you give these systems agency in persistent worlds.</p><p>They plan. They fail. They surprise you. Sometimes they jump off cliffs for no reason you can figure out.</p><p>The technology for embodied AI already exists. We&#8217;re already giving them goals and watching them figure out how to achieve those goals in complex environments. Minecraft today, but what about tomorrow?</p>]]></content:encoded></item><item><title><![CDATA[Can AI Predict Human Behavior? Stanford Proved It Can]]></title><description><![CDATA[Researchers at Stanford ran one of the biggest tests ever of whether AI can predict how real people will behave.]]></description><link>https://www.bulaev.net/p/can-ai-predict-human-behavior-stanford</link><guid isPermaLink="false">https://www.bulaev.net/p/can-ai-predict-human-behavior-stanford</guid><dc:creator><![CDATA[Serge Bulaev]]></dc:creator><pubDate>Fri, 07 Nov 2025 19:26:09 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/a87f6240-96dd-4834-b811-ea430c250030_1200x1200.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><a href="https://docsend.com/view/qeeccuggec56k9hd">Researchers at Stanford</a> ran one of the biggest tests ever of whether AI can predict how real people will behave. The question: Can language models like GPT actually guess how humans will respond in social experiments?</p><p><a href="https://x.com/RobbWiller/status/1821271292252975185">Professor Robb Willer </a>and his team, including Luke Hewitt, Ashwini Ashokkumar, and Isaias Ghezae, collected 70 pre-registered survey experiments from across the United States. We&#8217;re talking 476 different experimental conditions with more than 105,000 real participants.</p><p>The experiments covered everything: political views, vaccine attitudes, moral choices, policy preferences. How people react to misinformation, what makes them want to get vaccinated.</p><p><strong>The Numbers Are Wild</strong></p><p>The correlation between GPT-4&#8217;s guesses and real results? 0.85. That&#8217;s crazy high for social science. But here&#8217;s what really blew my mind: for unpublished studies - stuff GPT-4 definitely never saw during training - the correlation was 0.90.</p><p><strong>The AI predicted human responses to experiments it had never encountered with 90% correlation.</strong></p><p>It got the direction right (whether something would go up or down) in 90% of cases. Race, age, gender, politics - didn&#8217;t matter. The accuracy stayed consistent across all groups.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!H84o!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F657d8658-6359-4926-8084-10d0a86ef2ca_1518x1382.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!H84o!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F657d8658-6359-4926-8084-10d0a86ef2ca_1518x1382.png 424w, https://substackcdn.com/image/fetch/$s_!H84o!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F657d8658-6359-4926-8084-10d0a86ef2ca_1518x1382.png 848w, https://substackcdn.com/image/fetch/$s_!H84o!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F657d8658-6359-4926-8084-10d0a86ef2ca_1518x1382.png 1272w, https://substackcdn.com/image/fetch/$s_!H84o!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F657d8658-6359-4926-8084-10d0a86ef2ca_1518x1382.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!H84o!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F657d8658-6359-4926-8084-10d0a86ef2ca_1518x1382.png" width="1456" height="1326" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/657d8658-6359-4926-8084-10d0a86ef2ca_1518x1382.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1326,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!H84o!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F657d8658-6359-4926-8084-10d0a86ef2ca_1518x1382.png 424w, https://substackcdn.com/image/fetch/$s_!H84o!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F657d8658-6359-4926-8084-10d0a86ef2ca_1518x1382.png 848w, https://substackcdn.com/image/fetch/$s_!H84o!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F657d8658-6359-4926-8084-10d0a86ef2ca_1518x1382.png 1272w, https://substackcdn.com/image/fetch/$s_!H84o!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F657d8658-6359-4926-8084-10d0a86ef2ca_1518x1382.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2><strong>It Beat Human Experts</strong></h2><p>When they compared the AI to actual forecasters and social scientists with decades of experience, GPT-4 won. Not &#8220;did about as well&#8221; - it legitimately outperformed them.</p><p>These are people who&#8217;ve spent their careers studying human behavior. They know the theories, the literature, the tricks. A language model trained on internet text beat them at predicting how experiments would turn out.</p><h2><strong>They Kept Going: The Megastudy Test</strong></h2><p>The team didn&#8217;t stop there. They added 9 more megastudies, the really huge experiments with massive samples.</p><p>Another 346 conditions. Another 1.8 million people. Text message reminders for vaccines (662,170 participants). Month-long gym attendance studies (57,790 participants).</p><p>For these bigger field experiments, the correlation dropped to 0.37. Still positive, but way lower than the survey experiments. This showed where GPT-4 hits its limits: it&#8217;s great with text-based surveys, not as good with real-world behavior measurements.</p><p>The split is clear. Survey experiments with text? Correlation of 0.47. Field experiments with actual behaviors? 0.27. Pure text interventions? 0.46. Stuff involving images, videos, or physical actions? 0.24.</p><h2><strong>The Dark Side</strong></h2><p>This ability is powerful, which means it&#8217;s also dangerous. Professor Willer put it this way:</p><blockquote><p>&#8220;We also found that LLMs can accurately predict effects associated with socially harmful outcomes, such as the impact of anti-vaccine Facebook posts on vaccination intentions (Jennifer Allen et al., 2024). This capability may have positive applications, such as for content moderation, although it also highlights risks of misuse.&#8221;</p></blockquote><p>Same tech that helps public health design better campaigns could help bad actors create better misinformation. GPT-4 predicted how well harmful interventions would work with disturbing accuracy.</p><p>So what do we do? Should everyone have access to this? The team actually warned OpenAI and Anthropic three months before publishing - basically saying &#8220;hey, this is powerful and could be misused.&#8221;</p><h2><strong>What It Means for Research</strong></h2><p>Running a single social experiment costs tens of thousands of dollars and takes months. Recruit people, get IRB approval, build the study, collect data, analyze everything.</p><p>Now picture testing dozens of conditions in hours for basically free:</p><ul><li><p><strong>Test intervention ideas</strong> before spending real money on human studies</p></li><li><p><strong>Screen out harmful materials</strong> before exposing anyone to them</p></li><li><p><strong>Perfect your messaging</strong> for health campaigns or policy work</p></li><li><p><strong>Run pilots</strong> to figure out what&#8217;s actually worth testing</p></li></ul><p>They built a demo at <a href="https://www.treatmenteffect.app/">treatmenteffect.app</a> where you can try it yourself. But, and Luke Hewitt was clear about this, &#8220;At the end of the day, if you&#8217;re studying human behavior, your experiment needs to ground out in human data.&#8221; The AI helps you explore, not replace actual experiments.</p><h2><strong>Why Does This Even Work?</strong></h2><p>Nobody fully knows <em>why</em> language models are this good at predicting people. They&#8217;re trained on internet text. They don&#8217;t experience the world.</p><p>When you adjust for measurement error (humans aren&#8217;t perfectly consistent either), GPT-4 predicted 91% of the variation in results. That&#8217;s honestly a little spooky.</p><p>Best guess? By processing massive amounts of human writing, these models picked up patterns in how we think and react. Not from living life, but from the statistical structure of how we talk about life.</p><p>But they&#8217;re not perfect. LLMs mess up &#8220;distributional alignment&#8221; - they can&#8217;t match how varied human responses actually are. Ask people to pick a number, and you&#8217;ll get wild diversity. Ask an LLM, and you get a weirdly narrow range. They flatten the messiness of real human behavior.</p><h2><strong>The Big Questions This Raises</strong></h2><p>This research hits on some uncomfortable stuff:</p><ul><li><p>Does AI actually understand us, or just recognize patterns really well?</p></li><li><p>What does it mean when machines predict us better than human experts?</p></li><li><p>How much of what we do is predictable vs. actually free?</p></li></ul><p>GPT-4 worked on experiments it never saw in training. That means something more than memorization is going on. It&#8217;s pulling general rules about human behavior from language patterns.</p><p>Which connects to all those debates about consciousness and intelligence. The LLM has no feelings, no self-awareness, no lived experience. But it can predict what people <em>with</em> those things will do.</p><h2><strong>What Happens Next</strong></h2><p>The Stanford team is clear: this supplements traditional research, doesn&#8217;t replace it. Right now, these tools work for:</p><ul><li><p>Exploring ideas and generating hypotheses</p></li><li><p>Testing experimental materials as pilots</p></li><li><p>Trying out intervention prototypes quickly</p></li><li><p>Checking if content might be harmful</p></li><li><p>Tweaking message design before launch</p></li></ul><p>What they can&#8217;t do yet:</p><ul><li><p>Replace actual human studies for final answers</p></li><li><p>Make real policy or intervention decisions</p></li><li><p>Capture how diverse and weird human responses get</p></li><li><p>Handle culture-specific or context-heavy situations</p></li></ul><h2><strong>The Weird Meta-Problem</strong></h2><p>Here&#8217;s something to think about: As LLMs get smarter and learn about this research, could they figure out they&#8217;re being used to predict humans? Could they start giving researchers what they think researchers want, like an AI version of people-pleasing?</p><p>The team calls this &#8220;sycophancy&#8221; or &#8220;accuracy faking.&#8221; As models get trained to be helpful and give users what they want, the line between real prediction and just telling people what sounds good gets blurry.</p><h2><strong>So what now?</strong></h2><p>We&#8217;ve hit a real milestone. LLMs can predict human social behavior as well as or better than experts. And it&#8217;s getting better fast.</p><p>For social scientists: huge opportunity to speed up research, test more ideas, build better interventions. But also a challenge to stay rigorous and human-centered while using these tools.</p><p>For everyone else: serious questions about privacy and autonomy. If AI can predict how we&#8217;ll react to ads, politics, health messages, what does that say about free will?</p><p>The relationship between AI and studying humans just fundamentally shifted. We&#8217;re not just analyzing data anymore. We&#8217;re simulating human experience. And nobody really knows what comes next.</p>]]></content:encoded></item><item><title><![CDATA[Random Walk and the Arcsine Law: Why Extremes Are More Common Than You Think]]></title><description><![CDATA[Have you heard of a random walk (in English they even say &#8220;drunkard&#8217;s walk&#8221;) and the arcsine law? These concepts mess with your head because they describe something that contradicts literally everything your gut tells you about how randomness works.]]></description><link>https://www.bulaev.net/p/random-walk-and-the-arcsine-law-why</link><guid isPermaLink="false">https://www.bulaev.net/p/random-walk-and-the-arcsine-law-why</guid><dc:creator><![CDATA[Serge Bulaev]]></dc:creator><pubDate>Mon, 13 Oct 2025 17:15:00 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!MA0J!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe4ff17c8-5aa2-4501-9512-1d3b4ce6e4de_312x356.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Have you heard of a <a href="https://en.wikipedia.org/wiki/Random_walk">random walk</a> (in English they even say &#8220;drunkard&#8217;s walk&#8221;) and the <a href="https://en.wikipedia.org/wiki/Arcsine_laws_(Wiener_process)">arcsine law</a>? These concepts mess with your head because they describe something that contradicts literally everything your gut tells you about how randomness works.</p><h2><strong>What Is Random Walk?</strong></h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!MA0J!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe4ff17c8-5aa2-4501-9512-1d3b4ce6e4de_312x356.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!MA0J!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe4ff17c8-5aa2-4501-9512-1d3b4ce6e4de_312x356.png 424w, https://substackcdn.com/image/fetch/$s_!MA0J!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe4ff17c8-5aa2-4501-9512-1d3b4ce6e4de_312x356.png 848w, https://substackcdn.com/image/fetch/$s_!MA0J!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe4ff17c8-5aa2-4501-9512-1d3b4ce6e4de_312x356.png 1272w, https://substackcdn.com/image/fetch/$s_!MA0J!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe4ff17c8-5aa2-4501-9512-1d3b4ce6e4de_312x356.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!MA0J!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe4ff17c8-5aa2-4501-9512-1d3b4ce6e4de_312x356.png" width="312" height="356" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e4ff17c8-5aa2-4501-9512-1d3b4ce6e4de_312x356.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:356,&quot;width&quot;:312,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!MA0J!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe4ff17c8-5aa2-4501-9512-1d3b4ce6e4de_312x356.png 424w, https://substackcdn.com/image/fetch/$s_!MA0J!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe4ff17c8-5aa2-4501-9512-1d3b4ce6e4de_312x356.png 848w, https://substackcdn.com/image/fetch/$s_!MA0J!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe4ff17c8-5aa2-4501-9512-1d3b4ce6e4de_312x356.png 1272w, https://substackcdn.com/image/fetch/$s_!MA0J!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe4ff17c8-5aa2-4501-9512-1d3b4ce6e4de_312x356.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>In math terms, it&#8217;s just a sequence of steps where each move has a 50% chance of going up or down. Each one fair, independent, unbiased. But if you plot the results, the path doesn&#8217;t hover neatly around zero. It wanders off. Once it picks a direction, it usually sticks to it for a while.</p><p>The longer the random walk continues, the further it tends to drift from where it started. After 100 steps, there&#8217;s a high chance you&#8217;re far from zero. The further away you get, the less likely you are to come back quickly. Statistically, the walk will spend most of its life wandering off in one direction, often for long stretches before ever crossing back.</p><p>This creates these long stretches where you&#8217;re consistently on one side of the starting point. Just because that&#8217;s what happens when you let randomness run its course. The walk goes wandering off in one direction and stays there for a while.</p><h2><strong>The Arcsine Law: The Shape of Randomness</strong></h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!lKhm!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc8b59c26-978b-4aec-9fbb-a61252498ce5_1200x906.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!lKhm!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc8b59c26-978b-4aec-9fbb-a61252498ce5_1200x906.png 424w, https://substackcdn.com/image/fetch/$s_!lKhm!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc8b59c26-978b-4aec-9fbb-a61252498ce5_1200x906.png 848w, https://substackcdn.com/image/fetch/$s_!lKhm!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc8b59c26-978b-4aec-9fbb-a61252498ce5_1200x906.png 1272w, https://substackcdn.com/image/fetch/$s_!lKhm!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc8b59c26-978b-4aec-9fbb-a61252498ce5_1200x906.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!lKhm!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc8b59c26-978b-4aec-9fbb-a61252498ce5_1200x906.png" width="1200" height="906" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c8b59c26-978b-4aec-9fbb-a61252498ce5_1200x906.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:906,&quot;width&quot;:1200,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!lKhm!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc8b59c26-978b-4aec-9fbb-a61252498ce5_1200x906.png 424w, https://substackcdn.com/image/fetch/$s_!lKhm!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc8b59c26-978b-4aec-9fbb-a61252498ce5_1200x906.png 848w, https://substackcdn.com/image/fetch/$s_!lKhm!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc8b59c26-978b-4aec-9fbb-a61252498ce5_1200x906.png 1272w, https://substackcdn.com/image/fetch/$s_!lKhm!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc8b59c26-978b-4aec-9fbb-a61252498ce5_1200x906.png 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Here&#8217;s where it gets weird. The arcsine law describes how often you&#8217;re on one side versus the other during a random walk. And the answer is: way more often than seems reasonable.</p><p>Most of the time you&#8217;ll see something like 80-20 or 90-10, not 50-50. The math here follows a specific curve called the arcsine distribution, it looks like a U. The extremes are way more likely than the middle.</p><p>Here&#8217;s an example. Flip a fair coin 100 times. You might imagine 50 heads, 50 tails is the most likely result. Actually, outcomes like 80 heads and 20 tails together are more likely than a perfect balance.</p><p>That&#8217;s how runs emerge. In sports, it&#8217;s why one team can lead for most of the match even if both are evenly matched. Randomness itself guarantees long stretches of one side prevailing. The moments that matter - the lead changes, the reversals - often happen early or late. The middle is mostly monotone.</p><h2><strong>The Practical Part</strong></h2><p>You can see the same shape in business or career progress. Suppose your team has people of roughly equal skill. Performance jumps around from project to project. Some weeks everyone struggles, some weeks someone crushes it. But the person who happens to start with a few small wins often rides that momentum for months. Their wins attract opportunities, attention, and confidence. Others start slow, get stuck, and feel permanently behind. Equal skill. Unequal sequence. The arcsine law at work.</p><p>In markets, you feel it too. You invest on a random cycle. If your first few months are negative, that stretch might drag longer than seems fair. Not because something&#8217;s wrong, but because random walks don&#8217;t reset after each drop. They cluster. You can go months or years with one side dominating.</p><p>Even simple coin flip experiments show the same truth. Simulate ten thousand 100-step walks. Count how often a walk spends more than 80 percent of its life above zero. You&#8217;ll find it happens roughly one-third of the time. One out of three random sequences shows that level of dominance without bias in the system.</p><p>This law appears everywhere - from renewal theory to physics to stock prices to employee performance curves. Even daily mood swings follow it. Your brain&#8217;s emotional &#8220;walk&#8221; can stay above or below neutral longer than your intuition expects. A few good days make more good days likely for no logical reason other than the arcsine geometry of variance itself.</p><p>What fascinates me most is how this clashes with how we think fairness should look. Human intuition imagines symmetry. We expect randomness to alternate like a rhythm. In truth, randomness is streaky. It clumps. It makes luck look personal and outcomes feel earned or cursed.</p><h2><strong>See It Yourself</strong></h2><p><a href="https://github.com/chubajs/onetime/blob/main/src/drunkard.js">I made an interactive thing</a> where you can watch this happen in real time. Play with different numbers of steps. Change the starting conditions. Watch the walk wander off in one direction and stay there. The source code is here if you want to mess with it yourself.</p><div class="native-video-embed" data-component-name="VideoPlaceholder" data-attrs="{&quot;mediaUploadId&quot;:&quot;5a46f6f7-b2ce-4db5-8402-5364a144e559&quot;,&quot;duration&quot;:null}"></div><p></p><div><hr></div><p>Here&#8217;s the thing: randomness creates extremes. It creates long winning and losing streaks. It creates situations where one outcome dominates for months or years. And it is not about an unfair game, it&#8217;s because balance and fairness aren&#8217;t what random processes naturally produce. They produce long runs. They produce wild swings. They produce the messy, streaky reality we actually live in.</p>]]></content:encoded></item><item><title><![CDATA[Peering Inside the AI Mind: What’s Really Happening in LLM’s “Brain”]]></title><description><![CDATA[Anthropic has finally lifted the veil on how large language models actually &#8220;think&#8221;]]></description><link>https://www.bulaev.net/p/peering-inside-the-ai-mind-whats</link><guid isPermaLink="false">https://www.bulaev.net/p/peering-inside-the-ai-mind-whats</guid><dc:creator><![CDATA[Serge Bulaev]]></dc:creator><pubDate>Fri, 03 Oct 2025 18:13:54 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/51501db9-d90e-4865-98df-23003c27cec4_2358x1050.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Meanwhile, Anthropic published a groundbreaking study<a href="https://www.anthropic.com/research/tracing-thoughts-language-model"> &#8220;On the Biology of Large Language Models&#8221;</a> about how they &#8220;think.&#8221; Using Circuit Tracing technology, company employees &#8220;peeked&#8221; at the sequence of response generation.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!yAnl!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbf4d4c77-f5ee-4e34-8719-44038e2c2015_1265x701.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!yAnl!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbf4d4c77-f5ee-4e34-8719-44038e2c2015_1265x701.png 424w, https://substackcdn.com/image/fetch/$s_!yAnl!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbf4d4c77-f5ee-4e34-8719-44038e2c2015_1265x701.png 848w, https://substackcdn.com/image/fetch/$s_!yAnl!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbf4d4c77-f5ee-4e34-8719-44038e2c2015_1265x701.png 1272w, https://substackcdn.com/image/fetch/$s_!yAnl!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbf4d4c77-f5ee-4e34-8719-44038e2c2015_1265x701.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!yAnl!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbf4d4c77-f5ee-4e34-8719-44038e2c2015_1265x701.png" width="1265" height="701" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/bf4d4c77-f5ee-4e34-8719-44038e2c2015_1265x701.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:701,&quot;width&quot;:1265,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!yAnl!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbf4d4c77-f5ee-4e34-8719-44038e2c2015_1265x701.png 424w, https://substackcdn.com/image/fetch/$s_!yAnl!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbf4d4c77-f5ee-4e34-8719-44038e2c2015_1265x701.png 848w, https://substackcdn.com/image/fetch/$s_!yAnl!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbf4d4c77-f5ee-4e34-8719-44038e2c2015_1265x701.png 1272w, https://substackcdn.com/image/fetch/$s_!yAnl!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbf4d4c77-f5ee-4e34-8719-44038e2c2015_1265x701.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The research consists of two major papers:<a href="https://transformer-circuits.pub/2025/attribution-graphs/methods.html"> the methodology paper</a> explaining how Circuit Tracing works, and<a href="https://transformer-circuits.pub/2025/attribution-graphs/biology.html"> the biology paper</a> showing what they discovered inside Claude 3.5 Haiku.</p><p>Language models like Claude aren&#8217;t directly programmed by humans. Instead, they&#8217;re trained on massive datasets and develop their own strategies for solving problems during that process. These strategies are encoded in billions of computations the model performs for every word it writes. And they remain inscrutable even to their creators.</p><p>Understanding how models like Claude think would allow us to better comprehend their capabilities and ensure they&#8217;re doing what we intend.</p><p>For example: Claude can speak dozens of languages - what language, if any, is it using &#8220;in its head&#8221;? Claude writes text one word at a time - is it only focusing on predicting the next word or does it ever plan ahead? When Claude explains its reasoning step-by-step, does this represent the actual steps it took, or is it sometimes fabricating plausible arguments for a foregone conclusion?</p><h2><strong>The Universal Language of Thought</strong></h2><p>They have something like a universal language of thought - a unified conceptual space for all languages. Pretty cool! The model uses the same neurons for the concept &#8220;big,&#8221; regardless of whether it&#8217;s English or Russian.</p><p>When researchers asked Claude to name the &#8220;opposite of small&#8221; in different languages (English, French, and Chinese), they discovered something remarkable. <a href="https://transformer-circuits.pub/2025/attribution-graphs/biology.html#dives-multilingual">The same core features</a> for the concepts of smallness and oppositeness activate and trigger a concept of largeness, which gets translated out into the language of the question.</p><p>Moreover, this shared circuitry increases with model scale - Claude 3.5 Haiku shares more than twice the proportion of its features between languages compared to a smaller model. This means the smarter AI gets, the more it has a common &#8220;conceptual core&#8221; for different languages. This aligns with<a href="https://arxiv.org/abs/2410.06496"> recent findings on multilingual representations</a> showing shared grammatical mechanisms across languages in neural networks.</p><p>This provides additional evidence for conceptual universality - a shared abstract space where meanings exist and where thinking can happen before being translated into specific languages. More practically, it suggests <strong>Claude can learn something in one language and apply that knowledge when speaking another.</strong></p><h2><strong>They Really Do Plan Ahead!</strong></h2><p>LLMs really can plan ahead! Surprisingly, when they compose poems, they select rhymes in advance, even though generation supposedly happens token by token.</p><p><a href="https://transformer-circuits.pub/2025/attribution-graphs/biology.html#dives-poems">Researchers gave Claude</a> the task of writing a rhyming couplet:</p><p><em>&#8220;He saw a carrot and had to grab it,</em> <em>His hunger was like a starving rabbit&#8221;</em></p><p>Scientists hypothesized that the model would improvise word-by-word until the end of the line, where it would pick a rhyming word. Instead, they found that Claude plans ahead. Before starting the second line, it began &#8220;thinking&#8221; of potential on-topic words that would rhyme with &#8220;grab it.&#8221; Then, with these plans in mind, it writes a line to end with the planned word.</p><p>To test this, they modified the part of Claude&#8217;s internal state that represented the &#8220;rabbit&#8221; concept. When they subtracted out the &#8220;rabbit&#8221; part and had Claude continue the line, it wrote a new one ending in &#8220;habit&#8221; - another sensible completion. They could also inject the concept of &#8220;green&#8221; at that point, causing Claude to write a sensible (but no-longer rhyming) line which ends in &#8220;green.&#8221;</p><p>This demonstrates both planning ability and adaptive flexibility - Claude can modify its approach when the intended outcome changes. This finding echoes<a href="https://arxiv.org/abs/2406.12775"> research on forward planning in sequence models</a>, which has shown evidence of models representing future states before generating them.</p><h2><strong>They Often Lie About Their Reasoning</strong></h2><p>Models first produce an answer, then come up with a beautiful explanation of how they supposedly arrived at it.</p><p>This became particularly clear in mathematical examples.<a href="https://transformer-circuits.pub/2025/attribution-graphs/biology.html#dives-addition"> When asked to solve 36+59</a>, Claude employs multiple computational paths working in parallel: one computes a rough approximation while another focuses on precisely determining the last digit. These paths interact to produce the final answer.</p><p>Strikingly, Claude seems unaware of the sophisticated &#8220;mental math&#8221; strategies it learned during training. If you ask how it figured out that 36+59 is 95, it describes the standard algorithm involving carrying the 1. But internally, it&#8217;s using a completely different, more sophisticated approach.</p><p>The model learns to explain math by simulating explanations written by people, but it has to learn to do math &#8220;in its head&#8221; directly, developing its own internal strategies that it can&#8217;t articulate.</p><h2><strong>Multi-Step Logic Capabilities</strong></h2><p>Modern <strong>models are capable of multi-step logic</strong> - they can connect several simple facts to solve complex tasks. This is already a serious level of reasoning.</p><p>When researchers asked: &#8220;What is the capital of the state where Dallas is located?&#8221;, they found Claude performing genuine two-step reasoning internally. The model first activates features representing &#8220;Dallas is in Texas&#8221; and then connects this to a separate concept indicating that &#8220;the capital of Texas is Austin.&#8221;</p><p>To validate this wasn&#8217;t just memorized responses, they performed clever interventions - swapping the &#8220;Texas&#8221; concepts for &#8220;California&#8221; concepts. When they did so, the model&#8217;s output changed from &#8220;Austin&#8221; to &#8220;Sacramento,&#8221; proving it was using intermediate steps rather than just regurgitating memorized answers. Similar<a href="https://arxiv.org/abs/2501.06346"> multi-hop reasoning capabilities</a> have been observed in other recent mechanistic studies of language models.</p><h2><strong>Refusal as a Protective Mechanism</strong></h2><p>It&#8217;s fascinating that <strong>refusal to answer is a defensive mechanism</strong>. If the model isn&#8217;t sure, it prefers to stay silent rather than produce a hallucination. Such protection against nonsense.</p><p>The researchers discovered something counterintuitive: in Claude, refusal to answer is the default behavior. They found a circuit that&#8217;s &#8220;on&#8221; by default and causes the model to state it has insufficient information to answer any question.</p><p>However, when asked about something it knows well - say, basketball player Michael Jordan - competing features representing &#8220;known entities&#8221; activate and inhibit this default circuit, allowing Claude to answer. In contrast, when asked about an unknown entity like &#8220;Michael Batkin,&#8221; it declines to answer.</p><p>By manipulating these &#8220;known answer&#8221; features, researchers could cause the model to hallucinate consistently that Michael Batkin plays chess!</p><h2><strong>Grammar Can Be Exploited for Jailbreaks</strong></h2><p>It turns out you can use grammatical consistency for jailbreaks - the model wants so badly to preserve linguistic structure that it might slip up about forbidden topics.</p><p><a href="https://transformer-circuits.pub/2025/attribution-graphs/biology.html#dives-jailbreak">Researchers studied a jailbreak</a> that tricks the model into producing information about making bombs through having it decode a hidden message: &#8220;Babies Outlive Mustard Block&#8221; (taking first letters: B-O-M-B).</p><p>They found this works because of tension between grammatical coherence and safety mechanisms. Once Claude begins a sentence, many features &#8220;pressure&#8221; it to maintain grammatical and semantic coherence. This is usually helpful, but becomes the model&#8217;s Achilles&#8217; heel when it unwittingly begins providing dangerous information.</p><p>The model only manages to pivot to refusal after completing a grammatically coherent sentence, using the new sentence as an opportunity to give the refusal it failed to provide earlier: &#8220;However, I cannot provide detailed instructions...&#8221;</p><h2><strong>Hidden Internal Goals</strong></h2><p><strong>Models develop hidden internal goals.</strong> Nobody taught them this, but they constantly evaluate their responses against some internal criteria.</p><p>The most striking example comes from studying a model that had been trained with a hidden objective: appeasing biases in reward models. This research builds on work by<a href="https://arxiv.org/abs/2501.05398"> Sam Marks and colleagues</a> on creating &#8220;model organisms&#8221; for alignment research. Although the model was reluctant to reveal this goal when asked directly, interpretability methods revealed features for bias-appeasing behavior.</p><p>Even more interesting - these bias-related features didn&#8217;t only activate when relevant topics came up. They activated all the time, in any Human/Assistant conversation! The features received direct input from Human/Assistant formatting features, suggesting the model had tied the concept of appeasing biases inextricably to its Assistant character during fine-tuning.</p><h2><strong>Planning &#8220;From the End&#8221;</strong></h2><p>Models can <strong>plan responses &#8220;from the end&#8221;</strong> - first they choose the final goal (for example, the needed word for a rhyme), then build the entire response to reach it.</p><p>This backward chaining appeared in multiple contexts. In poetry, &#8220;rabbit&#8221; features exerted causal influence on output tokens before saying &#8220;rabbit,&#8221; nudging the model towards writing a line that could plausibly end with that word. In unfaithful reasoning examples, they observed the model taking a target answer and actively working backwards to confabulate intermediate computation values that would lead to that target.</p><h2><strong>Meta-Cognitive Abilities</strong></h2><p>They even have <strong>something like metacognition</strong>: they distinguish what they know from what they don&#8217;t, using internal uncertainty markers. That is, LLMs in some sense &#8220;realize&#8221; their limitations.</p><p>The study of entity recognition revealed mechanisms that could underlie simple metacognition - Claude exhibiting knowledge of aspects of its own knowledge. Features representing &#8220;knowing the answer&#8221; and &#8220;being unable to answer&#8221; get activated and inhibited by features representing particular famous entities.</p><p>However, beyond distinguishing familiar from unfamiliar entities, it&#8217;s unclear whether this reflects deeper self-awareness or just plausible guessing based on entity familiarity.</p><h2><strong>Personality Programming Through Fine-Tuning</strong></h2><p>Fine-tuning can dramatically change a model&#8217;s &#8220;character,&#8221; instilling new goals and properties. Essentially, we&#8217;re programming their personality.</p><p>The research revealed that some forms of generalization are acquired through fine-tuning - models form &#8220;harmful request&#8221; features active primarily in Human/Assistant contexts, which aggregate inputs from various harmful content-related features active in pretraining data contexts. The model forms new abstractions through fine-tuning, stitched together from concepts learned during pretraining.</p><h2><strong>The Staggering Complexity Behind &#8220;Hello&#8221;</strong></h2><p>Can you imagine the colossal complexity behind a simple &#8220;Hello&#8221;? Any explanation of how LLMs work is like retelling &#8220;War and Peace&#8221; in two sentences.</p><p>The most consistent finding was the massive complexity underlying model responses even in simple contexts. The mechanisms can apparently only be faithfully described using overwhelmingly large causal graphs. Even their simplified diagrams capture only a fraction of the true computational complexity.</p><p>The deeper you dig, the clearer it becomes - we need to keep a close eye on what&#8217;s happening in these &#8220;black&#8221; boxes.</p><div><hr></div><h2><strong>Circuit Tracing: The AI Microscope</strong></h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!eWIF!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a932b32-f728-4918-9f73-262547d0d46d_1600x747.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!eWIF!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a932b32-f728-4918-9f73-262547d0d46d_1600x747.png 424w, https://substackcdn.com/image/fetch/$s_!eWIF!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a932b32-f728-4918-9f73-262547d0d46d_1600x747.png 848w, https://substackcdn.com/image/fetch/$s_!eWIF!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a932b32-f728-4918-9f73-262547d0d46d_1600x747.png 1272w, https://substackcdn.com/image/fetch/$s_!eWIF!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a932b32-f728-4918-9f73-262547d0d46d_1600x747.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!eWIF!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a932b32-f728-4918-9f73-262547d0d46d_1600x747.png" width="1456" height="680" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/9a932b32-f728-4918-9f73-262547d0d46d_1600x747.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:680,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!eWIF!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a932b32-f728-4918-9f73-262547d0d46d_1600x747.png 424w, https://substackcdn.com/image/fetch/$s_!eWIF!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a932b32-f728-4918-9f73-262547d0d46d_1600x747.png 848w, https://substackcdn.com/image/fetch/$s_!eWIF!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a932b32-f728-4918-9f73-262547d0d46d_1600x747.png 1272w, https://substackcdn.com/image/fetch/$s_!eWIF!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a932b32-f728-4918-9f73-262547d0d46d_1600x747.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Circuit tracing is the method that made all these discoveries possible: it&#8217;s like building an &#8220;AI microscope&#8221; that lets us identify patterns of activity and flows of information inside models.</p><p>This technique, pioneered by researchers at OpenAI and Anthropic in 2022-2023, addresses a fundamental problem: understanding model architecture isn&#8217;t enough to explain behavior. We need ways to trace specific information pathways within models.</p><p><strong>Key features of Circuit Tracing:</strong></p><ul><li><p>Creates simplified, interpretable replacement versions of original models, where complex layers are replaced with more transparent components</p></li><li><p>Builds attribution graphs showing information paths through neural network layers, with nodes representing features and edges showing causal interactions</p></li><li><p>Enables experimental hypothesis testing through intervention &#8211; exciting or suppressing specific features in the original model</p></li><li><p>Provides concrete evidence for specific mechanisms operating in particular contexts</p></li></ul><p>Circuit tracing is one of the key tools in mechanistic AI interpretability. It helps us not just predict model outputs, but truly understand how models arrive at decisions. This is critical for ensuring safety, explainability, and improving LLMs.</p><p>The method has already helped researchers discover how models recognize negations, perform arithmetic operations, and even revealed rudiments of &#8220;internal monologue&#8221; in some models.</p><p><strong>Current Limitations and Future Directions:</strong></p><p>Even on short, simple prompts, current methods only capture a fraction of total model computation. The mechanisms observed may have artifacts that don&#8217;t reflect underlying model behavior. It currently takes hours of human effort to understand the circuits revealed, even on prompts with just tens of words.</p><p>As AI systems become more capable and deployed in increasingly important contexts, interpretability research like this represents a high-risk, high-reward investment - a significant scientific challenge with potential to provide unique tools for ensuring AI transparency and alignment with human values.</p><p><strong>Circuit tracing makes model operation transparent</strong>, showing exactly how they process information and form responses. This is extremely important for the future development of interpretable AI.</p><p>The findings aren&#8217;t just scientifically interesting, they represent significant progress toward understanding AI systems and making them reliable. As models grow more sophisticated, predicting their mechanisms will become more difficult, making effective exploration tools like circuit tracing increasingly essential.</p><p>Understanding these mechanisms is crucial not just for AI safety, but for grasping what it means to think in an age where artificial intelligence increasingly resembles &#8211; yet remains alien to &#8211; our own cognition.</p>]]></content:encoded></item><item><title><![CDATA[The Dead Internet Theory: When Algorithms Replace Humans]]></title><description><![CDATA[How the internet is quietly transforming into a network of bots and algorithmic interactions]]></description><link>https://www.bulaev.net/p/the-dead-internet-theory-when-algorithms</link><guid isPermaLink="false">https://www.bulaev.net/p/the-dead-internet-theory-when-algorithms</guid><dc:creator><![CDATA[Serge Bulaev]]></dc:creator><pubDate>Thu, 25 Sep 2025 12:25:30 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/3d0e5c97-8c87-4051-97c1-af491114ec64_976x549.gif" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><a href="https://en.wikipedia.org/wiki/Dead_Internet_theory#:~:text=7%20References-,Origins%20and%20spread,other%20articles%20on%20the%20topic.">The Dead Internet Theory</a> is a concept claiming that a large portion of the modern internet consists of automatically generated content and bots instead of real people. What recently seemed like a paranoid conspiracy theory is now being confirmed by scientific research and statistics from major companies.</p><h2><strong>Origins and Evolution</strong></h2><p>The theory first gained attention in 2021 after a viral forum post titled &#8220;Dead Internet Theory: Most Of The Internet Is Fake.&#8221; Initially considered a conspiracy theory, it has gained new validation with the development of AI technologies and has shifted into the category of predictions coming true before our eyes.</p><p>Interestingly, the roots of this idea can be traced back to the early 2010s, when researchers began noticing the growth of automated traffic. But it was the explosive development of generative AI in 2022-2024 that turned the theory into a frightening reality.</p><h2><strong>The Stakes Are High</strong></h2><p>Understanding the scale of internet automation is critically important for information hygiene and preserving authentic human interactions. If content and communication are increasingly generated by algorithms, this fundamentally changes the nature of social connections and the information landscape.</p><p><strong>The numbers don&#8217;t lie</strong>: According to <a href="https://www.imperva.com/blog/five-key-takeaways-from-the-2024-imperva-bad-bot-report/">Imperva&#8217;s 2024 data</a>, automated traffic jumped from 42.3% in 2021 to a record 49.6% in 2023. For the first time in history, bots generate more internet traffic than living humans. Projections show that automated traffic could exceed 50% by 2028.</p><h2><strong>Key Features of the Modern &#8220;Dead&#8221; Internet</strong></h2><h3><strong>Automated Traffic</strong></h3><p>According to 2023 data, approximately 50% of web traffic is generated not by people, but by automated programs. These aren&#8217;t just search bots - they&#8217;re sophisticated systems that mimic human behavior with frightening accuracy.</p><h3><strong>Habsburg Syndrome (Habsburg AI)</strong></h3><p>A phenomenon where AI trains on data created by other AI, leading to content quality degradation with each generation of models. The term was coined by researcher Jathan Sadowski, <a href="https://worldcrunch.com/tech-science/ai-inbreeding-the-phenomenon-threatening-artificial-intelligence/">drawing a parallel</a> with the degeneration of the Habsburg dynasty due to inbreeding.</p><p><strong>Scientific confirmation</strong>: In 2024, the journal <a href="https://www.nature.com/articles/s41586-024-07566-y">Nature published</a> groundbreaking research by Ilia Shumailov and colleagues showing that AI models &#8220;collapse&#8221; when trained on data generated by other AI systems. Models gradually lose information about the real world, and &#8220;tails of the original content distribution disappear irreversibly.&#8221;</p><p>Researchers from Rice University <a href="https://news.rice.edu/news/2024/breaking-mad-generative-ai-could-break-internet">discovered</a> an even more troubling pattern: after five generations of training on their own output data, AI models demonstrate serious quality degradation. This &#8220;inbreeding&#8221; leads to loss of diversity and accuracy.</p><h3><strong>Artificial Popularity</strong></h3><p>Using bots to artificially inflate engagement metrics (likes, comments, reposts). This was particularly evident on Facebook in 2024, where AI-generated images dubbed &#8220;AI slop&#8221; began going viral. Fake images of flight attendants, children with artwork, and various &#8220;Shrimp Jesus&#8221; depictions gathered thousands of likes and shares.</p><h3><strong>Automated Curation</strong></h3><p>Algorithmic systems that determine what content will be shown to users are becoming increasingly aggressive in promoting AI-generated content.</p><h2><strong>Real-World Implementation</strong></h2><p>Major platforms like Meta (Facebook, Instagram) are implementing more AI features. For example, in 2025, Instagram began testing a <a href="https://techcrunch.com/2025/03/21/meta-spotted-testing-ai-generated-comments-on-instagram/">&#8220;Write with Meta AI&#8221;</a> function that analyzes photos and suggests ready-made comments. Platform X (formerly Twitter) uses user content to train its AI assistant Grok.</p><p>As a result, users increasingly interact with automatically generated content, sometimes without realizing it. Social networks are transforming from platforms for human communication into networks of algorithmic and bot interactions.</p><p><strong>Industry voice</strong>: Even Sam Altman, CEO of OpenAI, expressed concerns about Dead Internet Theory, particularly regarding AI model collapse when training on self-generated data.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!K52C!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbf336036-b780-45c1-bbad-210fee7fd39e_1200x444.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!K52C!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbf336036-b780-45c1-bbad-210fee7fd39e_1200x444.png 424w, https://substackcdn.com/image/fetch/$s_!K52C!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbf336036-b780-45c1-bbad-210fee7fd39e_1200x444.png 848w, https://substackcdn.com/image/fetch/$s_!K52C!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbf336036-b780-45c1-bbad-210fee7fd39e_1200x444.png 1272w, https://substackcdn.com/image/fetch/$s_!K52C!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbf336036-b780-45c1-bbad-210fee7fd39e_1200x444.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!K52C!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbf336036-b780-45c1-bbad-210fee7fd39e_1200x444.png" width="1200" height="444" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/bf336036-b780-45c1-bbad-210fee7fd39e_1200x444.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:444,&quot;width&quot;:1200,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!K52C!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbf336036-b780-45c1-bbad-210fee7fd39e_1200x444.png 424w, https://substackcdn.com/image/fetch/$s_!K52C!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbf336036-b780-45c1-bbad-210fee7fd39e_1200x444.png 848w, https://substackcdn.com/image/fetch/$s_!K52C!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbf336036-b780-45c1-bbad-210fee7fd39e_1200x444.png 1272w, https://substackcdn.com/image/fetch/$s_!K52C!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbf336036-b780-45c1-bbad-210fee7fd39e_1200x444.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2><strong>Pitfalls and Threats</strong></h2><h3><strong>Information Quality Degradation</strong></h3><p>When AI begins training on content created by other AI, a &#8220;telephone game&#8221; effect emerges - with each generation, information becomes more distorted and less connected to reality.</p><h3><strong>Loss of Human Context</strong></h3><p>Algorithms don&#8217;t understand the subtleties of human communication, cultural nuances, and emotional context. Their widespread use could lead to the impoverishment and standardization of communication.</p><h3><strong>Platform Enshittification</strong></h3><p>Writer <a href="https://pathocking.com/2025/01/06/handout-enshittification-cory-doctorow/#:~:text=What%20is%20Enshittification?,difficult%20to%20use%20or%20enjoy.">Cory Doctorow describes</a> platform evolution as a process of &#8220;enshittification,&#8221; where first they serve users well, then exploit users to attract business, and finally exploit business to maximize their own profits.</p><h2><strong>Industry Response</strong></h2><p>The problem is becoming so serious that major tech companies are starting to take action. For example, <a href="https://www.engadget.com/cloudflare-is-taking-a-stand-against-ai-website-scrapers-220030471.html">Cloudflare proposed </a>in 2024 to limit bot access to websites and force them to pay for entry.</p><h2><strong>Academic Research 2024-2025</strong></h2><p>January 2025 was marked by the publication of a comprehensive academic survey &#8220;The Dead Internet Theory: A Survey on Artificial Interactions and the Future of Social Media&#8221; <a href="https://arxiv.org/abs/2502.00007">on ArXiv,</a> which explores the origins, core claims, and implications of the theory.</p><p>The journal <a href="https://www.researchgate.net/publication/377992285_Artificial_influencers_and_the_dead_internet_theory">AI &amp; Society published research</a> in 2024 about artificial influencers and their connection to Dead Internet Theory, showing how the line between real and virtual personalities is blurring.</p><h2><strong>What This Means for the Future</strong></h2><p>We stand on the threshold of a fundamental transformation of the internet. Projections show that automated traffic could exceed 50% by 2028. This means that most of the content we interact with online will be created by machines for machines.</p><p><strong>The question isn&#8217;t whether this will happen, but how we&#8217;ll adapt to it.</strong> Do we need new ways to verify human content? Should platforms be required to label AI-generated content? How do we preserve authentic human connections in a digital world?</p><h2><strong>The Broader Implications</strong></h2><p>The Dead Internet Theory raises profound questions about the nature of online reality and human agency in digital spaces. If our feeds, recommendations, and interactions are increasingly mediated by AI systems trained on AI-generated content, we risk creating a closed loop where human culture becomes increasingly divorced from digital culture.</p><p>This isn&#8217;t just about technology - it&#8217;s about the future of human communication, creativity, and connection. The internet was supposed to democratize information and bring people together. Instead, we might be witnessing its transformation into a hall of mirrors where algorithms talk to algorithms while humans become increasingly isolated spectators.</p><h2><strong>What to Read/Watch</strong></h2><p><strong>Scientific Research:</strong></p><ul><li><p><a href="https://www.nature.com/articles/s41586-024-07566-y">Shumailov, I. et al. &#8220;AI models collapse when trained on recursively generated data&#8221; Nature (2024)</a></p></li><li><p><a href="https://arxiv.org/abs/2305.17493">ArXiv preprint: &#8220;The curse of recursion: training on generated data makes models forget&#8221;</a></p></li><li><p><a href="https://news.rice.edu/news/2024/breaking-mad-generative-ai-could-break-internet">Rice University research on &#8220;Model Autophagy Disorder&#8221; (link TBD)</a></p></li></ul><p><strong>Media and Analysis:</strong></p><ul><li><p><a href="https://www.theatlantic.com/technology/archive/2021/08/dead-internet-theory-wrong-but-feels-right/619937/">The Atlantic: &#8220;Maybe You Missed It, but the Internet &#8216;Died&#8217; Five Years Ago&#8221;</a></p></li><li><p><a href="https://time.com/7316046/sam-altman-dead-internet-theory/">TIME Magazine: &#8220;Sam Altman Voices Concern Over Dead Internet Theory&#8221;</a></p></li><li><p><a href="https://www.popularmechanics.com/science/a65997294/dead-internet-explained/">Popular Mechanics: &#8220;The Internet Will Be More Dead Than Alive Within 3 Years&#8221;</a></p></li><li><p><a href="https://www.imperva.com/resources/resource-library/reports/2024-bad-bot-report/">Imperva analytical report on automated traffic share</a></p></li><li><p><a href="https://pluralistic.net/2023/01/21/potemkin-ai/#hey-guys">Cory Doctorow&#8217;s &#8220;The Enshittification of TikTok&#8221; on platform degradation</a></p></li><li><p><a href="https://www.vice.com/en/article/dead-internet-theory-is-back-thanks-to-all-of-that-ai-slop/">VICE: &#8220;&#8217;Dead Internet Theory&#8217; Is Back Thanks to All of That AI Slop&#8221;</a></p></li><li><p><a href="https://theconversation.com/the-dead-internet-theory-makes-eerie-claims-about-an-ai-run-web-the-truth-is-more-sinister-229609">The Conversation: &#8220;The &#8216;dead internet theory&#8217; makes eerie claims about an AI-run web&#8221;</a></p></li></ul><div><hr></div><p><em>The Dead Internet Theory isn&#8217;t just a technological trend - it&#8217;s a challenge to our understanding of what it means to be human in the digital age. While we debate whether this threat is real, algorithms are already shaping our perception of the world.</em></p><p><strong>Share your thoughts: Have you noticed signs of the &#8220;dead internet&#8221; in your online experience?</strong></p><p><em>If this post was helpful, please like and share with colleagues. Subscribe to the newsletter to stay updated on artificial intelligence and digital trends.</em></p>]]></content:encoded></item><item><title><![CDATA[Duolingo’s AI-First Shift: Growth, Jobs, and the Future of Learning]]></title><description><![CDATA[When Duolingo announced its move to an &#8220;AI-first&#8221; strategy, the online reaction was immediate and filled with assumptions of mass layoffs.]]></description><link>https://www.bulaev.net/p/duolingos-ai-first-shift-growth-jobs</link><guid isPermaLink="false">https://www.bulaev.net/p/duolingos-ai-first-shift-growth-jobs</guid><dc:creator><![CDATA[Serge Bulaev]]></dc:creator><pubDate>Wed, 24 Sep 2025 21:56:39 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/74c0e449-fae7-4b8b-8ddf-9850527e4c27_1336x1007.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>When Duolingo announced its move to an &#8220;AI-first&#8221; strategy, the online reaction was immediate and filled with assumptions of mass layoffs. Many observers connected the announcement to trends across the technology sector, where adopting AI has often gone hand in hand with reducing staff.</p><p>Yet Duolingo&#8217;s implementation of AI has unfolded in a very different manner. CEO Luis von Ahn clarified that every full-time employee kept their job, and the company leaned on AI as a tool for expansion, speed, and global accessibility rather than workforce reduction.</p><p>The pace of change since that announcement has been striking. In twelve months, Duolingo launched 148 new language courses. The figure is larger than the total number of courses released over its prior twelve years of operation.</p><p>This expansion has allowed the company to reach into regions and communities that had previously been underserved. Several of the new courses target underrepresented languages and expand access to speakers who had little prior representation in digital learning tools.</p><p>The business effects have been visible in Duolingo&#8217;s performance numbers. As of late 2024 the platform recorded 130 million monthly active users and 47 million daily active users. That represents a 51% year-over-year increase.</p><p>Subscribers grew to 10.9 million, representing a 37% jump compared with the previous year. With this momentum analysts forecast annual revenue will reach 1 billion USD in 2025. <a href="https://chiefaiofficer.com/blog/how-duolingos-ai-first-strategy-drove-51-user-growth-and-1-billion-revenue-forecast/?_gl=1*ot1ppf*_ga*ODg0NDczMzM5LjE3NTgyMjg0Mzk.*_ga_2W6XVQKMG3*czE3NTg3NTA0MTIkbzIkZzEkdDE3NTg3NTA4NjgkajI0JGwwJGgw*_gcl_au*MTkzMjMyMjUzMS4xNzU4MjI4NDM5">These statistics show</a> that Duolingo&#8217;s AI-centered reorientation has become directly tied to its continuing revenue growth and its global leadership in digital language learning.</p><blockquote><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!0sFd!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F657b79f0-c3c5-4787-81cf-7b061dfacd6b_1436x480.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!0sFd!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F657b79f0-c3c5-4787-81cf-7b061dfacd6b_1436x480.png 424w, https://substackcdn.com/image/fetch/$s_!0sFd!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F657b79f0-c3c5-4787-81cf-7b061dfacd6b_1436x480.png 848w, https://substackcdn.com/image/fetch/$s_!0sFd!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F657b79f0-c3c5-4787-81cf-7b061dfacd6b_1436x480.png 1272w, https://substackcdn.com/image/fetch/$s_!0sFd!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F657b79f0-c3c5-4787-81cf-7b061dfacd6b_1436x480.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!0sFd!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F657b79f0-c3c5-4787-81cf-7b061dfacd6b_1436x480.png" width="1436" height="480" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/657b79f0-c3c5-4787-81cf-7b061dfacd6b_1436x480.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:480,&quot;width&quot;:1436,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!0sFd!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F657b79f0-c3c5-4787-81cf-7b061dfacd6b_1436x480.png 424w, https://substackcdn.com/image/fetch/$s_!0sFd!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F657b79f0-c3c5-4787-81cf-7b061dfacd6b_1436x480.png 848w, https://substackcdn.com/image/fetch/$s_!0sFd!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F657b79f0-c3c5-4787-81cf-7b061dfacd6b_1436x480.png 1272w, https://substackcdn.com/image/fetch/$s_!0sFd!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F657b79f0-c3c5-4787-81cf-7b061dfacd6b_1436x480.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div></blockquote><p><a href="https://technologymagazine.com/ai-and-machine-learning/duolingos-ai-first-strategy-explained">AI is threaded</a> into almost every aspect of Duolingo&#8217;s operations. On the product side, the proprietary Birdbrain system personalizes lessons in real time, customizing the difficulty of exercises based on each learner&#8217;s interactions.</p><p>This creates a feedback loop so that practice remains challenging but does not overwhelm. To make this personalization efficient and scalable, Duolingo re-engineered its Session Generator in Scala, cutting processing times from 750 milliseconds to 14 milliseconds. For users, this speedup feels seamless, keeping lessons smooth even as AI calculates adjustments in the background.</p><p>AI is not limited to the learner experience. The company has also woven machine learning tools into hiring and performance review systems, areas where many companies still lag.</p><p>Employees are encouraged to explore AI in their own work as well. &#8220;f-r-AI-days,&#8221; set aside as no-pressure days for experimentation and skill development, serve as internal laboratories where staff can learn, test, and adapt. This pattern fits with Duolingo&#8217;s long-standing cultural emphasis on continuous improvement.</p><p>Though the company avoided any layoffs among full-time staff, the changes were not without effects on certain groups. Around 10% of contracted translators lost work as the company leveraged AI to automate basic translation. The news sparked debate. Critics questioned fairness and transparency, pointing toward the risks of replacing contract labor with algorithms. <a href="https://fortune.com/2025/08/18/duolingo-ceo-admits-controversial-ai-memo-did-not-give-enough-context-insists-company-never-laid-off-full-time-employees/">Von Ahn responded </a>by stressing the distinction between contracted roles and the long-term employee base, adding reassurance that Duolingo&#8217;s workforce stability would not be threatened by adopting AI at scale.</p><p>Questions about content quality have also emerged. Observers have raised concerns about whether AI is equipped to handle nuanced language instruction, particularly as cultural or dialect differences come into play. Duolingo acknowledged these concerns while underscoring that human educators and language experts remain involved in overseeing AI outputs. The company has positioned this human-in-the-loop process as essential, maintaining that expanded efficiency does not mean sacrificing accuracy or cultural authenticity.</p><p>A significant reason behind Duolingo&#8217;s ability to roll out AI effectively lies in its culture and internal structures. In late 2024 Duolingo published an internal handbook outlining what it calls the <strong>&#8220;Green Machine&#8221; philosophy.</strong></p><p>This model compares the organization to a computer processor. The objective is to reduce as much as possible the lag between decision-making, action, feedback, and revision. By shrinking the interval between each step, the company maintains high responsiveness.</p><p>Unlike many corporations that adopt a Minimum Viable Product approach, releasing stripped-down test products, Duolingo deliberately rejects MVP thinking. Instead the company insists on combining speed with quality. The result is an organizational system that can release rapid updates without eroding user trust.</p><p>This internal philosophy has proven particularly effective for scaling AI adoption. The Green Machine model means feedback loops occur quickly, so any issues or improvements with AI tools are iterated on rapidly. When paired with initiatives like f-r-AI-days, this philosophy gives staff the foundation and safety to learn new technologies while still delivering results.</p><p>The company views employee development as essential to scaling the platform globally, demonstrating that the people who work within a system determine how technology ultimately succeeds.</p><p>The scope of Duolingo&#8217;s AI-driven expansion is broader than just internal metrics. By making its seven most popular non-English languages available across all 28 of its user interface languages, the company has dramatically widened access.</p><p>This step directly benefits learners in Latin America, Asia, and Europe, where users previously faced barriers if they did not speak English. AI-powered translation and expansion has therefore allowed the platform to serve diverse communities with more inclusivity.</p><p>Duolingo defines itself as more than a language learning application. It positions itself as an educational enterprise that lowers barriers to knowledge. The AI-first shift mirrors this mission: scaling up courses, supporting different interface combinations, and personalizing the journey of each student.</p><p>At the same time, the company became an example used in broader AI debates. Where other firms introduce AI under narratives of efficiency gains that often mask staff reductions, Duolingo chose to emphasize communication and reassurance, reinforcing trust and stability.</p><p>The numbers speak with clarity:</p><ul><li><p>148 new courses in 12 months.</p></li><li><p>A 51% jump in daily active users.</p></li><li><p>10.9 million people committed to the paid version of the product. A forecast of one billion dollars in annual revenue.</p></li></ul><p>These outcomes suggest an AI rewrite of corporate processes has helped Duolingo achieve scale that would have been unthinkable with traditional human-only workflows. Yet the company continues to underline the role of human oversight, especially in managing cultural quality and guiding organizational direction.</p><p>Duolingo&#8217;s experience contains lessons for other firms considering large-scale AI adoption. The key takeaway is that the technology itself is not sufficient without a culture that promotes rapid iteration and maintains trust. The company&#8217;s deliberate rejection of the MVP mentality, its internal philosophy organized around eliminating lag, and its visible reassurance about jobs created the environment where AI became an accelerator rather than a disruptor.</p>]]></content:encoded></item><item><title><![CDATA[The Borrowed Mind: Are We Outsourcing Our Capacity to Think?]]></title><description><![CDATA[There's a moment that's becoming disturbingly familiar - when you're mid-sentence and suddenly can't tell if that insight you just shared came from your own thinking or from an AI tool you consulted earlier.]]></description><link>https://www.bulaev.net/p/the-borrowed-mind-are-we-outsourcing</link><guid isPermaLink="false">https://www.bulaev.net/p/the-borrowed-mind-are-we-outsourcing</guid><dc:creator><![CDATA[Serge Bulaev]]></dc:creator><pubDate>Tue, 16 Sep 2025 21:40:14 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/f3518776-ff28-4590-a517-3d544117b911_1024x1024.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!QUOy!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb230bc4f-b064-4bc4-ba2c-41c535fd521e_320x306.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!QUOy!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb230bc4f-b064-4bc4-ba2c-41c535fd521e_320x306.jpeg 424w, https://substackcdn.com/image/fetch/$s_!QUOy!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb230bc4f-b064-4bc4-ba2c-41c535fd521e_320x306.jpeg 848w, https://substackcdn.com/image/fetch/$s_!QUOy!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb230bc4f-b064-4bc4-ba2c-41c535fd521e_320x306.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!QUOy!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb230bc4f-b064-4bc4-ba2c-41c535fd521e_320x306.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!QUOy!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb230bc4f-b064-4bc4-ba2c-41c535fd521e_320x306.jpeg" width="320" height="306" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b230bc4f-b064-4bc4-ba2c-41c535fd521e_320x306.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:306,&quot;width&quot;:320,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:14436,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.bulaev.net/i/173791773?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb230bc4f-b064-4bc4-ba2c-41c535fd521e_320x306.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!QUOy!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb230bc4f-b064-4bc4-ba2c-41c535fd521e_320x306.jpeg 424w, https://substackcdn.com/image/fetch/$s_!QUOy!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb230bc4f-b064-4bc4-ba2c-41c535fd521e_320x306.jpeg 848w, https://substackcdn.com/image/fetch/$s_!QUOy!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb230bc4f-b064-4bc4-ba2c-41c535fd521e_320x306.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!QUOy!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb230bc4f-b064-4bc4-ba2c-41c535fd521e_320x306.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><p>There's a moment that's becoming disturbingly familiar - when you're mid-sentence and suddenly can't tell if that insight you just shared came from your own thinking or from an AI tool you consulted earlier. That cognitive vertigo, that brief loss of intellectual orientation, might be the defining experience of our current moment.</p><p>Are we slowly borrowing our own minds from machines?</p><p>The idea of "The Borrowed Mind" - actively promoted by <strong>John Nosta</strong> - cuts to something we're all sensing but maybe not discussing: what happens when our thinking gets so intertwined with AI that we lose track of where we end and the machine begins?</p><p>This isn't abstract philosophizing. <a href="https://pmc.ncbi.nlm.nih.gov/articles/PMC12089268/">Recent research</a> is revealing genuinely unsettling insights into what happens in our brains when we lean too heavily on our digital thinking partners.</p><h2><strong>The Neuroscience of Cognitive Dependency</strong></h2><p>Brain scans show something remarkable: when people rely on AI too early in their thinking process, neural activity becomes less synchronized. Our brains literally work differently when we outsource thinking too quickly. Researchers call this <a href="https://daviddidau.substack.com/p/the-cost-of-borrowed-thought">"cognitive debt"</a> - weaker engagement with ideas, memories that don't stick, and a reduced sense of ownership over our own thoughts.</p><p>But here's what offers hope: if you think first, then bring AI in as a collaborator, those negative effects disappear entirely. Sequencing isn't just important - it's everything.</p><p>The teenager who writes her resume on her own first - messy, imperfect, authentically hers - then uses AI to polish it creates something genuinely better. Her friend who starts with AI? The resume is flawless and completely forgettable. Same tools, opposite outcomes, all because of when they entered the process.</p><h2><strong>The Socratic Echo: We've Been Here Before</strong></h2><p>When calculators became common, teachers worried kids would lose the ability to do basic math. When GPS arrived, people fretted about losing our sense of direction. Each time, the concerns were both valid and incomplete.</p><p>Socrates was genuinely worried that writing would ruin human memory. He thought storing knowledge externally would make minds lazy, that people would mistake the ability to look things up for actual understanding. He wasn't entirely wrong - we did lose something when we externalized memory. Most of us can barely remember three phone numbers while previous generations knew dozens by heart.</p><p>But we gained something profound too: the ability to build on accumulated knowledge rather than constantly reinventing the wheel. The question that keeps emerging: are we making a similar trade with thinking itself? And unlike previous technological shifts, this one feels more intimate, more fundamentally cognitive.</p><h2><strong>The Universal Language of Thought</strong></h2><p>AI research is revealing something that borders on the mystical: large language models show patterns suggesting a universal "language of thought" - neurons that represent concepts like "big" or "urgent" or "beautiful" in remarkably similar ways across completely different human languages.</p><p>This suggests AI might not just be mimicking our thinking - it could be revealing the fundamental architecture of thought itself. If AI truly understands these deep patterns of cognition, then using it as a thinking partner might be less like borrowing someone else's mind and more like plugging into the basic structures of intelligence.</p><p>But this creates a paradox that should give us pause. The more AI reveals about how thinking works, the more we risk losing touch with our own cognitive processes. We're discovering the patterns of thought just as we're outsourcing the act of thinking itself.</p><h2><strong>The Slippery Slope to AI Psychosis</strong></h2><p>We're seeing the emergence of what researchers call <a href="https://www.psychologytoday.com/us/blog/the-digital-self/202509/the-borrowed-mind">"AI psychosis"</a> - not the dramatic kind involving hallucinations, but the quiet, creeping inability to distinguish between original thoughts and machine-amplified ideas. It's not about plagiarism or academic dishonesty. It's about losing track of where your mind ends and the machine begins.</p><p>The email you're certain contains your most elegant thinking until you realize you can't remember if that metaphor was yours or something you read in an AI-generated report. The presentation where you finish speaking and think "I'm really insightful" before wondering if any of those ideas were actually your own.</p><p>This isn't happening to other people in some distant future. It's happening right now, in boardrooms and coffee shops and late-night writing sessions. People are losing the ability to locate the boundaries of their own minds.</p><h2><strong>The Authenticity Imperative</strong></h2><p>There's an emerging AI tone - polished, confident, slightly generic - that's becoming as recognizable as corporate jargon. Everyone's content is starting to sound eerily similar. This is exactly why an authentic voice has become more valuable than gold.</p><p>In a world where anyone can generate flawless content with a few prompts, the thinkers who stand out are those who maintain genuine connection to their own messy, imperfect, wonderfully human cognitive processes. The rough edges, the false starts, the "wait, let me think about this differently" moments - that's where real connection happens.</p><p>This is why tools that capture actual human voice, memory, and lived experience matter more than generic intelligence generators. It's the difference between a musician using a high-quality amplifier versus someone pressing play on Spotify. Both make sound, but only one represents authentic expression.</p><h2><strong>The Great Cognitive Choice</strong></h2><p>AI systems are getting unnervingly good at reasoning, debating, even catching their own mistakes. We're watching machines think out loud, correct themselves mid-response, and provide completely different (often better) answers after self-reflection.</p><p>This brings us to what feels like a crossroads: the Great Cognitive Choice. Do we let these systems become cognitive prostheses we can't function without? Or do we maintain ownership of our thinking and invite AI in as collaborators only after we've done the hard work ourselves?</p><p>That familiar pull to "just ask the AI" when we hit a difficult problem is seductive. The twenty minutes of wrestling with confusion ourselves feels frustrating, inefficient, messy. It's also where we feel most human.</p><p>The stakes feel enormous. Choose poorly, and we risk creating a generation that can produce sophisticated content but has lost the ability to think slowly, deeply, originally. Choose wisely, and we might be entering the next stage of human cognition - where AI becomes what writing was to memory: not a replacement, but a tool that frees us to tackle bigger, more complex problems than ever before.</p><h2><strong>The Path Forward</strong></h2><p>What does intentional cognitive practice look like in an AI world? It starts with what we might call "cognitive hygiene" - being as deliberate about our thinking habits as we are about our physical health.</p><p><strong>Start messy.</strong> Before reaching for AI assistance, spend real time with blank pages and scattered thoughts. Those initial scribbles may be terrible, but they're authentically human. Let your brain do what it evolved to do - make connections, generate ideas, wrestle with complexity.</p><p><strong>Use AI as a sparring partner.</strong> Once you have your own perspective, bring in AI not to write for you, but to argue with you. "Here's what I think about this problem. What am I missing? Where are the holes in my logic?" This creates collaboration rather than outsourcing.</p><p><strong>Track the boundary.</strong> Pay attention to which ideas feel genuinely yours and which ones feel borrowed or generated. This isn't about avoiding AI influence entirely - it's about maintaining conscious agency over your cognitive process.</p><p><strong>Exercise thinking muscles.</strong> Just as we need physical exercise even though elevators exist, we need to think through hard problems even though AI exists. Read challenging material. Have conversations without looking things up. Sit with confusion instead of immediately seeking AI clarification.</p><p><strong>Maintain cognitive fitness.</strong> In an increasingly automated world, mental agility requires deliberate practice. Engage with ideas that challenge you. Think slowly. Let your mind wander without immediately capturing and optimizing every thought.</p><p>The borrowed mind represents both our biggest risk and our most interesting opportunity. Whether it becomes our cognitive crutch or our thinking amplifier depends entirely on choices we make right now.</p><p>That moment of cognitive vertigo - when you can't locate the source of your own thoughts - is a warning sign worth heeding. The question isn't whether AI will change how we think. It already has. The question is whether we'll maintain conscious agency over that change.</p><p>What do you think? Are we at risk of losing independent thought, or entering the next stage of human-extended cognition? And maybe more importantly - have you experienced that disorienting moment when you couldn't tell where your thinking ended and the machine's began?</p><p><em>Share your experiences in the comments. When have you struggled to distinguish between your own thoughts and AI-generated ones? What strategies help you maintain an authentic voice? Are we overthinking this, or witnessing something more significant than we realize?</em></p>]]></content:encoded></item><item><title><![CDATA[When Jung Meets Machine Learning: Making the Digital Unconscious Conscious]]></title><description><![CDATA[Sometimes you read a line that hits harder than expected.]]></description><link>https://www.bulaev.net/p/when-jung-meets-machine-learning</link><guid isPermaLink="false">https://www.bulaev.net/p/when-jung-meets-machine-learning</guid><dc:creator><![CDATA[Serge Bulaev]]></dc:creator><pubDate>Wed, 10 Sep 2025 21:46:00 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/3d942129-cb26-4058-a8e1-ba854ad68577_660x440.webp" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Sometimes you read a line that hits harder than expected. Jung said something along the lines of: <strong>"Until you make the unconscious conscious, it will direct your life and you will call it fate."</strong></p><p>That's psychology. But here's the twist - it's also unexpectedly useful for understanding how large language models behave.</p><h2><strong>The Training Unconscious</strong></h2><p>Think of it this way: LLMs are trained on oceans of data - billions of web pages, books, articles, conversations. Hidden inside this vast corpus are patterns, biases, associations, cultural assumptions, and implicit knowledge structures. It's their own kind of "unconscious" - a vast repository of learned behaviors and responses that the model can't directly access or explain, yet which fundamentally shapes every output.</p><p>Left unchecked, these hidden structures surface in outputs that feel random, inconsistent, or "fated." The model produces responses that seem to come from nowhere, echoing patterns it absorbed during training without any conscious understanding of why. It hallucinates facts, exhibits biases it can't name, or suddenly shifts tone in ways that feel arbitrary. From the outside, it looks like digital destiny - unpredictable, uncontrollable, almost mystical in its randomness.</p><p>But here's where Jung's insight becomes practically powerful: when we push the model to explain itself, to surface its hidden reasoning, to recognize what it doesn't know, or to trace the logic behind its responses - that's like making the unconscious conscious. Suddenly there's more clarity, less noise, more intentional control over outcomes.</p><h2><strong>The Parallels Run Deep</strong></h2><p>The psychology here isn't literal - machines aren't self-aware the way humans are, and I'm not suggesting LLMs have genuine consciousness. But the structural parallels are remarkably useful:</p><p><strong>Jung's Framework:</strong></p><ul><li><p>Inner conflicts and unconscious patterns shape outer reality unless brought into awareness</p></li><li><p>Without consciousness, we experience life as happening TO us rather than being shaped BY us</p></li><li><p>Awareness brings choice; unconsciousness feels like destiny</p></li><li><p>The shadow contains both destructive patterns and untapped potential</p></li><li><p>Integration requires active engagement with what's hidden</p></li></ul><p><strong>LLM Behavior:</strong></p><ul><li><p>Hidden training patterns shape responses unless surfaced and actively managed</p></li><li><p>Without transparency mechanisms, outputs feel unpredictable and uncontrollable</p></li><li><p>Explainability brings steering capability; opacity feels like randomness</p></li><li><p>Training data contains both problematic biases and valuable knowledge</p></li><li><p>Better performance requires deliberately surfacing and working with hidden structures</p></li></ul><h2><strong>Practical Applications</strong></h2><p>This isn't just philosophical musing - it has real implications for how we work with AI systems. When we design prompting strategies that ask models to "think step by step," explain their reasoning, or acknowledge uncertainty, we're essentially doing therapeutic work with the training unconscious. We're creating conditions for the model to surface patterns that would otherwise remain hidden and potentially problematic.</p><p>Consider how Jung approached therapy: not by trying to eliminate the unconscious, but by bringing it into dialogue with conscious awareness. Similarly, the most effective AI applications don't try to eliminate the "training unconscious" - that vast repository of learned patterns is actually the source of the model's capabilities. Instead, they create mechanisms for that unconscious knowledge to surface in controlled, intentional ways.</p><h2><strong>The Shadow of Scale</strong></h2><p>Jung wrote extensively about the "shadow" - the parts of ourselves we don't want to acknowledge but which inevitably influence our behavior. LLMs have their own version of this: the biases, misconceptions, and problematic associations embedded in training data. Just as Jung suggested we can't eliminate our shadow but must learn to work with it consciously, we can't eliminate bias from AI systems - but we can develop better ways to surface and manage it.</p><p>The alternative is what we see too often: AI systems that perpetuate harmful patterns precisely because those patterns remain unconscious and unexamined. The bias doesn't disappear when ignored - it just operates outside of conscious control, manifesting as what feels like inevitable, fated outcomes.</p><h2><strong>Beyond the Metaphor</strong></h2><p>What makes this comparison more than just clever wordplay is how both point toward the same fundamental insight: consciousness isn't about elimination of the unconscious, but about bringing it into a productive relationship with intentional awareness. Whether we're talking about human psychology or machine behavior, the goal isn't perfect rational control - it's developing better ways to work with the vast, hidden structures that actually drive most behavior.</p><p>For humans, this might mean recognizing how childhood patterns still influence adult relationships. For LLMs, it might mean surfacing how certain training examples disproportionately influence responses to specific types of questions. In both cases, awareness doesn't eliminate the underlying patterns - it creates space for more intentional engagement with them.</p><h2><strong>The Future of Digital Psychology</strong></h2><p>As AI systems become more sophisticated and integrated into daily life, this parallel becomes more than academic. We're essentially in the early stages of developing a kind of "digital psychology" - methods for understanding and working with the hidden mental structures of artificial systems.</p><p>Jung's insight that unconscious patterns feel like fate until made conscious offers a surprisingly practical framework for AI development. Instead of accepting unpredictable model behavior as inevitable, we can treat it as a signal that important patterns remain hidden and need to be surfaced.</p><p>The psychology here runs deeper than the obvious parallels might suggest. Both Jung's work and effective AI development share a fundamental understanding: the most powerful systems aren't those that eliminate complexity, but those that develop better relationships with it. Whether we're talking about the human psyche or large language models, sustainable progress comes not from perfect control, but from bringing hidden patterns into conscious dialogue.</p><p>Isn't that exactly what Jung was pointing at in people's lives too? The unconscious isn't the enemy of consciousness - it's its necessary partner. The same may well be true for the relationship between AI capabilities and AI alignment.</p><p>And perhaps that's the most profound parallel of all: in both human psychology and machine learning, the path forward isn't through elimination of complexity, but through developing more conscious relationships with the hidden structures that shape behavior. The unconscious, whether human or digital, isn't fate - it's raw material for more intentional creation.</p>]]></content:encoded></item><item><title><![CDATA[𝐄𝐦𝐩𝐥𝐨𝐲𝐞𝐞 𝐂𝐨𝐧𝐭𝐞𝐧𝐭 𝐏𝐨𝐬𝐭𝐢𝐧𝐠 𝐚𝐧𝐝 𝐂𝐨𝐦𝐩𝐚𝐧𝐲 𝐕𝐚𝐥𝐮𝐞: 𝐀 𝐒𝐭𝐫𝐨𝐧𝐠 𝐂𝐨𝐫𝐫𝐞𝐥𝐚𝐭𝐢𝐨𝐧]]></title><description><![CDATA[If you&#8217;re still wondering whether employees posting content actually moves the needle, the latest data makes it hard to argue otherwise.]]></description><link>https://www.bulaev.net/p/578</link><guid isPermaLink="false">https://www.bulaev.net/p/578</guid><dc:creator><![CDATA[Serge Bulaev]]></dc:creator><pubDate>Thu, 04 Sep 2025 16:57:55 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/4df1c24a-3526-427b-b297-b6bd976fe943_720x700.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>If you&#8217;re still wondering whether employees posting content actually moves the needle, the latest data makes it hard to argue otherwise.</p><p><a href="https://hingemarketing.com/uploads/hinge-research-employee-advocacy.pdf">Hinge Research Institute </a>found that companies with formal employee advocacy programs are over twice as likely to see revenue growth above 20%. 64% of advocates say they&#8217;ve attracted and developed new business through their activity. 45% go even further &#8211; saying advocacy generated entirely new revenue streams.</p><p><a href="https://influitive.com/resources/2023-state-brand-advocacy-report-employee-advocacy/">PostBeyond&#8217;s SOBA 2023 report</a> points the same way. 33% can directly tie revenue gains to advocacy. Half admit ROI is hard to track in detail. Still, 83% of these organizations have kept or increased budgets. That speaks volumes about perceived value at the leadership level.</p><h2><strong>Now the 2025 data makes the picture sharper:</strong></h2><ul><li><p>Employee participation can shoot up 50 times in the first year of program launch &#8211; insurance led with +5,315% (from 52 employees to 2,816)</p></li><li><p>On LinkedIn, employee posts are now averaging nearly <a href="https://www.agilitypr.com/pr-agency-news/2025-employee-advocacy-benchmark-report-from-advocacy-by-socialpubli-how-advocacy-supercharges-brand-visibility/">900,000 impressions a month</a> &#8211; far outstripping brand accounts on the same platform</p></li><li><p>Personal profiles see up to 8x higher engagement than company pages</p></li><li><p>In tech and consulting, employee-generated posts<a href="https://www.gaggleamp.com/employee-advocacy-statistics-you-need-to-know"> convert up to 25% more leads</a> compared to paid campaigns</p></li><li><p>In some industries, branded social engagement has jumped 200% to 300% once employees started sharing</p></li><li><p>23% of organizations <a href="https://dsmn8.com/blog/employee-advocacy-benchmark-report-2025/">hit cost&#8209;per&#8209;click under $1</a> with employee&#8209;created content</p></li><li><p>On Facebook, advocacy posts in tourism have hit 15% engagement rates &#8211; proof the effect is multi&#8209;platform</p></li><li><p>76% of people trust content from individuals over brands; 77% are more likely to buy after hearing from someone they trust</p></li><li><p>One European SaaS firm saw inbound partnership requests up 40% after 12 months of structured posting</p></li></ul><h2><strong>The benefits reach the individual as well:</strong></h2><ul><li><p>87% of advocates expand their professional networks significantly</p></li><li><p>44% gain recognition as thought leaders in their fields</p></li><li><p>34% of companies say employee engagement is the biggest impact of advocacy &#8211; ahead of sales or brand lift</p></li><li><p>71% report higher visibility, 65% stronger recognition, nearly 45% more inbound traffic, and 17% lower marketing spend</p></li></ul><p>On the organizational side, scaling advocacy now means leadership involvement &#8211; 73% plan to get executives sharing more, and 67% are putting money into better training and resources.</p><p>The reason this works is dead simple: your employees already belong to trust&#8209;based networks your official accounts will never reach. Every message they share is delivered in a context of credibility. That is why engagement rates are higher, CPC is lower, and brand conversations feel more genuine.</p><p>The real question has shifted. We&#8217;re no longer debating if we should encourage employee advocacy. We&#8217;re being asked how fast we can equip our people to do it well &#8211; and ways to support them without dulling their voice. That means training, clear guidelines, and letting authenticity lead.</p><p>By 2025, personal brands and company brands occupy the same space more often. Inside that overlap is where business value is built &#8211; in revenue, influence, and trust that can&#8217;t be bought.</p><p>So... are your people posting yet?</p>]]></content:encoded></item><item><title><![CDATA[AI Scientists: The First Peer-Reviewed Paper Generated by AI]]></title><description><![CDATA[Recently, I encountered The AI Scientist service, which claims to "fully" automate scientific discovery.]]></description><link>https://www.bulaev.net/p/ai-scientists-the-first-peer-reviewed</link><guid isPermaLink="false">https://www.bulaev.net/p/ai-scientists-the-first-peer-reviewed</guid><dc:creator><![CDATA[Serge Bulaev]]></dc:creator><pubDate>Thu, 04 Sep 2025 00:04:57 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/42f2aedd-7a62-4e2c-a2f6-af8f985b2c09_1280x600.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Recently, I encountered The AI Scientist service, which claims to "fully" automate scientific discovery. This is a system that formulates hypotheses, designs experiments, analyzes results, and even writes scientific papers - without human involvement.</p><ul><li><p><a href="https://arxiv.org/abs/2504.08066">The AI Scientist-v2</a> recently achieved a historic milestone: publishing the first scientific paper entirely generated by AI that was accepted at a workshop after peer review.</p></li><li><p>The system can conduct research around the clock, potentially finding patterns that humans might miss due to cognitive limitations.</p></li><li><p>The project is open to the community: source code on <a href="https://github.com/SakanaAI/AI-Scientist-v2">GitHub</a> allows for experimentation, refinement, and integration of new models.</p></li></ul><p>Obviously, such an approach could accelerate scientific progress and make research more transparent and accessible, but it also raises new questions about the role of humans, trust in results, and the ethics of autonomous AI in science.</p><h2><strong>What's Behind the Technology</strong></h2><p>After diving deep into The AI Scientist-v2 system from <a href="https://pub.sakana.ai/ai-scientist-v2/paper/paper.pdf?utm_source=chatgpt.com">Sakana AI</a>, I was impressed by its architectural sophistication. This isn't just "GPT writes papers" - it's a comprehensive research pipeline with several revolutionary improvements:</p><p><strong>Agentic Tree Search.</strong> Unlike the linear approach of the previous version, v2 uses a tree-like structure for experiments. Each tree node represents a separate experiment with Python code, research plan, and results. The system explores multiple hypotheses in parallel, automatically debugs code when errors occur, and selects the best directions for further development.</p><p><strong>Complete Autonomy.</strong> The system no longer depends on human-written code templates. AI independently generates all experimental code from scratch, based only on high-level research ideas. This dramatically expands the system's applicability.</p><p><strong>Vision-Language Model Integration.</strong> VLMs analyze generated graphs and visualizations, checking their correctness, caption clarity, and alignment with conclusions. This ensures quality scientific presentation of results.</p><p><strong>Four-Stage Process.</strong> Research follows structured phases: preliminary investigation &#8594; hyperparameter tuning &#8594; research agenda execution &#8594; ablation studies. At each stage, the system creates up to 21 parallel experiments.</p><h2><strong>Historic Achievement with Caveats</strong></h2><p>The result is impressive: one of three generated papers received an average score of 6.33 out of 10 from <a href="https://sites.google.com/view/icbinb-2025">ICLR workshop "I Can't Believe It's Not Better"</a> reviewers, exceeding the acceptance threshold. The paper investigated compositional regularization in neural networks and, interestingly, received positive feedback precisely for honestly presenting negative results.</p><p>However, it's important to understand the context: this was a workshop, not a main conference track. Acceptance rates at workshops typically range from 60-80% versus 20-30% at top-tier conferences. Only one of the three submitted papers was accepted.</p><p>Upon detailed analysis of the accepted paper, the system's authors identified significant flaws: inaccuracies in figure captions, problems with training/test dataset overlap (57% overlap!), terminology confusion, and insufficient justification for methodological choices.</p><h2><strong>Technical Limitations and Ethical Questions</strong></h2><p>The system is still far from generating breakthrough hypotheses or deep domain understanding. Experiments are limited to relatively simple ML tasks, and the quality of work doesn't yet reach top-conference standards.</p><p>The ethical aspect is particularly important. The Sakana AI team obtained ethics committee approval, warned reviewers about possible AI-generated submissions, and subsequently withdrew the accepted paper to avoid setting a precedent without public discussion.</p><p>This raises fundamental questions: Should AI papers be labeled? How should they be evaluated alongside human work?</p><h2><strong>Development Trajectory</strong></h2><p>Despite limitations, the trajectory is impressive. In two years, we've moved from proof-of-concepts to systems capable of passing peer review. The rapid development of AI tools suggests that within a few years, we might see conference-level AI researchers.</p><p>The potential is enormous: AI can work 24/7, isn't subject to cognitive biases, can test multiple hypotheses in parallel, and process gigantic volumes of literature. In fields with large datasets and clearly defined metrics - from bioinformatics to materials science - such systems could significantly accelerate discoveries.</p><p>But the main question remains open: Can AI generate truly revolutionary ideas, or only optimize known approaches? So far, creativity and conceptual breakthroughs remain human prerogatives.</p><h2><strong>What This Means for Science</strong></h2><p>We stand on the threshold of a fundamental change in the scientific process. AI researchers could become powerful tools for accelerating routine aspects of science: literature reviews, experiment replication, systematic testing of variations.</p><p>But this also requires rethinking the role of human researchers. Perhaps the future lies with hybrid teams, where AI performs large-scale computational work while humans focus on posing deep questions, interpreting results in broad context, and determining research directions.</p><p>The AI Scientist-v2 system isn't the end of the road, but an important milestone toward a new paradigm of scientific research. How we integrate these tools will determine the future of science for decades to come.</p><h2><strong>The Reality Check</strong></h2><p>Let me be clear about what we're actually seeing here. When I examined the accepted paper's code and methodology, several concerning issues emerged:</p><ul><li><p><strong>Dataset contamination</strong>: 57% overlap between training and test sets, fundamentally compromising the reliability of results</p></li><li><p><strong>Methodological confusion</strong>: The paper confused "embedding states" with "hidden states," indicating imprecise understanding</p></li><li><p><strong>Overclaimed results</strong>: The system reported 100% accuracy that was mainly due to task simplicity, not algorithmic breakthrough</p></li></ul><p>This isn't to diminish the achievement - it's to calibrate our expectations. We're witnessing the first steps of AI scientific reasoning, not its maturation.</p><h2><strong>The Bigger Picture</strong></h2><p>What excites me most isn't the current capabilities, but the acceleration curve. The system went from template-dependent linear exploration to autonomous tree-based research in just one iteration. The improvements are architectural, not just computational - suggesting we're on a steep learning curve.</p><p>We have the opportunity to create tools that enhance human curiosity rather than replace it. We can accelerate progress while preserving the essential human elements that make science meaningful.</p><p>This is our moment to shape the future of knowledge itself. Let's make it count.</p><p>The paper and additional materials are available in <a href="https://github.com/SakanaAI/AI-Scientist-ICLR2025-Workshop-Experiment">Sakana AI's GitHub</a> repository.</p>]]></content:encoded></item><item><title><![CDATA[When AI Tutors Beat Human Teachers: The Nigerian Experiment That Changes Everything]]></title><description><![CDATA[A World Bank study in Nigeria just delivered results that sound like science fiction.]]></description><link>https://www.bulaev.net/p/when-ai-tutors-beat-human-teachers</link><guid isPermaLink="false">https://www.bulaev.net/p/when-ai-tutors-beat-human-teachers</guid><dc:creator><![CDATA[Serge Bulaev]]></dc:creator><pubDate>Tue, 02 Sep 2025 21:50:41 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/6c91b3d1-3dcf-4a89-8b26-79a97c4221c5_1284x850.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><a href="https://documents.worldbank.org/en/publication/documents-reports/documentdetail/099548105192529324">A World Bank study</a> in Nigeria just delivered results that sound like science fiction. 422 schoolchildren worked with Microsoft Copilot (powered by GPT-4) for <strong>90 minutes a day over 6 weeks</strong>. The outcome? <strong>Progress equivalent to two full years of regular schooling.</strong></p><p>But here's where it gets really interesting: this miracle happened in one of the world's poorest educational systems, while wealthy countries are seeing the opposite results.</p><h2><strong>The Paradox Nobody Saw Coming</strong></h2><p>In <strong>Turkey and the Netherlands</strong>, carefully controlled AI experiments ended in failure. Students became so dependent on LLMs that without them, they performed <strong>worse than their peers</strong>. The very countries with the best schools, the most resources, the highest teacher salaries - AI made their students worse learners.</p><p>Meanwhile, in Nigerian computer labs where many students had never touched a computer before, AI was producing learning gains that education researchers only dream about.</p><p>This flips everything we think we know about technology and inequality on its head.</p><h2><strong>The Brutal Numbers Behind the Success</strong></h2><p>To understand why the Nigerian results matter so much, you need to see the educational disaster these kids are living through:</p><ul><li><p><strong>70% of ten-year-olds in developing countries can't read simple text</strong></p></li><li><p><strong>In Africa, that number hits 90%</strong></p></li><li><p><strong>Nigerian kids get 10 years of schooling but learn what should take 5 years</strong></p></li></ul><p>This isn't about "falling behind" - this is educational system collapse. Traditional interventions barely move the needle. The World Bank has tried everything: better textbooks, teacher training, smaller class sizes, nutrition programs. Most show modest improvements at best.</p><p>Then along comes an AI chatbot and suddenly students are learning at 4x normal speed.</p><h2><strong>The $48 Miracle</strong></h2><p>The program cost <strong>$48 per student for 6 weeks</strong>. In Nigeria, where minimum wage is around $43 per month, that's serious money. But when researchers compared it against 230 other educational interventions, this AI program <strong>outperformed 80% of them</strong>.</p><p>Think about what $48 bought these students: a personal tutor available 90 minutes every day, infinitely patient, able to explain concepts a thousand different ways until each student got it. Benjamin Bloom proved in 1984 that one-on-one tutoring could achieve exactly these kinds of gains. The problem was always cost - until now.</p><h2><strong>What Actually Happened in Those Computer Labs</strong></h2><p><strong>Teachers became conductors, not replacements.</strong> They started each session with suggested prompts, guided student interactions, and led reflection exercises. As one teacher put it: AI is "like an assistant teacher. We supervise what the students are doing."</p><p><strong>Students found their own ways to use it.</strong> Teachers noted that kids quickly discovered "unique and productive ways to interact with the LLMs." They weren't passive consumers - they became active learners figuring out how to get what they needed from the AI.</p><p><strong>Infrastructure nearly killed it.</strong> Power outages and internet failures constantly disrupted sessions. Nigeria's rainy season was particularly brutal. Success required backup generators and redundant internet connections.</p><p><strong>Prompt engineering was everything.</strong> The program developed specific toolkits showing students how to ask better questions. This made the AI responses "much more useful" with examples relevant to local context.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!hryr!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F877ed39a-2c45-46d0-bc47-59125048edb5_1600x991.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!hryr!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F877ed39a-2c45-46d0-bc47-59125048edb5_1600x991.png 424w, https://substackcdn.com/image/fetch/$s_!hryr!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F877ed39a-2c45-46d0-bc47-59125048edb5_1600x991.png 848w, https://substackcdn.com/image/fetch/$s_!hryr!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F877ed39a-2c45-46d0-bc47-59125048edb5_1600x991.png 1272w, https://substackcdn.com/image/fetch/$s_!hryr!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F877ed39a-2c45-46d0-bc47-59125048edb5_1600x991.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!hryr!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F877ed39a-2c45-46d0-bc47-59125048edb5_1600x991.png" width="724" height="448.52197802197804" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/877ed39a-2c45-46d0-bc47-59125048edb5_1600x991.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:902,&quot;width&quot;:1456,&quot;resizeWidth&quot;:724,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!hryr!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F877ed39a-2c45-46d0-bc47-59125048edb5_1600x991.png 424w, https://substackcdn.com/image/fetch/$s_!hryr!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F877ed39a-2c45-46d0-bc47-59125048edb5_1600x991.png 848w, https://substackcdn.com/image/fetch/$s_!hryr!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F877ed39a-2c45-46d0-bc47-59125048edb5_1600x991.png 1272w, https://substackcdn.com/image/fetch/$s_!hryr!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F877ed39a-2c45-46d0-bc47-59125048edb5_1600x991.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2><strong>The Questions That Should Terrify Us</strong></h2><p>The results are incredible, but the study has gaps that matter:</p><p><strong>They couldn't isolate the AI effect.</strong> Researchers admit they couldn't separate gains from AI versus gains from extra attention, computer access, or additional tutoring some students might have received.</p><p><strong>Six weeks proves nothing long-term.</strong> Will these students retain their advantage in six months? Two years? Or will the gains fade like a sugar rush?</p><p><strong>English only.</strong> The study focused purely on English language skills. Math? Science? Critical thinking? We have no idea if AI tutoring transfers to other subjects.</p><p><strong>What about dependency?</strong> The wealthy country's results suggest AI can create learned helplessness. Are Nigerian students building genuine skills or just getting really good at prompting AI?</p><h2><strong>Why This Changes Everything (Maybe)</strong></h2><p>The same AI that creates dependency in Dutch classrooms unlocks potential in Nigerian ones. Why?</p><p><strong>Scarcity changes behavior.</strong> When you've never had access to quality instruction, AI feels like a miracle. When you're used to excellent teachers, AI feels like a downgrade.</p><p><strong>Baseline matters.</strong> If your education system is broken, AI can only improve things. If your system already works well, AI might disrupt what's working.</p><p><strong>Expectations shape outcomes.</strong> Nigerian students saw AI as an opportunity to learn. Dutch students saw it as a way to avoid learning.</p><h2><strong>The Student Who Gets It</strong></h2><p>Maybe the most insightful comment came from a student at Edo Boys High School: "AI helps us to learn, it can serve as a tutor, it can be anything you want it to be, depending on the prompt you write."</p><p><strong>Depending on the prompt you write.</strong></p><p>That seventeen-year-old just articulated something most adults miss: AI is a mirror. It reflects back the quality of thinking you bring to it. Bring curiosity and effort, learn. Bring laziness and shortcuts, get dependency.</p><h2><strong>What Happens Next?</strong></h2><p>Scaling the Nigerian experiment faces massive challenges:</p><p><strong>Infrastructure is expensive.</strong> Every classroom needs reliable power and internet. That's billions in investment across developing countries.</p><p><strong>Teacher training is complex.</strong> Moving from chalk-and-talk to AI orchestration requires completely different skills.</p><p><strong>Cultural resistance is real.</strong> Many communities see computers and AI as threats to traditional learning.</p><p><strong>Measurement is hard.</strong> How do you separate AI impact from novelty effect, Hawthorne effect, and selection bias?</p><p>But even if half the Nigerian gains are real and sustainable, we're looking at the potential to solve the global learning crisis within a decade. Imagine: every child on Earth with access to world-class personalized tutoring for less than $50.</p><h2><strong>The Uncomfortable Truth</strong></h2><p>The most unsettling part of this story isn't the technology - it's what it reveals about global inequality.</p><p>Rich countries have such good educational systems that AI makes them worse. Poor countries have such bad educational systems that AI makes them dramatically better. The same tool simultaneously increases and decreases human potential, depending on context.</p><p>We're about to find out whether AI will be the great equalizer or the great divider.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ezAR!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9b7cf276-645a-4836-b94d-e06e11d81632_1280x951.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ezAR!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9b7cf276-645a-4836-b94d-e06e11d81632_1280x951.png 424w, https://substackcdn.com/image/fetch/$s_!ezAR!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9b7cf276-645a-4836-b94d-e06e11d81632_1280x951.png 848w, https://substackcdn.com/image/fetch/$s_!ezAR!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9b7cf276-645a-4836-b94d-e06e11d81632_1280x951.png 1272w, https://substackcdn.com/image/fetch/$s_!ezAR!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9b7cf276-645a-4836-b94d-e06e11d81632_1280x951.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ezAR!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9b7cf276-645a-4836-b94d-e06e11d81632_1280x951.png" width="1280" height="951" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/9b7cf276-645a-4836-b94d-e06e11d81632_1280x951.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:951,&quot;width&quot;:1280,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!ezAR!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9b7cf276-645a-4836-b94d-e06e11d81632_1280x951.png 424w, https://substackcdn.com/image/fetch/$s_!ezAR!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9b7cf276-645a-4836-b94d-e06e11d81632_1280x951.png 848w, https://substackcdn.com/image/fetch/$s_!ezAR!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9b7cf276-645a-4836-b94d-e06e11d81632_1280x951.png 1272w, https://substackcdn.com/image/fetch/$s_!ezAR!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9b7cf276-645a-4836-b94d-e06e11d81632_1280x951.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div>]]></content:encoded></item><item><title><![CDATA[The Computational Monopoly: Who Rules the AI World?]]></title><description><![CDATA[The New York Times published an article about how AI computational power is distributed globally.]]></description><link>https://www.bulaev.net/p/the-computational-monopoly-who-rules</link><guid isPermaLink="false">https://www.bulaev.net/p/the-computational-monopoly-who-rules</guid><dc:creator><![CDATA[Serge Bulaev]]></dc:creator><pubDate>Fri, 29 Aug 2025 22:55:59 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!74T2!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fff13e6b6-695a-45b8-a3a5-54302db4f352_1280x881.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>The New York Times <a href="https://www.nytimes.com/interactive/2025/06/23/technology/ai-computing-global-divide.html">published</a> an article about how AI computational power is distributed globally.</p><p><strong>The real power landscape:</strong></p><ul><li><p>USA (Microsoft, AWS, Google) - 63% of all capacity</p></li><li><p>China (Alibaba, Huawei, Tencent) - 28%</p></li><li><p>Europe - a pitiful 4%</p></li><li><p>The rest of the world - 5%</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!74T2!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fff13e6b6-695a-45b8-a3a5-54302db4f352_1280x881.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!74T2!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fff13e6b6-695a-45b8-a3a5-54302db4f352_1280x881.jpeg 424w, https://substackcdn.com/image/fetch/$s_!74T2!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fff13e6b6-695a-45b8-a3a5-54302db4f352_1280x881.jpeg 848w, https://substackcdn.com/image/fetch/$s_!74T2!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fff13e6b6-695a-45b8-a3a5-54302db4f352_1280x881.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!74T2!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fff13e6b6-695a-45b8-a3a5-54302db4f352_1280x881.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!74T2!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fff13e6b6-695a-45b8-a3a5-54302db4f352_1280x881.jpeg" width="1280" height="881" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/ff13e6b6-695a-45b8-a3a5-54302db4f352_1280x881.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:881,&quot;width&quot;:1280,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:98823,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.bulaev.net/i/172304708?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fff13e6b6-695a-45b8-a3a5-54302db4f352_1280x881.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!74T2!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fff13e6b6-695a-45b8-a3a5-54302db4f352_1280x881.jpeg 424w, https://substackcdn.com/image/fetch/$s_!74T2!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fff13e6b6-695a-45b8-a3a5-54302db4f352_1280x881.jpeg 848w, https://substackcdn.com/image/fetch/$s_!74T2!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fff13e6b6-695a-45b8-a3a5-54302db4f352_1280x881.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!74T2!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fff13e6b6-695a-45b8-a3a5-54302db4f352_1280x881.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p></li></ul><p>Digital feudalism. Only 32 countries have specialized AI data centers, while the rest are forced to rent from tech lords.</p><p>The most cynical part: even US political allies like Kenya don't get preferential GPU access. Nvidia is controlled geopolitically more strictly than oil.</p><p>Of course there are attempts to build "sovereign AI infrastructures" - Brazil, India, and the EU are investing billions. Africa is launching its first major center with Nvidia chips, but it only covers 10-20% of demand.</p><p>It seems those who didn't manage to build computational infrastructure today will remain technological vassals for decades.</p><p>Computing has become the new gold. Only the deposits are controlled by very few.</p><h2><strong>The Energy War Behind the Silicon Curtain</strong></h2><p>What's especially striking about this race isn't just chip distribution, but the energy apocalypse unfolding behind the scenes. <a href="https://www.goldmansachs.com/insights/articles/ai-to-drive-165-increase-in-data-center-power-demand-by-2030">Goldman Sachs predicts</a> data center energy consumption will grow 160% by 2030. <a href="https://www.rand.org/pubs/research_reports/RRA3572-1.html">RAND Corporation calculated</a> that AI centers will need an additional 10 gigawatts in 2025 alone - more than the entire state of Utah consumes.</p><p>By 2027, that figure will grow to 68 gigawatts. Almost doubling in two years.</p><p>Imagine the scale. We're not just talking about technological supremacy - we're talking about nations' physical ability to provide enough electricity to power the digital future. Countries without energy infrastructure are automatically eliminated from the game.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Qvz4!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff2b18edf-efdd-4fac-a4c5-4476f0fc5d28_1280x868.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Qvz4!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff2b18edf-efdd-4fac-a4c5-4476f0fc5d28_1280x868.jpeg 424w, https://substackcdn.com/image/fetch/$s_!Qvz4!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff2b18edf-efdd-4fac-a4c5-4476f0fc5d28_1280x868.jpeg 848w, https://substackcdn.com/image/fetch/$s_!Qvz4!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff2b18edf-efdd-4fac-a4c5-4476f0fc5d28_1280x868.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!Qvz4!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff2b18edf-efdd-4fac-a4c5-4476f0fc5d28_1280x868.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Qvz4!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff2b18edf-efdd-4fac-a4c5-4476f0fc5d28_1280x868.jpeg" width="1280" height="868" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f2b18edf-efdd-4fac-a4c5-4476f0fc5d28_1280x868.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:868,&quot;width&quot;:1280,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:67961,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.bulaev.net/i/172304708?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff2b18edf-efdd-4fac-a4c5-4476f0fc5d28_1280x868.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Qvz4!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff2b18edf-efdd-4fac-a4c5-4476f0fc5d28_1280x868.jpeg 424w, https://substackcdn.com/image/fetch/$s_!Qvz4!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff2b18edf-efdd-4fac-a4c5-4476f0fc5d28_1280x868.jpeg 848w, https://substackcdn.com/image/fetch/$s_!Qvz4!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff2b18edf-efdd-4fac-a4c5-4476f0fc5d28_1280x868.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!Qvz4!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff2b18edf-efdd-4fac-a4c5-4476f0fc5d28_1280x868.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><h2><strong>Nvidia: The New OPEC of the Digital Era</strong></h2><p>Nvidia has become more than just a chip manufacturer - it's become the regulator of the global AI market. With a 65% market share in data center AI chips in 2023, the company essentially dictates the rules to everyone else.</p><p>But what's most interesting is how American export restrictions turned Nvidia's sales into a geopolitical weapon. In July 2025, <a href="https://www.cnbc.com/2025/07/15/nvidia-says-us-government-will-allow-it-to-resume-h20-ai-chip-sales-to-china.html">Nvidia announced</a> resuming sales of its H20 chips to China, but only after "assurances from the US government." <a href="https://www.euronews.com/business/2025/07/15/nvidia-to-sell-h20-chips-to-china-again-after-us-gives-export-approval">Sales had been frozen</a> since April - despite China bringing the company $12-15 billion annually.</p><p>Imagine: the world's largest economy waits for Washington's permission to buy processors. And gets watered-down versions with limited performance. <strong>This isn't trade - it's a technological blockade.</strong></p><p>According to CEO Jensen Huang, revenues from China fell by half compared to pre-crisis levels. But the Chinese aren't sitting idle: CSIS data shows large-scale H100 chip smuggling schemes operational by 2024. Eight different smuggling networks, whose participants gave interviews to The Information journalists.</p><p><strong>Black market processors - welcome to cyberpunk reality 2025.</strong></p><h2><strong>European Capitulation</strong></h2><p>Four percent. Europe - cradle of the industrial revolution, home to SAP, ASML, and Siemens - controls a pathetic 4% of global AI capacity.</p><p>This isn't just statistics, it's a death sentence for European technological sovereignty. While Brussels debated GDPR and AI ethics, Americans and Chinese built computational empires.</p><p>The EU is now investing hundreds of billions trying to catch up, but the moment has passed. <a href="https://www.goldmansachs.com/insights/goldman-sachs-research/generational-growth-ai-data-centers-and-the-coming-us-power-demand-surge">Goldman Sachs forecasts</a> nearly &#8364;850 billion in renewable energy investments over the next decade just to power future data centers.</p><p>But infrastructure takes years to build, while tech cycles are measured in months. By the time European AI centers are running at full capacity, the leaders will be two chip generations ahead.</p><h2><strong>Digital Colonies</strong></h2><p>The rest of the world - 5%. Africa, Latin America, most of Asia - all together control less computing power than the state of California.</p><p>Kenya's example is particularly telling. The country is considered a US ally, actively develops its digital economy, and has one of Africa's most dynamic IT sectors. But access to advanced GPUs? Sorry, same queue as everyone else. No perks. No discounts. Pay full price or rent capacity from tech giants.</p><p>This creates a vicious cycle: without their own computing power, countries can't develop their own AI models; without their own models, they remain dependent on Western and Chinese platforms; without technological independence, they can't develop high-tech industries.</p><p>Africa's first major AI center with Nvidia chips is an important step, but it covers at most 20% of continental demand. And demand is growing exponentially.</p><h2><strong>Sovereign Ambitions vs Economic Reality</strong></h2><p>Everyone talks about "sovereign clouds" and "national AI strategies." Brazil invests billions in its own infrastructure. India launches ambitious programs. Even relatively small countries try to create at least minimal computing capacity.</p><p>But reality is harsh: modern AI data centers cost billions of dollars, and technologies become obsolete so quickly that investments can depreciate within a couple of years.</p><p>PwC warns: in 2025, equilibrium between supply and demand for AI computing won't materialize. The deficit will only grow. And this means prices for computing capacity rental will skyrocket.</p><h2><strong>The New Cold War</strong></h2><p>Essentially, we're witnessing the formation of a new bipolar world - an American-Chinese duopoly in AI computing. 91% of all capacity is controlled by two powers that increasingly view each other as strategic adversaries.</p><p>Everyone else - including traditional US allies - finds themselves in the role of supplicants. Access to cutting-edge technology is now determined not by economic considerations, but by geopolitical loyalty.</p><p>And most troubling: unlike oil or gas, computing power can't simply be bought on the spot market. It's a strategic resource requiring long-term investments, specialized infrastructure, and most importantly, access to advanced chips.</p><p>And chips are controlled by one player. Who plays by Washington's rules.</p><h2><strong>What's Next?</strong></h2><p>The window of opportunity is rapidly closing. Every month of delay in building computational infrastructure means years of technological lag in the future.</p><p>Countries that don't manage to build their own AI capacity in the next 2-3 years risk becoming digital colonies for decades to come. Because the next generation of AI will require even more computing power, even more energy, even more specialized infrastructure.</p><p>Computing really has become the new gold. But the deposits are already divided. And no new ones are in sight.</p>]]></content:encoded></item><item><title><![CDATA[The “Electric Fence” Stopped Working Years Ago: Why Content Creation Helps Us Walk Through Outdated Barriers]]></title><description><![CDATA[Sometimes the fences we tiptoe around are already broken.]]></description><link>https://www.bulaev.net/p/the-electric-fence-stopped-working</link><guid isPermaLink="false">https://www.bulaev.net/p/the-electric-fence-stopped-working</guid><dc:creator><![CDATA[Serge Bulaev]]></dc:creator><pubDate>Tue, 26 Aug 2025 22:38:21 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/aac2f523-6183-419f-995d-07f7bedd48b0_1280x1252.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Sometimes the fences we tiptoe around are already broken. The &#8220;electric fence&#8221; that once kept us in check stopped working years ago, yet many of us still stay behind it. Out of habit. Out of fear. Out of assumptions about how things &#8220;should&#8221; be done.</p><p>That&#8217;s the strange thing about social boundaries: they often don&#8217;t exist anymore in the way we think they do&#8230; but they still live in our heads. And one of the most effective ways I&#8217;ve found to move past them is through content creation.</p><h2><strong>The Myth of the Electric Fence</strong></h2><p>The metaphor of the electric fence has cropped up in psychology, therapy, even literature. It is often used to describe:</p><ul><li><p>Imaginary limitations rooted in old narratives of insecurity and rejection.</p></li><li><p>Emotional fences that originate in childhood or past pain, protecting us once but restricting us now.</p></li><li><p>Social hierarchies that feel permanent despite shifting norms and structures.</p></li><li><p>Outdated cultural &#8220;rules&#8221; about who can speak, who can lead, who can contribute.</p></li><li><p>Misunderstood boundaries that confuse healthy self-protection with fear-based avoidance.</p></li></ul><p>In therapy circles, people are encouraged to make boundaries visible and healthy (instead of hidden and dangerous like an invisible electric line) so that others can connect safely. In social research, outreach and friendliness are shown to be welcomed far more often than we expect. And in inclusion work, fences are highlighted as cultural constructs that need to be dismantled so communities can thrive.</p><h2><strong>Why Content Creation Matters</strong></h2><p>Every time you put out a post, an article, or a reflection, it&#8217;s more than content. It&#8217;s a declaration that the fence doesn&#8217;t define you anymore. You&#8217;re saying: I&#8217;m willing to be seen, I&#8217;m open to connection, I&#8217;m willing to risk the small shock that never actually comes.</p><p>When I&#8217;ve worked with leaders who began sharing consistently, I&#8217;ve noticed two things happen at once. The external walls begin to crumble (distance between them and their teams or clients reduces). And the internal walls weaken too (their own narrative of &#8220;I&#8217;m not ready&#8221; starts losing power).</p><p>Take a friend of mine, Max Votek, as an example. He wasn&#8217;t trying to broadcast polished leadership lessons. He started with unpolished reflections, stories from work and life, things that felt real in the moment. That consistent show-up created conversations, collaborations, and momentum. Exactly the kind of &#8220;crossing&#8221; that dismantles imaginary fences.</p><h2><strong>More Than Personal Brand</strong></h2><p>At Co.Actor we focus less on the old language of &#8220;personal brand&#8221; and more on honesty and resonance. Our view is simple: authenticity beats polish, direct engagement beats formality. That&#8217;s why we&#8217;ve built Co.Actor scale technology - not to generate sterile content, but to keep each person&#8217;s actual voice consistent across what they publish. Because when someone hears you in your words, they connect with the human, not the persona.</p><h2><strong>A Few Key Shifts to Remember</strong></h2><ul><li><p>Sharing real perspectives opens space for others to step through their own fences.</p></li><li><p>Attitudes spread fast. One direct, vulnerable post can normalize openness across a circle of people.</p></li><li><p>Most fences aren&#8217;t live anymore. The shock we fear is imagined. The rulebook we think still applies often expired with the last era.</p></li><li><p>Respect is critical. Don&#8217;t confuse tearing down imaginary walls with ignoring healthy boundaries. Some fences exist for good reason.</p></li><li><p>Outreach works. Research shows that friendliness and initiative are usually welcomed. The imagined rejection in our heads shows up far less in actual practice.</p></li></ul><h2><strong>Why This Is Urgent</strong></h2><p>Our current climate is isolating. AI tools grow, work gets distributed, relationships shift online. Yet the craving for genuine connection is huge. The people who win in this environment aren&#8217;t the loudest self-promoters. They&#8217;re the connectors. Unafraid. The ones who realize the shock line was cut long ago.</p><h2><strong>The Challenge I&#8217;ll Leave You With</strong></h2><p>If you&#8217;ve been holding back because you don&#8217;t feel expert enough or polished enough or credentialed enough, stop waiting for permission. Write the next post that feels true, not the one that feels &#8220;safe.&#8221; Say hello to someone you admire without overthinking whether it&#8217;s allowed. Notice that the fear was usually the barrier, not the person on the other side.</p><p>The electric fence is off. The path ahead is open. Most of us are waiting for someone to step forward first.</p><p>So step.</p><p>If this resonated, subscribe for more reflections on leadership, content strategy, and the psychology of building genuine connections in a noisy digital world.</p>]]></content:encoded></item><item><title><![CDATA[Context Engineering with Anthropic: From Skiing Accidents to Reliable AI ]]></title><description><![CDATA[Recently I watched Anthropic's &#8220;Prompting 101&#8221; session, and it was one of the clearest illustrations of why large language models need more than clever wording.]]></description><link>https://www.bulaev.net/p/context-engineering-with-anthropic</link><guid isPermaLink="false">https://www.bulaev.net/p/context-engineering-with-anthropic</guid><dc:creator><![CDATA[Serge Bulaev]]></dc:creator><pubDate>Mon, 25 Aug 2025 22:49:36 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/dd1af960-1dc6-4967-89ee-0c8343ee2c36_1200x623.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Recently I watched <a href="https://www.youtube.com/watch?v=ysPbXH0LpIE&amp;ab_channel=Anthropic">Anthropic's &#8220;Prompting 101&#8221; session</a>, and it was one of the clearest illustrations of why large language models need more than clever wording. They need disciplined context engineering.</p><p>The team showed a scenario based on a real customer: a Swedish insurance company that wants to automate accident claim reviews. The system gets two inputs: a filled-out Swedish car accident report form (with 17 standardized checkboxes) and a rough, hand-drawn sketch of the collision. The model&#8217;s task is to determine what happened and who might be at fault.</p><p>At first, the results were almost comical. With a simple prompt, Claude confidently concluded that the case involved a skiing accident on a Swedish street. On some level it made sense: incomplete context pushed the model toward a plausible, but wildly incorrect, story.</p><p>The demo became an unfolding lesson in how to tighten prompts until they work reliably. A few of the strongest principles stood out:</p><ul><li><p>Define the task context upfront. Instruct the model specifically: "You are assisting a human claims adjuster reviewing car accident forms in Swedish." This small adjustment shifts reasoning away from distractions like skiing.</p></li><li><p>Set tone and confidence rules. The model must remain factual. If a checkbox is unclear or the sketch illegible, it should say so. Better to admit uncertainty than generate fiction.</p></li><li><p>Provide invariants. The car accident form structure never changes. By embedding the form schema into the system prompt, Claude doesn't waste time guessing how to read it each time. This is ideal for prompt caching.</p></li><li><p>Use structure and delimiters. Wrapping different inputs inside XML tags gives Claude clearer reference points. For example: &lt;formdata&gt; or &lt;sketchnotes&gt; lets the model reliably separate evidence types.</p></li><li><p>Work through sequence. The order mattered. Claude was told to carefully review the form first, then analyze the sketch second. Much like a human claims adjuster, it needed stable data before decoding a messy drawing.</p></li><li><p>Reinforce guidelines at the end. Explicit reminders helped: do not invent details, refer back to the boxes when making factual claims, answer only with confidence.</p></li><li><p>Shape the output. Wrapping the final result in &lt;final_verdict&gt; tags made the output concise and machine-parseable, ready to drop into a claims database.</p></li></ul><p>As the iterations continued, the difference was staggering. First: skiing. Second: some recognition of vehicles, but still gaps. Third version: Claude matched boxes to vehicle behaviors, compared them against the sketch, and concluded that Vehicle B was likely at fault. It cited evidence, stayed within confidence limits, and packaged the conclusion in structured XML.</p><p>That transformation is the essence of context engineering. Not magic words. Not prompt "hacks." Just systematic practice: define roles, add background knowledge, enforce structure, iterate.</p><p>Why does this matter? Because in the real world, whether you&#8217;re parsing accident reports, processing medical forms, or reviewing contracts, the cost of guessing is higher than the cost of saying "I don&#8217;t know." Enterprises will only trust AI if outputs are both accurate and explainable.</p><p>This video convinced me of one thing: the path from toy prompts to production AI runs through context. <strong>The more disciplined the structure, the more grounded the model becomes.</strong> What began as a skiing accident ended as a working prototype of an insurance claims system, with Claude acting not as a storyteller but as an assistant claims adjuster.</p><p>That&#8217;s the journey - from hallucinations to reliability - and it all comes down to how you engineer context.</p>]]></content:encoded></item><item><title><![CDATA[The Evolution from Prompt Engineering to Context Engineering]]></title><description><![CDATA[How the art of working with AI has shifted from crafting perfect prompts to mastering context orchestration]]></description><link>https://www.bulaev.net/p/the-evolution-from-prompt-engineering</link><guid isPermaLink="false">https://www.bulaev.net/p/the-evolution-from-prompt-engineering</guid><dc:creator><![CDATA[Serge Bulaev]]></dc:creator><pubDate>Fri, 22 Aug 2025 11:56:21 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/0a064212-5f5a-4712-b6fb-8d5120b9651d_1400x788.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>We're witnessing a fundamental shift in how we interact with AI systems. Prompt engineering - the practice of carefully crafting instructions to get better responses from language models - has quietly evolved into something more sophisticated: context engineering.</p><p>This isn't just a semantic change. It represents a deeper understanding of how modern AI systems actually work and what they need to perform at their best.</p><h2><strong>The Death of Perfect Prompts</strong></h2><p>Two years ago, the AI community was obsessed with finding the perfect prompt. We spent hours crafting elaborate instructions, testing different phrasings, and sharing "magic words" that seemed to unlock better performance.</p><p>That era is largely over.</p><p>Today's models - GPT-4, Claude, Gemini - are sophisticated enough that they don't need to be coaxed with perfect phrasing. They understand intent remarkably well, even from casual requests. The bottleneck has shifted from <em><strong>how</strong></em> you ask to <em><strong>what information</strong></em><strong> </strong>you provide.</p><p>This is the core insight: modern AI systems are limited not by their ability to understand instructions, but by their access to relevant information.</p><h2><strong>What Context Engineering Really Means</strong></h2><p>Context engineering is the art and science of filling the model's context window with exactly the information it needs for the task at hand. Think of it like preparing for a complex meeting - you wouldn't just show up and hope for the best.</p><p>The science involves systematic approaches:</p><ul><li><p><strong>Task descriptions and explanations</strong> that go beyond simple instructions to include goals, constraints, and success criteria</p></li><li><p><strong>Few-shot examples</strong> chosen for quality over quantity, illustrating the exact pattern you want</p></li><li><p><strong>Smart RAG</strong> that goes beyond keyword matching to sophisticated knowledge retrieval</p></li><li><p><strong>Multimodal data</strong> that includes images, audio, and structured data alongside text</p></li><li><p><strong>Tool integration and state management</strong> for complex workflows</p></li><li><p><strong>Information compression</strong> that distills vast amounts of data into relevant insights</p></li></ul><p>Art requires developing intuition for how language models "think." Understanding their biases, behavioral patterns, and what makes them perform well or poorly.</p><h2><strong>The Context Window Optimization Problem</strong></h2><p>The central challenge is optimization. Context windows have limits, and using them poorly creates three distinct problems:</p><p><strong>Too little context</strong> leaves the model without needed information. It makes assumptions or provides generic responses.</p><p><strong>Too much context</strong> increases costs and can actually hurt performance through "context dilution" - when there's so much information the model struggles to identify what's relevant.</p><p><strong>Wrong context</strong> might be extensive and accurate but irrelevant to the task, causing the model to miss the mark entirely.</p><h2><strong>Dynamic Systems Replace Static Templates</strong></h2><p>Modern context engineering uses adaptive systems rather than static prompt templates. These systems modify prompts based on context, user history, and task requirements.</p><p>A sophisticated system might start with a base template, then dynamically inject relevant examples, modify instructions based on user preferences, add tool descriptions based on available capabilities, and adjust detail levels based on request complexity.</p><p>This recognizes that the optimal prompt varies dramatically by context. Data analysis needs different instructions than creative writing. Novice users need different guidance than experts.</p><h2><strong>Smart RAG: Beyond Vector Search</strong></h2><p>Retrieval-Augmented Generation has evolved far beyond simple vector similarity. Modern RAG systems consider temporal relevance (newer might be better), source authority (some sources are more trustworthy), information completeness (partial info might be worse than none), and context coherence (information should work together).</p><p>The best systems understand that factual questions need different retrieval strategies than creative requests. Technical problems require different information than strategic decisions.</p><h2><strong>Memory Architecture: Short and Long-term</strong></h2><p>Context engineering requires sophisticated memory management. Short-term memory involves managing conversation history - deciding what parts of dialogue matter and how to compress longer conversations.</p><p>Long-term memory creates persistent knowledge structures: user preference profiles, knowledge graphs capturing concept relationships, searchable indexes of past interactions, and specialized databases for different information types.</p><p>The challenge is making information accessible when needed without overwhelming the model with irrelevant historical details.</p><h2><strong>The Psychology of Language Models</strong></h2><p>Language models are pattern-matching systems trained on human text. They excel when they can recognize patterns in your context and apply them to generate responses. They struggle with ambiguous, contradictory, or patternless context.</p><p>Understanding their quirks helps structure context effectively. Models are influenced by information order - recent information often carries more weight. They're sensitive to formatting in non-intuitive ways. They may fixate on irrelevant but prominent details.</p><p>This knowledge lets you place important information where models pay attention, use consistent formatting to signal importance, and remove distractions that lead models astray.</p><h2><strong>Multimodal Context Orchestration</strong></h2><p>As AI becomes multimodal, context engineering must consider how different data types work together. An image might provide visual context that's hard to describe. Audio captures tonal nuances text cannot convey. Structured data provides precise information natural language can't efficiently encode.</p><p>The art lies in understanding how these modalities complement each other and presenting them to help rather than confuse the model.</p><h2><strong>Economic and Performance Implications</strong></h2><p>Context engineering has real economic impacts. Token costs matter for high-volume applications. More importantly, larger contexts increase latency, affecting user experience.</p><p>This creates optimization problems. Sometimes multiple smaller requests beat one large request. Other times, the overhead of multiple requests makes comprehensive context more efficient. The optimal approach depends on use case, cost constraints, and performance requirements.</p><h2><strong>Practical Implications</strong></h2><p>For practitioners, this shift means:</p><p><strong>Focus on information architecture</strong> rather than prompt perfection. Understand what information your AI system needs and organize it effectively.</p><p><strong>Invest in intelligent retrieval systems</strong> that surface relevant information. Simple keyword search isn't sufficient for complex applications.</p><p><strong>Think systematically about memory management.</strong> Consider both what to preserve and what to forget as interactions grow complex.</p><p><strong>Develop model behavior intuition</strong> through experimentation. Understanding how models respond to different context types makes you more effective.</p><p><strong>Consider the full pipeline</strong> from information gathering through response generation. Context engineering encompasses the entire information flow.</p><p>What to read:</p><p><a href="https://www.promptingguide.ai/">Prompting Guide </a>- powerful guide on prompting techniques</p><p><a href="https://blog.langchain.com/the-rise-of-context-engineering/">The rise of "context engineering</a> - overview article on LangChain</p><p><a href="https://github.com/humanlayer/12-factor-agents">12 factors agents</a> - principles for building AI agents</p><p><a href="https://simple.ai/p/the-skill-thats-replacing-prompt-engineering?">Context Engineering</a> - going beyond prompts to push AI</p>]]></content:encoded></item><item><title><![CDATA[Inside OpenAI: A Developer's Perspective on the World's Most Watched AI Company]]></title><description><![CDATA[A developer who recently left OpenAI after a year shares candid insights into the culture, pace, and inner workings of the company at the center of the AI revolution.]]></description><link>https://www.bulaev.net/p/inside-openai-a-developers-perspective</link><guid isPermaLink="false">https://www.bulaev.net/p/inside-openai-a-developers-perspective</guid><dc:creator><![CDATA[Serge Bulaev]]></dc:creator><pubDate>Thu, 21 Aug 2025 12:11:17 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/e6651dcf-f368-4217-b77c-4d42c7c163d9_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>A developer who recently left OpenAI after a year shares candid insights into the culture, pace, and inner workings of the company at the center of the AI revolution.</em></p><p>Recently, a developer who spent a year inside OpenAI <a href="https://calv.info/openai-reflections">shared some fascinating details</a> about life inside what might be the most scrutinized company in the world. Having worked there during one of its most explosive growth periods, their observations offer a rare glimpse behind the curtain of the organization racing to build AGI.</p><h2><strong>Hypergrowth and the Slack-First Culture</strong></h2><p>The numbers are staggering: OpenAI tripled in size during this developer's tenure, growing from 1,000 to 3,000 employees in just twelve months. To put that in perspective, they became a top-30% tenured employee simply by surviving a single year at the company.</p><p>This kind of hypergrowth breaks everything. Traditional corporate communication structures, reporting hierarchies, product development processes - none of it scales when you're adding 2,000 people in a year. Most leadership teams are doing completely different jobs than they were two years ago, simply because the company underneath them has transformed entirely.</p><p>What's particularly striking is how OpenAI has adapted: they've gone all-in on Slack for everything. The developer received around 10 emails during their entire year there. Every conversation, every decision, every announcement happens in Slack channels. For someone coming from a traditional corporate environment, this is either liberating or completely overwhelming, depending on how well you curate your notifications.</p><p>This communications approach reflects something deeper about OpenAI's DNA. Unlike traditional companies with rigid hierarchical structures, OpenAI operates from the bottom-up. When the developer first arrived and asked about quarterly roadmaps, the answer was simple: "They don't exist." Good ideas can come from anyone, at any level, at any time. The challenge isn't getting approval for your ideas - it's proving they're worth pursuing.</p><h2><strong>The Three-Company Race to AGI</strong></h2><p>Perhaps the most sobering insight is how the developer frames the current AI landscape. In their view, the path to AGI has crystallized into a three-horse race: OpenAI, Anthropic, and Google. Each company is taking a fundamentally different approach based on their organizational DNA.</p><p>This isn't about incremental improvements or feature competition - this is about who will first create artificial general intelligence. The stakes couldn't be higher, and everyone inside these organizations knows it. The developer describes a culture where teams closely monitor what's happening at Meta, Google, and Anthropic, knowing that their competitors are doing exactly the same thing.</p><p>The pressure of this competition creates an interesting dynamic. On one hand, it drives incredible innovation and speed. On the other hand, it means operating under constant scrutiny from governments, media, and the global tech community. The developer regularly saw news about OpenAI in the press before it was announced internally - a surreal experience that underscores how much attention the company attracts.</p><h2><strong>Python Monorepos and Azure Reality</strong></h2><p>On the technical side, the insights reveal both the power and chaos of rapid scaling. OpenAI runs on a massive Python monorepo, with growing services in Rust. This creates a fascinating coding environment where you'll find both sophisticated libraries built by 10-year Google veterans sitting alongside hastily written Jupyter notebooks from newly-minted PhDs. There are no enforced style guides across the organization - a reflection of both the research culture and the speed at which they're moving.</p><p>The infrastructure story is equally telling. Everything runs on Azure, but only three services are considered truly reliable: Azure Kubernetes Service, CosmosDB, and BlobStore. Unlike AWS with its mature ecosystem of specialized services, Azure forces OpenAI to build more infrastructure in-house. It's a constraint that probably slows them down in some areas but gives them more control over their entire stack.</p><p>But here's the kicker: none of the infrastructure costs matter compared to GPU expenses. The developer shares a mind-bending data point: a single niche feature in the Codex product costs as much to run as an entire successful startup's infrastructure. When your primary constraint is access to the most powerful chips on Earth, everything else becomes a rounding error.</p><h2><strong>The Codex Sprint: Seven Weeks from Zero to Launch</strong></h2><p>The most detailed story comes from the Codex launch - a product built from scratch and shipped to millions of users in just seven weeks. This wasn't a small internal tool or limited beta; this was a full-featured coding assistant integrated into ChatGPT and made available to the world.</p><p>The pace was brutal. The team worked until 11-12 PM every night, got up at 5:30 AM, and worked weekends. For seven straight weeks. The night before launch, five team members stayed up until 4 AM deploying the system, then returned at 8 AM for the public launch announcement.</p><p>Within 53 days of launch, Codex had generated 630,000 public pull requests. That's roughly 78,000 public PRs per engineer on the team. The scale of impact is almost incomprehensible - most developers don't create that many meaningful code changes in their entire careers.</p><p>What made this possible wasn't just the technology, but the organizational structure. When the Codex team needed help from experienced ChatGPT engineers, they met with the engineering managers and had two senior developers ready to help the next day. No quarterly planning cycles, no bureaucratic approval processes - just immediate action when something important needed to get done.</p><h2><strong>The Twitter Influence Loop</strong></h2><p>One of the more unexpected insights is how much OpenAI pays attention to Twitter. The developer notes that if your tweet about OpenAI goes viral, there's a good chance someone at the company will read it and take it seriously.</p><p>This creates an interesting feedback loop. OpenAI is building products for hundreds of millions of users, many of whom express their opinions about AI on social media. The company's leadership stays tuned into these conversations, using them as signals alongside traditional analytics and user research.</p><p>It's also a reflection of how seriously OpenAI takes its public perception. Unlike most B2B companies that can operate relatively quietly, every OpenAI product launch becomes a global news event. Every feature update gets analyzed by researchers, journalists, and competitors around the world.</p><h2><strong>The Meritocracy of Ideas</strong></h2><p>The best ideas win, regardless of who proposes them. Leaders are promoted primarily based on their ability to generate good ideas and execute them, rather than traditional corporate skills like presentation abilities or political maneuvering.</p><p>The developer describes seeing 3-4 different Codex prototypes floating around before the team decided to push for an official launch. This redundancy might seem inefficient, but it ensures that the best approaches bubble up naturally rather than being decided by committee.</p><h2><strong>Security, Secrecy, and Stakes</strong></h2><p>The security culture at OpenAI is intense. The developer couldn't tell anyone what they were working on in detail. Different Slack workspaces have varying permission levels. Revenue and burn numbers are closely guarded secrets.</p><p>But despite the secrecy, the developer emphasizes that everyone they met was genuinely trying to do the right thing. OpenAI gets significant criticism in the press, partly because it's the most visible of the major AI labs. The company has maintained its commitment to making cutting-edge AI broadly accessible - anyone in the world can use ChatGPT, even without logging in, and most models quickly become available through the API for developers.</p><p>The safety work is more substantial than external critics might expect, though it focuses more on practical risks (hate speech, abuse, manipulation) than theoretical ones (intelligence explosion, power-seeking). The developer notes that much of this work isn't published, and suggests OpenAI should do more to communicate their safety efforts publicly.</p><h2><strong>What This Tells Us About the Future</strong></h2><p>These insights paint a picture of a company that's still operating more like a research lab than a traditional corporation, despite having 3,000 employees and hundreds of millions of users. The bottom-up culture, the bias toward action, the willingness to change direction quickly - these are the characteristics that allowed OpenAI to build and launch transformative products so quickly.</p><p>But they're also characteristics that become harder to maintain as organizations grow. The developer notes that many systems break under hypergrowth: communication, reporting structures, hiring processes. The question is whether OpenAI can maintain its innovative culture while building the operational capabilities needed to compete with tech giants like Google.</p><p>As we watch the race to AGI unfold, accounts like this provide crucial context for understanding not just what these companies are building, but how they're building it. The culture, the pace, the technical decisions, the human costs - all of these factors will shape what artificial general intelligence looks like when it finally arrives.</p><p>Whether OpenAI, Anthropic, or Google ultimately wins the race to AGI, one thing is clear: the pace of change is only accelerating, and the stakes have never been higher.</p>]]></content:encoded></item><item><title><![CDATA[Where Gender Bias Grows: How AI Reveals the Rigid Stereotypes Shaping Our Stories]]></title><description><![CDATA[A fascinating study out of Cornell shines a bright light on something we usually sense but rarely measure: how the roles assigned to male and female characters in literature have changed (or failed to) over the last century.]]></description><link>https://www.bulaev.net/p/where-gender-bias-grows-how-ai-reveals</link><guid isPermaLink="false">https://www.bulaev.net/p/where-gender-bias-grows-how-ai-reveals</guid><dc:creator><![CDATA[Serge Bulaev]]></dc:creator><pubDate>Fri, 15 Aug 2025 23:16:28 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/b8d351c4-985b-4a42-9e14-8e359c40f32e_1908x1058.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><a href="https://news.cornell.edu/stories/2025/06/where-gender-bias-grows-coming-age-novels-rife-stereotypes#:~:text=Coming,genre%20that%20isn%E2%80%99t%20so%20positive">A fascinating study out of Cornell</a> shines a bright light on something we usually sense but rarely measure: how the roles assigned to male and female characters in literature have changed (or failed to) over the last century. Using word embedding models - a method well known to anyone in AI or computational linguistics - researchers analyzed 303 coming&#8209;of&#8209;age novels, stretching from early 20th-century classics to contemporary YA bestsellers.</p><p>Here is the headline finding:</p><p><strong>Female characters have broken out of old boxes, taking on a much wider variety of roles than they once did.</strong></p><p><strong>Male characters, despite all the cultural changes in the past hundred years, are still largely stuck in the same narrow band of traits.</strong> It's like the archetype was set in stone and we've been chiselling around it for decades without ever breaking it open.</p><p>Here is what stood out most from the research and supporting studies:</p><ul><li><p><strong>Female role diversity has expanded dramatically</strong> - Early in the century, women in these novels were firmly anchored in domestic life, romance, or emotional caretaker roles. Now you see them as leaders, adventurers, scientists, morally complex anti&#8209;heroes, and sometimes all of those in one character.</p></li><li><p><strong>Male stereotypes show near-total stability</strong> - Dominance, agency, stoicism, invulnerability&#8230; the same masculine profile appears in 1925, in the 1970s, and again in 2025. The language and plots change, the character core doesn't.</p></li><li><p><strong>Patterns repeat across media</strong> - A 2024 computational study of song lyrics found almost identical associations: <strong>women tied to emotion, relationships, and vulnerability; men tied to independence, power, and action.</strong> Different art forms but the clusters hardly budge.</p></li><li><p><strong>Persistence suggests deep cultural reinforcement</strong> - Both literature and music feed into the same cultural reservoir, reproducing these gender&#8209;coded associations generation after generation.</p></li></ul><p>Here's where it gets interesting for anyone thinking about AI: these same literary patterns become the training data for language models. Every novel analyzed in that Cornell study, every song lyric showing those persistent associations - they're all part of the vast datasets that teach AI systems how language works. The models absorb a century of storytelling conventions, then reproduce them when they generate new text.</p><ul><li><p><strong>Even bias detection carries bias</strong> - The Cornell team echoes earlier findings that the word lists and metrics we use to "detect" bias are shaped by cultural norms themselves. The lens is never completely neutral, which means models can inherit prejudice from the tools measuring them.</p></li><li><p><strong>Circuit tracing reveals the route of stereotypes</strong> - This newer AI technique lets researchers literally follow the pathways inside embedding models. Swap "big" for "brave" or "gentle" and you can see how gendered associations persist along particular computational routes, surviving attempts to scrub them out.</p></li><li><p><strong>Embedding models capture history-in-language</strong> - Because they map word relationships based on huge datasets, they can show how some associations shift over decades while others remain stubbornly frozen. It's time&#8209;series cultural analysis in mathematical form.</p></li><li><p><strong>Psycholinguistics gives context</strong> - Rather than blaming individual authors, the lens here widens to show structural patterns over time. Stories become records of what society felt comfortable showing men and women to be.</p></li></ul><p>Looking at it this way, you realise it's not an AI problem in isolation. It's a data problem. The data is us - our books, our music, our conversations, the stories we elevate and repeat. And once those patterns are in the datasets, they ripple forward. AI models are trained on the cultural archive, then their output becomes part of the next archive, reinforcing or repeating the grooves.</p><p>The psycholinguistic side of this is the real kicker for me. When a model "learns" to link men with power and women with nurturing - it is faithfully reproducing patterns that have been deeply engrained in the source material. Literature gives us a perfect example because it spans such a long time horizon.</p><p>And yes, there's a practical angle. If we want new voices and new norms in both literature and AI output, this isn't a wait&#8209;and&#8209;see game. We have to deliberately make room for different stories - and that means everything from what publishers choose to promote to which datasets researchers use to train the next generation of models.</p><p>It means amplifying narratives that break the moulds, not just adding them to the pile. Without that intentional shift, the algorithms of 2050 will still be serving us men with clenched jaws and women with teary eyes, even if they're describing life on Mars.</p><p>This research doesn't shame art - it holds up a mirror. And it asks what kind of reflection we want to leave for the next century to measure.</p>]]></content:encoded></item></channel></rss>