<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[Musings by Chris Hayduk]]></title><description><![CDATA[Reading widely and diving deeply]]></description><link>https://www.chrishayduk.com</link><image><url>https://substackcdn.com/image/fetch/$s_!UbBK!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F27ef8414-9281-48f9-b72c-6c8dc93c7664_799x799.png</url><title>Musings by Chris Hayduk</title><link>https://www.chrishayduk.com</link></image><generator>Substack</generator><lastBuildDate>Wed, 29 Apr 2026 19:13:08 GMT</lastBuildDate><atom:link href="https://www.chrishayduk.com/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Chris Hayduk]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[chrishayduk@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[chrishayduk@substack.com]]></itunes:email><itunes:name><![CDATA[Chris Hayduk]]></itunes:name></itunes:owner><itunes:author><![CDATA[Chris Hayduk]]></itunes:author><googleplay:owner><![CDATA[chrishayduk@substack.com]]></googleplay:owner><googleplay:email><![CDATA[chrishayduk@substack.com]]></googleplay:email><googleplay:author><![CDATA[Chris Hayduk]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[Reading Recap: Q1 2026]]></title><description><![CDATA[Brief reviews of the books and papers I read from Jan-March 2026]]></description><link>https://www.chrishayduk.com/p/reading-recap-q1-2026</link><guid isPermaLink="false">https://www.chrishayduk.com/p/reading-recap-q1-2026</guid><dc:creator><![CDATA[Chris Hayduk]]></dc:creator><pubDate>Sun, 19 Apr 2026 22:38:44 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!DH9B!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff66de1e8-703f-4367-b54e-cdd7bbffb9b2_1536x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!DH9B!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff66de1e8-703f-4367-b54e-cdd7bbffb9b2_1536x1024.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!DH9B!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff66de1e8-703f-4367-b54e-cdd7bbffb9b2_1536x1024.png 424w, https://substackcdn.com/image/fetch/$s_!DH9B!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff66de1e8-703f-4367-b54e-cdd7bbffb9b2_1536x1024.png 848w, https://substackcdn.com/image/fetch/$s_!DH9B!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff66de1e8-703f-4367-b54e-cdd7bbffb9b2_1536x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!DH9B!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff66de1e8-703f-4367-b54e-cdd7bbffb9b2_1536x1024.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!DH9B!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff66de1e8-703f-4367-b54e-cdd7bbffb9b2_1536x1024.png" width="1456" height="971" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f66de1e8-703f-4367-b54e-cdd7bbffb9b2_1536x1024.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:971,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2297740,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.chrishayduk.com/i/192876062?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff66de1e8-703f-4367-b54e-cdd7bbffb9b2_1536x1024.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!DH9B!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff66de1e8-703f-4367-b54e-cdd7bbffb9b2_1536x1024.png 424w, https://substackcdn.com/image/fetch/$s_!DH9B!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff66de1e8-703f-4367-b54e-cdd7bbffb9b2_1536x1024.png 848w, https://substackcdn.com/image/fetch/$s_!DH9B!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff66de1e8-703f-4367-b54e-cdd7bbffb9b2_1536x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!DH9B!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff66de1e8-703f-4367-b54e-cdd7bbffb9b2_1536x1024.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The idea for this series started with <a href="https://x.com/ChrisHayduk/status/2038059730476704025?s=20">a tweet I made</a> a few weeks ago.</p><p>Then, a couple of weeks later, <span class="mention-wrap" data-attrs="{&quot;name&quot;:&quot;Anna G&#225;t&quot;,&quot;id&quot;:5533222,&quot;type&quot;:&quot;user&quot;,&quot;url&quot;:null,&quot;photo_url&quot;:&quot;https://substackcdn.com/image/fetch/$s_!O5od!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F174b3cbe-5f37-4524-8d92-847b10416022_399x399.jpeg&quot;,&quot;uuid&quot;:&quot;3af6e801-8e1f-4fba-8263-bcb16527242b&quot;}" data-component-name="MentionToDOM"></span> posted a fantastic essay titled <a href="https://american-innocence.com/p/the-sovereign-reader">The Sovereign Reader</a> , which contained this wonderful quote:</p><blockquote><p>The truth is the more unique and personalized the books you read, the more original a thinker you will become. This, in short, is how you become you. I am a strong post-Hegelian believer in the personal duty of coming into our full being throughout our lives. Other than finding a fitting occupation and worthy life companions, cultivating your own mind is the prerequisite for building an existence for yourself that is truly yours.</p></blockquote><p>This quote connected with the ideas I&#8217;d been toying with in my earlier tweet &#8212; many of the most interesting people in the world have likely cultivated themselves and their minds in this sense, and thus, peering into their bookshelves would help us understand the inner self that they cultivated. And I would kill to see what they read, when they read it, and what they thought of it!</p><p>So I thought, &#8220;why not be the change you want to see in the world?&#8221; and decided to share what I&#8217;m reading much more often (and hopefully this inspires some people far more interesting than me to also share what they&#8217;re reading!)</p><p>The result is this article, the first in a quarterly update series in which I list, in chronological order, all the books I&#8217;ve completed this quarter. I&#8217;ll also list some of the notable papers I read in the quarter, though this list won&#8217;t be exhaustive, since my skim-to-deep-read ratio is much higher for papers than for books! Some of these papers and books will have brief reactions and/or reviews accompanying them if they especially stood out to me.</p><p>I hope you enjoy!</p><div><hr></div><h2>Books</h2><p><strong>#1. Claude Hopkins, </strong><em><strong>Scientific Advertising</strong></em></p><p>A pretty useful book capturing the earliest forms of approaches now taken for granted in the advertising industry &#8212; namely, taking a scientific approach to advertising through A/B testing aggressively and tracking conversions tied to each ad type, rather than trying to &#8220;plan&#8221; ad creative without customer feedback.</p><p>Interesting to read if you are in/adjacent to the ad space. Also contains useful nuggets on how to think about advertising. The one concept that really stuck with me is that advertising is <em>not </em>for building brand awareness, getting your name out there, or any other ancillary goal. Instead, it is just sales at scale. As a result, when writing your ads, you should imagine that you are standing in front of <em>one</em> of your target customers and trying to sell him or her your product. What would you say to this one person? How long would your pitch be? What would you do to legitimize yourself, your company, and your product to this one person? Think through these questions and write your ad based on your answers.</p><p><strong>#2. Alexandre Dumas, </strong><em><strong>The Count of Monte Cristo</strong></em></p><p>I don&#8217;t really know where to begin with this book because I loved it so much. Read it if you haven&#8217;t. Re-read it if you have. It hooks you from the first page and doesn&#8217;t let go for its full 1,300+ pages.</p><p>If there is one theme I can give you that stuck with me through the book, it is an interesting parallel to <em>Crime and Punishment</em>. The protagonists in both books, Edmond Dant&#232;s here and Raskolnikov in <em>Crime and Punishment</em>, have these ideas of grandeur, of being a superior man who does not need to obey the laws and customs of inferior men. But when they come into contact with human connection, those ideas fall apart. Dant&#232;s wavers in his plans to kill Albert de Morcerf after talking to Mercedes. He also resolves to save Valentine de Villefort after Maximillien Morrel reveals that he is in love with her. These contacts with love and social connection cause him to change his divine mission of Providence and remove some of the suffering he had planned for his enemies. Similarly, when Raskolnikov is alone, he frequently thinks of his superiority to others and hates those around him. But when he comes into contact with others, especially those who inspire tenderness in him, like Sonia Marmeladov or her father when he was dying, he behaves extremely altruistically and without any consideration for his ideals that he is above the common man, and acts in any way that he wants as long as it suits his aims.</p><p>To connect both to Nietzsche (who was explicitly influenced by <em>Crime and Punishment </em>and likely influenced by <em>The Count of Monte Cristo</em>) and his conception of the &#220;bermensch<em>,</em> both authors seem to assert that removing oneself from the rest of humanity is necessary to become the &#220;bermensch. </p><p>For Dostoevsky, this task is impossible and results in the destruction of the individual who attempts it. We find salvation only by connecting with humanity, not by setting ourselves apart. By contrast, for Dumas, this represents a genuine tension &#8212; Dant&#232;s <em>is</em> able to set himself apart and <em>does </em>act as a superior man, or &#220;bermensch. And this does bring him fulfillment in some ways (namely, avenging his wrongful imprisonment and his father&#8217;s death). However, this choice <em>also</em> precludes his happiness in other ways, as while acting as the Count, he must make choices that hurt those closest to him.</p><p><strong>#3. Dan Wang, </strong><em><strong>Breakneck: China&#8217;s Quest to Engineer the Future</strong></em></p><p>I found Wang&#8217;s now-famous framing of the US as &#8220;the lawyerly society&#8221; and China as &#8220;the engineering state&#8221; very illuminating &#8212; it is a neat little framework that helps you make sense of the many strengths and weaknesses of each country. However, I thought the most useful part of the book was not its thesis or the framework it introduced, but rather the stories Wang tells throughout. He lived in China from 2017 to 2023, experiencing the country&#8217;s rapid economic growth and the way this continually reshaped society firsthand. But he also experienced the early authoritarian turns instituted by Xi Jinping, as well as the COVID-19 pandemic and dramatic lockdown of Shanghai in 2022. Getting an insider&#8217;s view of Chinese society in this period was clarifying in a way that the many economic and historical analyses of China are not.</p><p>And, of course, <a href="https://danwang.co/">Dan Wang&#8217;s annual letters</a> are self-recommending for more of this type of writing. </p><p><strong>#4. Eric Berger, </strong><em><strong>Liftoff: Elon Musk and the Desperate Early Days That Launched SpaceX</strong></em></p><p>At this point, I&#8217;ve read a few books on Elon and listened to many podcasts about him or including him, so there wasn&#8217;t a lot in the way of surprises in <em>Liftoff</em>. What this book does well, however, is to impart the <em>feeling</em> of the life-or-death struggle that followed Elon around in those first few years at SpaceX. The company existed on the razor&#8217;s edge for essentially all of its early life, and we get to see the intensity with which Elon drove his employees (and himself) to ensure the company survived.</p><p><strong>#5. Ali Aminian, </strong><em><strong>Generative AI System Design Interview</strong></em></p><p>Fairly useful for interview prep, pairs well with the other books in the series (<em>Machine Learning System Design Interview </em>and <em>System Design Interview</em> Volumes I &amp; II).</p><p><strong>#6. Jacques Hadamard, </strong><em><strong>The Mathematician&#8217;s Mind: The Psychology of Invention in the Mathematical Field</strong></em></p><p>In this book, Jacques Hadamard asks the question, &#8220;How do the world&#8217;s greatest mathematicians and physicists actually make their field-defining discoveries? What is going on inside their heads when they solve these supremely difficult problems?&#8221; </p><p>Being an elite mathematician himself, Hadamard was able to interview some of the greatest minds of his era to learn the answer to this question. The book compiles a list of vignettes from Poincar&#233; to Einstein, probing the question from different angles, and essentially lands on the following answer &#8212; the rational, conscious mind is useful for getting us started on a problem and for verifying an answer once we have it, but the actual work of producing a novel insight is done by the unconscious mind. The discovery process he deduced from his observations, and that he asserts all great mathematical discoveries abide by, is as follows:</p><ol><li><p><strong>Preparation</strong> (primarily conscious) &#8212; the conscious mind focuses on a problem for an extended period of time, collecting relevant information and trying out several avenues for solution</p></li><li><p><strong>Incubation</strong> (primarily unconscious) &#8212; the unconscious mind, directed in its goals by the focus of the conscious mind in the Preparation stage, sets to work searching for high-level solutions. This is where the bulk of problem-solving and discovery is actually done. The unconscious mind is better at viewing the problem as a &#8220;whole&#8221; and at uncovering unexpected insights and connections than the conscious mind. The unconscious mind evaluates proposed solutions based on aesthetic criteria.</p></li><li><p><strong>Illumination</strong> (primarily unconscious) &#8212; an idea generated by the unconscious mind that satisfies the unconscious criteria springs forth into the conscious mind</p></li><li><p><strong>Verification</strong> (primarily conscious) &#8212; the conscious mind sets to work translating the unconscious&#8217;s idea into formal mathematical language and verifies that it is logically correct.</p></li></ol><p>This book inspired my recent essay, <a href="https://www.chrishayduk.com/p/the-unreasonable-effectiveness-of">The Unreasonable Effectiveness of LLMs in Mathematics</a>.</p><p><strong>#7. Stendhal, </strong><em><strong>The Red and The Black</strong></em></p><p>After reading <em>Crime and Punishment</em> last year, I&#8217;ve become slightly obsessed with the crisis in literature and philosophy that Napoleon presented after his meteoric rise and fall in the late 18th and early 19th centuries. I&#8217;ll start by giving a brief sketch of how I see the situation and the intellectual crisis that followed.</p><p>For the 300 or so years prior to Napoleon, philosophy had been building up to a notion of equality among people. The Enlightenment supercharged this, popularizing ideas about egalitarianism, self-determination, and natural rights. The American Revolution then put these abstract ideas into practice, and was quickly followed by the French Revolution, which sought to put an even more radical version of Enlightenment ideals into practice.</p><p>The Enlightenment&#8217;s ideas on political philosophy and these two revolutions were both rebelling against the prevailing traditional doctrines of the time. Power and the right to rule were seen as hereditary &#8212; God ordained a specific line to rule, and this right was passed down from father to eldest son in an unbroken chain. Moreover, the structure of society itself was seen as a crucial part of any nation. The peasants were meant to be peasants, the rulers were meant to rule. To go against this was to go against the natural order ordained by God.</p><p>These two groups were locked in an existential struggle in the late 1700s when Napoleon erupted onto the scene, forever changing the trajectory of the world. Through unparalleled competence, energy, and charisma,  he became emperor of France, and then nearly the emperor of Europe. He reshaped society at the high-level scale of large historic battles, as well as at the granular scale of laws, regulations, and standards. By the time of his final defeat, there was not one piece of European life that did not have Napoleon&#8217;s fingerprints on it.</p><p>His almost-mythic life refuted both groups&#8217; ideals.</p><p>The traditionalists asserted that humans <em>can</em> influence history, but that this is only the providence of hereditary kings (i.e., greatness comes from family lineage). The new ideas of the French Revolution hold that humans cannot influence history and that all humans are equal in this impotence &#8212; no one is inherently greater or more capable than anyone else. Napoleon refutes both &#8212; he <em>is</em> history, and his will controls the direction of events. But he is <em>not </em>a hereditary monarch; he is an upstart rising on his own merits. At once, he says to the monarchs, &#8220;Some men are destined to rule, but you are not those men. You don&#8217;t deserve your lot,&#8221; and says to the egalitarians, &#8220;You are not equal to me. None could accomplish what I have done&#8221;. </p><p>However, he also synthesized both positions. The egalitarians are right that monarchs are not predestined to rule &#8212; Napoleon shows that there are others with greater merit than the hereditary kings of Europe. He also validates the traditionalists that people are not equal &#8212; he is clearly more fit to rule than those who preceded him in the French Revolution.</p><p>This is what caused the crisis in literature and philosophy of the 19th century &#8212; whether you&#8217;re an egalitarian or a traditionalist, how do you explain the problem of Napoleon? And what does the fact that something like him occurred imply about the world?</p><p>Now, with all that background context out of the way, Stendhal tackles the problem of Napoleon from a sociological angle. He observes that Napoleon arose in a time of profound change, with France submerged in chaos due to the Revolution, and asks, &#8220;What happens when a Napoleonic figure arises in a society that is totally resistant to change?&#8221;</p><p>His answer is the character Julien Sorel, an extremely intelligent, Napoleon-obsessed youth from the lower class of the fictional town of Verri&#232;res, set during the Bourbon Restoration. Like his idol, Sorel possesses a formidable mind and intense ambition to rise from his lowly starting place in society. But unlike his idol, Sorel&#8217;s France does not reward competence and ambition. Instead, it places a premium on playing social roles and playing them well. As a result, Sorel&#8217;s ambition, which in an earlier age would have been channeled into great acts of heroism on the battlefield (or, in a future age, into founding a great company), was instead channeled into playing the roles that society valued in order to advance. </p><p>Thus, we see Sorel become a priest despite not believing in God, start affairs with various upper-class women without loving them, and tutor Latin despite not enjoying teaching Latin. He does these things because they are the legible paths by which a member of the lower class can aspire to move towards the upper class (but never fully join it).</p><p>Sorel is acutely aware of the incongruity between Napoleon&#8217;s life and his own &#8212; he frequently ruminates that his own quest for power, which involves lying and assuming social roles, is pathetic compared to Napoleon&#8217;s world-changing deeds. This incongruity forms the heart of the book and demonstrates the necessity of favorable cultural and societal circumstances for a Napoleonic figure to be fully actualized.</p><p><strong>#8. Fyodor Dostoevsky, </strong><em><strong>Notes from Underground</strong></em></p><p>Dostoevsky remains a writer whose prose is fairly unenjoyable for me to read, but whose themes and ideas sit with me for long after I have completed the book. <em>Notes from Underground</em> will likely require several re-reads for me to fully absorb its points, but on my first pass, I came away with the following major takeaway: Dostoevsky strongly opposes the systematization of humanity.</p><p>In the first part of the book, the Underground Man spends much of his time asserting that the utilitarian ethic, namely, converting the good to a function that must be maximized, does not lead to fulfillment. And, even more strikingly, he says that elevating this conception of life to the highest good would necessitate the loss of free will. If we are optimizing a value function, there is a set of actions that maximize it. If we could wave a magic wand and have this set of actions revealed to us, then it would be congruous with our stated highest principle to follow these actions to the letter and incongruous to assert any form of free will. Hence, the highest form of &#8220;good&#8221; in the utilitarian perspective necessitates a loss of free will, a reduction of humans to automatons. The Underground Man states this and then also observes that, if you were to reveal this value function to people, they would act against it just to spite it and assert their free will. And if you forced humans to live this way and renounce their free will, it would be spiritual torture. Hence, the rational, utility-maximization approach is anti-human.</p><p>In the second part of the book, the Underground Man seems to espouse a philosophy that, on its surface, is diametrically opposed to utility maximization. Instead, he seems to be living his life through the most exaggerated form of performative Romanticism. He constantly makes irrational decisions based on fleeting passions that overtake him. However, he is very self-conscious about these actions &#8212; he does not act immediately upon feeling the emotion. Instead, he thinks that he <em>should </em>take action x in response to feeling y and ruminates on the fact that he <em>didn&#8217;t </em>take action x in the moment (sometimes for months at a time). It is only after berating himself that he then takes his &#8220;passionate&#8221; action in response to his emotions. In this way, the Underground Man is forcing himself to exist within a system that constrains his behavior &#8212; it is different and less legible than the utilitarian perspective he railed against in the first part of the book, but it nonetheless serves to constrain his freedom of will substantially, and leads to a similar spiritual sickness.</p><p>In both cases &#8212; through the Underground Man&#8217;s words in part 1 and through his actions in part 2 &#8212; we see the failures of forcing humanity into a behavioral box. As far as I can tell, Dostoevsky aimed to show that by constraining human behavior, we harm ourselves and become something like the Underground Man.</p><p><strong>#9. Sally Smith Hughes, </strong><em><strong>Genentech: The Beginnings of Biotech</strong></em></p><p>A really fun history of the first major biotech company and the recombinant DNA technology that enabled it. It&#8217;s interesting to see the obstacles that Herb Boyer and Bob Swanson needed to overcome in order to make Genentech a reality &#8212; from raising capital from highly skeptical investors to convincing scientists that joining a company could be both scientifically meaningful and monetarily rewarding, the pair had to navigate so many issues that we now take for granted in startups. The contrast is stark with today&#8217;s AI boom, where investors will hand out billions to newly founded AI labs without the faintest hint of a business plan or even a differentiating idea.</p><div><hr></div><h2>Papers</h2><p><em>For the papers I don't have much to say about beyond the contribution itself, I'll just give the challenge and the solution. That is, I&#8217;ll state the main challenge the paper addresses, followed by a brief description of the solution it proposes.</em></p><p><strong>#1. Chai et al.,</strong> <em><strong>MultiPath: Multiple Probabilistic Anchor Trajectory Hypotheses for Behavior Prediction </strong></em><strong>(2019)</strong></p><p><strong>Challenge: </strong>Predicting future motion for agents in a scene is challenging because the set of potential outcomes is highly uncertain and multi-modal. Any approach that predicts only one future trajectory presents inherent problems. For example, how many times do we need to sample from this model to be confident that a pedestrian won&#8217;t jaywalk?</p><p><strong>Solution: </strong>The paper introduces MultiPath, an approach that uses a fixed set of anchor trajectories (namely, the modes extracted from the dataset). The model then learns to output a probability distribution over these anchors, as well as an offset Gaussian that measures the deviation from the anchor (captured by the mean) and the position&#8217;s uncertainty (captured by the covariance). This setup allows us to model two kinds of uncertainty:</p><ol><li><p><em>Intent uncertainty</em>, captured by the distribution over anchor trajectories</p></li><li><p><em>Control uncertainty</em>, captured by the probabilistic distribution over the agent&#8217;s state at each time step, conditioned on selecting a particular anchor.</p></li></ol><p><strong>#2. Nayakanti et al., </strong><em><strong>Wayformer: Motion Forecasting via Simple &amp; Efficient Attention Networks </strong></em><strong>(2022)</strong></p><p><strong>Challenge:</strong> Self-driving cars receive inputs of many different modalities (velocities, positions, images, LiDAR, road graphs, traffic light states, etc.) and types (static vs. dynamic). Many solutions rely on combining results from modality-specific models, leading to highly complex and brittle systems.</p><p><strong>Solution:</strong> Introduces Wayformer, a family of attention-based architectures for motion forecasting that are simple and homogeneous. Wayformer offers a compact model architecture comprising an attention-based scene encoder and a decoder. Ablations showed that early fusion of modalities in this architecture performed best. Achieved SOTA on Waymo Open Dataset.</p><p><strong>#3. Gulino et al., </strong><em><strong>Waymax: An Accelerated, Data-Driven Simulator for Large-Scale Autonomous Driving Research </strong></em><strong>(2023)</strong></p><p><strong>Challenge:</strong> Simulation is crucial for generating larger training datasets and evaluating autonomous vehicle planning software. However, existing simulations are typically very expensive to run, may not reliably match real-world data when they generate scenes, and may not exhibit realistic agent behavior.</p><p><strong>Solution:</strong> Waymax solves these problems with simulation by:</p><ol><li><p>Improving speed through writing the simulator in JAX. This allows Waymax to be run on hardware accelerators (GPUs/TPUs)</p></li><li><p>Initializing simulations from real-world data in the Waymo Open Dataset. This grounds simulations in real trajectories.</p></li><li><p>Using agent models to control the dynamic objects in the scene.</p></li></ol><p><strong>#4. Jumper et al., </strong><em><strong>Highly accurate protein structure prediction with AlphaFold</strong></em><strong> (2021)</strong></p><p>This was not my first time through this seminal paper (or the second or the third&#8230;), but it <em>was </em>the first time I went through it with the depth required to <a href="https://github.com/ChrisHayduk/minAlphaFold2/tree/main">reimplement AlphaFold2 from scratch</a> (note: still very much a work in progress!).</p><p>The AlphaFold2 paper itself is a relatively easy read, providing a brief overview of how the model works, how it fits within the frameworks of previous approaches, and the unprecedented performance it achieves. But it does leave a sizeable gap between a high-level understanding of what the model is doing and a <em>true</em> understanding of its inner workings.</p><p>For that, we need to dive into the 62-page supplementary paper (a full 5 times longer than the main paper!). </p><p>I think this piece is where the ingenuity of the AlphaFold team, as well as their clarity of insight, really shines. Each component of the model, labeled as an algorithm in the paper, is explained using a combination of text, diagrams, and complete pseudocode. In addition, the paper clearly outlines how each algorithm is stitched together in higher-level functions, such as the AlphaFold2 training loop or inference pass. And if that&#8217;s not enough, it even goes into great detail on the training setup &amp; parameters, ablation studies to determine what parts of the network are truly necessary, and visualizations of the attention matrices to see what the model has truly &#8220;learned&#8221;.</p><p>To me, the clarity of the supplementary paper felt as if it were screaming out to me that it <em>needed</em> a code implementation that was just as clear. One that was not intended for production use or included dozens of Dockerfiles, deployment scripts, and ancillary supporting files, but instead emphasized clarity, pedagogy, and alignment with the paper's algorithm breakdown.</p><p>This, along with Andrej Karpathy&#8217;s minGPT, was the inspiration for my <a href="https://github.com/ChrisHayduk/minAlphaFold2/tree/main">minAlphaFold2</a> project. The goal of this project is to align the model exactly with the supplementary paper&#8217;s structure and pseudocode, using only PyTorch primitives, with an emphasis on simplicity, clarity, and explanatory comments. Once the code portion of the project is complete, stay on the lookout for an &#8220;Annotated AlphaFold2&#8221;, in the same vein as Harvard&#8217;s <a href="https://nlp.seas.harvard.edu/annotated-transformer/">Annotated Transformer</a> project.</p><p>AlphaFold2 (along with GPT-3) represents one of the two largest breakthroughs in the history of artificial intelligence. And yet the pedagogical material available for AF2 is essentially non-existent when compared to the pedagogical material available for GPT-3-style models. I hope to do a small part in changing this.</p><p><strong>#5. Zhu et al., </strong><em><strong>Scaling Latent Reasoning via Looped Language Models </strong></em><strong>(2025)</strong></p><p>So my brief speculation tied to this paper caused quite a stir on ML Twitter:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!1LZT!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fabe70d9a-6c56-4dad-b9c5-d0470d7f988a_1206x1642.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!1LZT!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fabe70d9a-6c56-4dad-b9c5-d0470d7f988a_1206x1642.png 424w, https://substackcdn.com/image/fetch/$s_!1LZT!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fabe70d9a-6c56-4dad-b9c5-d0470d7f988a_1206x1642.png 848w, https://substackcdn.com/image/fetch/$s_!1LZT!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fabe70d9a-6c56-4dad-b9c5-d0470d7f988a_1206x1642.png 1272w, https://substackcdn.com/image/fetch/$s_!1LZT!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fabe70d9a-6c56-4dad-b9c5-d0470d7f988a_1206x1642.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!1LZT!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fabe70d9a-6c56-4dad-b9c5-d0470d7f988a_1206x1642.png" width="1206" height="1642" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/abe70d9a-6c56-4dad-b9c5-d0470d7f988a_1206x1642.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1642,&quot;width&quot;:1206,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:696747,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.chrishayduk.com/i/192876062?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fabe70d9a-6c56-4dad-b9c5-d0470d7f988a_1206x1642.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!1LZT!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fabe70d9a-6c56-4dad-b9c5-d0470d7f988a_1206x1642.png 424w, https://substackcdn.com/image/fetch/$s_!1LZT!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fabe70d9a-6c56-4dad-b9c5-d0470d7f988a_1206x1642.png 848w, https://substackcdn.com/image/fetch/$s_!1LZT!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fabe70d9a-6c56-4dad-b9c5-d0470d7f988a_1206x1642.png 1272w, https://substackcdn.com/image/fetch/$s_!1LZT!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fabe70d9a-6c56-4dad-b9c5-d0470d7f988a_1206x1642.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>This was not meant to be taken as fact (or to blow up the way it did)! I was mostly just speculating based on the patterns I saw in the benchmark scores for Claude Mythos and some inferences on Anthropic&#8217;s compute budget. But here I&#8217;ll give a brief explanation of <em>why </em>it seems plausible to me that this architectural choice has made its way into the frontier models.</p><p>First, I&#8217;d like to point out (and thank you to <a href="https://x.com/ChrisHayduk/status/2042952450097778958?s=20">Kalomaze on Twitter</a> for setting this straight), that I <em>don&#8217;t</em> think the frontier models are using the main contribution from this paper (namely, the ability for the model to choose how many times it loops based on a prediction head that predicts the probability it has already arrived at the correct token). My claim is now a bit weaker &#8212; that the models <em>may</em> be trained using weight tying. That is, the model loops a deterministic number of times, feeding the last layer&#8217;s output embedding back into the input layer. This essentially increases the model's depth without increasing its total parameter count. As such, I would expect the training process to look more like recycling in AlphaFold2 (see supplementary paper linked above) rather than looping in the ByteDance paper.</p><p>Anyway, with that context out of the way, I&#8217;ll give a brief overview of <em>why</em> I think looping seems plausible for frontier models. First, we&#8217;ll start with the benchmark scores.</p><p>Since looping increases model depth without increasing parameter count, what you end up with is a decoupling of the model&#8217;s knowledge (i.e., what it has memorized) and the model&#8217;s reasoning (i.e., its ability to manipulate that knowledge).</p><p>To explain this phenomenon, I&#8217;ll include a screenshot from the ByteDance paper along with a brief explanation of this image from my piece <a href="https://www.chrishayduk.com/p/the-unreasonable-effectiveness-of">The Unreasonable Effectiveness of LLMs in Mathematics</a>:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!-Pyl!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9a5cc39-917c-42de-a08a-23e717f17b21_2170x1094.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!-Pyl!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9a5cc39-917c-42de-a08a-23e717f17b21_2170x1094.png 424w, https://substackcdn.com/image/fetch/$s_!-Pyl!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9a5cc39-917c-42de-a08a-23e717f17b21_2170x1094.png 848w, https://substackcdn.com/image/fetch/$s_!-Pyl!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9a5cc39-917c-42de-a08a-23e717f17b21_2170x1094.png 1272w, https://substackcdn.com/image/fetch/$s_!-Pyl!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9a5cc39-917c-42de-a08a-23e717f17b21_2170x1094.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!-Pyl!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9a5cc39-917c-42de-a08a-23e717f17b21_2170x1094.png" width="1456" height="734" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c9a5cc39-917c-42de-a08a-23e717f17b21_2170x1094.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:734,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:392162,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.chrishayduk.com/i/192876062?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9a5cc39-917c-42de-a08a-23e717f17b21_2170x1094.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!-Pyl!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9a5cc39-917c-42de-a08a-23e717f17b21_2170x1094.png 424w, https://substackcdn.com/image/fetch/$s_!-Pyl!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9a5cc39-917c-42de-a08a-23e717f17b21_2170x1094.png 848w, https://substackcdn.com/image/fetch/$s_!-Pyl!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9a5cc39-917c-42de-a08a-23e717f17b21_2170x1094.png 1272w, https://substackcdn.com/image/fetch/$s_!-Pyl!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9a5cc39-917c-42de-a08a-23e717f17b21_2170x1094.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><blockquote><p>On the left-hand side of the image [above], the authors compared the bits of knowledge memorized by different train models (y-axis) to the number of parameters of those models (x-axis). For each model size, they trained a looped and a non-looped model variant, represented by large and small circles, respectively. You may not have noticed that there are two circle sizes on the graph because they basically overlap exactly for all model sizes. Thus, adding looping does <em>not</em> allow the model to learn more information.</p><p>Now, if we turn our attention to the right-hand side of the image [above], we can see the performance of a baseline transformer and a looped variant at different sizes on a reasoning-heavy benchmark. The numbers jump off the page &#8212; at each fixed model size, the looped model <em>vastly </em>outperforms the non-looped model. More astoundingly, a two-layer looped model <em>substantially </em>outperforms the largest non-looped model (the 12-layer variant in the first row).</p></blockquote><p>I think we see this dynamic playing out in the benchmark results for Claude Mythos, captured in the Twitter screenshot I shared above. For benchmarks that rely more heavily on stored internal knowledge (e.g., GPQA and MMLU, which evaluate models on multiple-choice questions spanning a wide range of subjects), the improvement over Claude Opus 4.6 is fairly modest. We do see a larger improvement in Humanity&#8217;s Last Exam without tools, though these questions are designed to be more difficult than GPQA and MMLU, and as such, there is some level of gains that can be attributed to improved reasoning. (I also want to state here that, of course, Mythos is likely a larger model than Opus 4.6, so it will also have more knowledge memorized &#8212; the debate is on how much compute/capacity was devoted to reasoning vs. improved knowledge).</p><p>The benchmark that really jumped off the screen for me was GraphWalks BFS. This benchmark uses randomly generated sets of nodes and edges as input to the LLM. It then asks the LLM to perform various searches on this graph (e.g., find node X&#8217;s parent, start a breadth-first search from node Y and return a list of nodes you reach after 2 iterations, etc). These tasks, along with the fact that the nodes &amp; edges don&#8217;t represent anything semantically meaningful, mean that this benchmark skews the <em>most</em> towards reasoning and away from memorized world knowledge out of any benchmark listed for Mythos. And, as we can see in the image, it is by far the biggest jump from Opus 4.6 to Mythos. </p><p>To me, this suggests that Mythos represents a relatively modest parameter scale-up and a much more aggressive depth scale-up, which would align with Mythos being a looped transformer.</p><p>Now for my second reason &#8212; compute constraints.</p><p>At this point, it&#8217;s no secret that the major labs are majorly compute-constrained. TSMC and NVIDIA have their capacities booked out years in advance, and the explosion in demand coming from coding models has only made this problem worse in recent months. As such, any innovation that can reduce the number of GPUs required to serve each batch of requests is of great importance to the labs &#8212; if a batch can be squeezed into fewer GPUs at equivalent levels of performance, the lab can serve more simultaneous batches and thus drive more revenue at a fixed level of compute.</p><p>Since a looped transformer results in more reasoning depth at a fixed parameter count, what you get is a smarter model that is able to fit in a smaller amount of GPU RAM. Exactly aligning with this critical constraint for the model providers.</p><p>Again, this is all speculation! I may be entirely off base! But these are the two major reasons that led me to strongly suspect that some form of looping is being used in Claude Mythos.</p><div><hr></div><p>If you&#8217;ve made it this far, thank you for reading! This piece ended up being much longer than I had initially anticipated. As Joan Didion famously said, &#8220;I don't know what I think until I write it down&#8221;, and writing this article definitely revealed things I didn&#8217;t realize I had thought about these books and papers.</p><p>If there&#8217;s one thing I can leave you with, let it be this &#8212; please share what you&#8217;re reading!</p>]]></content:encoded></item><item><title><![CDATA[The Unreasonable Effectiveness of LLMs in Mathematics]]></title><description><![CDATA[A journey into the mathematician's unconscious]]></description><link>https://www.chrishayduk.com/p/the-unreasonable-effectiveness-of</link><guid isPermaLink="false">https://www.chrishayduk.com/p/the-unreasonable-effectiveness-of</guid><dc:creator><![CDATA[Chris Hayduk]]></dc:creator><pubDate>Mon, 13 Apr 2026 12:02:39 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Nw8q!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8586c61b-e4c1-424f-b385-3ec012f9fad3_1536x1024.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Nw8q!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8586c61b-e4c1-424f-b385-3ec012f9fad3_1536x1024.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Nw8q!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8586c61b-e4c1-424f-b385-3ec012f9fad3_1536x1024.jpeg 424w, https://substackcdn.com/image/fetch/$s_!Nw8q!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8586c61b-e4c1-424f-b385-3ec012f9fad3_1536x1024.jpeg 848w, https://substackcdn.com/image/fetch/$s_!Nw8q!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8586c61b-e4c1-424f-b385-3ec012f9fad3_1536x1024.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!Nw8q!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8586c61b-e4c1-424f-b385-3ec012f9fad3_1536x1024.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Nw8q!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8586c61b-e4c1-424f-b385-3ec012f9fad3_1536x1024.jpeg" width="1456" height="971" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8586c61b-e4c1-424f-b385-3ec012f9fad3_1536x1024.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:971,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;Image&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Image" title="Image" srcset="https://substackcdn.com/image/fetch/$s_!Nw8q!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8586c61b-e4c1-424f-b385-3ec012f9fad3_1536x1024.jpeg 424w, https://substackcdn.com/image/fetch/$s_!Nw8q!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8586c61b-e4c1-424f-b385-3ec012f9fad3_1536x1024.jpeg 848w, https://substackcdn.com/image/fetch/$s_!Nw8q!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8586c61b-e4c1-424f-b385-3ec012f9fad3_1536x1024.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!Nw8q!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8586c61b-e4c1-424f-b385-3ec012f9fad3_1536x1024.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Image from @alexwei_&#8217;s <a href="https://x.com/alexwei_/status/1946477742855532918">announcement post</a> on X</figcaption></figure></div><p>In July 2024, DeepMind unveiled AlphaProof &#8212; an AlphaZero-inspired agent that constructs mathematical arguments in Lean, a programming language for proofs. It broke new ground in mathematical performance, achieving a silver medal in the 2024 International Math Olympiad.</p><p>One year later, in July 2025, OpenAI announced that they had achieved a gold medal in the 2025 International Math Olympiad using a raw LLM &#8212; no reinforcement learning in Lean space, no translation between natural language and formal proof languages. In the span of a few weeks, this same model would go on to add a gold medal at the International Olympiad in Informatics and a 2nd place finish at the AtCoder World Tour Finals to its achievements.</p><p>How is it possible that a general LLM, one that operates in natural language and would be just as comfortable answering questions about lasagna recipes in ChatGPT as it is scoring a gold medal in the IMO, could defeat a model that was custom-made to solve math problems by thinking directly in proof space? And is the current LLM architectural paradigm enough for us to solve mathematics?</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.chrishayduk.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.chrishayduk.com/subscribe?"><span>Subscribe now</span></a></p><h2>A Brief Explanation of AlphaProof</h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!eBNs!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1e50e52d-6ce8-4da2-985b-885e64fb9437_1773x1102.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!eBNs!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1e50e52d-6ce8-4da2-985b-885e64fb9437_1773x1102.png 424w, https://substackcdn.com/image/fetch/$s_!eBNs!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1e50e52d-6ce8-4da2-985b-885e64fb9437_1773x1102.png 848w, https://substackcdn.com/image/fetch/$s_!eBNs!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1e50e52d-6ce8-4da2-985b-885e64fb9437_1773x1102.png 1272w, https://substackcdn.com/image/fetch/$s_!eBNs!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1e50e52d-6ce8-4da2-985b-885e64fb9437_1773x1102.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!eBNs!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1e50e52d-6ce8-4da2-985b-885e64fb9437_1773x1102.png" width="1456" height="905" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1e50e52d-6ce8-4da2-985b-885e64fb9437_1773x1102.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:905,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;Fig. 1: AlphaProof core reasoning components.&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Fig. 1: AlphaProof core reasoning components." title="Fig. 1: AlphaProof core reasoning components." srcset="https://substackcdn.com/image/fetch/$s_!eBNs!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1e50e52d-6ce8-4da2-985b-885e64fb9437_1773x1102.png 424w, https://substackcdn.com/image/fetch/$s_!eBNs!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1e50e52d-6ce8-4da2-985b-885e64fb9437_1773x1102.png 848w, https://substackcdn.com/image/fetch/$s_!eBNs!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1e50e52d-6ce8-4da2-985b-885e64fb9437_1773x1102.png 1272w, https://substackcdn.com/image/fetch/$s_!eBNs!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1e50e52d-6ce8-4da2-985b-885e64fb9437_1773x1102.png 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">The core reasoning components of AlphaProof. Figure 1 in the AlphaProof paper</figcaption></figure></div><p>AlphaProof is an LLM- and reinforcement-learning-based approach to mathematical proof generation published by DeepMind in November 2025 (with its initial announcement in July 2024).<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> The model was inspired by AlphaZero, the successor model to AlphaGo &amp; AlphaGo Zero, which taught itself to play chess, shogi, and Go purely through reinforcement learning (RL) from self-play</p><p>The AlphaProof research team began by translating mathematical problems from natural language into Lean, a formal proof language that allows users to build mathematical arguments through explicit axioms, theorems, and deductive steps. Proofs in Lean are built up one step at a time by applying actions (called tactics) to change the current proof state. Lean guarantees that each step must be logically rigorous and consistent &#8212; if not, the proof won&#8217;t compile. </p><p>To generate a sufficiently large dataset to learn from, a Gemini-based LLM (known as the formalizer) was trained to translate natural-language mathematical statements into Lean (see the figure below for an example). </p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!UCtO!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc41e829e-79bf-46f7-9c85-95fc4ac7ea93_4087x1707.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!UCtO!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc41e829e-79bf-46f7-9c85-95fc4ac7ea93_4087x1707.jpeg 424w, https://substackcdn.com/image/fetch/$s_!UCtO!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc41e829e-79bf-46f7-9c85-95fc4ac7ea93_4087x1707.jpeg 848w, https://substackcdn.com/image/fetch/$s_!UCtO!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc41e829e-79bf-46f7-9c85-95fc4ac7ea93_4087x1707.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!UCtO!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc41e829e-79bf-46f7-9c85-95fc4ac7ea93_4087x1707.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!UCtO!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc41e829e-79bf-46f7-9c85-95fc4ac7ea93_4087x1707.jpeg" width="1456" height="608" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c41e829e-79bf-46f7-9c85-95fc4ac7ea93_4087x1707.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:608,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;Extended Data Fig. 2: An example of the auto-formalization process.&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Extended Data Fig. 2: An example of the auto-formalization process." title="Extended Data Fig. 2: An example of the auto-formalization process." srcset="https://substackcdn.com/image/fetch/$s_!UCtO!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc41e829e-79bf-46f7-9c85-95fc4ac7ea93_4087x1707.jpeg 424w, https://substackcdn.com/image/fetch/$s_!UCtO!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc41e829e-79bf-46f7-9c85-95fc4ac7ea93_4087x1707.jpeg 848w, https://substackcdn.com/image/fetch/$s_!UCtO!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc41e829e-79bf-46f7-9c85-95fc4ac7ea93_4087x1707.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!UCtO!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc41e829e-79bf-46f7-9c85-95fc4ac7ea93_4087x1707.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">The formalization system translates natural language math statements to valid Lean code. Extended Data Fig. 2 in the AlphaProof paper</figcaption></figure></div><p>With this setup in place, constructing math becomes a game-like RL environment, where the state is the current proof status, the set of actions is the set of possible Lean tactics, and the reward for each action is -1 (encouraging shorter proofs, since we aim to maximize the cumulative reward).</p><p>AlphaProof gets its tactics from the combination of a prover agent and a search algorithm inspired by AlphaZero. The prover agent is a 3-billion-parameter encoder-decoder transformer model that suggests tactics to apply next (given the current prompt state) and estimates their expected cumulative return (that is, what I would expect my total reward to be starting from the proof state that this action will take me to and continuing until I complete the problem). The tree search algorithm explores sequences of actions suggested by the prover agent and evaluates their results.</p><p>The prover agent learns by training on the Lean problems generated by the formalizer. It, along with the tree search algorithm, generates attempted proofs and receives a learning signal depending on whether a valid proof is found or the agent times out during the search.</p><p>The prover agent + tree search algorithm approach allows us to scale along two axes: training time for the prover agent and test-time computation for the tree search algorithm. This dual-scaling enables strong performance on held-out IMO problems, as shown in the table below. When we increase tree search time from 2 TPU minutes per problem to 12 TPU hours per problem, we see validation accuracy jump from 33.2% to 43%.7%</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!1Dw3!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b6a7fe7-0540-4e0b-a912-75542492181f_1830x356.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!1Dw3!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b6a7fe7-0540-4e0b-a912-75542492181f_1830x356.png 424w, https://substackcdn.com/image/fetch/$s_!1Dw3!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b6a7fe7-0540-4e0b-a912-75542492181f_1830x356.png 848w, https://substackcdn.com/image/fetch/$s_!1Dw3!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b6a7fe7-0540-4e0b-a912-75542492181f_1830x356.png 1272w, https://substackcdn.com/image/fetch/$s_!1Dw3!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b6a7fe7-0540-4e0b-a912-75542492181f_1830x356.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!1Dw3!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b6a7fe7-0540-4e0b-a912-75542492181f_1830x356.png" width="1456" height="283" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/5b6a7fe7-0540-4e0b-a912-75542492181f_1830x356.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:283,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:48441,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.chrishayduk.com/i/192973763?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b6a7fe7-0540-4e0b-a912-75542492181f_1830x356.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!1Dw3!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b6a7fe7-0540-4e0b-a912-75542492181f_1830x356.png 424w, https://substackcdn.com/image/fetch/$s_!1Dw3!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b6a7fe7-0540-4e0b-a912-75542492181f_1830x356.png 848w, https://substackcdn.com/image/fetch/$s_!1Dw3!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b6a7fe7-0540-4e0b-a912-75542492181f_1830x356.png 1272w, https://substackcdn.com/image/fetch/$s_!1Dw3!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b6a7fe7-0540-4e0b-a912-75542492181f_1830x356.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">AlphaProof performance on the held-out IMO validation set. Extracted from Table 1 in the AlphaProof paper</figcaption></figure></div><p>However, from the above, we can see that the paper uses another scaling axis &#8212; namely, test-time RL (TTRL). </p><p>The way this works is that, for challenging problems, a variant generator can create hundreds of thousands of distinct yet similar formal problems for the proof network to continue training on. The prover will then learn from these similar examples, updating its weights as it gets rewards for completing proofs. After TTRL has been executed, the prover agent now &#8220;knows&#8221; substantially more about the problem area adjacent to the problem we actually care about, thereby improving its accuracy.</p><p>Thus, to achieve the highest levels of performance, the AlphaProof system needed to overfit on new data that was extremely similar to the questions asked. It did <em>not </em>succeed out of the box at the IMO, even after training on ~80 million formal problems. In particular, on the IMO holdout set (as seen in the table above), AlphaProof required a 4-order-of-magnitude increase in compute budget to go from a 33.2% to a 58.3% success rate. The final 4.4 percentage points of performance (from 53.9% to 58.3%) required an additional order of magnitude in compute (from 3,000 TPU minutes per problem to 30,000 TPU minutes per problem). The paper explicitly states that &#8220;each of these solutions required 2&#8211;3 days of (test-time RL) TTRL, demonstrating substantial problem-specific adaptation at inference.&#8221;</p><p><strong>The main takeaway of all this is the following:</strong> AlphaProof used a highly math-specific RL environment, along with supporting models (e.g., a formalization system and a variant generator), and a formal proof language to achieve its groundbreaking theorem-proving results. It did indeed set a state-of-the-art in mathematical benchmarks such as the IMO; however, despite the scaffold already being highly specific to the problem, it was still insufficient, with the model still requiring multiple TPU days of test-time RL and hundreds of TPU days of tree-search time per problem to achieve optimal performance. Thus, although highly performant, AlphaProof was <em>not</em> a general-purpose theorem-proving system &#8212; it still required substantial overfitting and custom software to be useful.</p><h2>GPT-5.x and <em>The Mathematician&#8217;s Mind</em></h2><p>While AlphaProof performed well with its highly-custom scaffold and approach to overfitting problems, OpenAI&#8217;s IMO Gold model blew it out of the water without all of these math-specific bells and whistles. </p><p>How is this possible? How can a general-purpose system perform so much better than an application-specific system? And why have all the large labs moved away from these application-specific systems in the intervening ~2 years since AlphaProof&#8217;s release?</p><p>The answer lies in the subconscious mind.</p><p>Jacques Hadamard, one of the great mathematicians of the 20th century, published <em>The Mathematician&#8217;s Mind: The Psychology of Invention in the Mathematical Field </em>in 1945. Building on Henri Poincar&#233;&#8217;s 1908 lecture titled L&#8217;invention Math&#233;matique (Mathematical Invention), he interviewed several of the greatest living mathematicians &amp; physicists at the time (including George Polya, Claude L&#233;vi-Strauss, and Albert Einstein) and assessed how their process of mathematical discovery <em>felt</em> from a phenomenological perspective.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a> He wanted to answer the following question: What is occurring in the minds of our greatest mathematicians when they make new discoveries?</p><p>The answer Hadamard found, synthesizing results from these interviews, was that mathematical discoveries arise from the interplay between conscious and unconscious processes. And, in particular, the idea for the proof tends to start in the subconscious.</p><blockquote><p>&#8230;let us remember that every mental work and especially the work of discovery implies the cooperation of the unconscious, be it the superficial or (fairly often) the more or less remote one; that, inside of that unconscious (resulting from a preliminary conscious work), there is that starting of ideas which Poincar&#233; has compared to a projection of atoms and which can be more or less scattered ; that concrete representations are generally used by the mind for the maintenance and synthesis of combinations. This carries, in the first place, the consequence that, <strong>strictly speaking, there is hardly any completely logical discovery. Some intervention of intuition issuing from the unconscious is necessary at least to initiate the logical work. </strong>[<em>emphasis mine</em>]</p></blockquote><p>Hadamard asserts that <em>no</em> mathematical discovery is purely logical. The unconscious mind, in all cases that he examined, played a crucial role in the development of rigorous mathematical arguments. This role, and the handoffs between the subconscious and conscious minds, were distilled by Hadamard into the following framework for mathematical discovery:</p><ol><li><p><strong>Preparation</strong> (primarily conscious) &#8212; the conscious mind focuses on a problem for an extended period of time, collecting relevant information and trying out several avenues for solution</p></li><li><p><strong>Incubation</strong> (primarily unconscious) &#8212; the unconscious mind, directed in its goals by the focus of the conscious mind in the Preparation stage, sets to work searching for high-level solutions. This is where the bulk of problem-solving and discovery is actually done. The unconscious mind is better at viewing the problem as a &#8220;whole&#8221; and at uncovering unexpected insights and connections than the conscious mind. The unconscious mind evaluates proposed solutions based on aesthetic criteria.</p></li><li><p><strong>Illumination</strong> (primarily unconscious) &#8212; an idea generated by the unconscious mind that satisfies the unconscious criteria springs forth into the conscious mind </p></li><li><p><strong>Verification</strong> (primarily conscious) &#8212; the conscious mind sets to work translating the unconscious&#8217;s idea into formal mathematical language and verifies that it is logically correct.</p></li></ol><p>Thus, we can see that the unconscious mind is actually responsible for <em>generating </em>the proof structure. The conscious mind just sets the scene during the Preparation stage and verifies that suggested proof structure during the Verification stage. But it does <em>not</em> produce the critical insights regarding the proof&#8217;s structure that actually lead to the discovery itself. As a result, we can say that the discovery process is decidedly not rigorous.</p><p>AlphaProof&#8217;s downfall was that it was designed to act like the rigorous conscious mind &#8212; every step must be constructed rigorously and logically consistent with previous steps, and the proof is built up step by step. In this manner, AlphaProof acts at a <em>local</em> level with the prover trying to find the best next incremental step in a proof, whereas the mathematician&#8217;s unconscious mind works at a <em>global</em> level by identifying a full proof sketch all at once. Only then does the conscious mind fill in the rigorous details. </p><p>Based on this insight, we see that an AlphaProof-like system is most closely aligned with the Preparation and Verification stages in Hadamard&#8217;s framework. It completely omits the Incubation and Illumination phases, where the actual work of discovery occurs.</p><p>OpenAI&#8217;s gold-medal-winning model, as a standard LLM deployed for mathematical use cases, represents a new paradigm in mathematical reasoning that relaxes AlphaProof&#8217;s constraints on only following rigorous thought. It enables messier, higher-level reasoning in the language space rather than in the Lean-verified proof space. The system can think through and mentally pressure-test high-level approaches &amp; proof sketches, rather than being forced to focus exclusively on granular proof steps, as in AlphaProof.</p><p>To demonstrate this, I provided GLM 5.1 with question C4 (a medium difficulty combinatorics problem) from the 2024 International Math Olympiad. By using an open-source model, we can see the full reasoning trace and how its thought process leads to its final answer.</p><p>In the excerpt provided below of the model&#8217;s reasoning trace, you can see how messy and non-rigorous the thinking truly is. The model is jumping around in conceptual space, backtracking frequently, and probing different high-level directions.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!vVjp!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F04ca1c14-5269-44a4-885b-b38b14c620fe_1958x1074.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!vVjp!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F04ca1c14-5269-44a4-885b-b38b14c620fe_1958x1074.png 424w, https://substackcdn.com/image/fetch/$s_!vVjp!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F04ca1c14-5269-44a4-885b-b38b14c620fe_1958x1074.png 848w, https://substackcdn.com/image/fetch/$s_!vVjp!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F04ca1c14-5269-44a4-885b-b38b14c620fe_1958x1074.png 1272w, https://substackcdn.com/image/fetch/$s_!vVjp!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F04ca1c14-5269-44a4-885b-b38b14c620fe_1958x1074.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!vVjp!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F04ca1c14-5269-44a4-885b-b38b14c620fe_1958x1074.png" width="1456" height="799" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/04ca1c14-5269-44a4-885b-b38b14c620fe_1958x1074.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:799,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:281487,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.chrishayduk.com/i/192973763?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F04ca1c14-5269-44a4-885b-b38b14c620fe_1958x1074.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!vVjp!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F04ca1c14-5269-44a4-885b-b38b14c620fe_1958x1074.png 424w, https://substackcdn.com/image/fetch/$s_!vVjp!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F04ca1c14-5269-44a4-885b-b38b14c620fe_1958x1074.png 848w, https://substackcdn.com/image/fetch/$s_!vVjp!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F04ca1c14-5269-44a4-885b-b38b14c620fe_1958x1074.png 1272w, https://substackcdn.com/image/fetch/$s_!vVjp!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F04ca1c14-5269-44a4-885b-b38b14c620fe_1958x1074.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Reasoning excerpt from GLM 5.1&#8217;s attempt to solve question C4 from IMO 2024.</figcaption></figure></div><p>We can see from the above example that this brings us closer in line with the Incubation and Illumination phases of Hadamard&#8217;s process &#8212; the LLM is able to search for high-level solutions and make jumps that are <em>not</em> logically rigorous and are <em>not </em>incremental (in the sense that they do not need to be at the scale of a single tactic in Lean). The model is free to jump around in natural language between several, potentially unrelated concepts. It is only forced to converge on a rigorous, verified solution at the end of its answer once one of the high-level solutions satisfies its aesthetic sense.</p><p>This relaxation of constraints on the thinking process also explains how the system can generalize across domains &#8212; the model does not need all the math-specific scaffolding and so can learn a more general process of discovery. This process, as outlined by Hadamard, is not limited to math (as evidenced by Einstein's inclusion in the group). In fact, it seems to me that discovery in <em>any</em> domain that can culminate in a Verification step will follow this overarching process. And we do see this in the success that OpenAI&#8217;s model enjoyed at the IOI and AtCoder competitions shortly following its gold medal at the IMO.</p><p>So, is math (and all other easily verified domains) solved with this paradigm? </p><p>Not quite. While OpenAI&#8217;s gold medal LLM represents a meaningful improvement over AlphaProof in bringing AI proof systems into alignment with how mathematicians actually think, there is still a glaring gap: the model is required to reason in language space and produce one token at a time. Although this allows for broader, less rigorous thought than AlphaProof, it is still far more constrained than Hadamard&#8217;s description of the unconscious. Specifically, Hadamard asserts that, in addition to being non-rigorous, unconscious thought is often not even interpretable. He says that at this stage, all mathematicians think without language or precise symbols, and many do not even use clear images. The ideas themselves are vague, amorphous, and global. Hadamard says:</p><blockquote><p>Practically all of [the mathematicians interviewed]&#8230; avoid not only the use of mental words but also, just as I do, the mental use of algebraic or any other precise signs; also as in my case, they use vague images.</p></blockquote><p>This differs substantially from LLMs, in which each thought has a fixed amount of computation applied to it, after which that thought must then be made concrete in the form of a token. As a result, current systems are limited in how closely they adhere to Hadamard&#8217;s discovery framework. This fundamentally limits how creative they can be &#8212; the more we force thoughts to be interpretable and legible, the less unexpected and amorphous they will be, and thus the less they will approximate the unconscious mind.</p><h2>The Reasoning Paradigm of the Future</h2><p>As Hadamard astutely observed, true discovery occurs at the subconscious level rather than through rational, conscious processes. When moving from AlphaProof to OpenAI&#8217;s IMO Gold model, we took one step towards the unconscious &#8212; from the uber-rational realm of Lean and step-by-step proof construction into the messy, unrigorous world of natural-language thought. </p><p>Although this thought is messy, it is still articulable and thus operates at a level above what we call the subconscious. Reasoning in natural language allows for more flexibility than reasoning in Lean, but we know that mathematicians don&#8217;t reason in language <em>at all</em> during the Incubation and Illumination steps.</p><p>If the bottleneck is the token itself &#8212; the forced collapse of each thought into a discrete, legible symbol &#8212; then the natural question is whether we can let models reason <em>before</em> that collapse happens. What would it look like for an LLM to think in a medium richer than language?</p><p><strong>For LLMs, the level below token-based thought is embedding-based thought.</strong> A model's actual computation lives in embedding space before it's projected down into the vocabulary &#8212; just as the mathematician's thoughts live in the unconscious during Incubation before being projected into the conscious mind during Illumination. Each token is a lossy compression of a much higher-dimensional internal state.</p><p>Recall Hadamard's observation that mathematicians reason in vague images rather than precise signs. An embedding is, in a loose sense, exactly that: a vague, high-dimensional representation that has not yet been forced into a precise sign.</p><p>Thus, rather than forcing the model to output its reasoning in token space, we can allow it to reason in this much higher-dimensional embedding space found within the model's internals. This allows for much more abstract and &#8220;unconscious&#8221; thought than standard token-based reasoning does, aligning more closely with Hadamard&#8217;s concept of unconscious thought.</p><p>ByteDance implemented this idea in their November 2025 paper titled &#8220;Scaling Latent Reasoning via Looped Language Models,&#8221; which developed a model family called Ouro.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a> This model family is what is known as a &#8220;looped language model&#8221; &#8212; that is, rather than outputting a token once the language model reaches the last layer in its computation graph, it loops back on itself by ingesting the embedding from layer N as new input for layer 1 (see leftmost diagram in the image below). </p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Cbd3!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F57fb767e-e4a9-4115-972e-b5b6725e60b4_2004x1002.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Cbd3!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F57fb767e-e4a9-4115-972e-b5b6725e60b4_2004x1002.png 424w, https://substackcdn.com/image/fetch/$s_!Cbd3!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F57fb767e-e4a9-4115-972e-b5b6725e60b4_2004x1002.png 848w, https://substackcdn.com/image/fetch/$s_!Cbd3!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F57fb767e-e4a9-4115-972e-b5b6725e60b4_2004x1002.png 1272w, https://substackcdn.com/image/fetch/$s_!Cbd3!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F57fb767e-e4a9-4115-972e-b5b6725e60b4_2004x1002.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Cbd3!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F57fb767e-e4a9-4115-972e-b5b6725e60b4_2004x1002.png" width="1456" height="728" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/57fb767e-e4a9-4115-972e-b5b6725e60b4_2004x1002.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:728,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:450015,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.chrishayduk.com/i/192973763?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F57fb767e-e4a9-4115-972e-b5b6725e60b4_2004x1002.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Cbd3!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F57fb767e-e4a9-4115-972e-b5b6725e60b4_2004x1002.png 424w, https://substackcdn.com/image/fetch/$s_!Cbd3!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F57fb767e-e4a9-4115-972e-b5b6725e60b4_2004x1002.png 848w, https://substackcdn.com/image/fetch/$s_!Cbd3!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F57fb767e-e4a9-4115-972e-b5b6725e60b4_2004x1002.png 1272w, https://substackcdn.com/image/fetch/$s_!Cbd3!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F57fb767e-e4a9-4115-972e-b5b6725e60b4_2004x1002.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Performance is less &#8220;spiky&#8221; for the looped language model. Figure 1 in the Looped Language Models paper</figcaption></figure></div><p>As a result, the model can learn to reason deeply about tokens <em>without first converting them to language space. </em>Moreover, the ByteDance team integrated this reasoning process directly into pretraining and implemented a mechanism for the model to determine when to stop looping during inference based on a predicted exit probability. This allows the model to more closely integrate its reasoning capabilities with its world knowledge (since it is trained during the pretraining stage), as well as to learn to modulate its reasoning based on the difficulty of the next token. </p><p>And, most importantly, this brings the model&#8217;s reasoning process much closer to Hadamard&#8217;s core Incubation and Illumination steps.</p><p>With this setup and its proximity to Hadamard&#8217;s Discovery process, a very general-purpose approach to reasoning in verifiable domains, we would expect the model to perform better across a number of benchmarks. And that is <em>precisely </em>what we see. Not only does the model outperform much larger models, but its performance is also less &#8220;spiky&#8221; &#8212; that is, it performs consistently well across many benchmarks rather than experiencing isolated &#8220;spikes&#8221; in performance on certain benchmarks (see the central and rightmost figures in the image above).</p><p>We can see further evidence that this approach improves reasoning in the image below. On the left-hand side of the image below, the authors compared the bits of knowledge memorized by different train models (y-axis) to the number of parameters of those models (x-axis). For each model size, they trained a looped and a non-looped model variant, represented by large and small circles, respectively. You may not have noticed that there are two circle sizes on the graph because they basically overlap exactly for all model sizes. Thus, adding looping does <em>not</em> allow the model to learn more information.</p><p>Now, if we turn our attention to the right-hand side of the image below, we can see the performance of a baseline transformer and a looped variant at different sizes on a reasoning-heavy benchmark. The numbers jump off the page &#8212; at each fixed model size, the looped model <em>vastly </em>outperforms the non-looped model. More astoundingly, a two-layer looped model <em>substantially </em>outperforms the largest non-looped model (the 12-layer variant in the first row).</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!QKls!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F52ca2a66-f66c-470b-b669-98d14d0203fd_2488x1262.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!QKls!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F52ca2a66-f66c-470b-b669-98d14d0203fd_2488x1262.png 424w, https://substackcdn.com/image/fetch/$s_!QKls!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F52ca2a66-f66c-470b-b669-98d14d0203fd_2488x1262.png 848w, https://substackcdn.com/image/fetch/$s_!QKls!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F52ca2a66-f66c-470b-b669-98d14d0203fd_2488x1262.png 1272w, https://substackcdn.com/image/fetch/$s_!QKls!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F52ca2a66-f66c-470b-b669-98d14d0203fd_2488x1262.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!QKls!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F52ca2a66-f66c-470b-b669-98d14d0203fd_2488x1262.png" width="1456" height="739" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/52ca2a66-f66c-470b-b669-98d14d0203fd_2488x1262.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:739,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:478018,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.chrishayduk.com/i/192973763?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F52ca2a66-f66c-470b-b669-98d14d0203fd_2488x1262.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!QKls!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F52ca2a66-f66c-470b-b669-98d14d0203fd_2488x1262.png 424w, https://substackcdn.com/image/fetch/$s_!QKls!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F52ca2a66-f66c-470b-b669-98d14d0203fd_2488x1262.png 848w, https://substackcdn.com/image/fetch/$s_!QKls!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F52ca2a66-f66c-470b-b669-98d14d0203fd_2488x1262.png 1272w, https://substackcdn.com/image/fetch/$s_!QKls!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F52ca2a66-f66c-470b-b669-98d14d0203fd_2488x1262.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Looping the language models doesn&#8217;t increase memorization capacity, but it does improve the ability to manipulate the memorized knowledge. Figure 6 in the Looped Language Models paper.</figcaption></figure></div><p>Hence, from the above, we can see that applying an embedding-space loop to language models, thus bringing them closer to Hadamard&#8217;s Discovery process, substantially improves reasoning ability. <em>And </em>this reasoning ability seems to generalize better across benchmarks than standard RL on top of language models. As a result, I expect <em>all </em>frontier LLMs to adopt this paradigm if they haven&#8217;t already.</p><p>Stepping back a bit, there is something quietly counterintuitive about the trajectory of the last two years. AlphaProof tried to make machines do math by forcing them to be maximally rigorous, with every proof built up from one logically rigorous step to the next. It was, in many ways, the purest expression of what we used to think mathematical reasoning <em>was</em>. And yet it hit a wall.</p><p>The lesson of OpenAI&#8217;s IMO Gold model, and now of looped language models, is that progress has come not from tightening the constraints on machine reasoning but from loosening them. First, we let models think messily in natural language. Now we are letting them think in a medium that isn&#8217;t even language. Each step has moved further from the clean, legible, step-by-step ideal. And each step has improved the models.</p><p>But loosening the constraints on Incubation only gets us half of Hadamard&#8217;s cycle. I&#8217;ve spent most of this piece arguing that LLMs are getting better at Incubation and Illumination, and that looped models push them further in that direction. But this leaves out the fact that current LLMs are weak at Verification. They hallucinate proofs, confidently assert false lemmas, and have no reliable internal signal for when their reasoning has gone off the rails. AlphaProof&#8217;s step-by-step rigor was, for all its limitations, a genuine Verification engine &#8212; every tactic was checked by Lean before the proof could proceed. Standard LLMs lack such a mechanism and are thus often prone to hallucinating results. The IMO Gold result is remarkable precisely because the model managed to produce verifiable outputs despite lacking a built-in verifier.</p><p>Thus, the real frontier is synthesis. AlphaProof gave us a formalizer and a verifier but no unconscious. Looped language models are giving us an unconscious but no verifier. Hadamard&#8217;s framework is designed to be a loop &#8212; when Verification fails, the mathematician returns to Preparation with new information about why it failed, which reshapes the next round of Incubation. The system that finally closes this loop will be the first one to actually instantiate Hadamard&#8217;s full cycle rather than just its generative half.</p><p>This, I think, is the actual research agenda for the next few years, and it has two halves that have to advance together. On one side, we need to keep deepening the unconscious by pushing latent-space reasoning further, scaling up looped architectures, and figuring out how to train them with the kinds of reinforcement signals that sharpen aesthetic judgment. On the other, we need to build formalizers and verifiers for a broader range of fields, so that the unconscious can receive concrete feedback from a Verification step. Neither half is sufficient on its own, as the unconscious without a verifier can drift into hallucinations, while the verifier without an unconscious is too rigid to make true intellectual leaps. A synthesis of both is required.</p><p>What this looks like concretely depends on the field. For mathematics, both halves are nearly in place. The unconscious is arriving via looped language models, and the verifier already exists in the form of Lean. All we need to do is use AlphaProof&#8217;s Formalizer to convert the Looped Language Model output to Lean, and we&#8217;re good to go.</p><p>The more difficult question is what we do in fields that don&#8217;t already have a Lean. The unconscious half seems to generalize quite well, judging by the less spiky benchmark results from the Looped Transformer paper. But the verifier half does not generalize. Instead, it has to be built field by field. And what counts as a verifier in one domain looks nothing like what counts as a verifier in another.</p><p>In physics, building the verifier means investing in experiment automation. That is, robotic labs that can run, measure, and report on experiments proposed by a latent-reasoning model, closing the loop between hypothesis and data without a human in the middle. In biology, the same logic points to high-throughput assay platforms that can test a model&#8217;s proposed interventions at scale, with the same feedback loop into the unconscious. In economics and the social sciences, the verifier half has to be weaker, such as large-scale simulation environments, prediction markets, or structured forecasting tournaments that can at least provide a noisy verification signal where a clean one is impossible. But even a noisy verifier may be enough to close the loop, so long as the unconscious half is there to propose hypotheses worth testing in the first place.</p><p>Hadamard told us in 1945 that the real work of discovery happens in a place that is neither rigorous nor articulable, and that the conscious mind&#8217;s job is to set the stage and then check the work afterward. Looped language models are finally giving us the unconscious that Hadamard envisioned. The next paradigm wires the verifier into the output of the looped language models, closing the loop on Hadamard&#8217;s Discovery process. The fields where we can build both halves are the fields where Hadamard&#8217;s full cycle will finally run on a machine and where we should expect AI progress to accelerate most dramatically in the coming years.</p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p><a href="https://www.nature.com/articles/s41586-025-09833-y">Olympiad-level formal mathematical reasoning with reinforcement learning</a></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>Hadamard, J. (2020). <em>The Mathematician&#8217;s Mind: The Psychology of Invention in the Mathematical Field</em>. Princeton University Press.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p><a href="https://arxiv.org/abs/2510.25741">Scaling Latent Reasoning via Looped Language Models</a></p></div></div>]]></content:encoded></item><item><title><![CDATA[2025 Year in Books]]></title><description><![CDATA[A new tradition]]></description><link>https://www.chrishayduk.com/p/2025-year-in-books</link><guid isPermaLink="false">https://www.chrishayduk.com/p/2025-year-in-books</guid><dc:creator><![CDATA[Chris Hayduk]]></dc:creator><pubDate>Fri, 16 Jan 2026 21:35:58 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!uWH1!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb342a63-b6ec-4dbf-8bf7-1205baff6c9e_5504x3072.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!uWH1!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb342a63-b6ec-4dbf-8bf7-1205baff6c9e_5504x3072.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!uWH1!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb342a63-b6ec-4dbf-8bf7-1205baff6c9e_5504x3072.jpeg 424w, https://substackcdn.com/image/fetch/$s_!uWH1!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb342a63-b6ec-4dbf-8bf7-1205baff6c9e_5504x3072.jpeg 848w, https://substackcdn.com/image/fetch/$s_!uWH1!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb342a63-b6ec-4dbf-8bf7-1205baff6c9e_5504x3072.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!uWH1!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb342a63-b6ec-4dbf-8bf7-1205baff6c9e_5504x3072.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!uWH1!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb342a63-b6ec-4dbf-8bf7-1205baff6c9e_5504x3072.jpeg" width="1456" height="813" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/db342a63-b6ec-4dbf-8bf7-1205baff6c9e_5504x3072.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:813,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:9421470,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.chrishayduk.com/i/184692771?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb342a63-b6ec-4dbf-8bf7-1205baff6c9e_5504x3072.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!uWH1!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb342a63-b6ec-4dbf-8bf7-1205baff6c9e_5504x3072.jpeg 424w, https://substackcdn.com/image/fetch/$s_!uWH1!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb342a63-b6ec-4dbf-8bf7-1205baff6c9e_5504x3072.jpeg 848w, https://substackcdn.com/image/fetch/$s_!uWH1!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb342a63-b6ec-4dbf-8bf7-1205baff6c9e_5504x3072.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!uWH1!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb342a63-b6ec-4dbf-8bf7-1205baff6c9e_5504x3072.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>I really enjoyed the format of <span class="mention-wrap" data-attrs="{&quot;name&quot;:&quot;Corin Wagen&quot;,&quot;id&quot;:9321224,&quot;type&quot;:&quot;user&quot;,&quot;url&quot;:null,&quot;photo_url&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/7368e707-6fa8-4446-8376-c151fc2966b3_925x925.jpeg&quot;,&quot;uuid&quot;:&quot;20fd1a07-872d-4b13-a004-3063da3632d7&quot;}" data-component-name="MentionToDOM"></span>&#8217;s book review lists, so I&#8217;m going to make it a yearly tradition to summarize my reading for the year in a similar manner. I&#8217;ll list the completed books in chronological order below. For books that were notable (whether notably bad, notably good, or containing some notable idea), I&#8217;ll include a short blurb with my thoughts and my major takeaways.</p><p><strong>#1. Sean McMeekin, </strong><em><strong>Stalin&#8217;s War: A New History of World War II</strong></em></p><p>A history of World War II that starts by asking the following questions:</p><ol><li><p>Which leaders (if any) were in power pre-, during, and post-war in order to really influence the course of events (the build-up, the actual fighting, and the treaty negotiations afterward)?</p></li><li><p>Which leaders (if any) benefited the most from the war?</p></li><li><p>Which leaders (if any) had a vested interest in the European powers going to war?</p></li><li><p>Which leaders (if any) successfully bent the leaders of the other powers to their will?</p></li></ol><p>McMeekin asserts that, when we stop to consider these questions, there is only one leader who satisfies all four: Joseph Stalin.</p><p>Starting from this point, the book outlines a contrarian history of the war, centering on Stalin's role in instigating the war between Germany and the Western Powers, attempting to contain Germany through resource constraints, and, finally, manipulating the Western Powers to achieve his war aims as cheaply as possible.</p><p>I&#8217;m not enough of an expert on WWII history to refute or support McMeekin&#8217;s claims, but it certainly made for one of the most interesting history reads I&#8217;ve ever had.</p><p><strong>#2. Nicolas Cole, </strong><em><strong>The Art and Business of Online Writing: How to Beat the Game of Capturing and Keeping Attention</strong></em></p><p><strong>#3. Amaury De Riencourt, </strong><em><strong>The Coming Caesars</strong></em></p><p>Published in 1957, this book outlines the progression of American government from its original, firmly aristocratic character (we didn&#8217;t even vote for president in the beginning of the Republic!) to its then-new mass democratic flavor. De Riencourt draws parallels between this progression and the progression of the Greco-Roman world leading up to the end of the Roman Republic and the beginning of the Roman Empire. He uses this parallel to analyze the trajectory of the US and predict the arrival of an eventual American Caesar. </p><p>Unfortunately, this book has become extremely relevant in recent years.</p><p><strong>#4. Niall Ferguson, </strong><em><strong>The Pity of War: Explaining World War I</strong></em></p><p>Another contrarian historical perspective, this time blaming the United Kingdom for the outbreak of World War I. It argues that had the UK decisively demonstrated its willingness to enter the war before its outbreak, or had instead committed to staying out of it, the total devastation for everyone would have been significantly reduced.</p><p>Given the death and destruction wrought by World War I, as well as the knock-on consequences of the war (the Spanish Flu pandemic, the Great Depression, &amp; finally World War II), it&#8217;s hard to argue that the UK&#8217;s decision to support France was optimal in the long run. All of the European powers would likely have been better off if the UK had left France to fend for itself and it had lost to Germany more swiftly. It hardly seems like the counterfactual could be any worse than what Europe endured from 1914 to 1945. However, I&#8217;m not sure that I&#8217;m totally convinced that the war was England&#8217;s <em>fault</em> in the sense that Ferguson asserts. As with McMeekin&#8217;s book, I&#8217;ll need to do more reading on the WWI period to really evaluate these claims.</p><p><strong>#5. Samo Burja, </strong><em><strong>Great Founder Theory</strong></em></p><p>Some books aim to inform, diving into great detail on a specific concept or event, analyzing all of its minutiae from every angle. Other books aim to enlighten, expounding a big idea that attempts to fundamentally shift your worldview.</p><p>This book is one that aims to enlighten.</p><p>Burja&#8217;s big idea is a sort of re-skinning of the Great Man of History ideal &#8212; except, this time, it incorporates institutional design as its centerpiece. </p><p>His theory basically goes as follows: we can measure an individual's &#8220;greatness&#8221; by the effect they have on shaping the world. Any shaping that is limited to the individual&#8217;s lifetime is inherently highly constrained, and thus, they cannot be considered a Great Founder. (Note that, in Burja&#8217;s view, this means that Napoleon is not a Great Founder for his military campaigns, but he <em>is </em>a Great Founder for things like standardizing units of measurement, creating conditions that led to growing nationalism throughout Europe, and reorganizing the French military structure to support the first foray into total war). In order to ensure that the individual continues shaping the world long after their death, they must create some mechanism by which to transfer their worldview down through the generations. These are our institutions, and their founders are those who have made the world we live in today.</p><p>The remainder of the book examines the implications of this theory from various angles.</p><p><strong>#6. Amaury De Riencourt, </strong><em><strong>The American Empire</strong></em></p><p><strong>#7. Peter Attia, </strong><em><strong>Outlive: The Science and Art of Longevity</strong></em></p><p><strong>#8. Alonso Cueto, </strong><em><strong>La Hora Azul</strong></em></p><p><strong>#9. Chris Miller, </strong><em><strong>Chip War: The Fight for the World&#8217;s Most Critical Technology</strong></em></p><p>I was late in reading this one, but it remains just as relevant (if not more so) today.</p><p>I wish I had read it earlier because it quite clearly lays out the case for investing in Intel as the fab of the future in the US.</p><p><strong>#10. William L. Hamilton, </strong><em><strong>Graph Representation Learning</strong></em></p><p><strong>#11. Patrick McGee, </strong><em><strong>Apple in China: The Capture of the World&#8217;s Greatest Company</strong></em></p><p>One of the best business books I&#8217;ve ever read. It treats the history of Apple as a history of its manufacturing division, centering on its ability to churn out massive volumes of some of the most advanced devices in the world. By centering the manufacturing story, China is inevitably centered as well &#8212; at first as an apprentice, later as a partner, and finally as an uneasy frenemy of sorts.</p><p>This book will teach you so much about Apple, about manufacturing, and about China. I cannot recommend it enough.</p><p><strong>#12. Jeff Pepper, </strong><em><strong>Mulan, Woman Warrior: An Easy-to-Read Story in Simplified Chinese and Pinyin, 240 Word Vocabulary</strong></em></p><p>I started learning a bit of Mandarin in 2025, and so far it&#8217;s been great fun! I typically like to learn languages through reading (I&#8217;ve used this technique before in <a href="https://www.chrishayduk.com/p/one-year-of-spanish-immersion-using-refold-dbd5d55206fe">Spanish</a> and, to a lesser extent, in French). So, to get started on my journey, I looked for some simple yet enjoyable books I could work my way through. </p><p>This led me to discover <a href="https://imagin8press.com/">Jeff Pepper</a>, who has simplified many culturally relevant Chinese texts to make them accessible to beginner and intermediate audiences. These include <em>The Journey to the West</em>, various Chinese folktales, a biography of Confucius, and more. <em>Mulan, Woman Warrior </em>is one of the easier of his published books (and tells the same story as my favorite childhood movie!), and so I found it extremely enjoyable to read. I&#8217;m very much looking forward to working through <em>The Journey to the West</em> this year.</p><p><strong>#13. Dan Brown, </strong><em><strong>The Secret of Secrets</strong></em></p><p>Gemini defines &#8220;narrative thrust&#8221; as follows:</p><blockquote><p>Narrative thrust is the compelling force that drives a story forward, creating momentum that pulls the reader from one scene to the next, keeping them engaged and eager to discover what happens next</p></blockquote><p>Narrative thrust covers a multitude of sins. Dan Brown, for all his faults, knows how to tell a story with thrust. <em>The Secret of Secrets </em>is no different.</p><p><strong>#14. Fyodor Dostoevsky (translated by Michael R. Katz), </strong><em><strong>Crime and Punishment</strong></em></p><p>This is a book that I very much did not enjoy reading, but cannot stop thinking about now that I have read it. </p><p>For those who have read Dostoevsky, I&#8217;m sure the reasons why I didn&#8217;t enjoy the experience of reading the book are obvious &#8212; the stilted prose, the confined narrative arc, and the oppressive, suffocating atmosphere that permeates every scene. But for those same readers, the reasons I can&#8217;t stop thinking about it must be equally as obvious &#8212; the psychological drama, the probing questions into the nature of morality, and the battle between logical rationality and subconscious thought.</p><p>Two themes really struck me throughout this book. </p><p>The first is Raskolnikov&#8217;s ability to reason himself into positions that his subconscious refuses to comply with. To him and other characters in the book, his reasoning is flawless when he expounds on his life philosophy (which, to summarize, is that he is a Napoleonic figure, an example of a higher man, and, as such, is unbound by the constraints of standard morality). In particular, when he is alone and able to ruminate, he frequently finds himself reinforcing this idea in his head. However, whenever confronted with real human connection throughout the story, his Ubermensch idealization breaks down and gives way to a more Christian morality. In addition, there are moments, even when Raskolnikov is alone and avoiding human connection, where thoughts bubble up from his subconscious, questioning his crime. This dichotomy between what he wants to be (the Ubermensch) and what he subconsciously knows he is (the Christian) tears him apart internally throughout the novel.</p><p>The second theme concerns the book's ending. After admitting to his crimes, Raskolnikov is imprisoned in Siberia. At the beginning of his imprisonment, he continues his oscillation between feeling himself an Ubermensch and hating himself for it. Only after a year of isolation, during which Sonya Marmeladova, the daughter of a man Raskolnikov helped earlier in the novel, fell ill and could no longer visit him, did Raskolnikov repent and release himself from his delusions. In a sense, the imprisonment in Siberia acted as a death &amp; resurrection for him, after which he had shed the sins of his past life and resurrected as a moral, human agent rather than an aspiring Ubermensch. </p><p>Both themes capture the centrality of human connection &amp; emotion in our cognition &#8212; Raskolnikov can&#8217;t maintain the behavior implied by his reasoning when he comes into contact with others, and he only fully gives up his delusions when he experiences the pain of briefly losing his last human connection, Sonya, during her illness. This tension between self-overcoming and human connection is worth keeping in mind, particularly for those of us who&nbsp;<em>do&nbsp;</em>aspire to more and, in doing so, shed a bit of our natural humanity.</p><p>There is another theme I&#8217;d like to explore, but I&#8217;ll refrain from going on too much of a tangent here. I&#8217;ll just say that there are also really interesting parallels between the second theme I mentioned, the life of Edmond Dantes from <em>The Count of Monte Cristo</em>, and the life of Napoleon. I hope to explore these parallels in-depth in a future article.</p><p><strong>#15. Tae Kim, </strong><em><strong>The Nvidia Way: Jensen Huang and the Making of a Tech Giant</strong></em></p><p>A really fun history of today&#8217;s most important company. Lots of great insights into how Jensen works.</p><p><strong>#16. Harry M. Schey, </strong><em><strong>Div, Grad, Curl, and All That: An Informal Text on Vector Calculus</strong></em></p><p>This book builds up the machinery of multivariable calculus through the lens of electromagnetism and physics applications. I&#8217;ve been trying to go back and review some math (e.g., calculus, linear algebra, etc.) from a more physics- and geometry-forward perspective so that I can improve my intuitions for the machinery of those topics. <em>Div, Grad, Curl, and All That</em> was great for that purpose.</p><p><strong>#17. Michio Kaku, </strong><em><strong>Quantum Supremacy</strong></em></p><p>Yes, I&#8217;m aware that many of Kaku&#8217;s books devolve into quackery. Yes, I&#8217;m aware that <a href="https://scottaaronson.blog/?p=7321">Scott Aaronson</a> called this &#8220;the worst book about quantum computing&#8230; that I&#8217;ve ever encountered&#8221;. And yes, it does live up to that illustrious billing.</p><p>And yet, I still find myself feeling invigorated to learn more and explore new areas every time I read one of Kaku&#8217;s books. I&#8217;m not sure what this says about me, but here we are.</p><p><strong>#18. Richard P. Feynman, </strong><em><strong>Surely You&#8217;re Joking, Mr. Feynman! Adventures of a Curious Character</strong></em></p><p>Surprisingly, Feynman&#8217;s autobiography focuses quite little on his actual work in physics. Instead, we get to peer into the creative and whimsical mind at play of one of the titans of 20th-century physics. Expect lots of stories of mischief and side quests, like lockpicking at Los Alamos or learning to play the drums to join a band during Carnival in Brazil.</p><p><strong>#19. Dwarkesh Patel, </strong><em><strong>The Scaling Era: An Oral History of AI, 2019&#8211;2025</strong></em></p><p>This book cuts up many of Dwarkesh&#8217;s interviews and groups them together in chapters organized by theme (e.g., &#8220;Chapter 1: Scaling&#8221;, &#8220;Chapter 2: Evals&#8221;, etc). </p><p>Even if you&#8217;ve listened to all of these podcasts previously (as I have), the book still provides a lot of value through its theme-based organization and the extensive footnotes/marginalia that Dwarkesh has included to clarify topics and provide additional detail.</p><p>Overall, a nice book to quickly assess where we are and where we&#8217;re going in AI.</p><p><strong>#20. Tyler Cowen and Daniel Gross, </strong><em><strong>Talent: How to Identify Energizers, Creatives, and Winners Around the World</strong></em></p><p>Finding above-average (or even great) employees is not too difficult &#8212; we broadly know how to conduct interviews to find people who are competent in their role and conscientious in their execution. </p><p>But how do you find people who are truly exceptional?</p><p>In <em>Talent</em>, Cowen and Gross aim to answer this question by covering interesting, non-standard interview techniques to help quickly assess whether someone is an outlier among outliers. Very useful if you are a venture capitalist, startup founder, or anyone else aiming to find people who are truly at the tails of the distribution.</p><div><hr></div><p>Overall, this was a bit of a lighter year of reading than I usually have. In 2026, I&#8217;m aiming to increase both the volume and the complexity of the reading I&#8217;m doing &#8212; hopefully, there will be many more math, physics, and deep learning books in next year&#8217;s list than were in this year&#8217;s. </p><p>In addition, I really like learning history (both political and business) through biographies, and I got away from that a bit in 2025. I&#8217;m going to work on incorporating many more of those in 2026. </p><p>Lastly, my reading over the last few years has slanted <em>heavily </em>towards non-fiction, but, like Raskolnikov, I&#8217;ve been missing the human connection and insights that great fiction can bring. My goal for 2026 is to explore many more of the great books, particularly the seminal novels of the 19th century.</p>]]></content:encoded></item><item><title><![CDATA[A Tale of Two Futures]]></title><description><![CDATA[America is building Skynet. China is building The Jetsons.]]></description><link>https://www.chrishayduk.com/p/a-tale-of-two-futures</link><guid isPermaLink="false">https://www.chrishayduk.com/p/a-tale-of-two-futures</guid><dc:creator><![CDATA[Chris Hayduk]]></dc:creator><pubDate>Sun, 11 Jan 2026 23:07:46 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!TYv0!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6ac0eb39-fc7b-4b08-9d86-b29cd9fa21ef_5504x3072.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!TYv0!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6ac0eb39-fc7b-4b08-9d86-b29cd9fa21ef_5504x3072.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!TYv0!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6ac0eb39-fc7b-4b08-9d86-b29cd9fa21ef_5504x3072.jpeg 424w, https://substackcdn.com/image/fetch/$s_!TYv0!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6ac0eb39-fc7b-4b08-9d86-b29cd9fa21ef_5504x3072.jpeg 848w, https://substackcdn.com/image/fetch/$s_!TYv0!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6ac0eb39-fc7b-4b08-9d86-b29cd9fa21ef_5504x3072.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!TYv0!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6ac0eb39-fc7b-4b08-9d86-b29cd9fa21ef_5504x3072.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!TYv0!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6ac0eb39-fc7b-4b08-9d86-b29cd9fa21ef_5504x3072.jpeg" width="1456" height="813" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/6ac0eb39-fc7b-4b08-9d86-b29cd9fa21ef_5504x3072.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:813,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:10015133,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.chrishayduk.com/i/184223287?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6ac0eb39-fc7b-4b08-9d86-b29cd9fa21ef_5504x3072.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!TYv0!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6ac0eb39-fc7b-4b08-9d86-b29cd9fa21ef_5504x3072.jpeg 424w, https://substackcdn.com/image/fetch/$s_!TYv0!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6ac0eb39-fc7b-4b08-9d86-b29cd9fa21ef_5504x3072.jpeg 848w, https://substackcdn.com/image/fetch/$s_!TYv0!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6ac0eb39-fc7b-4b08-9d86-b29cd9fa21ef_5504x3072.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!TYv0!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6ac0eb39-fc7b-4b08-9d86-b29cd9fa21ef_5504x3072.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The US and China are engaged in a two-country race towards an AI-powered future. Both countries have directed funding, policy initiatives, and talent towards the AI sector at levels matched only by the internet buildout of the 1990s or railroad construction of the 1880s. But both of these countries are building towards diametrically opposed futures. </p><p>One country has taken the view that it is on the path to artificial superintelligence (ASI) &#8212; the point at which AI will be more effective than the most capable humans at every task. The view of their prominent AI labs is that once ASI is achieved, there will be a runaway intelligence explosion, with the AI rapidly improving itself and reaching unthinkable levels of intelligence. This genius AI will then be able to solve our most pressing problems in mathematics, physics, philosophy, and more with minimal difficulty.</p><p>The other country has taken the view that it is not peak intelligence that matters, but rather the distribution of intelligence. It aims to develop intelligent AI models (though not superintelligent) that are fast and cheap enough to be embedded in machines across the economy. It wants household robots in every home, talking cars, and refrigerators that can do the grocery shopping.</p><p>One future is top-down, the other is bottom-up. One is centralized, the other is decentralized. One results in gains accruing to the select few corporations, the other results in gains throughout the entire economy. One is Skynet, the other is The Jetsons.</p><p>The great irony of the current AI race situation is that the centralized SkyNet future belongs to the democratic United States, and the decentralized Jetsons future belongs to authoritarian China.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.chrishayduk.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Musings by Chris Hayduk is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h2>The Skynet Future</h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!9mVt!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6c6b0ca-a9ec-4961-9d18-5ffc475e7c45_5504x3072.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!9mVt!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6c6b0ca-a9ec-4961-9d18-5ffc475e7c45_5504x3072.jpeg 424w, https://substackcdn.com/image/fetch/$s_!9mVt!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6c6b0ca-a9ec-4961-9d18-5ffc475e7c45_5504x3072.jpeg 848w, https://substackcdn.com/image/fetch/$s_!9mVt!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6c6b0ca-a9ec-4961-9d18-5ffc475e7c45_5504x3072.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!9mVt!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6c6b0ca-a9ec-4961-9d18-5ffc475e7c45_5504x3072.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!9mVt!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6c6b0ca-a9ec-4961-9d18-5ffc475e7c45_5504x3072.jpeg" width="1456" height="813" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e6c6b0ca-a9ec-4961-9d18-5ffc475e7c45_5504x3072.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:813,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:8615005,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.chrishayduk.com/i/184223287?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6c6b0ca-a9ec-4961-9d18-5ffc475e7c45_5504x3072.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!9mVt!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6c6b0ca-a9ec-4961-9d18-5ffc475e7c45_5504x3072.jpeg 424w, https://substackcdn.com/image/fetch/$s_!9mVt!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6c6b0ca-a9ec-4961-9d18-5ffc475e7c45_5504x3072.jpeg 848w, https://substackcdn.com/image/fetch/$s_!9mVt!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6c6b0ca-a9ec-4961-9d18-5ffc475e7c45_5504x3072.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!9mVt!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6c6b0ca-a9ec-4961-9d18-5ffc475e7c45_5504x3072.jpeg 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Skynet is the fictional AI developed by Cyberdyne Systems in the original Terminator movie. It&#8217;s a large, powerful AI system that, once deployed, rapidly increases its intelligence to the point of becoming self-aware. When it becomes self-aware, it becomes self-interested and decides that the best way to preserve its existence is to eliminate <em>all </em>humans. Skynet then appropriates the nuclear codes of the United States and launches these nuclear weapons in an attempt to eliminate the human race from the face of the Earth. Small groups of humans survive the nuclear fallout, and they live in a post-apocalyptic world fighting robots directed by Skynet that are attempting to exterminate humanity once and for all.</p><p>This vision is very apocalyptic, but I think it captures the sentiment of the US AI scene better than any other popular depiction of AI. From the initial ambitions of Skynet running the United States to its intelligence explosion to its final destruction of humanity, these views are not outlandish in Silicon Valley. In fact, they may be the norm among AI researchers and AI lab CEOs. And these views have substantial implications for how AI research is playing out in the United States.</p><p>The CEOs of AI research labs are explicitly building towards this form of superintelligence. Crucially, these labs view pushing the intelligence frontier of these models as the core goal of their research. They explicitly look to benchmarks that measure intelligence on extremely difficult tasks (such as FrontierMath, GDPVal, and ARC-AGI-2) as the core metrics that they&#8217;re optimizing against. Their goal is to produce a &#8220;country of geniuses in a datacenter&#8221;, as Dario Amodei put it in his article &#8220;Machines of Loving Grace&#8221;.  Amodei believes that, once achieved, artificial superintelligence could compress 50-100 years of biological research into 5-10 years. </p><p>Moreover, many of the prominent figures in AI view the path to superintelligence as a race. They believe that as we move closer to superintelligence, we will be able to achieve an automated AI researcher that can analyze its own codebase and improve it rapidly. These algorithmic gains from AI producing its own code improvements will result in an intelligence explosion, such that the first lab to produce an automated AI researcher will immediately gain an insurmountable lead in intelligence over the other labs. As such, not only do the AI lab leaders <em>believe</em> this Skynet scenario, but they also view it as a race that must be won at all costs. The very existence of their companies in their minds depends upon reaching the superintelligent AI first. To understand the pace of AI investment and the amount of that investment that becomes allocated to training ever-larger and more capable models (rather than more cost-efficient or broadly distributed models), you need to internalize that the prominent figures in AI strongly believe this to be the true state of the world. </p><p>To race towards superintelligence, massive increases in the two core inputs to AI training are needed: data and compute. As a result, the AI labs have invested hundreds of billions of dollars into massive compute scale-outs, data acquisitions, and RL environment development. New data centers are coming online that consume gigawatts of electricity to train larger models with higher parameter counts. In addition, these new models are fed data and trained in RL environments produced by hired PhDs and leading experts across math, computer science, finance, and more. </p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!75Fj!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c886bd8-a15e-41f2-8503-097d3a5cefef_1920x1080.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!75Fj!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c886bd8-a15e-41f2-8503-097d3a5cefef_1920x1080.png 424w, https://substackcdn.com/image/fetch/$s_!75Fj!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c886bd8-a15e-41f2-8503-097d3a5cefef_1920x1080.png 848w, https://substackcdn.com/image/fetch/$s_!75Fj!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c886bd8-a15e-41f2-8503-097d3a5cefef_1920x1080.png 1272w, https://substackcdn.com/image/fetch/$s_!75Fj!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c886bd8-a15e-41f2-8503-097d3a5cefef_1920x1080.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!75Fj!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c886bd8-a15e-41f2-8503-097d3a5cefef_1920x1080.png" width="1456" height="819" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/9c886bd8-a15e-41f2-8503-097d3a5cefef_1920x1080.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:819,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:119304,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.chrishayduk.com/i/184223287?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c886bd8-a15e-41f2-8503-097d3a5cefef_1920x1080.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!75Fj!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c886bd8-a15e-41f2-8503-097d3a5cefef_1920x1080.png 424w, https://substackcdn.com/image/fetch/$s_!75Fj!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c886bd8-a15e-41f2-8503-097d3a5cefef_1920x1080.png 848w, https://substackcdn.com/image/fetch/$s_!75Fj!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c886bd8-a15e-41f2-8503-097d3a5cefef_1920x1080.png 1272w, https://substackcdn.com/image/fetch/$s_!75Fj!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c886bd8-a15e-41f2-8503-097d3a5cefef_1920x1080.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>To afford the energy, compute, and data to produce these models that push benchmark metrics on FrontierMath and GDPVal, more and more centralization is encouraged in AI research. As shown in the chart above, the cost of building a frontier data center has been increasing exponentially, from around $7 billion in 2022 (when ChatGPT was first released) to a projected $106 billion in 2027. Hence, if there is a fixed amount of private &amp; public funding available to AI companies, it is in investors&#8217; interest to allocate that funding to a small number of companies so that they can afford the requisite frontier data centers to train these models. With funding too widely distributed, no single player would be able to produce a model that outperformed the state-of-the-art on these frontier benchmarks.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!vyXb!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff976423b-3d9f-43f1-9b0d-fdf825a5359b_1920x1080.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!vyXb!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff976423b-3d9f-43f1-9b0d-fdf825a5359b_1920x1080.png 424w, https://substackcdn.com/image/fetch/$s_!vyXb!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff976423b-3d9f-43f1-9b0d-fdf825a5359b_1920x1080.png 848w, https://substackcdn.com/image/fetch/$s_!vyXb!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff976423b-3d9f-43f1-9b0d-fdf825a5359b_1920x1080.png 1272w, https://substackcdn.com/image/fetch/$s_!vyXb!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff976423b-3d9f-43f1-9b0d-fdf825a5359b_1920x1080.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!vyXb!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff976423b-3d9f-43f1-9b0d-fdf825a5359b_1920x1080.png" width="728" height="409.5" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f976423b-3d9f-43f1-9b0d-fdf825a5359b_1920x1080.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:false,&quot;imageSize&quot;:&quot;normal&quot;,&quot;height&quot;:819,&quot;width&quot;:1456,&quot;resizeWidth&quot;:728,&quot;bytes&quot;:242767,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.chrishayduk.com/i/184223287?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff976423b-3d9f-43f1-9b0d-fdf825a5359b_1920x1080.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:&quot;center&quot;,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!vyXb!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff976423b-3d9f-43f1-9b0d-fdf825a5359b_1920x1080.png 424w, https://substackcdn.com/image/fetch/$s_!vyXb!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff976423b-3d9f-43f1-9b0d-fdf825a5359b_1920x1080.png 848w, https://substackcdn.com/image/fetch/$s_!vyXb!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff976423b-3d9f-43f1-9b0d-fdf825a5359b_1920x1080.png 1272w, https://substackcdn.com/image/fetch/$s_!vyXb!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff976423b-3d9f-43f1-9b0d-fdf825a5359b_1920x1080.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>We can observe this trend more clearly in the training compute cost of frontier models from 2012 through 2025. We can see that model training costs rapidly increased from roughly $3 million in 2022 to over $300 million in 2025. In addition, the trend line shows this cost increasing at a rate of 0.5 orders of magnitude per year, indicating that the cost of a single training run next year (2027) will increase to about $3 billion. Projected forward to the end of the decade (2030), a training run would cost roughly $100 billion (with the data centers powering such a run likely costing in excess of $1 trillion).</p><p>Given all of this, we can think of the American AI ecosystem as a bet on increasing centralization and increasing scale. It is a bet on a benevolent Skynet future &#8212; leveraging unprecedented resources to build a single, massive AI model capable of solving the most difficult problems in science, technology, politics, and philosophy. </p><h2>The Jetsons Future</h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!JCQV!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F93b633a2-3b1f-4c7c-9b0e-6ff0d60a2ae0_5504x3072.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!JCQV!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F93b633a2-3b1f-4c7c-9b0e-6ff0d60a2ae0_5504x3072.jpeg 424w, https://substackcdn.com/image/fetch/$s_!JCQV!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F93b633a2-3b1f-4c7c-9b0e-6ff0d60a2ae0_5504x3072.jpeg 848w, https://substackcdn.com/image/fetch/$s_!JCQV!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F93b633a2-3b1f-4c7c-9b0e-6ff0d60a2ae0_5504x3072.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!JCQV!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F93b633a2-3b1f-4c7c-9b0e-6ff0d60a2ae0_5504x3072.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!JCQV!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F93b633a2-3b1f-4c7c-9b0e-6ff0d60a2ae0_5504x3072.jpeg" width="1456" height="813" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/93b633a2-3b1f-4c7c-9b0e-6ff0d60a2ae0_5504x3072.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:813,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:8652965,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.chrishayduk.com/i/184223287?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F93b633a2-3b1f-4c7c-9b0e-6ff0d60a2ae0_5504x3072.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!JCQV!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F93b633a2-3b1f-4c7c-9b0e-6ff0d60a2ae0_5504x3072.jpeg 424w, https://substackcdn.com/image/fetch/$s_!JCQV!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F93b633a2-3b1f-4c7c-9b0e-6ff0d60a2ae0_5504x3072.jpeg 848w, https://substackcdn.com/image/fetch/$s_!JCQV!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F93b633a2-3b1f-4c7c-9b0e-6ff0d60a2ae0_5504x3072.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!JCQV!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F93b633a2-3b1f-4c7c-9b0e-6ff0d60a2ae0_5504x3072.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The Jetsons is an animated sitcom from the 1960s that depicts a future defined not by a single technological breakthrough but by the accumulation of countless small conveniences &#8212; flying cars, household robots, and apartments that predict each family&#8217;s needs. The Jetsons&#8217; future is not one of transcendence but of leisure &#8212; technology has not produced a godlike intelligence but has instead seeped into every object, automating away the drudgery of daily life. The show&#8217;s vision is one of abundance through distribution: no single machine is particularly impressive, but the sheer proliferation of helpful machines has transformed the nature of work and home life entirely.</p><p>The Jetsons aired the same year as the Cuban Missile Crisis. Sixty years later, it is the CCP &amp; China, not the United States, that is building toward its vision.</p><p>Chinese AI labs burst onto the scene in early 2025 with the release of DeepSeek-R1. Unlike its American counterparts, the notable aspect of DeepSeek&#8217;s model was <em>not</em> its raw performance &#8212; it was strong, but it lagged behind the frontier. The truly impressive aspect of DeepSeek-R1 was that it performed similarly to frontier models at a fraction of the cost, both in terms of serving cost and training cost. For example, as I detailed in another article (<a href="https://www.chrishayduk.com/p/open-source-llms-are-eating-the-world">Open Source LLMs Are Eating the World</a>), DeepSeek-R1 was able to perform nearly as well as OpenAI o1 (the frontier model at the time) on the MMLU Pro benchmark, while only costing $6.75 to run the full benchmark suite compared to o1&#8217;s $75. This represented an 11x drop in the cost to serve the model at roughly equivalent performance levels.</p><p>The success of DeepSeek-R1 has sparked a wave of innovation in open-source Chinese AI. Various companies have entered the fray, including Alibaba with its Qwen series, Z.ai with its GLM series, and Moonshot AI with its Kimi series. Each of these three core competitors, along with DeepSeek, has steadily pushed the cost of economically useful intelligence towards zero. </p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!oZCI!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F37587488-3401-4e68-b293-1088e5357ca0_990x509.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!oZCI!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F37587488-3401-4e68-b293-1088e5357ca0_990x509.png 424w, https://substackcdn.com/image/fetch/$s_!oZCI!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F37587488-3401-4e68-b293-1088e5357ca0_990x509.png 848w, https://substackcdn.com/image/fetch/$s_!oZCI!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F37587488-3401-4e68-b293-1088e5357ca0_990x509.png 1272w, https://substackcdn.com/image/fetch/$s_!oZCI!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F37587488-3401-4e68-b293-1088e5357ca0_990x509.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!oZCI!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F37587488-3401-4e68-b293-1088e5357ca0_990x509.png" width="990" height="509" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/37587488-3401-4e68-b293-1088e5357ca0_990x509.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:509,&quot;width&quot;:990,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:54656,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.chrishayduk.com/i/184223287?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F37587488-3401-4e68-b293-1088e5357ca0_990x509.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!oZCI!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F37587488-3401-4e68-b293-1088e5357ca0_990x509.png 424w, https://substackcdn.com/image/fetch/$s_!oZCI!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F37587488-3401-4e68-b293-1088e5357ca0_990x509.png 848w, https://substackcdn.com/image/fetch/$s_!oZCI!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F37587488-3401-4e68-b293-1088e5357ca0_990x509.png 1272w, https://substackcdn.com/image/fetch/$s_!oZCI!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F37587488-3401-4e68-b293-1088e5357ca0_990x509.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Produced by GPT 5.2 Pro using MMLU Pro benchmark results</figcaption></figure></div><p>In addition, the speed of innovation has compressed the time delta in model performance between closed-source AI labs and their Chinese open-source equivalents. The chart above shows the catch-up time required for open-source AI models to match the performance of closed-source models at different performance thresholds on the MMLU-Pro benchmark. We can see that earlier performance levels, such as the 60% threshold, required roughly two-thirds of a year before open-source AI could match closed-source AI. However, recent performance thresholds have been achieved in far less time, between a quarter to a third of a year. Chinese AI labs have ramped up their investments in data centers and energy. They now have access to purchase NVIDIA H200 chips, and the Chinese chip ecosystem is maturing more quickly than expected. As a result, we should likely expect this time gap to continue to compress rather than to expand. </p><p>The implications here are genuinely massive. This chart shows that open-source AI models from China can match the performance of our leading closed-source models in less than 6 months, often at an order of magnitude lower cost. </p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Wq9P!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3f6107d3-497b-434c-a47f-92354ccbddd2_1172x940.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Wq9P!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3f6107d3-497b-434c-a47f-92354ccbddd2_1172x940.jpeg 424w, https://substackcdn.com/image/fetch/$s_!Wq9P!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3f6107d3-497b-434c-a47f-92354ccbddd2_1172x940.jpeg 848w, https://substackcdn.com/image/fetch/$s_!Wq9P!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3f6107d3-497b-434c-a47f-92354ccbddd2_1172x940.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!Wq9P!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3f6107d3-497b-434c-a47f-92354ccbddd2_1172x940.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Wq9P!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3f6107d3-497b-434c-a47f-92354ccbddd2_1172x940.jpeg" width="1172" height="940" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3f6107d3-497b-434c-a47f-92354ccbddd2_1172x940.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:940,&quot;width&quot;:1172,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;Image&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Image" title="Image" srcset="https://substackcdn.com/image/fetch/$s_!Wq9P!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3f6107d3-497b-434c-a47f-92354ccbddd2_1172x940.jpeg 424w, https://substackcdn.com/image/fetch/$s_!Wq9P!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3f6107d3-497b-434c-a47f-92354ccbddd2_1172x940.jpeg 848w, https://substackcdn.com/image/fetch/$s_!Wq9P!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3f6107d3-497b-434c-a47f-92354ccbddd2_1172x940.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!Wq9P!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3f6107d3-497b-434c-a47f-92354ccbddd2_1172x940.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>In parallel with these advancements in open-source AI, China has been experiencing two massive buildouts over the last five years. The first has been a huge increase in energy generation, specifically in solar and nuclear energy. These rapid increases in energy generation, which frequently exceed the rest of the world combined, for example, in installed solar capacity, are enabling manufacturing at scales never before seen. In particular, the improvements to solar and batteries that are occurring and will continue to occur due to technological and cost improvements driven by this increased manufacturing will allow energy not just to be more plentiful overall, but also to be more local to specific needs. What this means practically is that energy will become local and mobile, allowing for various devices to become much more energy-intensive than they have been previously through the use of improved batteries and local solar panels. This will allow devices to include the substantial onboard compute required to run the top open-source AI models.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!MY6h!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff8a39805-ed7b-4ba4-b413-be04b84cb56b_1024x622.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!MY6h!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff8a39805-ed7b-4ba4-b413-be04b84cb56b_1024x622.png 424w, https://substackcdn.com/image/fetch/$s_!MY6h!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff8a39805-ed7b-4ba4-b413-be04b84cb56b_1024x622.png 848w, https://substackcdn.com/image/fetch/$s_!MY6h!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff8a39805-ed7b-4ba4-b413-be04b84cb56b_1024x622.png 1272w, https://substackcdn.com/image/fetch/$s_!MY6h!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff8a39805-ed7b-4ba4-b413-be04b84cb56b_1024x622.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!MY6h!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff8a39805-ed7b-4ba4-b413-be04b84cb56b_1024x622.png" width="1024" height="622" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f8a39805-ed7b-4ba4-b413-be04b84cb56b_1024x622.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:622,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;Q&amp;A: The global 'trade war' over China's booming EV industry - Carbon Brief&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Q&amp;A: The global 'trade war' over China's booming EV industry - Carbon Brief" title="Q&amp;A: The global 'trade war' over China's booming EV industry - Carbon Brief" srcset="https://substackcdn.com/image/fetch/$s_!MY6h!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff8a39805-ed7b-4ba4-b413-be04b84cb56b_1024x622.png 424w, https://substackcdn.com/image/fetch/$s_!MY6h!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff8a39805-ed7b-4ba4-b413-be04b84cb56b_1024x622.png 848w, https://substackcdn.com/image/fetch/$s_!MY6h!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff8a39805-ed7b-4ba4-b413-be04b84cb56b_1024x622.png 1272w, https://substackcdn.com/image/fetch/$s_!MY6h!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff8a39805-ed7b-4ba4-b413-be04b84cb56b_1024x622.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The second key build-out has been in advanced manufacturing. China was <em>already</em> the world's manufacturing center before this push into advanced manufacturing. But now it has moved up the value chain and very quickly has gone from a laggard in key industries to dominating them. For instance, five years ago, China was largely irrelevant in the electric car market. Now its electric car companies are leading the world and outselling Tesla. Several leading companies are now making strong pushes into humanoid robots and are setting the pace in that category. </p><p>This confluence of advanced manufacturing in robotics &amp; battery-powered vehicles, increased energy generation (specifically solar), and open-source AI that is energy- and compute-efficient will allow China to develop a truly intelligence-powered economy. The major factors will be in place to have human-level AI embedded into a large share of both consumer and industrial products. </p><p>With this approach, China aims to enable a new level of general abundance with household robots, self-driving electric cars, self-directed delivery drones, and household appliances that can make decisions for themselves, such as a refrigerator that can detect when you&#8217;re running low on specific supplies and order them agentically. </p><p>And crucially, this future does <em>not</em> depend on reaching superintelligence. As I&#8217;ve detailed in my other article, <a href="https://www.chrishayduk.com/p/open-source-llms-are-eating-the-world">Open Source LLMs Are Eating the World</a>, many economically relevant tasks operate in a task-saturation regime. That is, once the models exceed some threshold level of performance, future increases in model scale and training compute do not make meaningful differences in task-level performance. Moreover, models today are <em>already</em> capable of performing many economically viable tasks, such as coding complex apps, serving as customer support agents, and more. Hence, this broad deployment of cheap AI in physical goods will deliver returns quite quickly.</p><p>Making this intelligence cheap and abundant through energy- and compute-efficient open source AI will unlock massive economic value across the spectrum. There isn&#8217;t much doubt about this. The Jetsons future is clearly within reach. However, there <em>is</em> a question mark whether we will reach the benevolent Skynet future.</p><h2>Consequences of the Divide</h2><p>US labs are betting on transcendent intelligence. Chinese labs are betting on abundant intelligence. Who wins depends on which future actually arrives.</p><p>Four scenarios are possible. Only one favors the American approach.</p><h3>The Scenario Matrix</h3><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!b0jh!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F765ab45d-a2f6-41c7-945d-93828f81c4c1_1820x686.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!b0jh!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F765ab45d-a2f6-41c7-945d-93828f81c4c1_1820x686.png 424w, https://substackcdn.com/image/fetch/$s_!b0jh!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F765ab45d-a2f6-41c7-945d-93828f81c4c1_1820x686.png 848w, https://substackcdn.com/image/fetch/$s_!b0jh!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F765ab45d-a2f6-41c7-945d-93828f81c4c1_1820x686.png 1272w, https://substackcdn.com/image/fetch/$s_!b0jh!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F765ab45d-a2f6-41c7-945d-93828f81c4c1_1820x686.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!b0jh!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F765ab45d-a2f6-41c7-945d-93828f81c4c1_1820x686.png" width="1456" height="549" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/765ab45d-a2f6-41c7-945d-93828f81c4c1_1820x686.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:549,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:133144,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.chrishayduk.com/i/184223287?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F765ab45d-a2f6-41c7-945d-93828f81c4c1_1820x686.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!b0jh!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F765ab45d-a2f6-41c7-945d-93828f81c4c1_1820x686.png 424w, https://substackcdn.com/image/fetch/$s_!b0jh!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F765ab45d-a2f6-41c7-945d-93828f81c4c1_1820x686.png 848w, https://substackcdn.com/image/fetch/$s_!b0jh!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F765ab45d-a2f6-41c7-945d-93828f81c4c1_1820x686.png 1272w, https://substackcdn.com/image/fetch/$s_!b0jh!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F765ab45d-a2f6-41c7-945d-93828f81c4c1_1820x686.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>From the above matrix, we see that the American bet requires threading a needle: superintelligence must be achievable, it must trigger a runaway intelligence explosion, <em>and</em> that explosion must translate into massive economic returns unconstrained by physical bottlenecks.</p><p>Remove any link in that chain, and the calculus shifts.</p><p><strong>If superintelligence arrives but can&#8217;t escape physical constraints:</strong> The leading lab pulls ahead on benchmarks, but drug discovery still bottlenecks at FDA trials. Robotics still bottlenecks at manufacturing. Most economically useful tasks don&#8217;t require superintelligence anyway. Open-source competitors deliver similar real-world value at a fraction of the cost.</p><p><strong>If superintelligence arrives but no takeoff occurs:</strong> Energy infrastructure takes years to build. Training runs take months. Even with a superintelligent AI optimizing your codebase, the results manifest slowly enough for competitors to close the gap. No insurmountable lead materializes.</p><p><strong>If superintelligence never arrives:</strong> Intelligence gains follow a sigmoid curve&#8212;rapid improvement, then diminishing returns. At that plateau, the race shifts from &#8220;who&#8217;s smartest&#8221; to &#8220;who&#8217;s cheapest and most distributed.&#8221; China wins that race.</p><h3>The Asymmetric Bet</h3><p>The Jetsons future requires no miracles. Cheap, capable AI embedded in robots, vehicles, and appliances delivers value whether or not superintelligence is possible. China&#8217;s bet pays off in three of four scenarios.</p><p>The Skynet future requires everything to go right. Superintelligence must be reachable, takeoff must occur, and physical constraints must not bind. America&#8217;s bet pays off in one scenario.</p><h3>The Implication</h3><p>We see from the scenarios enumerated above that the US AI Lab approach is decisively dominant in only one of them. In all other scenarios, Chinese open-source AI is able to keep pace with the closed-source frontier, and, in doing so, it is guaranteed to bring about the Jetsons future that China is building towards. The American benevolent Skynet future is far from guaranteed. </p><p>In sum, to ensure that the United States broadly benefits from the AI revolution it has itself started, we need to take a page from the Chinese AI playbook. We must ensure that, even if superintelligence is out of reach, we will have cheap, abundant intelligence suffusing the economy. We must ensure that our portable energy infrastructure (i.e., solar panels and batteries), our robotics manufacturing capabilities, and our open source AI efforts are sufficient to power a truly intelligent economy. The failure to do so may cede technological leadership in the 21st century to the CCP &amp; China. </p><p></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.chrishayduk.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Musings by Chris Hayduk is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Open Source LLMs Are Eating the World]]></title><description><![CDATA[We are benchmarking LLMs incorrectly to predict economic utility]]></description><link>https://www.chrishayduk.com/p/open-source-llms-are-eating-the-world</link><guid isPermaLink="false">https://www.chrishayduk.com/p/open-source-llms-are-eating-the-world</guid><dc:creator><![CDATA[Chris Hayduk]]></dc:creator><pubDate>Fri, 09 Jan 2026 20:15:04 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!_XZh!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F57385725-8fc4-46fa-a04e-e3e08d16b487_1596x1136.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>The default way we evaluate large language models is fundamentally misaligned with how they create economic value. We track frontier capabilities across broad benchmarks (such as ARC-AGI-2, FrontierMath, and SWE-Bench Verified) and implicitly assume that whoever leads on these metrics captures the most value. </p><p>However, this view assumes that what matters is the model&#8217;s maximum intelligence across a broad range of tasks &#8212; the &#8220;PhD intelligence for all&#8221; mantra repeated by the large labs. </p><p>For practical company building, this framing is wrong, and understanding why reveals a structural advantage for open source models that holds regardless of when (or whether) we achieve AGI.</p><h2>I. Introduction: The Benchmarking Problem</h2><p>The standard narrative goes something like this: general model value scales with general performance. A model that scores higher on a diverse battery of benchmarks is more valuable than one that scores lower, and the companies training the most capable models will capture the lion&#8217;s share of economic returns.</p><p>But this misses how value actually gets created in practice. Companies don&#8217;t build products that require uniformly excellent performance across all possible tasks. They build for specific use cases: contract analysis, customer support, code generation, and medical documentation. Revenue comes from solving customer problems, and customer problems are specific. The &#8220;average&#8221; benchmark performance that frontier labs optimize for doesn&#8217;t map to any real product; it&#8217;s just an abstraction that obscures the actual economics.</p><p>For any given application, what matters is whether your model is good enough at <em>this particular thing</em>, not whether it can solve PhD-level mathematics problems or write publishable research. A legal tech startup needs strong performance on contract reasoning and citation accuracy. A customer support platform needs reliability in intent classification and tone. Neither benefits from improvements to the model&#8217;s ability to prove novel theorems.</p><p>Moreover, the relationship between capability and value is S-curved. Early capability improvements unlock entirely new use cases: a model that goes from 40% to 70% accuracy on a task might cross the threshold from &#8220;useless&#8221; to &#8220;useful with human oversight.&#8221; But a model that goes from 92% to 96% often delivers no additional value, because the human workflow was already designed around spot-checking outputs, and the bottleneck has shifted elsewhere to latency, cost, integration complexity, or user experience.</p><p>This is the crux of the argument: once a model clears the capability threshold for a given task, further intelligence improvements face rapidly diminishing returns. The contract analysis tool that&#8217;s &#8220;good enough&#8221; for lawyers to trust with first-pass review doesn&#8217;t become twice as valuable when the underlying model gets twice as capable. It just becomes overprovisioned.</p><h2>II. The Task Saturation Phenomenon</h2><p>For any specific task, <strong>the marginal value of model capability saturates at some threshold.</strong> Beyond a certain point, users cannot meaningfully distinguish between a model of size X and a model of size n&#215;X for any n &#8805; 1.</p><p>Consider what&#8217;s happened to standard benchmarks over the past few years. The chart below tracks top scores on benchmarks like ARC, MMLU, Winograd, HellaSwag, GSM8K, and TruthfulQA against their human baselines:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!7ZLe!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fea090f68-5a7b-4c59-a68c-2f74fffa13dc_1256x780.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!7ZLe!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fea090f68-5a7b-4c59-a68c-2f74fffa13dc_1256x780.png 424w, https://substackcdn.com/image/fetch/$s_!7ZLe!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fea090f68-5a7b-4c59-a68c-2f74fffa13dc_1256x780.png 848w, https://substackcdn.com/image/fetch/$s_!7ZLe!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fea090f68-5a7b-4c59-a68c-2f74fffa13dc_1256x780.png 1272w, https://substackcdn.com/image/fetch/$s_!7ZLe!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fea090f68-5a7b-4c59-a68c-2f74fffa13dc_1256x780.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!7ZLe!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fea090f68-5a7b-4c59-a68c-2f74fffa13dc_1256x780.png" width="1256" height="780" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/ea090f68-5a7b-4c59-a68c-2f74fffa13dc_1256x780.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:780,&quot;width&quot;:1256,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:113058,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.chrishayduk.com/i/184028048?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fea090f68-5a7b-4c59-a68c-2f74fffa13dc_1256x780.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!7ZLe!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fea090f68-5a7b-4c59-a68c-2f74fffa13dc_1256x780.png 424w, https://substackcdn.com/image/fetch/$s_!7ZLe!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fea090f68-5a7b-4c59-a68c-2f74fffa13dc_1256x780.png 848w, https://substackcdn.com/image/fetch/$s_!7ZLe!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fea090f68-5a7b-4c59-a68c-2f74fffa13dc_1256x780.png 1272w, https://substackcdn.com/image/fetch/$s_!7ZLe!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fea090f68-5a7b-4c59-a68c-2f74fffa13dc_1256x780.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The pattern is consistent: rapid improvement followed by convergence toward (and often slightly beyond) human-level performance. Once a benchmark is effectively &#8220;solved,&#8221; additional capability improvements deliver zero marginal value for tasks that the benchmark measures. A model scoring 95% on MMLU isn&#8217;t twice as useful for MMLU-adjacent tasks as one scoring 90%. For most practical purposes, they&#8217;re equivalent.</p><h2>III. Reframing the Analysis: Cost at Fixed Performance</h2><p>If capability saturates for specific tasks, then the relevant question isn&#8217;t &#8220;which model is most capable?&#8221; but rather &#8220;which model solves my task at the lowest cost?&#8221;</p><p>Once we fix a performance threshold (the point at which a task is effectively solved) we can track how the cost to achieve that threshold evolves over time. The a16z team did exactly <a href="https://a16z.com/llmflation-llm-inference-cost/">this analysis for MMLU scores</a>:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!_XZh!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F57385725-8fc4-46fa-a04e-e3e08d16b487_1596x1136.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!_XZh!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F57385725-8fc4-46fa-a04e-e3e08d16b487_1596x1136.png 424w, https://substackcdn.com/image/fetch/$s_!_XZh!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F57385725-8fc4-46fa-a04e-e3e08d16b487_1596x1136.png 848w, https://substackcdn.com/image/fetch/$s_!_XZh!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F57385725-8fc4-46fa-a04e-e3e08d16b487_1596x1136.png 1272w, https://substackcdn.com/image/fetch/$s_!_XZh!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F57385725-8fc4-46fa-a04e-e3e08d16b487_1596x1136.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!_XZh!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F57385725-8fc4-46fa-a04e-e3e08d16b487_1596x1136.png" width="1456" height="1036" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/57385725-8fc4-46fa-a04e-e3e08d16b487_1596x1136.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1036,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:279620,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.chrishayduk.com/i/184028048?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F57385725-8fc4-46fa-a04e-e3e08d16b487_1596x1136.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!_XZh!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F57385725-8fc4-46fa-a04e-e3e08d16b487_1596x1136.png 424w, https://substackcdn.com/image/fetch/$s_!_XZh!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F57385725-8fc4-46fa-a04e-e3e08d16b487_1596x1136.png 848w, https://substackcdn.com/image/fetch/$s_!_XZh!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F57385725-8fc4-46fa-a04e-e3e08d16b487_1596x1136.png 1272w, https://substackcdn.com/image/fetch/$s_!_XZh!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F57385725-8fc4-46fa-a04e-e3e08d16b487_1596x1136.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The trend line shows roughly a 10&#215; cost reduction every year for a fixed capability level. But the more important pattern is <em>which models</em> sit on that cost frontier over time. Early in a capability tier&#8217;s lifecycle, closed-source models from frontier labs define the frontier. But within months, open-source alternatives emerge at dramatically lower price points.</p><p>Look at the progression for MMLU &gt; 83: GPT-4 at $45 per million tokens, then GPT-4o at ~$10, then Claude 3.5 Sonnet at ~$10, and finally Llama 3.1 70B pushing costs down toward $0.50. The same pattern plays out for every capability threshold: closed-source models solve the task first, and then open-source models quickly make it cheaper. </p><p>Thus, if we imagine a fixed benchmark score as a proxy for the threshold at which a task is &#8220;solved&#8221;, we see that closed source models have historically had a payoff horizon of roughly one year before open source models made </p><h2>IV. Case Study: MMLU Pro Replication Speed</h2><p>MMLU Pro extends the original MMLU benchmark by increasing the number of multiple-choice options from 4 to 10, introducing misleading distractors, and emphasizing reasoning-heavy questions. It&#8217;s a harder benchmark, which allows us to separate out the performance levels of recently released models.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!mz2q!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd676c533-3d7d-4f19-8f4b-a3e29c484839_1460x1534.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!mz2q!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd676c533-3d7d-4f19-8f4b-a3e29c484839_1460x1534.png 424w, https://substackcdn.com/image/fetch/$s_!mz2q!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd676c533-3d7d-4f19-8f4b-a3e29c484839_1460x1534.png 848w, https://substackcdn.com/image/fetch/$s_!mz2q!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd676c533-3d7d-4f19-8f4b-a3e29c484839_1460x1534.png 1272w, https://substackcdn.com/image/fetch/$s_!mz2q!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd676c533-3d7d-4f19-8f4b-a3e29c484839_1460x1534.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!mz2q!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd676c533-3d7d-4f19-8f4b-a3e29c484839_1460x1534.png" width="1456" height="1530" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d676c533-3d7d-4f19-8f4b-a3e29c484839_1460x1534.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1530,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:139344,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.chrishayduk.com/i/184028048?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd676c533-3d7d-4f19-8f4b-a3e29c484839_1460x1534.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!mz2q!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd676c533-3d7d-4f19-8f4b-a3e29c484839_1460x1534.png 424w, https://substackcdn.com/image/fetch/$s_!mz2q!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd676c533-3d7d-4f19-8f4b-a3e29c484839_1460x1534.png 848w, https://substackcdn.com/image/fetch/$s_!mz2q!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd676c533-3d7d-4f19-8f4b-a3e29c484839_1460x1534.png 1272w, https://substackcdn.com/image/fetch/$s_!mz2q!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd676c533-3d7d-4f19-8f4b-a3e29c484839_1460x1534.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Benchmark results available <a href="https://www.vals.ai/benchmarks/mmlu_pro">here</a></figcaption></figure></div><p>Consider the 83% performance threshold. That is, models that answered at least 83% of questions correctly:</p><ul><li><p><strong>OpenAI o1</strong> was the first model to reach this level and did so upon its release on December 5, 2024. Its API pricing was $15 per million input tokens and $60 per million output tokens. <strong>The total cost to run the benchmark was $75.</strong></p></li><li><p><strong>DeepSeek R1</strong> was the first open source model to reach this level when it launched on January 20, 2025, priced at roughly $1.485 per million input tokens and $5.94 per million output tokens. <strong>The total cost to run the benchmark was $6.75</strong></p></li></ul><p>That&#8217;s an order of magnitude cost reduction in one month for equivalent task performance. If we want to be generous and use the release date of o1-preview, this still results in a time horizon of only 4 months before DeepSeek matched its performance with an open source model costing an order of magnitude less.</p><p>To drive the point home further still, DeepSeek V3.2 came out on December 1, 2025, and again achieved the 83% performance threshold, but this time at a&nbsp;<strong>two orders of magnitude reduction in cost</strong>&nbsp;when compared with OpenAI o1. Specifically, <strong>the total cost to run the benchmark was only $2.24.</strong></p><p>Thus, for a fixed level of performance, we see the price drop from $75 to $6.75 to $2.24 over the course of a single year. <strong>As a result, I argue that&nbsp;</strong><em><strong>any</strong></em><strong>&nbsp;task solved by a closed-source model will see enterprise buyers transition to cheaper open-source models within 6 months to one year.</strong></p><p>And there&#8217;s reason to expect this pace to accelerate. As Huawei and SMIC close the gap with NVIDIA and TSMC, and now that NVIDIA potentially regains the ability to sell H200 chips in China, the Chinese open-source labs will have access to better hardware while maintaining their cost structure advantages. We may be looking at only a couple of months between closed-source frontier releases and open-source replication with substantial cost reduction.</p><h2>V. The AGI-Agnostic Conclusion</h2><p>What I think makes this view most compelling is that it doesn&#8217;t depend on AGI being decades away.</p><p>The conventional case for open source often rests on an assumption that we&#8217;re approaching a capability plateau. That is, that base model improvements will slow down, shifting competition to fine-tuning, cost, and vertical specialization. This assumes that the vision of the future espoused by the US AI labs, predicated on artificial superintelligence (ASI) and runaway intelligence explosions, is wrong, while China&#8217;s view of commoditized intelligence is correct. That may well be true, but it&#8217;s a bet on a particular trajectory of AI progress.</p><p>The task saturation argument is stronger because it&#8217;s agnostic to the AGI timeline. When you&#8217;re building a company, you&#8217;re typically building for a specific use case. That means you&#8217;re operating in the saturation regime, not the model scale-up regime. Even if frontier models continue improving rapidly and the AI-2027 timeline plays out, the task your company is built around has a capability threshold beyond which additional model intelligence doesn&#8217;t matter.</p><p>And once you&#8217;re in the saturation regime, the only dimension of competition that matters is cost. Open source wins on cost, systematically and structurally, because open-source economics allow for lower margins and broader distribution.</p><p>The practical takeaway for company builders is this: bias toward open source, and do so for cost reasons rather than capability bets.</p><p>If you&#8217;re building an AI-native product, ask yourself: what capability threshold does my use case actually require? Chances are, that threshold is either already achieved by current open-source models or will be within 6-12 months of a closed-source model first reaching it. Build your infrastructure and workflows around the assumption that you&#8217;ll be running on open-source models, even if you start with closed-source APIs for speed to market.</p><p>The benchmark that matters isn&#8217;t &#8220;which model is smartest.&#8221; It&#8217;s &#8220;which model solves my task cheaply enough.&#8221; And open source is destined to systematically win that competition through relentless cost deflation.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.chrishayduk.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Musings by Chris Hayduk is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Knowledge Is Power Law Distributed]]></title><description><![CDATA[Being a generalist is easier than you think]]></description><link>https://www.chrishayduk.com/p/knowledge-is-power-law-distributed</link><guid isPermaLink="false">https://www.chrishayduk.com/p/knowledge-is-power-law-distributed</guid><dc:creator><![CDATA[Chris Hayduk]]></dc:creator><pubDate>Tue, 06 Jan 2026 02:40:50 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!yyFn!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffaa7c998-fa2a-4625-9144-3f0d13781e6d_5504x3072.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!yyFn!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffaa7c998-fa2a-4625-9144-3f0d13781e6d_5504x3072.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!yyFn!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffaa7c998-fa2a-4625-9144-3f0d13781e6d_5504x3072.jpeg 424w, https://substackcdn.com/image/fetch/$s_!yyFn!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffaa7c998-fa2a-4625-9144-3f0d13781e6d_5504x3072.jpeg 848w, https://substackcdn.com/image/fetch/$s_!yyFn!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffaa7c998-fa2a-4625-9144-3f0d13781e6d_5504x3072.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!yyFn!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffaa7c998-fa2a-4625-9144-3f0d13781e6d_5504x3072.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!yyFn!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffaa7c998-fa2a-4625-9144-3f0d13781e6d_5504x3072.jpeg" width="1456" height="813" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/faa7c998-fa2a-4625-9144-3f0d13781e6d_5504x3072.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:813,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:8005586,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.chrishayduk.com/i/183600317?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffaa7c998-fa2a-4625-9144-3f0d13781e6d_5504x3072.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!yyFn!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffaa7c998-fa2a-4625-9144-3f0d13781e6d_5504x3072.jpeg 424w, https://substackcdn.com/image/fetch/$s_!yyFn!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffaa7c998-fa2a-4625-9144-3f0d13781e6d_5504x3072.jpeg 848w, https://substackcdn.com/image/fetch/$s_!yyFn!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffaa7c998-fa2a-4625-9144-3f0d13781e6d_5504x3072.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!yyFn!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffaa7c998-fa2a-4625-9144-3f0d13781e6d_5504x3072.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Pictured above: British polymath Thomas Young. &#8220;The man who knew everything&#8221;</figcaption></figure></div><p>There is a common meme called &#8220;the last person who knew all of X&#8221;. Some variants include:</p><ul><li><p>&#8220;The last person who knew everything&#8221; &#8212; often referring to Thomas Young (1773-1829)</p></li><li><p>&#8220;The last person who knew all of math&#8221; &#8212; often referring to David Hilbert (1862-1943)</p></li><li><p>&#8220;The last person to know all of physics&#8221; &#8212; often referring to Enrico Fermi (1901-1954)</p></li></ul><p>The implication of these statements is that it is now impossible (or at least unheard of) to know everything in these domains. And the dates above show that this phenomenon of the generalist ended somewhere between 100 and 200 years ago, depending on how broadly we want to define the field of study. Thus, we have had around a century of specialization, with the idea that &#8220;becoming a true generalist is impossible&#8221; largely accepted as fact, whether explicitly or implicitly, by academics everywhere. </p><p>The march from high school diploma to undergraduate degree to PhD results in ever-greater degrees of specialization and an ever-narrowing aperture through which to view knowledge. The benefit commonly attributed to the narrowing of our education is that it helps us to reach the frontier of knowledge and thus make novel contributions more quickly. Its proponents say that no one can know everything anymore, and thus it is a waste of time to branch out to fields other than your chosen field of study. </p><p>However, my view is that we are artificially constraining our best and brightest by continuing to espouse this narrative that it&#8217;s impossible to be an effective generalist. Useful knowledge is <em>not </em>broadly distributed &#8212; it is concentrated in a handful of topics contained in only a handful of subjects. You can sense this intuitively if you walk through the aisles of your nearby Barnes &amp; Noble: there are tens of thousands of books spread across the shelves, and the books on the shelves change every few weeks or months. But how many of those titles will be remembered 1 year from now? 10 years from now? What about 100 years from now? It would be a safe bet that no knowledge or insight contained in any new release will be foundational moving forward. By contrast, the knowledge contained in the core of useful books can be reapplied across different sets of facts in disparate fields to solve problems more effectively and creatively than many specialists.</p><p><strong>Specifically, I claim that this useful core consists of roughly 300 books of the 170 million unique titles in existence (roughly 0.00017% of all books), and that this core will afford you ~99% of the value of all knowledge ever produced</strong>. That is, if you are ~98-99th percentile 18-year-old high school graduate (roughly corresponding to ~1500+ on the SAT), then you should be able to get to the forefront of knowledge in the core academic subjects (or have the tools necessary to rapidly reach the forefront of non-core subjects) in about 300 books.</p><p>In other words, knowledge is highly power law distributed.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!L5mm!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F941e5611-6565-4ce2-bdd3-b9c9b7b48b0c_3178x1690.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!L5mm!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F941e5611-6565-4ce2-bdd3-b9c9b7b48b0c_3178x1690.png 424w, https://substackcdn.com/image/fetch/$s_!L5mm!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F941e5611-6565-4ce2-bdd3-b9c9b7b48b0c_3178x1690.png 848w, https://substackcdn.com/image/fetch/$s_!L5mm!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F941e5611-6565-4ce2-bdd3-b9c9b7b48b0c_3178x1690.png 1272w, https://substackcdn.com/image/fetch/$s_!L5mm!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F941e5611-6565-4ce2-bdd3-b9c9b7b48b0c_3178x1690.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!L5mm!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F941e5611-6565-4ce2-bdd3-b9c9b7b48b0c_3178x1690.png" width="1456" height="774" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/941e5611-6565-4ce2-bdd3-b9c9b7b48b0c_3178x1690.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:774,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:340781,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.chrishayduk.com/i/183600317?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F941e5611-6565-4ce2-bdd3-b9c9b7b48b0c_3178x1690.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!L5mm!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F941e5611-6565-4ce2-bdd3-b9c9b7b48b0c_3178x1690.png 424w, https://substackcdn.com/image/fetch/$s_!L5mm!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F941e5611-6565-4ce2-bdd3-b9c9b7b48b0c_3178x1690.png 848w, https://substackcdn.com/image/fetch/$s_!L5mm!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F941e5611-6565-4ce2-bdd3-b9c9b7b48b0c_3178x1690.png 1272w, https://substackcdn.com/image/fetch/$s_!L5mm!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F941e5611-6565-4ce2-bdd3-b9c9b7b48b0c_3178x1690.png 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">The value of books represented as a power law distribution.</figcaption></figure></div><p>The core subjects that need to be learned to realize this outsized payoff consist of:</p><ul><li><p>Physics</p></li><li><p>Mathematics</p></li><li><p>Computer Science</p></li><li><p>History</p></li><li><p>Philosophy</p></li><li><p>Literature (with the canon changing depending on your cultural context)</p></li></ul><p>It is quite possible, within the constraints of our 300-book total, to reach the level of knowledge of a 2nd or 3rd year PhD student at an elite institution in each of these subjects. This is possible for two reasons: </p><ol><li><p>Each of these fields has a small core of topics that are explored from many different angles to form the many subtopics of that field. For example, symmetries &amp; action principles in physics, linearity &amp; topology in mathematics, or computability &amp; complexity in computer science.</p></li><li><p>Each of the core topics in these fields provides insight into multiple other fields in the list. Learning mathematics aids both physics and computer science (and, to an extent, philosophy). Learning philosophy aids both history and literature. And so on. </p></li></ol><p>Moreover, once you have reached this level of knowledge across each of these core subjects, you will be able to dive into any subfield of these fields or any field not listed in this core and rapidly reach the frontier of research. This is only possible because other fields use the core insights of these fields as their fundamental scaffolding. Chemistry, for instance, is ultimately the study of valence electrons governed by the laws of quantum mechanics; biology is the complex chemical machinery of life; economics is the application of multivariable calculus, game theory, and psychology; and political science is the ongoing, real-world stress test of history and moral philosophy. When you own the "root nodes" of the knowledge graph, the specialized nodes are often just specific parameters applied to general frameworks you have already mastered. Consequently, the transition from this generalist foundation to a specialist frontier is not a climb up a new mountain, but a lateral step onto a bridge you have already built, allowing you to perceive structural similarities between disciplines that the siloed expert remains blind to.</p><p>Once you accept this framework, you see that learning becomes much more about <em>curation</em> rather than exhaustively studying all that has been published. It is the <em>via negativa</em> &#8212; choosing not to read a vast majority of writing so that you can instead learn a vast majority of knowledge. Thus, the barrier to becoming a modern Thomas Young is no longer cognitive capacity, nor is it the "impossibility" of the expanding universe of knowledge. The barrier is the discipline to ignore the noise of the 99.9999% in favor of the signal of the 0.0001%, and the patience to master those root nodes before attempting to climb the branches. The library of the generalist is small, but it is heavy.</p><p>End note: As an example of what this curation looks like, here is a condensed roadmap I am using to go from no knowledge of physics to a graduate school level <a href="https://docs.google.com/spreadsheets/d/1vFhJny3V2W8snY5Sex2emuQ1cRPeTiYmqK6pTwD2GMk/edit?usp=sharing">[link]</a></p>]]></content:encoded></item><item><title><![CDATA[It's Time For Google to Acquire Intel]]></title><description><![CDATA[Breaking the Nvidia-TSMC monopoly]]></description><link>https://www.chrishayduk.com/p/its-time-for-google-to-acquire-intel</link><guid isPermaLink="false">https://www.chrishayduk.com/p/its-time-for-google-to-acquire-intel</guid><dc:creator><![CDATA[Chris Hayduk]]></dc:creator><pubDate>Thu, 25 Sep 2025 13:02:32 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!POjm!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6f0c6e63-7ae3-404b-9c1a-fbbf8778b9d7_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!POjm!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6f0c6e63-7ae3-404b-9c1a-fbbf8778b9d7_1024x1024.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!POjm!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6f0c6e63-7ae3-404b-9c1a-fbbf8778b9d7_1024x1024.png 424w, https://substackcdn.com/image/fetch/$s_!POjm!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6f0c6e63-7ae3-404b-9c1a-fbbf8778b9d7_1024x1024.png 848w, https://substackcdn.com/image/fetch/$s_!POjm!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6f0c6e63-7ae3-404b-9c1a-fbbf8778b9d7_1024x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!POjm!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6f0c6e63-7ae3-404b-9c1a-fbbf8778b9d7_1024x1024.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!POjm!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6f0c6e63-7ae3-404b-9c1a-fbbf8778b9d7_1024x1024.png" width="1024" height="1024" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/6f0c6e63-7ae3-404b-9c1a-fbbf8778b9d7_1024x1024.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1024,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;Generated Image September 24, 2025 - 11:01PM.png&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Generated Image September 24, 2025 - 11:01PM.png" title="Generated Image September 24, 2025 - 11:01PM.png" srcset="https://substackcdn.com/image/fetch/$s_!POjm!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6f0c6e63-7ae3-404b-9c1a-fbbf8778b9d7_1024x1024.png 424w, https://substackcdn.com/image/fetch/$s_!POjm!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6f0c6e63-7ae3-404b-9c1a-fbbf8778b9d7_1024x1024.png 848w, https://substackcdn.com/image/fetch/$s_!POjm!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6f0c6e63-7ae3-404b-9c1a-fbbf8778b9d7_1024x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!POjm!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6f0c6e63-7ae3-404b-9c1a-fbbf8778b9d7_1024x1024.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Nvidia made headlines this week when it announced it would invest up to $100&#8239;billion into OpenAI and help deploy at least 10&#8239;GW of AI infrastructure. The move, frequently memed as an &#8220;infinite money glitch,&#8221; with capital and revenue cycling between Nvidia and OpenAI (see the image below), effectively ensures a substantial fraction of Nvidia&#8217;s GPUs will land in OpenAI&#8209;aligned datacenters (via leasing or outright purchases). </p><p>This comes on the heels of OpenAI&#8217;s <strong>&gt;</strong>$300&#8239;billion &#8220;Stargate&#8221; build&#8209;out with Oracle, which targets <strong>~</strong>4.5&#8239;GW of capacity, further tightening the market for top&#8209;end accelerators.</p><p>And that&#8217;s before accounting for OpenAI&#8217;s ongoing expansion on Microsoft Azure, where the relationship now runs under a right&#8209;of&#8209;first&#8209;refusal model for new capacity rather than blanket exclusivity, still conferring practical priority on Azure deployments while allowing OpenAI to add capacity with other partners. </p><p><strong>Netting this out:</strong> through the end of the decade, OpenAI has assembled an envelope of roughly 10&#8211;15&#8239;GW of Nvidia&#8209;powered capacity across Oracle, Microsoft, and other partners with overlap between these footprints, so think of this as a shared umbrella rather than purely additive numbers. For context, independent analyses estimate <strong>~</strong>10&#8239;GW of additional AI data&#8209;center power could be needed globally in 2025 alone; in other words, OpenAI&#8217;s program is on the scale of a full year of incremental world AI build&#8209;out.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!kCHD!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe695baec-b55a-495b-a17a-a1a31bc5ec51_1004x632.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!kCHD!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe695baec-b55a-495b-a17a-a1a31bc5ec51_1004x632.jpeg 424w, https://substackcdn.com/image/fetch/$s_!kCHD!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe695baec-b55a-495b-a17a-a1a31bc5ec51_1004x632.jpeg 848w, https://substackcdn.com/image/fetch/$s_!kCHD!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe695baec-b55a-495b-a17a-a1a31bc5ec51_1004x632.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!kCHD!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe695baec-b55a-495b-a17a-a1a31bc5ec51_1004x632.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!kCHD!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe695baec-b55a-495b-a17a-a1a31bc5ec51_1004x632.jpeg" width="1004" height="632" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e695baec-b55a-495b-a17a-a1a31bc5ec51_1004x632.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:632,&quot;width&quot;:1004,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;Logos of OpenAI, NVIDIA, and Oracle arranged in a triangular pattern with arrows labeled \&quot;$100 billion\&quot; connecting them. Text at the top reads \&quot;THE INFINITE MONEY GLITCH\&quot; in red.&quot;,&quot;title&quot;:&quot;Logos of OpenAI, NVIDIA, and Oracle arranged in a triangular pattern with arrows labeled \&quot;$100 billion\&quot; connecting them. Text at the top reads \&quot;THE INFINITE MONEY GLITCH\&quot; in red.&quot;,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Logos of OpenAI, NVIDIA, and Oracle arranged in a triangular pattern with arrows labeled &quot;$100 billion&quot; connecting them. Text at the top reads &quot;THE INFINITE MONEY GLITCH&quot; in red." title="Logos of OpenAI, NVIDIA, and Oracle arranged in a triangular pattern with arrows labeled &quot;$100 billion&quot; connecting them. Text at the top reads &quot;THE INFINITE MONEY GLITCH&quot; in red." srcset="https://substackcdn.com/image/fetch/$s_!kCHD!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe695baec-b55a-495b-a17a-a1a31bc5ec51_1004x632.jpeg 424w, https://substackcdn.com/image/fetch/$s_!kCHD!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe695baec-b55a-495b-a17a-a1a31bc5ec51_1004x632.jpeg 848w, https://substackcdn.com/image/fetch/$s_!kCHD!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe695baec-b55a-495b-a17a-a1a31bc5ec51_1004x632.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!kCHD!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe695baec-b55a-495b-a17a-a1a31bc5ec51_1004x632.jpeg 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Courtesy of SemiAnalysis&#8217;s <a href="https://x.com/dylan522p/status/1970346183827783756">Dylan Patel</a></figcaption></figure></div><p>The above data implies Nvidia GPU availability will tighten <em>substantially</em> for other frontier&#8209;model players&#8212;Anthropic (primarily on AWS), xAI, Meta, and Google DeepMind&#8212;raising effective prices and lead times and forcing harder choices about model cadence, context windows, and training tokens.</p><p>Google has been trying to break out of this Nvidia-dominated mold for years through the development of its own AI&#8209;specialized TPUs for training and inference. But these in-house designed chips still pass through chokepoints that Nvidia heavily influences, especially TSMC wafers and advanced packaging. By the end of 2025, analysts expect Nvidia to be <strong>~</strong>20%+ of TSMC revenue (second only to Apple), and the CoWoS&#8209;class packaging and HBM ecosystems remain binding constraints even as capacity expands. TSMC&#8217;s allocation is fundamentally contractual, driven by prepays and take&#8209;or&#8209;pay deals, and it will be reluctant to shift meaningful share away from Nvidia while demand remains red&#8209;hot.</p><p>To escape the straitjacket created by the Nvidia&#8209;OpenAI alignment, Google should buy Intel (or a substantial portion of it), fund the High&#8209;NA EUV ramp, and prepare to manufacture TPUs on Intel fabs as that capacity comes online. That gives Google end&#8209;to&#8209;end control of its AI training infrastructure&#8212;chip architectures, training software, chip manufacturing, and data center buildout&#8212;and a guaranteed runway independent of Nvidia&#8217;s queue. </p><p>Recent events make this even more urgent. Nvidia just disclosed a $5&#8239;billion Intel investment at $23.28/share (roughly 5% of Intel&#8217;s outstanding shares), alongside a product pact in which Intel will build x86 SoCs integrating Nvidia RTX GPU chiplets for PCs and collaborate on custom data&#8209;center CPUs&#8212;clear evidence that Intel&#8217;s roadmap can be steered by anchor customers. Intel is also now soliciting an Apple investment, according to Bloomberg/Reuters reporting.</p><p>Given the quickly changing dynamics around Intel, Google must act quickly and decisively. For example, a $25&#8239;billion purchase at $35/share would buy on the order of ~714&#8239;million shares, implying ~16%&#8211;17% of Intel based on ~4.37&#8239;billion shares outstanding&#8212;placing Google ahead of both the U.S. government (~10%) and Nvidia (~4&#8211;5%) as the largest shareholder. That level of ownership could anchor governance and direct capex toward TPU&#8209;critical fabs and packaging lines.</p><p>In practice, this looks like the following:</p><ol><li><p><strong>A minority stake + board influence</strong> sufficient to align Intel Foundry&#8217;s roadmap to TPU requirements</p></li><li><p><strong>A TPU-only supply compact</strong>: multi-year, take-or-pay wafer and advanced packaging commitments, with right-of-first-allocation during shortages and pricing bands tied to verifiable tool/packaging milestones. </p></li><li><p><strong>Parallel open&#8209;market TPU SKUs</strong> to keep utilization high and de&#8209;risk capex&#8212;turning Google&#8217;s silicon into a software&#8209;first, capacity&#8209;priced product.</p></li></ol><p>#3 is the longest shot, but perhaps the most enticing benefit of the investment. This would open up a second profit engine to fuel Google&#8217;s growth over the next decade, especially as its Search business comes under threat from AI-search competitors (such as OpenAI&#8217;s search-enabled offerings). In fact, Nvidia&#8217;s data&#8209;center business is now running at an annualized ~$160&#8239;billion revenue pace, which is comparable to Google&#8217;s Search cash cow. Thus, the addition of the TPU revenue line provides substantial growth opportunities and a potential hedge against Google&#8217;s eroding search moat.</p><p>If this plan works, Google gets scheduling certainty, lower $/token, a faster model cadence independent of Nvidia&#8217;s allocation calendar, and another revenue stream that could potentially reach the level of Google Search. If it stumbles, the downside is capped at a financial position that should still appreciate if Intel&#8217;s foundry inflects. Either way, for $25 billion, Google can buy its way out of the Nvidia-TSMC duopoly and into the driver&#8217;s seat of AI compute.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.chrishayduk.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Musings by Chris Hayduk is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[The Strategic Implications of GPT-5 for OpenAI]]></title><description><![CDATA[OpenAI shifts away from the enterprise and toward the consumer]]></description><link>https://www.chrishayduk.com/p/the-strategic-implications-of-gpt</link><guid isPermaLink="false">https://www.chrishayduk.com/p/the-strategic-implications-of-gpt</guid><dc:creator><![CDATA[Chris Hayduk]]></dc:creator><pubDate>Fri, 08 Aug 2025 15:39:23 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!_Syy!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e6943e8-d8d3-4e51-b3f2-e941d5d890db_1536x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!_Syy!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e6943e8-d8d3-4e51-b3f2-e941d5d890db_1536x1024.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!_Syy!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e6943e8-d8d3-4e51-b3f2-e941d5d890db_1536x1024.png 424w, https://substackcdn.com/image/fetch/$s_!_Syy!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e6943e8-d8d3-4e51-b3f2-e941d5d890db_1536x1024.png 848w, https://substackcdn.com/image/fetch/$s_!_Syy!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e6943e8-d8d3-4e51-b3f2-e941d5d890db_1536x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!_Syy!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e6943e8-d8d3-4e51-b3f2-e941d5d890db_1536x1024.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!_Syy!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e6943e8-d8d3-4e51-b3f2-e941d5d890db_1536x1024.png" width="1456" height="971" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/5e6943e8-d8d3-4e51-b3f2-e941d5d890db_1536x1024.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:971,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2364241,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.chrishayduk.com/i/170448919?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e6943e8-d8d3-4e51-b3f2-e941d5d890db_1536x1024.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!_Syy!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e6943e8-d8d3-4e51-b3f2-e941d5d890db_1536x1024.png 424w, https://substackcdn.com/image/fetch/$s_!_Syy!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e6943e8-d8d3-4e51-b3f2-e941d5d890db_1536x1024.png 848w, https://substackcdn.com/image/fetch/$s_!_Syy!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e6943e8-d8d3-4e51-b3f2-e941d5d890db_1536x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!_Syy!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e6943e8-d8d3-4e51-b3f2-e941d5d890db_1536x1024.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Image courtesy of GPT-5</figcaption></figure></div><p>After years of anticipation and hype, GPT-5 is finally out. And the results are decidedly mixed. GPT-5 is undoubtedly a great model &#8212; it is #1 across the board on LMArena, sets new highs in SWE-Bench and a host of other coding tasks, and performs great across a range of math benchmarks. However, the expectations for GPT-5 were that it would blow the competition out of the water. Instead, it has made incremental improvements across all of these benchmarks, and is highly likely to be passed over whenever Google releases its next Gemini model in short order (or just when Gemini 2.5 Deep Think gets benchmarked!) </p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!FLx3!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4abf679a-4d93-45c9-85a5-6435c84b0f6f_750x469.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!FLx3!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4abf679a-4d93-45c9-85a5-6435c84b0f6f_750x469.png 424w, https://substackcdn.com/image/fetch/$s_!FLx3!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4abf679a-4d93-45c9-85a5-6435c84b0f6f_750x469.png 848w, https://substackcdn.com/image/fetch/$s_!FLx3!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4abf679a-4d93-45c9-85a5-6435c84b0f6f_750x469.png 1272w, https://substackcdn.com/image/fetch/$s_!FLx3!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4abf679a-4d93-45c9-85a5-6435c84b0f6f_750x469.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!FLx3!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4abf679a-4d93-45c9-85a5-6435c84b0f6f_750x469.png" width="750" height="469" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/4abf679a-4d93-45c9-85a5-6435c84b0f6f_750x469.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:469,&quot;width&quot;:750,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;OpenAI botches the charts in GPT-5 introduction &#8211; FlowingData&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="OpenAI botches the charts in GPT-5 introduction &#8211; FlowingData" title="OpenAI botches the charts in GPT-5 introduction &#8211; FlowingData" srcset="https://substackcdn.com/image/fetch/$s_!FLx3!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4abf679a-4d93-45c9-85a5-6435c84b0f6f_750x469.png 424w, https://substackcdn.com/image/fetch/$s_!FLx3!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4abf679a-4d93-45c9-85a5-6435c84b0f6f_750x469.png 848w, https://substackcdn.com/image/fetch/$s_!FLx3!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4abf679a-4d93-45c9-85a5-6435c84b0f6f_750x469.png 1272w, https://substackcdn.com/image/fetch/$s_!FLx3!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4abf679a-4d93-45c9-85a5-6435c84b0f6f_750x469.png 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Note that Claude Opus 4.1 scored a 74.5% on SWE-bench just days earlier, so GPT-5 performance is virtually the same (also what in the world is going on with charts at OpenAI???)</figcaption></figure></div><p>The largest gains to GPT-5 came in a less performance-based metric: it seems that, for this release, OpenAI highly prioritized reducing hallucinations and sycophancy in model output.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!dBF7!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe84b3dbd-444c-4eba-9bf5-0b69ffeb50e2_2048x1089.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!dBF7!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe84b3dbd-444c-4eba-9bf5-0b69ffeb50e2_2048x1089.png 424w, https://substackcdn.com/image/fetch/$s_!dBF7!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe84b3dbd-444c-4eba-9bf5-0b69ffeb50e2_2048x1089.png 848w, https://substackcdn.com/image/fetch/$s_!dBF7!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe84b3dbd-444c-4eba-9bf5-0b69ffeb50e2_2048x1089.png 1272w, https://substackcdn.com/image/fetch/$s_!dBF7!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe84b3dbd-444c-4eba-9bf5-0b69ffeb50e2_2048x1089.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!dBF7!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe84b3dbd-444c-4eba-9bf5-0b69ffeb50e2_2048x1089.png" width="1456" height="774" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e84b3dbd-444c-4eba-9bf5-0b69ffeb50e2_2048x1089.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:774,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;GPT-5 Benchmark Scores | ml-news &#8211; Weights &amp; Biases&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="GPT-5 Benchmark Scores | ml-news &#8211; Weights &amp; Biases" title="GPT-5 Benchmark Scores | ml-news &#8211; Weights &amp; Biases" srcset="https://substackcdn.com/image/fetch/$s_!dBF7!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe84b3dbd-444c-4eba-9bf5-0b69ffeb50e2_2048x1089.png 424w, https://substackcdn.com/image/fetch/$s_!dBF7!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe84b3dbd-444c-4eba-9bf5-0b69ffeb50e2_2048x1089.png 848w, https://substackcdn.com/image/fetch/$s_!dBF7!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe84b3dbd-444c-4eba-9bf5-0b69ffeb50e2_2048x1089.png 1272w, https://substackcdn.com/image/fetch/$s_!dBF7!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe84b3dbd-444c-4eba-9bf5-0b69ffeb50e2_2048x1089.png 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>So you may not notice large performance differences between GPT-5 and the leading models from other labs (or even when compared to OpenAI&#8217;s o3 model), but you likely will notice that the model is much less likely to make things up and say things that are flat out wrong just to produce an answer.</p><p>In addition, you will likely notice in the ChatGPT model picker that all of the previous models are gone: now there&#8217;s only GPT-5. This is another one of GPT-5&#8217;s main contributions &#8212; it greatly simplifies the model selection process. GPT-5 is more of a system than a model, dynamically routing requests to faster LLMs (analogous to GPT 4o) or slower, thinking LLMs (analogous to o3) depending on the complexity of the request.</p><p>(The two above points are important; we&#8217;ll come back to those later).</p><p>Likely in response to some widespread dismay at the performance benchmarks, Sam Altman tweeted the following after the GPT-5 announcement:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!b1QM!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F05ed997d-58c9-4adb-bb36-36bdc952c7a2_2160x2160.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!b1QM!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F05ed997d-58c9-4adb-bb36-36bdc952c7a2_2160x2160.png 424w, https://substackcdn.com/image/fetch/$s_!b1QM!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F05ed997d-58c9-4adb-bb36-36bdc952c7a2_2160x2160.png 848w, https://substackcdn.com/image/fetch/$s_!b1QM!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F05ed997d-58c9-4adb-bb36-36bdc952c7a2_2160x2160.png 1272w, https://substackcdn.com/image/fetch/$s_!b1QM!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F05ed997d-58c9-4adb-bb36-36bdc952c7a2_2160x2160.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!b1QM!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F05ed997d-58c9-4adb-bb36-36bdc952c7a2_2160x2160.png" width="1456" height="1456" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/05ed997d-58c9-4adb-bb36-36bdc952c7a2_2160x2160.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1456,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:464257,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.chrishayduk.com/i/170448919?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F05ed997d-58c9-4adb-bb36-36bdc952c7a2_2160x2160.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!b1QM!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F05ed997d-58c9-4adb-bb36-36bdc952c7a2_2160x2160.png 424w, https://substackcdn.com/image/fetch/$s_!b1QM!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F05ed997d-58c9-4adb-bb36-36bdc952c7a2_2160x2160.png 848w, https://substackcdn.com/image/fetch/$s_!b1QM!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F05ed997d-58c9-4adb-bb36-36bdc952c7a2_2160x2160.png 1272w, https://substackcdn.com/image/fetch/$s_!b1QM!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F05ed997d-58c9-4adb-bb36-36bdc952c7a2_2160x2160.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><strong>&#8220;we can release much, much smarter models&#8221;</strong></p><p>It seems Altman is asserting that OpenAI deliberately <em>chose</em> to release a model below the company&#8217;s capabilities, barely edging out its competitors (and likely not even edging out Google&#8217;s leading model) on most performance metrics. Instead, they deliberately <em>chose </em>to focus on reducing hallucinations and streamlining model selection as the main contributions of GPT-5. Why would OpenAI do this? Why not continue setting the benchmark for LLM model performance, as they&#8217;ve done since the ye olde days of GPT-2?</p><p>Because the strategic focus of the company has clearly shifted.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.chrishayduk.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.chrishayduk.com/subscribe?"><span>Subscribe now</span></a></p><h1>Market Dynamics Affecting OpenAI</h1><p>ChatGPT is a consumer application with <a href="https://www.techloy.com/chatgpt-is-on-track-to-reach-700-million-weekly-active-users/">700 million weekly active users</a>. And it is absolutely trouncing the competition in consumer adoption. The ChatGPT app in the Apple App Store has 3.3 million reviews, compared to just 377,000 for the Gemini app and 23,000 for the Claude app. This suggests that ChatGPT has a mobile install base that is 10x the size of Google Gemini and 100x the size of Anthropic&#8217;s Claude. Moreover, in June 2025, openai.com had 1.12 billion visits, while gemini.google.com had 265 million and Claude had 113 million &#8212; again suggesting a lead of at least an order of magnitude for OpenAI over its competitors in the consumer chat space (source: <a href="https://www.semrush.com/">https://www.semrush.com/</a>).</p><p>By contrast, according to <a href="https://menlovc.com/perspective/2025-mid-year-llm-market-update/">a Menlo Ventures report</a> from July 2025, Anthropic is actually the leading market provider for enterprise LLM API usage, with 32% market share vs. OpenAI&#8217;s 25% market share in mid-2025. Google is also growing and not far behind OpenAI, at 20% market share, up from 12% in 2024. OpenAI&#8217;s enterprise position has also been trending strongly negative, cratering from 50% market share in 2023 down to its current position of 25% market share in mid-2025.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ruow!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F39abed8f-1c2d-409f-ba6e-d5933542aea7_2560x1320.webp" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ruow!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F39abed8f-1c2d-409f-ba6e-d5933542aea7_2560x1320.webp 424w, https://substackcdn.com/image/fetch/$s_!ruow!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F39abed8f-1c2d-409f-ba6e-d5933542aea7_2560x1320.webp 848w, https://substackcdn.com/image/fetch/$s_!ruow!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F39abed8f-1c2d-409f-ba6e-d5933542aea7_2560x1320.webp 1272w, https://substackcdn.com/image/fetch/$s_!ruow!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F39abed8f-1c2d-409f-ba6e-d5933542aea7_2560x1320.webp 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ruow!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F39abed8f-1c2d-409f-ba6e-d5933542aea7_2560x1320.webp" width="1456" height="751" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/39abed8f-1c2d-409f-ba6e-d5933542aea7_2560x1320.webp&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:751,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;Enterprise LLM API market share by usage&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Enterprise LLM API market share by usage" title="Enterprise LLM API market share by usage" srcset="https://substackcdn.com/image/fetch/$s_!ruow!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F39abed8f-1c2d-409f-ba6e-d5933542aea7_2560x1320.webp 424w, https://substackcdn.com/image/fetch/$s_!ruow!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F39abed8f-1c2d-409f-ba6e-d5933542aea7_2560x1320.webp 848w, https://substackcdn.com/image/fetch/$s_!ruow!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F39abed8f-1c2d-409f-ba6e-d5933542aea7_2560x1320.webp 1272w, https://substackcdn.com/image/fetch/$s_!ruow!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F39abed8f-1c2d-409f-ba6e-d5933542aea7_2560x1320.webp 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>So we see the market dynamics pressing on OpenAI as a company &#8212; absolutely dominant positioning on the consumer side of the market, with a weak (and steadily weakening) position on the enterprise side of the market. This leaves OpenAI with a choice &#8212; double down on its success on the consumer side of the market, or attempt to win in the highly competitive enterprise space. This choice mainly comes down to where the company has a moat that can result in durable profit margins.</p><h1>How Market Dynamics Affect the Models</h1><p>First, let&#8217;s explore the dynamics of the consumer market. Consumers, by and large, make buying decisions not based on performance or objective metrics, but instead based on &#8220;vibes&#8221;. You can improve the vibes for a consumer by improving brand positioning (i.e., make the consumer feel a certain emotion from using your product) or by improving the user experience (UX) and user interface (UI) of your product (i.e., make the product more enjoyable for the user).</p><p>OpenAI&#8217;s lead in consumer usage stems primarily from precisely these areas, with its extremely strong branding and UI/UX improvements in its chat interface versus the competition. OpenAI had a strong, multi-year lead due to its first-mover advantage, providing its &#8220;ChatGPT&#8221; brand with significant mindshare in the consumer base. In addition, since the release of the original ChatGPT, OpenAI has focused strongly on the web and mobile chat experience. With features like memory and ChatGPT Projects, OpenAI has introduced a high level of personalization for users of the app, thereby creating a high switching cost moat &#8212; if you switch to Claude or Gemini, you can&#8217;t take ChatGPT&#8217;s memories or projects with you. This instantly makes the competing consumer apps less appealing to users in the same way that users of Spotify are reluctant to shift over to Apple Music once they have built up a library of playlists that they enjoy.</p><p>Hence, to improve consumer market share, OpenAI will need to continually pull the two levers of brand positioning and UI/UX improvements. Models can&#8217;t really improve brand positioning much, as that is more a function of marketing, so the main pressure on the model side of the equation will come from the UI/UX push. This pressure results in making models that are <strong>simpler and more enjoyable to use.</strong></p><p>Now we&#8217;ll shift our view to the enterprise market. Businesses, unlike consumers, strictly focus on return on investment when allocating capital expenditures. These ROI calculations will essentially have four inputs when it comes to LLMs: </p><ol><li><p>The API cost per million tokens</p></li><li><p>The number of tokens needed to solve a task</p></li><li><p>The value of that task</p></li><li><p>The performance of the LLM on that task</p></li></ol><p>We can then model the ROI of using an LLM as follows:</p><div class="latex-rendered" data-attrs="{&quot;persistentExpression&quot;:&quot;\\begin{aligned}\n\\text{value generated} &amp;= (\\text{value of task}) \\times (\\text{performance of LLM on that task}) \\\\\n\\text{cost} &amp;= (\\text{API cost per million tokens}) \\times (\\text{number of tokens needed to solve the task}) \\\\\n\\text{net profit} &amp;= \\text{value generated} - \\text{cost} \\\\\n\\text{ROI} &amp;= \\frac{\\text{net profit}}{\\text{cost}}\n\\end{aligned}&quot;,&quot;id&quot;:&quot;AVQSRYKDYM&quot;}" data-component-name="LatexBlockToDOM"></div><p>So, from the above, we can see that the only levers that LLM providers can pull to improve the ROI calculations for a company are:</p><ol><li><p>Decrease the cost per million tokens for the API</p></li><li><p>Decrease the number of tokens needed to solve the task</p></li><li><p>Increase LLM performance on the task</p></li></ol><p>Given the pressure from open source contributions (e.g., DeepSeek, Kimi, and Qwen), closed source model providers will never be able to compete on #1. #2 also runs counter to the current scaling of AI models &#8212; to increase test-time compute (and thus make the LLM useful for more difficult tasks), we by definition have to increase the number of tokens used. Hence, LLM providers competing in the enterprise have started to converge on #3 &#8212; improving LLM performance on the given task.</p><p>Now, <strong>there are two additional levers that an LLM provider can pull to improve the performance of the model on a specific task:</strong></p><ol><li><p><strong>Make the model smarter overall</strong></p></li><li><p><strong>Customize the model for that task</strong></p></li></ol><p>Broadly, Google has taken the first approach, with Gemini models consistently leading the pack in intelligence (particularly the new Gemini 2.5 Pro Deep Think model). OpenAI would struggle mightily to compete along this dimension because Google has such massive advantages in terms of scale &#8212; it has access to ridiculous amounts of compute and has indexed virtually all of the world&#8217;s data. Having a lead in algorithms is not a durable moat due to the speed of diffusion of inventions in Silicon Valley, and since model performance is a function of algorithms, data, and compute, Google will maintain a decisive lead here.</p><p>Meanwhile, Anthropic has taken the second approach, specializing its models for code using targeted reinforcement learning and building the Claude Code agentic harness. This is the lowest-hanging fruit for specialized models, given that this is the domain in which today&#8217;s LLMs perform best. Since Anthropic already has a large lead here, this then leaves OpenAI with two choices: find a less obvious niche for which it can start customizing its models, or compete directly with Anthropic in the coding space, where its competitor already has a large advantage.</p><p>From the above analysis, we can see that OpenAI has a large lead in the consumer market with durable moats, and that in order to improve those moats, OpenAI would need to improve the UI/UX of its models by making them simpler and more enjoyable to use. By contrast, to compete in the enterprise market, OpenAI would need to either produce the smartest model (where it is at a disadvantage compared to Google) or start customizing its models for targeted use cases (where it is at a disadvantage compared to Anthropic in the most obvious market of coding agents). </p><h1>Conclusion - GPT-5 as AI for the Common Man</h1><p>Now, let&#8217;s wrap up this argument.</p><p>We have already seen that OpenAI has a large and commanding lead in the consumer market, with a low and shrinking market share in the enterprise market. Now we have also shown that it has solid, defensible moats in the consumer market, and it is at a strong technical disadvantage in the enterprise market. We have also established that prioritizing consumers means improving model UI/UX, while prioritizing enterprise means improving model performance and specialization. Lastly, from the opening paragraphs, we have established that OpenAI deliberately did not make the highest-performing model possible. </p><p>Instead, they prioritized reducing hallucinations and streamlining the model selection process in ChatGPT. Both of these changes significantly improve the consumer experience, as confabulations can destroy consumer trust and erode brand advantages, while the old model picker with nearly 10 different models intimidated new users and caused high cognitive load when using the app.</p><p>As such, the logical conclusion is that <strong>OpenAI has chosen to prioritize consumers</strong> <strong>over enterprise, and GPT-5 is the result of this</strong>.</p><p>Hence, over the coming years, don&#8217;t expect OpenAI to consistently lead in model performance as they have over the past 3 years. Instead, look for continuing improvements in the usage experience of ChatGPT. If you want to find the best models overall or the best coding models, you&#8217;ll probably need to look to Google and Anthropic, respectively.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.chrishayduk.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Musings by Chris Hayduk is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Managing Civilizational Tail Risks]]></title><description><![CDATA[401k for Kids and Communist Revolutions]]></description><link>https://www.chrishayduk.com/p/managing-civilizational-tail-risks</link><guid isPermaLink="false">https://www.chrishayduk.com/p/managing-civilizational-tail-risks</guid><dc:creator><![CDATA[Chris Hayduk]]></dc:creator><pubDate>Wed, 16 Jul 2025 14:37:43 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!wBXO!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F23cb6006-b47c-4b44-822f-53dc65eb383a_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!wBXO!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F23cb6006-b47c-4b44-822f-53dc65eb383a_1024x1024.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!wBXO!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F23cb6006-b47c-4b44-822f-53dc65eb383a_1024x1024.png 424w, https://substackcdn.com/image/fetch/$s_!wBXO!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F23cb6006-b47c-4b44-822f-53dc65eb383a_1024x1024.png 848w, https://substackcdn.com/image/fetch/$s_!wBXO!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F23cb6006-b47c-4b44-822f-53dc65eb383a_1024x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!wBXO!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F23cb6006-b47c-4b44-822f-53dc65eb383a_1024x1024.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!wBXO!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F23cb6006-b47c-4b44-822f-53dc65eb383a_1024x1024.png" width="1024" height="1024" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/23cb6006-b47c-4b44-822f-53dc65eb383a_1024x1024.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1024,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1918864,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.chrishayduk.com/i/168124612?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F23cb6006-b47c-4b44-822f-53dc65eb383a_1024x1024.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!wBXO!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F23cb6006-b47c-4b44-822f-53dc65eb383a_1024x1024.png 424w, https://substackcdn.com/image/fetch/$s_!wBXO!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F23cb6006-b47c-4b44-822f-53dc65eb383a_1024x1024.png 848w, https://substackcdn.com/image/fetch/$s_!wBXO!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F23cb6006-b47c-4b44-822f-53dc65eb383a_1024x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!wBXO!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F23cb6006-b47c-4b44-822f-53dc65eb383a_1024x1024.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>While listening to <span class="mention-wrap" data-attrs="{&quot;name&quot;:&quot;Dwarkesh Patel&quot;,&quot;id&quot;:4281466,&quot;type&quot;:&quot;user&quot;,&quot;url&quot;:null,&quot;photo_url&quot;:&quot;https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb715ffd1-f7d7-4755-af88-c48efe647f5b_400x400.jpeg&quot;,&quot;uuid&quot;:&quot;3d2e7af2-161d-4604-a589-8f14dc9d7459&quot;}" data-component-name="MentionToDOM"></span>&#8217;s <a href="https://www.dwarkesh.com/p/stephen-kotkin?r=day7j&amp;utm_campaign=post&amp;utm_medium=web">excellent podcast</a> with <a href="https://en.wikipedia.org/wiki/Stephen_Kotkin">Stephen Kotkin</a> (go listen to it if you haven&#8217;t yet. I&#8217;ll wait.), Kotkin made a very interesting point on the dynamics that led to the Communist revolutions in Russia and China. I&#8217;ll provide an excerpt below (emphasis mine):</p><blockquote><p>[In Russia and China], you have this peasant land hunger. <strong>The peasants are often without their own holdings</strong>. They work on someone else's property, or their holdings are so small that if there's a little bit of bad weather let alone a massive drought, they're on the verge of starvation. <strong>Subsistence level agriculture is not politically stable&#8230; </strong></p><p>You need to deal with the peasant land hunger so that it becomes a stabilizing political force. You have the peasants get the land and then they have a piece of the status quo and want to retain the system, versus <strong>the peasants not having the land and they want to overthrow the system to get the land&#8230;</strong></p><p>So <strong>the peasants had their own revolution in 1917 and 1918</strong>, which  was not about the socialist parties. <strong>It's not about the Bolsheviks, it's not about Lenin, it's about the peasants seizing the land.</strong> But that creates an intense radicalism that becomes the platform for the socialists in the cities to gain and hold power in the system. You don't have that in the German case.</p></blockquote><p>Kotkin&#8217;s claim here is that a large, landless peasant class provides a powder keg, and the Communist movements in Russia and China were smart enough and opportunistic enough to capitalize on this latent instability in the system. By contrast, in Germany, the peasants were predominantly landowners, which provided the stability to weather the Communist movements. Meanwhile, in England, peasants as a class were starting to give way to the urban factory worker, who was able to progressively win voting rights and the right to organize into labor unions. </p><p>This result with a Red Russia &amp; China, a fascist Germany, and a capitalist UK seems obvious in retrospect, but for an observer in pre-WWI Europe, this would actually be an extremely surprising result. It was often assumed by Communist intellectuals (including the leaders of the Communist movement) that the global Communist Revolution would <em>begin</em> in England and Germany, which were perceived as the ripest environments for labor to overthrow the capitalist class.</p><p>So <em>why</em> did these dynamics play out so counterintuitively? To understand this, let&#8217;s analyze the sociological model implicit in Kotkin&#8217;s explanation for the causes of the Russian and Chinese revolutions. Broadly, I think Kotkin&#8217;s framing of the issue rests on separating the populace into two groups</p><ol><li><p><strong>The Dampeners</strong> represent the social classes who have a stake in the prevailing order&#8212;they have this stake because they stand to <em>gain </em>from political and economic stability and to <em>lose</em> from political and economic instability. In the tsarist Russian context, this would include the tsar and his bureaucrats, as well as the large landowners. </p></li><li><p><strong>The Powder Kegs </strong>represent the social classes who feel they do <em>not</em> have a stake in the prevailing order&#8212;that is, they feel they do <em>not</em> stand to <em>gain </em>from political and economic stability and, similarly, do <em>not</em> stand to <em>lose</em> from political and economic instability. In the tsarist Russian context, as we&#8217;ve discussed, this group is most easily seen in the landless peasant class.</p></li></ol><p>The Dampeners, then, can be thought of as any groups that are currently benefiting and, importantly, <em>feel</em> as though they are benefiting from the current economic, political, and social structure. They represent the vanguard of the current society, defending its interests since the society&#8217;s interests are <em>their</em> interests as well. That is, their natural inclination is to<em> dampen </em>any dramatic oscillations in their society in order to defend their current interests and their perceived upside. The Powder Kegs, on the other hand, can be thought of as disenfranchised groups that do not have a stake in the nation&#8217;s future. Due to their lack of downside in the face of societal collapse and lack of upside if society remains stable, this class is incentivized to flip the game board over and try to start over. However, in many cases, they will remain latent without an outside spark to <em>light</em> the powder keg.</p><p>In this framing, we can think of the Powder Kegs as generators of civilizational tail risks.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.chrishayduk.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Musings by Chris Hayduk is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h2>An Explanation of Tail Risks &amp; Non-Ergodicity</h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!_0zX!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F26eb4da9-8689-48b9-95af-b2aedfbdd04a_640x358.webp" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!_0zX!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F26eb4da9-8689-48b9-95af-b2aedfbdd04a_640x358.webp 424w, https://substackcdn.com/image/fetch/$s_!_0zX!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F26eb4da9-8689-48b9-95af-b2aedfbdd04a_640x358.webp 848w, https://substackcdn.com/image/fetch/$s_!_0zX!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F26eb4da9-8689-48b9-95af-b2aedfbdd04a_640x358.webp 1272w, https://substackcdn.com/image/fetch/$s_!_0zX!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F26eb4da9-8689-48b9-95af-b2aedfbdd04a_640x358.webp 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!_0zX!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F26eb4da9-8689-48b9-95af-b2aedfbdd04a_640x358.webp" width="640" height="358" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/26eb4da9-8689-48b9-95af-b2aedfbdd04a_640x358.webp&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:358,&quot;width&quot;:640,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;Tail Risk - What It Is and How To Hedge Against It&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Tail Risk - What It Is and How To Hedge Against It" title="Tail Risk - What It Is and How To Hedge Against It" srcset="https://substackcdn.com/image/fetch/$s_!_0zX!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F26eb4da9-8689-48b9-95af-b2aedfbdd04a_640x358.webp 424w, https://substackcdn.com/image/fetch/$s_!_0zX!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F26eb4da9-8689-48b9-95af-b2aedfbdd04a_640x358.webp 848w, https://substackcdn.com/image/fetch/$s_!_0zX!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F26eb4da9-8689-48b9-95af-b2aedfbdd04a_640x358.webp 1272w, https://substackcdn.com/image/fetch/$s_!_0zX!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F26eb4da9-8689-48b9-95af-b2aedfbdd04a_640x358.webp 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>In the finance world, tail risks are low-probability events that cause immense losses (i.e., the left side of the normal distribution in the above image). Tail risks in finance and many other domains also tend to have the following characteristics that make them especially pernicious:</p><ol><li><p>They tend to more closely follow a fat-tailed distribution&#8212;that is, low probability events occur <em>more often</em> than would be expected by a standard normal distribution modeling the event (e.g., 1-day stock returns)</p></li><li><p>The impact of the tail risk is typically <em>much</em> higher than the impact of a more typical data point. For example, a single day when the stock market crashes can wipe out years&#8217; worth of gains.</p></li></ol><p>Given the above, in fat-tailed domains with high-impact left-tail events, we need to be especially careful about defending against (<em>ahem</em>, dampening the effects of) tail risks. And society is <em>the </em>fat-tailed domain. As difficult as forecasting in finance is, forecasting future political and social developments using history might as well be astrology. It&#8217;s highly qualitative, subject to opinion, and, as a result, suffers from human status quo biases (or &#8220;nothing ever happens,&#8221; as the Twitterverse says). Hence, we almost definitely underestimate low-probability events in human society. Thus, the distribution is fat-tailed. In addition, we know that the risk of tail events in human society is immense&#8212;just look at the Bronze Age Collapse or Europe after the fall of the Roman Empire. Thus, the distribution of impactful societal events is fat-tailed <em>with</em> extremely high-impact tail events.</p><p>You may now be thinking, &#8220;Okay, these tail events are definitely a concern. But they are, by definition, low probability, so do we really need to worry about them?</p><p>Enter non-ergodicity.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!zg_C!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6d00797-ae7d-4c2a-99a3-dcded15859b4_602x509.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!zg_C!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6d00797-ae7d-4c2a-99a3-dcded15859b4_602x509.png 424w, https://substackcdn.com/image/fetch/$s_!zg_C!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6d00797-ae7d-4c2a-99a3-dcded15859b4_602x509.png 848w, https://substackcdn.com/image/fetch/$s_!zg_C!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6d00797-ae7d-4c2a-99a3-dcded15859b4_602x509.png 1272w, https://substackcdn.com/image/fetch/$s_!zg_C!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6d00797-ae7d-4c2a-99a3-dcded15859b4_602x509.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!zg_C!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6d00797-ae7d-4c2a-99a3-dcded15859b4_602x509.png" width="602" height="509" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c6d00797-ae7d-4c2a-99a3-dcded15859b4_602x509.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:509,&quot;width&quot;:602,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;A simple explanation of Ergodicity in finance: Part I | by Manu H | Medium&quot;,&quot;title&quot;:&quot;A simple explanation of Ergodicity in finance: Part I | by Manu H | Medium&quot;,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="A simple explanation of Ergodicity in finance: Part I | by Manu H | Medium" title="A simple explanation of Ergodicity in finance: Part I | by Manu H | Medium" srcset="https://substackcdn.com/image/fetch/$s_!zg_C!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6d00797-ae7d-4c2a-99a3-dcded15859b4_602x509.png 424w, https://substackcdn.com/image/fetch/$s_!zg_C!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6d00797-ae7d-4c2a-99a3-dcded15859b4_602x509.png 848w, https://substackcdn.com/image/fetch/$s_!zg_C!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6d00797-ae7d-4c2a-99a3-dcded15859b4_602x509.png 1272w, https://substackcdn.com/image/fetch/$s_!zg_C!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6d00797-ae7d-4c2a-99a3-dcded15859b4_602x509.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">This image captures the notion of spatial (ensemble) averages vs temporal (time) averages well.</figcaption></figure></div><p>To discuss non-ergodicity, we&#8217;ll start by intuitively defining ergodicity. Let&#8217;s imagine a volume, such as a box, and some time-dependent process, like the motion of a particle in the box. Now, suppose we take two distinct measures of the box:</p><ol><li><p>We look to take a particular point in time and compute the average position of all particles in the box.</p></li><li><p>We fix our attention on <em>one particle</em> and track its trajectory over time. We then compute the average of this one particle&#8217;s position based on the data we record as we follow it around the box</p></li></ol><p>In the first measure, we compute a&nbsp;<strong>spatial average&#8212;</strong>we look across the box, identify the positions of all particles within it, and then compute the average position of these particles.</p><p>In the second measure, we compute a&nbsp;<strong>temporal average</strong>&#8212;we focus on&nbsp;<em>a single particle over time</em>, identify all positions it visits, and then average&nbsp;these positions.</p><p>Mathematically, we can identify the spatial average as the standard expected value of a continuous random variable that you should be used to if you&#8217;ve taken an intro course in statistics or probability:</p><div class="latex-rendered" data-attrs="{&quot;persistentExpression&quot;:&quot;\\begin{align*}\n&amp;\\mathbb{E}[\\vec{x}] = \\int_{\\vec{x}} \\vec{x} \\cdot f(\\vec{x})\\\\\n&amp;\\text{where:}\\\\\n&amp;\\vec{x} \\text{ is a vector representing possible 3D coordinates}\\\\\n&amp;f \\text{ is the probability density function of particle positions}\n\\end{align*}&quot;,&quot;id&quot;:&quot;TKXSKGCEXP&quot;}" data-component-name="LatexBlockToDOM"></div><p>Meanwhile, the temporal average is given by the following:</p><div class="latex-rendered" data-attrs="{&quot;persistentExpression&quot;:&quot;\\begin{align*}\n&amp;\\lim_{T \\to \\infty} \\frac{1}{T} \\int_0^T g(t) dt\\\\\n&amp;\\text{where:}\\\\\n&amp;g \\text{ is the function describing the motion of a single particle}\\\\\n&amp;T \\text{ is the maximum time along the path}\n\\end{align*}&quot;,&quot;id&quot;:&quot;SXLEZTQTXF&quot;}" data-component-name="LatexBlockToDOM"></div><p>If the above two notions of an average are equal for a given random process, then we say that it is <strong>ergodic</strong>. Continuing our physical intuition, this equality is saying that, if we follow the motion of one particle in our box, it will visit all of the locations in the box in the <em>exact </em>proportions that we would expect to find all of the particles at a <em>fixed </em>time. That is, we can use spatial and temporal averages interchangeably because <em>they are the same for ergodic processes</em>. The particle motion example was chosen deliberately, as Brownian motion (i.e., the random motion of particles in a volume) is one of the canonical physical processes exhibiting ergodic behavior.</p><p>Now, non-ergodic processes are random processes in which the spatial average and the temporal average are <em>not </em>equal. The behavior of one data point over time does <em>not </em>equal the average behavior across <em>all</em> data points at one instant in time.</p><p>Why do we care to make this distinction? Because whenever people talk about averages, they almost always default to the spatial average and then go on to use these averages to make predictions about particular data points. For example, people will often say, &#8220;The average yearly return for the S&amp;P 500 is 10%, and therefore I can model my portfolio as gaining 10% per year.&#8221; I hope now the issue with this thinking is clear: <strong>our hypothetical stock allocator has computed a spatial average and then applied it to one data point over time without checking if the process is ergodic!</strong> And, spoiler alert, stock returns are a canonical example of non-ergodic processes, so this reasoning would lead to dangerously faulty conclusions!</p><p>Let&#8217;s now explore exactly <em>why</em> the combination of underestimating tail risks and confusing ergodic and non-ergodic processes is so dangerous with a concrete example.</p><h2>Tail Risks, Non-Ergodicity, &amp; Society</h2><p>We&#8217;ll start by building a simple model of the real GDP growth of a standard developed economy that has a latent probability of collapse. We can model this as a discrete-time process, where each year is treated as a single period. In any given year, we will estimate a 0.1% chance of ruin (i.e., the hypothetical country fails and its real GDP goes to 0). This can represent governmental collapse, nuclear war, incurable diseases that wipe out the population, etc. In the vast majority of years (with a probability of 99.9%), we&#8217;ll forecast that the country&#8217;s real GDP growth is normally distributed with a mean gain of 2.5% and a standard deviation of 0.5 percentage points.</p><p>This setup combines the main concepts of tail risks and non-ergodicity that we mentioned previously:</p><ol><li><p><strong>Tail risks &#8212; </strong>we have a 0.1% chance of ruin in any given year. So our risk is extremely rare but also very impactful, just as we&#8217;ve defined tail risks.</p></li><li><p><strong>Non-ergodicity &#8212; </strong>when a country fails (i.e., its GDP goes to 0), it doesn&#8217;t have the chance to start over. Hence, if we follow the failed country over time, after it fails, it doesn&#8217;t have the chance to explore the full state space (and hence the temporal average will differ from the spatial average).</p></li></ol><p>These two effects, when combined, produce <em>very</em> shocking effects, even though at first glance it might look like we only have a 0.1% chance of catastrophe.</p><p>We&#8217;ll now run this simulation for 500 years and observe the effects. In this simulation, every country will start with a GDP of $1.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Y35z!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F06738049-be61-4f0f-9b35-3da71597aae0_1000x600.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Y35z!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F06738049-be61-4f0f-9b35-3da71597aae0_1000x600.png 424w, https://substackcdn.com/image/fetch/$s_!Y35z!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F06738049-be61-4f0f-9b35-3da71597aae0_1000x600.png 848w, https://substackcdn.com/image/fetch/$s_!Y35z!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F06738049-be61-4f0f-9b35-3da71597aae0_1000x600.png 1272w, https://substackcdn.com/image/fetch/$s_!Y35z!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F06738049-be61-4f0f-9b35-3da71597aae0_1000x600.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Y35z!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F06738049-be61-4f0f-9b35-3da71597aae0_1000x600.png" width="1000" height="600" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/06738049-be61-4f0f-9b35-3da71597aae0_1000x600.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:600,&quot;width&quot;:1000,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:67965,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.chrishayduk.com/i/168124612?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F06738049-be61-4f0f-9b35-3da71597aae0_1000x600.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Y35z!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F06738049-be61-4f0f-9b35-3da71597aae0_1000x600.png 424w, https://substackcdn.com/image/fetch/$s_!Y35z!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F06738049-be61-4f0f-9b35-3da71597aae0_1000x600.png 848w, https://substackcdn.com/image/fetch/$s_!Y35z!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F06738049-be61-4f0f-9b35-3da71597aae0_1000x600.png 1272w, https://substackcdn.com/image/fetch/$s_!Y35z!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F06738049-be61-4f0f-9b35-3da71597aae0_1000x600.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>I want to start by directing your attention to the left side of the graph. You&#8217;ll see that <strong>a full</strong> <strong>39.7% of countries in this simulation collapsed, with their GDPs going to 0. </strong>Our initial intuition that we had a 0.1% chance of catastrophe was <em>incredibly </em>wrong&#8212;we have a 0.1% chance of catastrophe in any given year, but the chance that society fails in <em>any</em> year increases as time advances! Rather than saying &#8220;we have a 0.1% chance of catastrophe&#8221;, which sounds relatively benign and pushes people towards complacency, it would be far more accurate to say &#8220;we have a 40% chance of catastrophe in the next 500 years&#8221;, which sounds much more dire.</p><p>Now, take a look at the legend in the top-right corner of the graph. Here, you can consider the mean to be the spatial average we&#8217;ve been discussing&#8212;hence, in the vast majority of conversations where this data is getting discussed, someone would likely say &#8220;the average country over this 500-year sample saw its GDP grow from $1 to $12,032.61, a gain of 1,203,160%&#8221;. Hence, the reported &#8220;average&#8221;, the spatial average, actually doesn&#8217;t capture the experience of living in these conditions at all! You would think, based on seeing an average gain of over 1 million percent, that all or most countries developing under these conditions would be on the march to utopia. However, these countries only have barely better than a coin flip&#8217;s chance of not collapsing into complete chaos in this 500-year sample!</p><p>To better understand the dynamics of this process, we need to instead look at the temporal average. We can do this by computing the geometric mean of returns for each individual path and then averaging this set of geometric returns. This yields an average yearly <em>loss </em>of 39.5%, with an expected ending value of $0, aligning much more closely with the collapse dynamics we see in the histogram. This brings us to an even broader insight that can be made here, which we can see in the plot below.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ZZ4H!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F605a0ac2-3e2b-4c6c-a604-2d5d8e19a279_1000x600.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ZZ4H!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F605a0ac2-3e2b-4c6c-a604-2d5d8e19a279_1000x600.png 424w, https://substackcdn.com/image/fetch/$s_!ZZ4H!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F605a0ac2-3e2b-4c6c-a604-2d5d8e19a279_1000x600.png 848w, https://substackcdn.com/image/fetch/$s_!ZZ4H!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F605a0ac2-3e2b-4c6c-a604-2d5d8e19a279_1000x600.png 1272w, https://substackcdn.com/image/fetch/$s_!ZZ4H!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F605a0ac2-3e2b-4c6c-a604-2d5d8e19a279_1000x600.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ZZ4H!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F605a0ac2-3e2b-4c6c-a604-2d5d8e19a279_1000x600.png" width="1000" height="600" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/605a0ac2-3e2b-4c6c-a604-2d5d8e19a279_1000x600.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:600,&quot;width&quot;:1000,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:63560,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.chrishayduk.com/i/168124612?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F605a0ac2-3e2b-4c6c-a604-2d5d8e19a279_1000x600.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!ZZ4H!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F605a0ac2-3e2b-4c6c-a604-2d5d8e19a279_1000x600.png 424w, https://substackcdn.com/image/fetch/$s_!ZZ4H!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F605a0ac2-3e2b-4c6c-a604-2d5d8e19a279_1000x600.png 848w, https://substackcdn.com/image/fetch/$s_!ZZ4H!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F605a0ac2-3e2b-4c6c-a604-2d5d8e19a279_1000x600.png 1272w, https://substackcdn.com/image/fetch/$s_!ZZ4H!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F605a0ac2-3e2b-4c6c-a604-2d5d8e19a279_1000x600.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>As time moves forward in our sample, <strong>more and more of the countries are collapsing</strong>. Taking this to the limit, eventually, all societies in this simulation will collapse. Even more generally, when the probability of catastrophic ruin is greater than 0, <strong>any society</strong> <strong>will collapse given enough time.</strong></p><p>To tie things back into the beginning of our discussion, what Kotkin has identified in his analysis of the latent instability in the 20th-century Russian and Chinese societies is that <strong>the presence of disenfranchised peasant groups increased the yearly chance of ruin for these societies, and thus led to their demises. </strong>Combining with our analyses above, we see that when one or more Powder Keg groups exist in a society (making the probability of ruin non-zero), <strong>the long-run survival probability of that civilization is exactly 0. </strong>The more such Powder Kegs exist, and the more groups that appeal to these groups, the greater the probability of collapse in a given year.</p><p>On the flip side, if we re-run the simulation with an order of magnitude lower probability of ruin (i.e., a 0.01% risk of collapse per year), we see a <em>huge </em>reduction in risk in the 500-year window. Only 4.7% of societies in this scenario collapse, reducing our risk of ruin by 35 percentage points.</p><p>Given these analyses and the conclusions we&#8217;ve drawn, it would seem that identifying potential Powder Kegs and implementing policies to let off some steam are some of <em>the </em>most important actions a government can undertake. If we can identify these groups and pacify them, we can substantially improve the survival probability of our society over large timescales and avoid outright collapse. The late Russian tsars and Chinese emperors failed to identify and pacify their Powder Keg group. The Communist parties in those countries, however, did not overlook the peasant Powder Keg group and were thus able to leverage them to bring the country&#8217;s risk of ruin to reality.</p><p>With this lesson in mind, let&#8217;s look at a Powder Keg group that is roughly analogous to the landless peasants in tsarist Russia and examine some ways we can defuse the impending explosion.</p><h2>A 21st Century Powder Keg &amp; Its Potential Solutions</h2><p>To identify our Powder Keg, let&#8217;s return to our definition of the group: we need to find a social class that doesn&#8217;t have a stake in the prevailing political order. That is, <a href="https://www.theguardian.com/society/ng-interactive/2025/jul/13/first-time-us-homebuyers-low">we</a> <a href="https://news.gallup.com/poll/266807/percentage-americans-owns-stock.aspx">need</a> <a href="https://planning.org/foresight/trend/9310309/#:~:text=Millennials%20and%20Gen%20Z%20combined,Moyo%20Studio%2FGetty%20Images.">to</a> <a href="https://smartasset.com/financial-advisor/wealth-by-generation">find</a> <a href="https://www.axios.com/2024/06/22/gen-z-homeownership-rent-prices">a</a> <a href="https://investors.redfin.com/news-events/press-releases/detail/1032/redfin-reports-gen-zs-homeownership-rate-stagnated-in">group</a> that doesn&#8217;t stand to gain from the country&#8217;s success since it doesn&#8217;t own a stake in the country, and also has the <a href="https://www.reddit.com/r/FirstTimeHomeBuyer/comments/17j9n44/is_gen_z_screwed/">subjective</a> <a href="https://www.reddit.com/r/GenZ/comments/1aqegwp/i_shocked_my_dad_yesterday_when_i_told_him_most/">feeling</a> that it will <em>never </em>achieve such a stake in the current system. Hmm&#8230; where can we find such a group?</p><p>If you clicked on any of the myriad links in the previous paragraph, I think it&#8217;s clear that Gen Z fits the bill&#8212;they are currently dealing with extremely high housing costs that have put home ownership out of reach for many members of that generation (at least, that is, in the cities with the jobs they&#8217;d want to work and places they&#8217;d want to live). Most importantly, this generation <em>feels</em> like it will never get better. There is a very strong, pervasive sense that society is fundamentally set up so that they won&#8217;t succeed. We can see this dynamic starting to play out politically, with <a href="https://now.tufts.edu/2024/11/12/young-voters-shifted-toward-trump-still-favored-harris-overall">Gen Z shifting strongly to Trump</a> during the presidential election, who openly espoused the need to tear down the current economic and social order to &#8220;Make America Great Again&#8221;. We are also seeing this in New York City, where Gen Z voters are turning out in strong numbers <a href="https://www.amny.com/news/gen-z-voters-mamdani-primary-victory-2025/">to vote for Zohran Mamdani</a>, the once&#8209;underdog&#8209;turned&#8209;favorite mayoral candidate who has some <a href="https://thehill.com/homenews/house/5377426-mike-lawler-zohran-mamdani-capitalism/">fairly openly held anti-capitalist views.</a></p><p>Whether or not the issues Gen Z faces require a radical overthrow of the prevailing social order, much of Gen Z certainly <em>feels </em>that this is the case. And that is enough for a radical group (like the Communists did in 20th-century Russia and China) to leverage this group as the bedrock for a revolution. Which means, to protect our civilization from collapse, we have to resolve the issues they&#8217;re facing and give them (and future generations) a stake in the current social order.</p><p>The way I see it, having a stake in American society as it&#8217;s currently constructed essentially amounts to two things:</p><ol><li><p>Owning a home</p></li><li><p>Owning a piece of the stock market</p></li></ol><p>Owning a home is the traditional view of success in the United States, and so the inability to buy causes immense stress to younger generations. It makes them feel shut out from the typical path to prosperity that their parents and grandparents have followed. And, as we&#8217;ve discussed, simply <em>feeling</em> that there is no hope is enough to provide a latent layer of unrest, regardless of its truth. By providing them with a pathway to home ownership, we could reduce the anxiety and anger directed at the US social order. In addition, home ownership typically connects people to a community more than renting does, as they can form more permanent roots, which could also help to assuage the social isolation felt by many in the younger generations.</p><p>Similarly, the more modern view of success amounts to having significant capital in the US stock market. Gen Z generally feels resentment towards US companies and CEOs, mostly because they feel that they are being &#8220;robbed&#8221; or taken advantage of by these companies. This feeling is likely augmented due to their lack of ownership in the stock market&#8212;if they owned a slice of all of these corporations, it would be much harder to feel robbed since Gen Z would, in turn, be getting richer from their success.</p><p>So we now have a two-pronged approach&#8212;give young people a vested interest in US corporations and a vested interest in the housing market. Both of these assets will reduce the desire for societal disruption and instead incentivize stability and growth, since younger people would now be directly harmed by disruption and directly benefit from growth.</p><p>What would policies targeting these two areas look like?</p><h3>Housing Policy</h3><p>To defuse the Gen Z Powder Keg through housing, we must confront the sacred cows of American urban planning and economic policy head-on, adopting measures that prioritize radical supply-side interventions over the palliative demand-side subsidies that have only inflated bubbles and entrenched resentment. The goal here is not mere affordability tweaks but a structural overhaul that floods the market with housing units, crashing prices to levels where ownership becomes a default expectation rather than a distant dream, thereby converting potential revolutionaries into vested stakeholders who crave stability to protect their newfound equity.</p><p><strong>First, deregulation stands as the linchpin</strong>: easing zoning laws, building codes, and permitting processes to unleash denser, faster development. In cities like San Francisco or New York, where Byzantine zoning ordinances have artificially constricted supply, we could emulate the Houston model on steroids. Imagine upzoning entire neighborhoods to allow mid-rise apartments and mixed-use developments without the endless environmental impact studies or community vetoes that currently serve as de facto barriers erected by incumbent homeowners (our modern Dampeners) to preserve their scarcity-driven windfalls. By slashing approval timelines from years to months, we'd not only multiply housing stock exponentially but also incentivize innovation in modular construction and prefabrication, driving down costs through economies of scale. This isn't about anarchic sprawl; it's about recognizing that the current regulatory thicket is a form of crony capitalism, protecting legacy interests at the expense of intergenerational equity. The contrarian insight? Such deregulation would disproportionately benefit high-density urban cores, where Gen Z wants to live, fostering vibrant, walkable communities that counteract social atomization, turning isolated renters into networked homeowners with skin in the civic game.</p><p>Complementing this, <strong>implement a land value tax (LVT) while eliminating traditional property taxes.</strong> Drawing from Georgist economics, an LVT taxes the unimproved value of land itself, not the structures or improvements upon it. This shifts the fiscal burden onto speculative land hoarders and absentee owners, who currently profit from scarcity without contributing to productivity, and rewards those who build densely and efficiently. In practice, this would discourage vacant lots in prime areas (a plague in places like Los Angeles) and encourage vertical development, as the tax incentivizes maximizing output per square foot. Eliminating property taxes, which penalize improvements, removes the disincentive to renovate or expand, further accelerating supply. The net effect? Housing prices plummet as land speculation collapses, making entry-level ownership feasible even for baristas or entry-level coders. Critically, this policy is revenue-neutral or even positive for governments, funding infrastructure without the distortionary effects of income or sales taxes. </p><p>Next, <strong>a targeted reduction in union power within the construction sector to slash labor input costs.</strong> Unions, while historically vital for worker protections, have in many blue states metastasized into rent-seeking guilds that inflate wages far above market rates through prevailing wage laws and project labor agreements, adding 20-30% premiums to public projects alone. By reforming or outright repealing these mandates (perhaps via federal preemption in interstate commerce-affected builds), we could introduce competitive bidding from non-union labor, including skilled immigrants under expanded visa programs. This isn't anti-labor animus; it's a recognition that artificially high costs perpetuate the very inequality unions purport to fight, locking out younger workers from affordable homes while padding the pensions of Boomer incumbents. The result? Faster builds at lower prices, with spillover effects into related industries. Tie this to broader labor market fluidity (e.g., right-to-work expansions), and we create a virtuous cycle where construction booms absorb underemployed Gen Z, giving them immediate income upside alongside future ownership stakes.</p><p>Finally, <strong>ramp up domestic energy production to crater energy input costs, which permeate every stage of housing development from raw materials extraction to HVAC installation</strong>. Fracking, nuclear deregulation (streamlining NRC approvals to Yucca Mountain levels of efficiency), and offshore drilling expansions could flood the market with cheap BTUs, reversing the green energy mandates that have jacked up costs via intermittent renewables and grid unreliability. By prioritizing baseload abundance, we not only halve construction energy bills but also enable all-electric homes that are cheaper to operate, further lowering the ownership threshold. In our tail risk framework, this policy acts as a multiplier: cheaper energy stabilizes the broader economy, reducing the probability of ruin from exogenous shocks like oil crises, while directly empowering the housing supply surge that integrates Gen Z as Dampeners.</p><p>Implemented holistically, these policies could halve median home prices in high-demand areas within a decade, per supply elasticity models, transforming Gen Z's latent rage into conservative impulses. After all, nothing quells revolutionary fervor like a mortgage and rising property values.</p><h3>Stock Market Policy</h3><p>Shifting to the equity side, the objective is to forge a direct umbilical cord between young Americans and the productive engine of capitalism: the stock market. Rather than the anemic 401(k) matches or tax-advantaged IRAs that presuppose stable employment and disposable income, we need audacious, universal mechanisms that inject ownership stakes from cradle to career, preempting the feeling of disenfranchisement that fuels Powder Kegs.</p><p>The recently enacted MAGA (Money Accounts for Growth and Advancement) accounts (rebranded as Trump Accounts in the final text of the One Big Beautiful Bill) represent a strong starting point in this. These accounts automatically seed every eligible newborn with a $1,000 federal contribution (under a pilot for births from 2024 to 2028), invested in diversified funds tracking U.S. equity indices like the S&amp;P 500, with tax-advantaged growth and additional contributions capped at $5,000 annually from taxable entities or unlimited from tax-exempt organizations. Access is tiered&#8212;no withdrawals until age 18, then partial for qualified purposes like education, entrepreneurship, or homebuying, expanding at 25 and fully unrestricted at 30&#8212;with favorable long-term capital gains taxation on qualified distributions. These accounts instill from infancy a visceral sense of ownership in corporate America, equating personal gain to national prosperity. To secure broader political alignment and ensure the program's longevity beyond electoral cycles, however, a further rebranding is essential, shedding the partisan connotations of "MAGA" or "Trump Accounts" in favor of a neutral, inclusive name like "American Future Funds" or "National Prosperity Accounts." This move transcends left-right divides, framing the initiative as a non-ideological investment in collective stability; by inviting buy-in from all societal factions, we mitigate the risk of future repeals or sabotage, transforming it into an enduring institution that unites rather than polarizes, much like Social Security evolved from its contentious origins into a bipartisan bedrock.</p><p>However, these accounts don&#8217;t go far enough: the $1,000 seed is paltry in the face of compounding's long horizons (barely scratching $10,000 by age 18 at historical 7% real returns), eligibility skews toward newborns and young children (leaving Gen Z and millennials largely sidelined without retroactive catch-ups), contribution limits throttle broader participation, and the delayed access risks alienating a generation already primed for immediate disruption over deferred gratification. To truly dampen tail risks, we must expand them aggressively: scale the initial deposit to $5,000, extend automatic enrollment with catch-up lumps ($10,000+) for 18-35-year-olds funded by reallocating corporate tax windfalls from buybacks to "citizen shares," eliminate age-based restrictions in favor of penalty-enforced long-term holds, and broaden investment mandates to include total-market ETFs for true diversification. This supercharged version subverts the Marxist narrative of "exploitation by capitalists" by making everyone a micro-capitalist, diluting class warfare into shared prosperity. In tail risk terms, widespread equity ownership dampens systemic shocks; if a critical mass holds stakes, political pressures shift from "eat the rich" to "protect the market," reducing the annual ruin odds by fostering a Dampener supermajority.</p><p>To amplify this foundation even further, redirect any universal basic income (UBI) pilots or proposals straight into mandatory stock purchases for recipients under an expanded "Universal Basic Equity" (UBEI) framework, rather than cash disbursements that dissipate into consumption traps. Imagine channeling equivalent funds (say, $1,000 monthly equivalents) into the augmented MAGA/Trump Accounts or similar diversified portfolios, perhaps via Vanguard-esque vehicles with algorithmic rebalancing. This forces skin-in-the-game; recipients become vested in corporate efficiency and innovation, as their "income" derives from dividends and appreciation rather than zero-sum transfers.</p><p>Collectively, these stock policies embed a profound incentive alignment: young people gain from US economic ascendance, viewing disruptions&#8212;like tariffs or regulations&#8212;as direct threats to their wealth. This vested interest in continued success transmutes Powder Kegs into Dampeners, slashing civilizational ruin probabilities and extending our society's survival horizon indefinitely.</p><h2>Conclusion</h2><p>In revisiting Kotkin's insights on the peasant-driven upheavals that toppled empires in Russia and China, we've unearthed a timeless sociological truth: societies teeter on the edge of ruin when disenfranchised masses, our Powder Kegs, accumulate without recourse, their latent volatility amplified by opportunistic sparks. Through the lens of tail risks and non-ergodicity, we've modeled this as an inexorable march toward collapse in any system where the annual probability of catastrophe exceeds zero, where spatial averages seduce us into complacency while temporal realities reveal the fragility of civilizational paths. Gen Z, as the 21st-century analog to those landless peasants, embodies this threat not through malice but through a profound sense of exclusion from the American dream's twin pillars: homeownership and equity participation.</p><p>The policies outlined&#8212;radical housing deregulation paired with Georgist taxation, union reforms, and energy abundance to crash barriers to entry; alongside expansions of rebranded, universal stock accounts and equity-directed UBI to democratize capital&#8212;are strategic dampeners designed to integrate the alienated into the status quo's vanguard. By granting tangible stakes in stability, we slash that ruin probability, extending our society's viable horizon from inevitable doom to indefinite prosperity. This represents a first attempt at pragmatic ergodic engineering: in a fat-tailed world, the true revolutionaries are those who preempt the keg's ignition, ensuring that history's surprises favor continuity over cataclysm. Ignore this at our peril; embrace it, and we might just forge history&#8217;s first immortal civilization.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.chrishayduk.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Musings by Chris Hayduk is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[On Implicit and Explicit Learning]]></title><description><![CDATA[Some insights from my time learning languages]]></description><link>https://www.chrishayduk.com/p/on-implicit-and-explicit-learning</link><guid isPermaLink="false">https://www.chrishayduk.com/p/on-implicit-and-explicit-learning</guid><dc:creator><![CDATA[Chris Hayduk]]></dc:creator><pubDate>Wed, 09 Jul 2025 23:32:33 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!FnuP!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa745d1ec-6c4b-48de-bd1f-305c43f16313_1358x1920.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>I&#8217;ve been an off-and-on language learner for the last decade, and, after a brief hiatus over the last couple of years, I have returned to study Spanish and Mandarin in earnest.</p><p>Returning to the language learning grind has me reflecting on different modes of learning, their interplay, and how people often get them wrong. To provide us a jumping-off point, I&#8217;m going to define two broad modes of learning:</p><ol><li><p><strong>Explicit Learning</strong>, which we can consider any learning that requires teaching rules or procedures. This type of learning is often accompanied by a textual explanation of the procedure and some examples to practice applying that procedure. Most subjects in school are taught this way: e.g., learning a new approach in math or working through a grammar workbook in Spanish class.</p></li><li><p><strong>Implicit Learning,</strong> which we can consider any learning that occurs as a mostly unconscious process. A helpful rule of thumb is that any learned material that requires a response in seconds falls under the purview of implicit learning - at that time scale, there isn&#8217;t enough time to work through a procedure (and hence to use the tools of explicit learning).</p></li></ol><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.chrishayduk.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Musings by Chris Hayduk is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>Now, anyone who has learned a language in school (or read my not-so-subtle leading example in the Explicit Learning section) knows that explicit learning is the <em>modus operandi</em> of foreign language instruction in the United States. Virtually the whole process occurs in a grammar textbook where the student is shown some concept, such as the conjugation of verbs ending in -ar in the present tense in Spanish. The student then reads through the &#8220;logic&#8221; of how this grammar concept works and works through a set of examples in order to &#8220;apply&#8221; the procedure.</p><p>I hope the scare quotes I used above made it abundantly clear how ridiculous I think it is to make language learning logical or procedural. In fact, there is <em>no</em> logic to language grammar since it did not grow out of a logical process! Language grew organically alongside humans, and so it has all the same messy, non-logical characteristics that we see in all things biological and anthropological. This is why every grammar &#8220;rule&#8221; that is taught has dozens, if not hundreds, of exceptions! We&#8217;re applying a method of learning suitable for math to a completely unrelated field in language learning that categorically does <em>not </em>follow a logical structure with teachable procedures!</p><p>(Trigger warning for anyone traumatized by this teaching method: I am about to show an example of one of these accursed grammar worksheets.)</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!FnuP!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa745d1ec-6c4b-48de-bd1f-305c43f16313_1358x1920.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!FnuP!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa745d1ec-6c4b-48de-bd1f-305c43f16313_1358x1920.jpeg 424w, https://substackcdn.com/image/fetch/$s_!FnuP!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa745d1ec-6c4b-48de-bd1f-305c43f16313_1358x1920.jpeg 848w, https://substackcdn.com/image/fetch/$s_!FnuP!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa745d1ec-6c4b-48de-bd1f-305c43f16313_1358x1920.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!FnuP!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa745d1ec-6c4b-48de-bd1f-305c43f16313_1358x1920.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!FnuP!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa745d1ec-6c4b-48de-bd1f-305c43f16313_1358x1920.jpeg" width="1358" height="1920" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/a745d1ec-6c4b-48de-bd1f-305c43f16313_1358x1920.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1920,&quot;width&quot;:1358,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:169400,&quot;alt&quot;:&quot;19 Spanish Verb Worksheets - Free PDF at worksheeto.com&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="19 Spanish Verb Worksheets - Free PDF at worksheeto.com" title="19 Spanish Verb Worksheets - Free PDF at worksheeto.com" srcset="https://substackcdn.com/image/fetch/$s_!FnuP!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa745d1ec-6c4b-48de-bd1f-305c43f16313_1358x1920.jpeg 424w, https://substackcdn.com/image/fetch/$s_!FnuP!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa745d1ec-6c4b-48de-bd1f-305c43f16313_1358x1920.jpeg 848w, https://substackcdn.com/image/fetch/$s_!FnuP!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa745d1ec-6c4b-48de-bd1f-305c43f16313_1358x1920.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!FnuP!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa745d1ec-6c4b-48de-bd1f-305c43f16313_1358x1920.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Don&#8217;t say I didn&#8217;t warn you.</figcaption></figure></div><p>Now you might be thinking: &#8220;Okay, so easy fix &#8212; we just move all of language learning and any other similar fields to implicit learning. Explicit learning remains reserved for the logical subjects of math and physics. QED.&#8221; In this paradigm, <em>all</em> of the non-rigorous subjects would eschew explicit learning, and <em>all </em>of the rigorous subjects would<em> </em>focus <em>only </em>on explicit learning.</p><p>However, despite this aligning more with the natural contours of each subject, we would be leaving potential learning speed on the table by sticking only to implicit or explicit learning. I will give two examples to illustrate the point:</p><ul><li><p><strong>Language Learning (Implicit Learning Paradigm) &#8212; </strong>Suppose you are learning a foreign language and, like the good language learning student we&#8217;ve outlined thus far, you&#8217;re doing it fully through immersion (that is, implicit learning). You avoid all grammar explanations and worksheets, and instead spend all of your time listening to podcasts, reading books, and watching YouTube videos in your target language. If you come across a particular grammar construction that you don&#8217;t understand, you just try your best to understand it in context and keep immersing. With this setup, it might take <em>dozens of hours</em> of immersion to understand the grammar concept, particularly for rare constructions or variants of the same construction that look unrelated on the surface. Meanwhile, reading about the grammar topic and doing some exercises might take under an hour and prime you for your immersion.</p></li><li><p><strong>Mathematics (Explicit Learning Paradigm) &#8212; </strong>Now suppose that you&#8217;re a math student and, like the good math student we&#8217;ve outlined thus far, you&#8217;re doing it fully through working through derivations and proofs (that is, explicit learning). You avoid any form of memorization or intuition building, and instead spend all of your time focusing on logically deriving all of your course content. Now, as the course moves on, it will assume knowledge that you learned earlier in the semester. However, since you didn&#8217;t bother to do any memorizing or intuition-building, you don&#8217;t remember those earlier concepts with any sort of automaticity! You need to rederive all the knowledge from scratch, and, after the course material builds up beyond a certain point, the cognitive load of this derivation becomes too much, and you start to fall behind.</p></li></ul><p>As the two above examples show, even in subjects that are skewed towards one of the learning paradigms, the other still has an important role to play. </p><p>In implicit-dominant subjects, we can use explicit learning to identify or study patterns that we would like to notice <em>during </em>the implicit skill development. This would be like a language learner studying a grammar concept (explicit learning) to improve their understanding during immersion (implicit learning), or a basketball player breaking down defensive coverages in film (explicit learning) so he can make better reads during scrimmages and games (implicit learning).</p><p>In an explicit-domain subject, we can use implicit learning to memorize and internalize concepts and procedures without worrying about their logical derivation. This would be like a mathematics student memorizing common integrals (implicit learning) to make it easier to understand the logical derivations of concepts that use integration, such as probability distributions (explicit learning).</p><p>Thus, broken down this way, the problem of learning any subject can be reduced to a three-part problem:</p><ol><li><p>Where can you use explicit learning techniques in this subject?</p></li><li><p>Where can you use implicit learning techniques in this subject?</p></li><li><p>What should the balance and interplay of these learning paradigms be in this subject?</p></li></ol><p>If you nail these three questions for your particular subject, then you will have achieved optimal learning speed and efficiency.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.chrishayduk.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Musings by Chris Hayduk is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Gemini 2.5 Pro: How Data + Compute Moats Beat Algorithmic Tweaks]]></title><description><![CDATA[Gemini 2.5 Pro and Google's path to AI supremacy]]></description><link>https://www.chrishayduk.com/p/google-takes-the-lead-in-the-ai-race</link><guid isPermaLink="false">https://www.chrishayduk.com/p/google-takes-the-lead-in-the-ai-race</guid><dc:creator><![CDATA[Chris Hayduk]]></dc:creator><pubDate>Mon, 14 Apr 2025 21:08:30 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!6Mpj!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F67bced44-f871-4628-8285-a1ce227b72cc_1536x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>The race towards Artificial General Intelligence (AGI) and state-of-the-art AI models is often framed around breakthrough algorithms and novel architectures. However, a deeper analysis reveals that the true drivers of durable leadership lie elsewhere. While algorithmic innovation is crucial, the path to AI supremacy is increasingly paved with massive datasets and unparalleled computational power. When viewed through this lens, Google DeepMind emerges not just as a competitor, but as the likely frontrunner.</p><h3>The Trifecta of AI Progress: Algorithms, Compute, and Data</h3><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!6Mpj!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F67bced44-f871-4628-8285-a1ce227b72cc_1536x1024.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!6Mpj!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F67bced44-f871-4628-8285-a1ce227b72cc_1536x1024.png 424w, https://substackcdn.com/image/fetch/$s_!6Mpj!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F67bced44-f871-4628-8285-a1ce227b72cc_1536x1024.png 848w, https://substackcdn.com/image/fetch/$s_!6Mpj!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F67bced44-f871-4628-8285-a1ce227b72cc_1536x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!6Mpj!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F67bced44-f871-4628-8285-a1ce227b72cc_1536x1024.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!6Mpj!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F67bced44-f871-4628-8285-a1ce227b72cc_1536x1024.png" width="1456" height="971" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/67bced44-f871-4628-8285-a1ce227b72cc_1536x1024.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:971,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2138408,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.chrishayduk.com/i/161334401?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F67bced44-f871-4628-8285-a1ce227b72cc_1536x1024.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!6Mpj!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F67bced44-f871-4628-8285-a1ce227b72cc_1536x1024.png 424w, https://substackcdn.com/image/fetch/$s_!6Mpj!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F67bced44-f871-4628-8285-a1ce227b72cc_1536x1024.png 848w, https://substackcdn.com/image/fetch/$s_!6Mpj!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F67bced44-f871-4628-8285-a1ce227b72cc_1536x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!6Mpj!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F67bced44-f871-4628-8285-a1ce227b72cc_1536x1024.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Training large-scale AI models hinges on three interdependent pillars:</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.chrishayduk.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Musings by Chris Hayduk! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><ol><li><p><strong>Algorithms:</strong> These are the recipes, the architectures (like Transformers, Mixture-of-Experts), and the training methodologies (loss functions, optimization techniques) that dictate how effectively models learn patterns and relationships from data. Efficient algorithms extract more "knowledge" per unit of data and compute.</p></li><li><p><strong>Compute:</strong> This represents the raw processing power, typically measured in FLOPs (Floating Point Operations Per Second), required to execute the vast number of calculations involved in training deep neural networks. It's the energy input transforming potential into a trained artifact.</p></li><li><p><strong>Data:</strong> This is the raw material &#8211; the text, images, code, audio, video, and other modalities &#8211; from which the model learns the structure of the world, language, and reasoning. The quality, quantity, and diversity of data fundamentally shape the model's capabilities.</p></li></ol><p>These factors exhibit strong interplay. An algorithmic leap, like the transition from RNNs/LSTMs to Transformers for sequence modeling, unlocked the potential to effectively utilize vastly larger datasets and compute budgets. Before Transformers, training on web-scale text data with massive parameter counts often hit diminishing returns due to limitations in handling long-range dependencies and parallelization. The Transformer architecture, with its self-attention mechanism, was significantly more scalable, allowing marginal increases in data and compute to translate into tangible performance gains once more. The performance wasn't just better; the <em>scaling properties</em> improved.</p><h3>The Illusion of Algorithmic Moats</h3><p>Recent history is replete with examples emphasizing algorithmic prowess. The excitement around models like DeepSeek-R1, achieving remarkable performance with comparatively modest training resources, underscores the power of efficient architectures (like Mixture-of-Experts) and optimized training strategies. It proves that clever algorithms <em>can</em> significantly improve the compute/data-to-performance ratio.</p><p>However, as I argued previously in <em><a href="https://www.chrishayduk.com/p/on-algorithmic-moats-and-the-path">On Algorithmic Moats and the Path to AGI</a></em>, algorithms alone do not constitute a sustainable competitive advantage in the current AI landscape. Why?</p><ol><li><p><strong>Talent Mobility:</strong> The AI research community is fluid. Top researchers frequently move between major labs like Google DeepMind, OpenAI, Anthropic, and Meta, carrying conceptual knowledge and insights about successful (and unsuccessful) architectural experiments and training techniques. While NDAs exist, the fundamental <em>ideas</em> diffuse rapidly.</p></li><li><p><strong>Open Source and Publication:</strong> Key players like Meta (LLaMA series) and innovative teams like DeepSeek often open-source their models and research. Academic institutions and arXiv ensure rapid dissemination of novel techniques. This accelerates the entire field but levels the playing field algorithmically. A breakthrough published today can be replicated and built upon by competitors within months, if not weeks.</p></li></ol><p>Therefore, relying solely on being the <em>first</em> to discover the next architectural tweak is a fragile strategy. Being a fast follower, capable of rapidly implementing and scaling proven algorithmic advances discovered elsewhere, might be just as effective, <em>provided</em> you possess advantages in the other two factors.</p><h3>The Real Moats: Data and Compute Scale</h3><p>If algorithms are becoming increasingly commoditized, what provides a durable edge? The answer lies in the factors that are far harder to replicate: <strong>data and compute.</strong></p><p><strong>Why Scale Matters:</strong> The principle of scaling laws in deep learning empirically demonstrates that model performance often improves predictably, following a power law, as model size, dataset size, and training compute increase. While we've seen impressive results from smaller, efficient models, we are likely still far from the point of diminishing returns for many complex reasoning and multimodal tasks. Reaching the next plateau of AI capability will almost certainly require scaling data and compute far beyond current levels.</p><p><strong>Why They Are Moats:</strong></p><ol><li><p><strong>Non-Portability:</strong> Unlike algorithmic knowledge, engineers cannot easily take petabytes of proprietary, curated internal data or access to tens of thousands of specialized accelerators (like TPUs or GPUs) with them when they change jobs.</p></li><li><p><strong>High Barrier to Entry:</strong> Building world-class compute infrastructure (data centers, custom silicon, high-speed interconnects) and accumulating diverse, high-quality datasets at the scale required represents billions of dollars in capital expenditure and years, often decades, of cumulative effort and investment. This is not something startups or even well-funded competitors can easily replicate overnight.</p></li><li><p><strong>Synergistic Flywheels:</strong> Access to vast compute allows for more ambitious experiments and training larger models. These improved models, when deployed, can generate new, valuable interaction data, which feeds back into further model improvements, creating a virtuous cycle that is difficult for competitors with lesser resources to match.</p></li></ol><p><strong>Gemini 2.5 Pro: A Glimpse of the Advantage</strong></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!kvQw!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbc4a2e5a-b1ad-497c-8f16-5702bb3c315b_1466x1352.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!kvQw!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbc4a2e5a-b1ad-497c-8f16-5702bb3c315b_1466x1352.jpeg 424w, https://substackcdn.com/image/fetch/$s_!kvQw!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbc4a2e5a-b1ad-497c-8f16-5702bb3c315b_1466x1352.jpeg 848w, https://substackcdn.com/image/fetch/$s_!kvQw!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbc4a2e5a-b1ad-497c-8f16-5702bb3c315b_1466x1352.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!kvQw!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbc4a2e5a-b1ad-497c-8f16-5702bb3c315b_1466x1352.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!kvQw!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbc4a2e5a-b1ad-497c-8f16-5702bb3c315b_1466x1352.jpeg" width="1456" height="1343" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/bc4a2e5a-b1ad-497c-8f16-5702bb3c315b_1466x1352.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1343,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;r/singularity - Gemini 2.5 Pro benchmarks released&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="r/singularity - Gemini 2.5 Pro benchmarks released" title="r/singularity - Gemini 2.5 Pro benchmarks released" srcset="https://substackcdn.com/image/fetch/$s_!kvQw!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbc4a2e5a-b1ad-497c-8f16-5702bb3c315b_1466x1352.jpeg 424w, https://substackcdn.com/image/fetch/$s_!kvQw!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbc4a2e5a-b1ad-497c-8f16-5702bb3c315b_1466x1352.jpeg 848w, https://substackcdn.com/image/fetch/$s_!kvQw!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbc4a2e5a-b1ad-497c-8f16-5702bb3c315b_1466x1352.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!kvQw!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbc4a2e5a-b1ad-497c-8f16-5702bb3c315b_1466x1352.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Gemini 2.5 Pro Experimental, recently released by Google, offers a glimpse into how these interacting factors of <strong>data </strong>and <strong>compute</strong> will lead to a durable advantage in Google&#8217;s AI model performance. Despite OpenAI and DeepSeek releasing highly performant thinking models months in advance of Google (representing a large lead in algorithmic innovations), Gemini 2.5 Pro has managed to score #1 across the board in Chatbot Arena and across a wide range of benchmarks.</p><p>While Google describes Gemini 2.5 Pro partly through algorithmic concepts like "thinking models," the sheer breadth and depth of its capabilities, validated by both benchmarks and human preference, strongly suggest that these algorithms are being scaled and refined using computational resources and data diversity that few, if any, competitors can match. The "significantly enhanced base model" (as described by Google) is almost certainly a product of larger parameter counts trained for longer durations on more diverse data, enabled by Google's vertical integration of hardware (TPUs) and software within their hyper-scale data centers.</p><h3>Google's Unassailable Advantage</h3><p>This brings us to Google. When assessing data and compute advantages, Google stands in a league of its own.</p><p><strong>1. Data Dominance:</strong></p><ul><li><p><strong>Breadth and Modality:</strong> Google possesses arguably the most diverse and extensive collection of multimodal data on the planet. Consider the sources:</p><ul><li><p><strong>Google Search:</strong> Billions of daily queries provide unparalleled insight into human intent, language variation, and real-time information needs (text, images, implicit semantics).</p></li><li><p><strong>YouTube:</strong> The world's largest video platform offers vast amounts of video, audio, transcripts, comments, and multilingual content &#8211; crucial for multimodal understanding.</p></li><li><p><strong>Android:</strong> Interaction data from billions of devices provides insights into user behavior, application usage, and sensor inputs (potentially anonymized and aggregated).</p></li><li><p><strong>Google Maps:</strong> Geospatial data, satellite imagery, Street View imagery, reviews, and real-time traffic information.</p></li><li><p><strong>Gmail, Docs, Workspace:</strong> While respecting user privacy is paramount, Google potentially has access (for internal R&amp;D, aggregated/anonymized analysis, or opt-in features) to colossal amounts of text, code, and collaborative data reflecting professional and personal communication patterns.</p></li><li><p><strong>Google Books:</strong> A massive corpus of digitized text spanning centuries.</p></li><li><p><strong>Chrome:</strong> Web interaction data (aggregated and anonymized) reflecting how users navigate and consume information online.</p></li></ul></li><li><p><strong>Scale and Freshness:</strong> The sheer volume is staggering, but equally important is the constant influx of <em>new</em> data, keeping datasets fresh and reflecting current events, language evolution, and emerging trends. This continuous stream is vital for maintaining model relevance and accuracy.</p></li></ul><p><strong>2. Compute Superiority:</strong></p><ul><li><p><strong>Custom Silicon (TPUs):</strong> Google made a strategic bet on custom AI accelerators years ago with its Tensor Processing Units (TPUs). Now in their 7th generation, TPUs are designed specifically for large-scale ML training and inference, offering potentially significant advantages in performance-per-watt and performance-per-dollar <em>for Google's specific workloads and scale</em> compared to general-purpose GPUs. This vertical integration allows hardware and software co-design for optimal efficiency.</p></li><li><p><strong>Infrastructure Mastery:</strong> Google operates some of the world's most sophisticated and efficient data centers. Decades of experience in distributed systems (MapReduce, Borg/Kubernetes, Spanner) translate into an unparalleled ability to orchestrate and execute massively parallel training jobs reliably and efficiently across thousands of accelerators. This isn't just about owning chips; it's about the networking fabric, power delivery, cooling, and system software that make large-scale training feasible.</p></li><li><p><strong>Capital Investment:</strong> Google has the financial resources to sustain and expand this infrastructure lead, continuously investing billions in data centers and next-generation TPUs.</p></li></ul><h3>Conclusion: The Inevitable Frontrunner?</h3><p>While the AI race is far from over, and competitors like OpenAI and Anthropic continue to innovate, the fundamental dynamics favor players with entrenched advantages in data and compute. Algorithmic breakthroughs will continue to happen across the ecosystem, but they diffuse quickly. The ability to <em>scale</em> these algorithms using proprietary data and custom-built, hyper-scale infrastructure is the real differentiator.</p><p>Google's unparalleled data ecosystem, harvested across its diverse product portfolio, combined with its long-term investment in custom TPUs and mastery of planetary-scale computing, creates a formidable moat. Gemini 2.5 Pro is likely just an early indicator of what this integrated advantage can produce. As the demands for data and compute continue to escalate on the path to more capable AI, Google's lead in these foundational resources positions it strongly to outpace the competition and ultimately define the next era of artificial intelligence.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.chrishayduk.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Musings by Chris Hayduk! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[The Foundation Model Trap]]></title><description><![CDATA[Why AI Model Companies Are More Like Airlines than Like Cereal Companies]]></description><link>https://www.chrishayduk.com/p/the-foundation-model-trap</link><guid isPermaLink="false">https://www.chrishayduk.com/p/the-foundation-model-trap</guid><dc:creator><![CDATA[Chris Hayduk]]></dc:creator><pubDate>Wed, 05 Mar 2025 20:14:07 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!JK_8!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F240b429a-11df-478c-be6d-88ea2e904cc3_1024x768.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!JK_8!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F240b429a-11df-478c-be6d-88ea2e904cc3_1024x768.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!JK_8!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F240b429a-11df-478c-be6d-88ea2e904cc3_1024x768.jpeg 424w, https://substackcdn.com/image/fetch/$s_!JK_8!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F240b429a-11df-478c-be6d-88ea2e904cc3_1024x768.jpeg 848w, https://substackcdn.com/image/fetch/$s_!JK_8!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F240b429a-11df-478c-be6d-88ea2e904cc3_1024x768.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!JK_8!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F240b429a-11df-478c-be6d-88ea2e904cc3_1024x768.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!JK_8!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F240b429a-11df-478c-be6d-88ea2e904cc3_1024x768.jpeg" width="1024" height="768" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/240b429a-11df-478c-be6d-88ea2e904cc3_1024x768.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:768,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:214162,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.chrishayduk.com/i/158463759?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F240b429a-11df-478c-be6d-88ea2e904cc3_1024x768.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!JK_8!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F240b429a-11df-478c-be6d-88ea2e904cc3_1024x768.jpeg 424w, https://substackcdn.com/image/fetch/$s_!JK_8!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F240b429a-11df-478c-be6d-88ea2e904cc3_1024x768.jpeg 848w, https://substackcdn.com/image/fetch/$s_!JK_8!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F240b429a-11df-478c-be6d-88ea2e904cc3_1024x768.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!JK_8!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F240b429a-11df-478c-be6d-88ea2e904cc3_1024x768.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Harvey Sawikin recently wrote a great article analyzing the AI industry through a very Munger-like lens: will AI turn out more like the cereal industry (where there are many competitors with very healthy profit margins) or more like the airline industry (where competition compresses profit margins to near 0). </p><p>This idea has major implications for the major AI labs training foundation models today, such as OpenAI, Anthropic, and xAI. In this article I'll attempt to flesh out my understanding of this cereal vs. airline distinction and discuss why the airline scenario is more likely for the foundation model providers.</p><p>Before we dive in, you can find Harvey&#8217;s article below. I highly recommend giving it a read before continuing here.</p><div class="embedded-post-wrap" data-attrs="{&quot;id&quot;:156375904,&quot;url&quot;:&quot;https://harveysawikin.substack.com/p/ai-companies-cereals-or-airlines&quot;,&quot;publication_id&quot;:2339794,&quot;publication_name&quot;:&quot;Harvey&#8217;s Substack&quot;,&quot;publication_logo_url&quot;:&quot;https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbbce7176-1f83-43cd-8bbf-67a1bbfb334f_429x429.png&quot;,&quot;title&quot;:&quot;AI Companies: Cereals or Airlines?&quot;,&quot;truncated_body_text&quot;:&quot;In the post The Munger Games, inspired by my first-ever attendance at the Berkshire annual meeting and purchase and reading of Poor Charlie&#8217;s Almanack, I promised more commentary on Charlie Munger&#8217;s book once I&#8217;d reflected on it. One insight that stuck with me has come to the fore lately as I&#8217;ve tried to get my head around AI &#8211; an effort that isn&#8217;t theo&#8230;&quot;,&quot;date&quot;:&quot;2025-02-03T13:46:19.862Z&quot;,&quot;like_count&quot;:6,&quot;comment_count&quot;:3,&quot;bylines&quot;:[{&quot;id&quot;:32105441,&quot;name&quot;:&quot;Harvey Sawikin&quot;,&quot;handle&quot;:&quot;harveysawikin&quot;,&quot;previous_name&quot;:null,&quot;photo_url&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/65baf26e-5c5c-4a55-9377-22234d39af84_429x600.jpeg&quot;,&quot;bio&quot;:&quot;Fund manager, ex-lawyer, ex-novelist, art collector, writing about investing, marriage, culture, and random topics.&quot;,&quot;profile_set_up_at&quot;:&quot;2023-02-04T11:25:24.829Z&quot;,&quot;publicationUsers&quot;:[{&quot;id&quot;:2360843,&quot;user_id&quot;:32105441,&quot;publication_id&quot;:2339794,&quot;role&quot;:&quot;admin&quot;,&quot;public&quot;:true,&quot;is_primary&quot;:false,&quot;publication&quot;:{&quot;id&quot;:2339794,&quot;name&quot;:&quot;Harvey&#8217;s Substack&quot;,&quot;subdomain&quot;:&quot;harveysawikin&quot;,&quot;custom_domain&quot;:null,&quot;custom_domain_optional&quot;:false,&quot;hero_text&quot;:&quot;Welcome to my Substack, where I will write about investing (especially value investing), marriage and family, culture, and random topics. All subscription money will be donated to The Human Fund (sorry, that's for Seinfeld fans). Really, to a good cause.&quot;,&quot;logo_url&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/bbce7176-1f83-43cd-8bbf-67a1bbfb334f_429x429.png&quot;,&quot;author_id&quot;:32105441,&quot;theme_var_background_pop&quot;:&quot;#D10000&quot;,&quot;created_at&quot;:&quot;2024-02-10T21:01:10.289Z&quot;,&quot;email_from_name&quot;:null,&quot;copyright&quot;:&quot;Harvey Sawikin&quot;,&quot;founding_plan_name&quot;:null,&quot;community_enabled&quot;:true,&quot;invite_only&quot;:false,&quot;payments_state&quot;:&quot;enabled&quot;,&quot;language&quot;:null,&quot;explicit&quot;:false,&quot;is_personal_mode&quot;:false}}],&quot;is_guest&quot;:false,&quot;bestseller_tier&quot;:null}],&quot;utm_campaign&quot;:null,&quot;belowTheFold&quot;:false,&quot;type&quot;:&quot;newsletter&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="EmbeddedPostToDOM"><a class="embedded-post" native="true" href="https://harveysawikin.substack.com/p/ai-companies-cereals-or-airlines?utm_source=substack&amp;utm_campaign=post_embed&amp;utm_medium=web"><div class="embedded-post-header"><img class="embedded-post-publication-logo" src="https://substackcdn.com/image/fetch/$s_!UkOl!,w_56,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbbce7176-1f83-43cd-8bbf-67a1bbfb334f_429x429.png"><span class="embedded-post-publication-name">Harvey&#8217;s Substack</span></div><div class="embedded-post-title-wrapper"><div class="embedded-post-title">AI Companies: Cereals or Airlines?</div></div><div class="embedded-post-body">In the post The Munger Games, inspired by my first-ever attendance at the Berkshire annual meeting and purchase and reading of Poor Charlie&#8217;s Almanack, I promised more commentary on Charlie Munger&#8217;s book once I&#8217;d reflected on it. One insight that stuck with me has come to the fore lately as I&#8217;ve tried to get my head around AI &#8211; an effort that isn&#8217;t theo&#8230;</div><div class="embedded-post-cta-wrapper"><span class="embedded-post-cta">Read more</span></div><div class="embedded-post-meta">a year ago &#183; 6 likes &#183; 3 comments &#183; Harvey Sawikin</div></a></div><p>Okay, so let&#8217;s start with differentiating <em>why</em> cereals allow for competition with healthy profit margins, whereas airlines are a rough business for all involved.</p><p>Cereals have different flavors, so consumer preferences for a certain flavor can cause some degree of demand inelasticity. From a firm perspective, rather than chasing the flavor and profit margins of another firm's cereal, the more profitable long-term strategy is to specialize in a different flavor and reap your own healthy profit margins.</p><p>By contrast, the main service that airlines provide is transporting you from Point A to Point B. There isn't really an "experience" to speak of that differentiates airlines from one another (particularly for non-business class flyers), so the calculus for a consumer then comes down to only two factors: speed and cost. Speed can be achieved through two means: faster planes (which hasn't happened in decades) and more direct flights. Airlines are incentivized to provide direct flights between major cities/transit hubs because, if they did not, then any travelers going between major hubs (say, NYC and London) would choose the other airlines which did have those direct flights. Thus, we can assume that most major airlines will have direct flights between most major transit hubs/cities within a certian distance of each other. Hence, for any two airlines that have a direct flight between a fixed pair of cities, the <em>only</em> way to compete is on price, since this will be the <em>only</em> criterion differentiating the airlines for consumers. Thus, a small difference in price will lead to nearly all consumers choosing the cheaper options. This inherently <em>must</em> drive down profit margins as airlines seek to charge the lowest possible price while still maintaining profitability.</p><p>Translating this argument to AI, we see two potential paths forward:</p><ol><li><p><strong>Cereal Mode:</strong> We know that the data input to a model during its training process basically determines its behavior on the other end - what it acts like, what tasks it's good at, etc. Access to different types of data may thus give rise to different "flavors" of AI models, providing varying skill profiles and personalities. In this scenario, we could imagine that OpenAI provides the best chat experience (due to its large dataset of user chats), while Grok might provide the best news aggregation and summarization (due to its up-to-the-second Twitter data). This may provide enough distinction to allow each AI company to charge healthy profits margins on their respective foundation models.</p></li><li><p><strong>Airline Mode:</strong> In this case, maybe the data on the margins provided by chat interactions, Twitter, etc. doesn't move the needle much in terms of model behavior and capabilities. Perhaps the web-scale pretraining data drowns out the idosyncracies across each AI lab's datasets, leaving each lab's state-of-the-art AI models performing roughly identically. In this case, the only way to compete would be on the API pricing, with consumers rapidly moving to the cheapest option available that can perform the given task.</p></li></ol><p>Based on trends of the last few months, I think Airline Mode is looking more and more likely. The <a href="https://lmarena.ai/leaderboard">Chatbot Arena leaderboard</a> shows that all of the leading models from the main labs perform roughly similarly to each other (Grok 3 and GPT-4.5 are even currently within 1 Elo point of each other as of this writing!). And DeepSeek was able to reproduce OpenAI o1 in the span of a couple months (R1's Elo is actually 11 points higher than o1's). We're seeing <em>more</em> convergence between models over the last couple of years, not less. </p><p>Given that, unless a lab gets to AGI and the idea of recursive self-improvement leading to a permanent advantage turns out to be true, I don't see how foundation model training can provide durable, healthy profit margins without a significant change in business model for these companies.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.chrishayduk.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Musings by Chris Hayduk! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Understanding DeepSeek Part II: DeepSeek-V2]]></title><description><![CDATA[Compressing the key-value matrix]]></description><link>https://www.chrishayduk.com/p/understanding-deepseek-part-i-deepseek</link><guid isPermaLink="false">https://www.chrishayduk.com/p/understanding-deepseek-part-i-deepseek</guid><dc:creator><![CDATA[Chris Hayduk]]></dc:creator><pubDate>Wed, 05 Mar 2025 13:31:45 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!bv9y!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe7327a5e-c128-42a0-8f23-04b24d8b8fc1_1043x890.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h1>Summary</h1><p>DeepSeek-V2, released in June 2024, built off the success of DeepSeek's previous papers to set a new standard for training and inference efficiency. The core changes made to DeepSeek-V2 that set it apart from prior open source models occur in two core components of the transformer architecture: the attention block and the feed-forward network (see below image).</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!bv9y!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe7327a5e-c128-42a0-8f23-04b24d8b8fc1_1043x890.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!bv9y!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe7327a5e-c128-42a0-8f23-04b24d8b8fc1_1043x890.png 424w, https://substackcdn.com/image/fetch/$s_!bv9y!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe7327a5e-c128-42a0-8f23-04b24d8b8fc1_1043x890.png 848w, https://substackcdn.com/image/fetch/$s_!bv9y!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe7327a5e-c128-42a0-8f23-04b24d8b8fc1_1043x890.png 1272w, https://substackcdn.com/image/fetch/$s_!bv9y!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe7327a5e-c128-42a0-8f23-04b24d8b8fc1_1043x890.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!bv9y!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe7327a5e-c128-42a0-8f23-04b24d8b8fc1_1043x890.png" width="1043" height="890" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e7327a5e-c128-42a0-8f23-04b24d8b8fc1_1043x890.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:890,&quot;width&quot;:1043,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:179138,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.chrishayduk.com/i/158419950?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe7327a5e-c128-42a0-8f23-04b24d8b8fc1_1043x890.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!bv9y!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe7327a5e-c128-42a0-8f23-04b24d8b8fc1_1043x890.png 424w, https://substackcdn.com/image/fetch/$s_!bv9y!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe7327a5e-c128-42a0-8f23-04b24d8b8fc1_1043x890.png 848w, https://substackcdn.com/image/fetch/$s_!bv9y!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe7327a5e-c128-42a0-8f23-04b24d8b8fc1_1043x890.png 1272w, https://substackcdn.com/image/fetch/$s_!bv9y!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe7327a5e-c128-42a0-8f23-04b24d8b8fc1_1043x890.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The two key changes can be summarized as follows:</p><p>1.  <strong>Feed-Forward Network Optimization:</strong> DeepSeekMoE architecture</p><blockquote><p>Mixture of experts (MoE) layers are a drop-in replacement for the feed-forward layer in the standard transformer architecture. Prior to DeepSeekMoE, most MoE architectures functioned by splitting the feed-forward layer into several large feed-forward layers. Each input token would then "choose" 1 or 2 of these parallel feed-forward layers, also known as "experts", for its own computation. This architecture had one key problem - namely, each expert needed to learn large amounts of redundant information, since processing any token on any topic requires understanding of grammar, semantics, etc. DeepSeek solved this redundancy problem, thereby greatly increased the learning efficiency of the MoE architecture, through three key innovations. These included: more numerous, finer-grained experts; separating experts into shared and routing experts; and load balancing tokens across experts and devices. For more details on these innovations, see the previous blog post in the series.</p></blockquote><p>2. <strong>Attention Layer Optimization:</strong> Multi-head Latent Attention (MLA) </p><blockquote><p>Multi-head attention, described in detail in my other post, utilizes three matrices to produce new representations of the input tokens: the Query, Key, and Value matrices. Each of these matrices has dimension n x d, where n is the maximum length of the sequence and d is the dimension of the vector representing each token in the sequence. Standard transformers cache the Key and Value matrices for every layer fully in-memory at inference time, improving speed but resulting in large memory overhead. DeepSeek-V2's solution is to compress the Key and Value matrices at each layer into a single latent vector. At inference time, only this vector needs to be cached, substantially reducing memory requirements.</p></blockquote><p>Since we already described the DeepSeekMoE architecture in detail in the previous blog post of this series, this post will focus primarily on multi-head latent attention. We'll start by describing the problem it aims to solve, then move on to describing the intuition behind MLA's solution, and finally dive into the concrete math describing the method. We'll then end this post by discussing the effects that the combination of DeepSeekMoE and multi-head latent attention has on training and inference efficiency. Let's dive in!</p><div><hr></div><p><strong>Note: </strong>This post is part of the &#8220;Understanding DeepSeek&#8221; Series:</p><ol><li><p><a href="https://www.chrishayduk.com/p/understanding-deepseek-part-i-deepseekmoe">Understanding DeepSeek Part I: DeepSeekMoE</a></p></li><li><p>[This article] Understanding DeepSeek Part II: DeepSeek-V2</p></li><li><p>[Upcoming] Understanding DeepSeek Part III: DeepSeekMath</p></li><li><p>[Upcoming] Understanding DeepSeek Part IV: DeepSeek-Prover-V1.5</p></li><li><p>[Upcoming] Understanding DeepSeek Part V: DeepSeek-V3</p></li><li><p>[Upcoming] Understanding DeepSeek Part VI: DeepSeek-R1</p></li><li><p>[Upcoming] Understanding DeepSeek Part VII: Implications for the AI Industry and the World</p></li></ol><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.chrishayduk.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Musings by Chris Hayduk! Subscribe to make sure you don&#8217;t miss any new posts in this seris!</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><div><hr></div><h1>The Memory Efficiency Problem with Standard Multi-Head Attention</h1><p>Standard multi-head attention, at its core, solves the problem of deciding how to update our understanding of one concept, given a set of other, potentially-related concepts. In the case of language modeling, we want to update our understanding of a particular token using the understanding of the other tokens present in the sequence. To accomplish this, at each attention layer in a transformer, the model learns to parametrize three key matrices: the Query, Key, and Value matrices. These three matrices work together to identify the most relevant portions of the sequence for each token, and then to update each token's representation based on the relevant portions that were found. I won't cover the full details of how this is done here, but you can reference <a href="https://www.chrishayduk.com/p/a-primer-on-multi-head-causal-self">my other blog post</a> for more information.</p><p>Now, when language models are producing output at inference, we essentially need to place the transformer in a while loop. Until the transformer outputs an &#8220;End of Sequence&#8221; token, we&#8217;ll feed the input sequence into the transformer to produce the next token. Then, appending that next token to the input sequence, we&#8217;ll feed the newly elongated input sequence back into the transformer, repeating the process. </p><p>The key insight that enables caching here is the following: since modern LLMs are causal, meaning future tokens cannot influence previous tokens, by adding a new token to the end of the input sequence, we do not change the representation of any of the previous tokens. Hence, we do not need to recompute the hidden representations for the previous tokens, since these will be identical. </p><p>The only token for which we need to compute a new representation is the next token in the sequence (that is, the one token that doesn&#8217;t exist yet)! Another core insight coming from this observation is that we only need the key and value vectors for each previous token to compute the new token&#8217;s representation. Since the previous tokens&#8217; representations do not change, we don&#8217;t need to use the other tokens as "queries&#8221; to update their representations. However, we do need their key and value vectors so that we can &#8220;query&#8221; these vectors with the new token's query vector. </p><p>The above observations then give us a road map for caching values in the transformer in order to limit the number of computations we perform and speed up inference time. In particular, we must cache the Key and Value matrices at each hidden layer so that we can use these to compute the hidden representation for the new token. </p><p>Now let&#8217;s compute the memory requirements to store these cached values for Llama 3.3 70B, a state-of-the-art open source model at the time of writing.  (In practice, Llama 3.3 uses Grouped-Query Attention, which actually reduces caching requirements. For the sake of simplicity, we'll assume it uses standard attention here.)</p><p>Llama 3.3 has 80 attention layers. Each key and value vector in these attention layers has a dimension of 8192. And Llama 3.3 has a maximum context length of 128,000 tokens. </p><p>If Llama 3.3 is used in the default floating point 16 (FP16) mode, then each stored number will take up 2 bytes (16 bits). Hence, a single vector consisting of 8192 floating-point numbers will take up 16,384 bytes, or equivalently 16.384 kilobytes. For each cached token in our input, we need to store both a key vector *and* a value vector at each layer. Hence, at every layer, a cached token will require two vectors, totaling 32.768 KB in memory. Since there are 80 such layers, the cost to cache one token is thus 80 * 32.768 KB = 2621.44 KB (equivalently, 2.62 MB).</p><p>Now, suppose our input is 10,000 tokens long and we are producing the next token in the sequence. To cache the necessary data for the previous tokens, we need 10,000 * 2.62 MB = 26,200 MB (equivalently, 26.2 GB). </p><p>If our input uses the full Llama 3.3 context length of 128,000 tokens, the required space is 128,000 * 2.62 MB = 335,360 MB (equivalently, 335.36 GB). </p><p>As can be seen by the above example, memory requirements for the cache expand quickly as the input length increases. This makes it incredibly difficult to serve models with long context windows. In order to solve this problem with the standard transformer architecture, DeepSeek introduced Multi-head Latent Attention (MLA). </p><h1>Multi-head Latent Attention (MLA)</h1><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!BuXX!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F71b33121-208e-406d-82a4-faf03be131b4_1530x677.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!BuXX!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F71b33121-208e-406d-82a4-faf03be131b4_1530x677.png 424w, https://substackcdn.com/image/fetch/$s_!BuXX!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F71b33121-208e-406d-82a4-faf03be131b4_1530x677.png 848w, https://substackcdn.com/image/fetch/$s_!BuXX!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F71b33121-208e-406d-82a4-faf03be131b4_1530x677.png 1272w, https://substackcdn.com/image/fetch/$s_!BuXX!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F71b33121-208e-406d-82a4-faf03be131b4_1530x677.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!BuXX!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F71b33121-208e-406d-82a4-faf03be131b4_1530x677.png" width="1456" height="644" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/71b33121-208e-406d-82a4-faf03be131b4_1530x677.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:644,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:186879,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.chrishayduk.com/i/158419950?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F71b33121-208e-406d-82a4-faf03be131b4_1530x677.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!BuXX!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F71b33121-208e-406d-82a4-faf03be131b4_1530x677.png 424w, https://substackcdn.com/image/fetch/$s_!BuXX!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F71b33121-208e-406d-82a4-faf03be131b4_1530x677.png 848w, https://substackcdn.com/image/fetch/$s_!BuXX!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F71b33121-208e-406d-82a4-faf03be131b4_1530x677.png 1272w, https://substackcdn.com/image/fetch/$s_!BuXX!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F71b33121-208e-406d-82a4-faf03be131b4_1530x677.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>In order to overcome these memory efficiency issues, DeepSeek created the Multi-head Latent Attention layer. This layer modifies standard multi-head attention (depicted on the left side of the above image) by compressing the key and value matrices into a <em>single vector.</em> In practice, this looks like the following:</p><div class="latex-rendered" data-attrs="{&quot;persistentExpression&quot;:&quot;\\begin{align}\n    \\mathbf{c}^{KV}_t &amp;= W^{DKV} \\mathbf{h}_t, \\\\[8pt]\n    \\mathbf{k}^{C}_t &amp;= W^{UK} \\mathbf{c}^{KV}_t, \\\\[8pt]\n    \\mathbf{v}^{C}_t &amp;= W^{UV} \\mathbf{c}^{KV}_t, \\\\[8pt]\n    \\text{where} \\quad \n    \\mathbf{c}^{KV}_t &amp;\\text{ is the compressed latent vector for keys and values,}\\\\[5pt]\n\td_c \\; (&amp;<< d_n n_h) \\text{ denotes the KV compression dimension,} \\\\[5pt]\n    W_{DKV} &amp;\\in \\mathbb{R}^{d_c \\times d} \\text{ is the down-projection matrix} \\\\[5pt]\n    W^{UK}, &amp;W^{UV} \\in \\mathbb{R}^{d_h n_h \\times d_c} \\text{ are the up-projection matrices for the keys and values, respectively} \\\\[5pt]\n\\end{align}&quot;,&quot;id&quot;:&quot;BBBVCXUDSM&quot;}" data-component-name="LatexBlockToDOM"></div><p>That is, our model now must learn three additional matrices per layer - one down-projection matrix and two up-projection matrices. By learning these three matrices, we now no longer need to store the entire Key and Value matrices when caching previously-computed tokens. Instead, we can store the compressed latent vectors for each layer, where the compressed latent vector at a layer contains <em>all</em> of the information needed to produce the full Key &amp; Value matrices. </p><p>Thus, if we have L layers, we only need to store d_c * L values (d_c numbers in each latent vector and L latent vectors total, one per layer). </p><p>Let's take the example of Llama 3.3 that we illustrated above to see how much this gains us - previously, caching the full Key and Value matrices for the full 128,000 token context length of Llama 3.3 required 335.36 GB. Now, instead of caching the full matrices, let's imagine we've augmented Llama 3.3 to use MLA. DeepSeek sets the dimension of the latent vector to four times the hidden dimension, so we will use 32,768 as the dimension for our latent vector here. Hence, each vector takes up 0.06554 MB. Then, to cache one latent vector at each of Llama 3.3's 80 layers corresponds to using 80 * 0.06554 MB = 5.243 GB.</p><p>This is a <em>substantial</em> reduction from the initial requirement of 335.36 GB for standard attention, demonstrating the efficiency gains that can be driven using this approach.</p><h1>Training and Inference Efficiency</h1><p>DeepSeek-V2 introduces significant efficiency improvements in both training and inference compared to its predecessor, DeepSeek 67B, primarily through innovations in its architecture&#8212;especially the Multi-head Latent Attention (MLA). By compressing the Key and Value matrices into a single latent vector, MLA dramatically reduces memory consumption during inference. The reduction of the KV cache by approximately 93.3% translates directly into substantial gains in maximum generation throughput, allowing DeepSeek-V2 to achieve throughput levels up to 5.76 times greater than those observed in DeepSeek 67B. These optimizations enable DeepSeek-V2 to handle much longer contexts (up to 128K tokens) efficiently, positioning it as one of the most practical choices among large-scale language models for real-world applications where large-context inference is critical.</p><p>Additionally, the integration of DeepSeekMoE into the Feed-Forward Network layers synergizes well with MLA, enabling significant computational savings without sacrificing model performance. By activating only a fraction (21B) of its total parameters (236B), DeepSeek-V2 demonstrates economical training by saving 42.5% of training costs compared to dense models of similar scale. Thus, MLA plays a critical role not only in inference-time efficiency but also in making the pretraining phase more cost-effective.</p><h1>Results and Key Takeaways</h1><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ThVp!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F92d1cf29-86c1-4262-8f90-bcf584dc3273_1441x794.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ThVp!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F92d1cf29-86c1-4262-8f90-bcf584dc3273_1441x794.png 424w, https://substackcdn.com/image/fetch/$s_!ThVp!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F92d1cf29-86c1-4262-8f90-bcf584dc3273_1441x794.png 848w, https://substackcdn.com/image/fetch/$s_!ThVp!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F92d1cf29-86c1-4262-8f90-bcf584dc3273_1441x794.png 1272w, https://substackcdn.com/image/fetch/$s_!ThVp!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F92d1cf29-86c1-4262-8f90-bcf584dc3273_1441x794.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ThVp!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F92d1cf29-86c1-4262-8f90-bcf584dc3273_1441x794.png" width="1441" height="794" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/92d1cf29-86c1-4262-8f90-bcf584dc3273_1441x794.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:794,&quot;width&quot;:1441,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:193908,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.chrishayduk.com/i/158419950?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F92d1cf29-86c1-4262-8f90-bcf584dc3273_1441x794.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!ThVp!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F92d1cf29-86c1-4262-8f90-bcf584dc3273_1441x794.png 424w, https://substackcdn.com/image/fetch/$s_!ThVp!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F92d1cf29-86c1-4262-8f90-bcf584dc3273_1441x794.png 848w, https://substackcdn.com/image/fetch/$s_!ThVp!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F92d1cf29-86c1-4262-8f90-bcf584dc3273_1441x794.png 1272w, https://substackcdn.com/image/fetch/$s_!ThVp!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F92d1cf29-86c1-4262-8f90-bcf584dc3273_1441x794.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The innovative Multi-head Latent Attention layer significantly enhances the practical deployability of DeepSeek-V2. Compared to traditional Multi-Head Attention, MLA achieves superior inference performance while simultaneously overcoming the KV cache bottleneck. With its novel low-rank joint compression strategy, MLA significantly reduces inference memory overhead, making DeepSeek-V2 particularly suited for high-throughput, real-time applications requiring extensive context management.</p><p>Empirical evaluations on various benchmarks illustrate the clear strengths of DeepSeek-V2, even when compared against other leading open-source models of the time. Notably, DeepSeek-V2 consistently achieved top-tier performance on benchmarks such as MMLU, math reasoning tasks, and coding challenges, highlighting the architectural advantages introduced by MLA. Moreover, these enhancements enabled DeepSeek-V2 to be trained and served at a fraction of the cost of comparably performing dense models (see the above image).</p><p>All in all, Multi-head Latent Attention represented another significant milestone for DeepSeek on the path towards highly optimized training and inference that marked their revolution with DeepSeek-R1 and DeepSeek-V3. The next blog post in this series will dive into the new innovations introduced for DeepSeek-V3, building upon the foundations laid here and forming the base model used to train DeepSeek's state-of-the-art reasoning model.</p><div><hr></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.chrishayduk.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Musings by Chris Hayduk! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[A Primer on Multi-Head Causal Self-Attention]]></title><description><![CDATA[The neural network layer that kicked off the LLM craze]]></description><link>https://www.chrishayduk.com/p/a-primer-on-multi-head-causal-self</link><guid isPermaLink="false">https://www.chrishayduk.com/p/a-primer-on-multi-head-causal-self</guid><dc:creator><![CDATA[Chris Hayduk]]></dc:creator><pubDate>Sat, 01 Feb 2025 00:32:37 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/d3efecf9-3e49-410d-b69e-5cf3ceecc999_1026x1148.webp" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Lately, I've been writing quite a few series that center around the transformer architecture. For many of those blog posts, I struggle to decide whether I should include the background information necessary to understand attention (greatly increasing the length of the blog post) or I should assume the reader already knows this information (limiting the reach of my audience). Thus, this post is intended to be a compromise between the two positions, allowing me to link this post as background reading in any future blog post that requires knowledge of the nuts and bolts of the attention architecture.</p><p>This will be a "living" blog post, in that it will be edited and expanded upon as my own understanding of the architecture grows and deepens. If there are any radically large changes that I make, I will re-email the post out to subscribers for their review. Otherwise, feel free to check back periodically to see how the article has changed!</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.chrishayduk.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Subscribe to get notified of updates to this primer.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h1>The Basic Terminology of Multi-Head Causal Self-Attention</h1><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!HgXR!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F594a8d67-c26a-433c-91d0-2448c6d3de74_635x347.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!HgXR!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F594a8d67-c26a-433c-91d0-2448c6d3de74_635x347.png 424w, https://substackcdn.com/image/fetch/$s_!HgXR!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F594a8d67-c26a-433c-91d0-2448c6d3de74_635x347.png 848w, https://substackcdn.com/image/fetch/$s_!HgXR!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F594a8d67-c26a-433c-91d0-2448c6d3de74_635x347.png 1272w, https://substackcdn.com/image/fetch/$s_!HgXR!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F594a8d67-c26a-433c-91d0-2448c6d3de74_635x347.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!HgXR!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F594a8d67-c26a-433c-91d0-2448c6d3de74_635x347.png" width="635" height="347" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/594a8d67-c26a-433c-91d0-2448c6d3de74_635x347.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:347,&quot;width&quot;:635,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:40706,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!HgXR!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F594a8d67-c26a-433c-91d0-2448c6d3de74_635x347.png 424w, https://substackcdn.com/image/fetch/$s_!HgXR!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F594a8d67-c26a-433c-91d0-2448c6d3de74_635x347.png 848w, https://substackcdn.com/image/fetch/$s_!HgXR!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F594a8d67-c26a-433c-91d0-2448c6d3de74_635x347.png 1272w, https://substackcdn.com/image/fetch/$s_!HgXR!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F594a8d67-c26a-433c-91d0-2448c6d3de74_635x347.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The standard attention block used in first-generation LLMs like GPT-2 and GPT-3 is <strong>multi-head causal self-attention</strong>. </p><p>The goal of this variant of attention, like any attention variant, is to learn how to update a vector using other context vectors in order to accomplish some goal. In the case of language modeling, our vectors represent tokens, which you can think of as roughly analogous to words. The goal of these vector updates is to accurately predict the next word in the sentence. It is called <strong>causal</strong> because this type of attention ensures that each word can only update itself using previous words in the sentence - that is, it can't look ahead and update itself using words that haven't been written yet! It is called <strong>self-attention</strong> because the things that each word is paying attention to are the other words in the sentence. There is no outside data or context involved here. And finally, it is termed <strong>multi-head</strong> because, at each attention layer, we have multiple attention operations occurring in parallel. These parallel attention operators are referred to as "heads".</p><p>To produce the results of attention, each attention head takes as input a sequence of tokens represented as vectors. These vectors are passed through three feed-forward networks per head in parallel, projecting each token's vector into three new vectors. These new vectors are commonly referred to as the query, key, and value vectors. These query, key, and value vectors are then used to update the vector representations of the words in our sentence, improving the model's understanding of the concepts contained in the sentence.</p><p>Let's take a look at how this is done in practice.</p><h1>The Mathematics of Causal Self-Attention</h1><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!IwgS!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ff99c21-1666-4108-869a-102bb1bec947_925x739.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!IwgS!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ff99c21-1666-4108-869a-102bb1bec947_925x739.png 424w, https://substackcdn.com/image/fetch/$s_!IwgS!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ff99c21-1666-4108-869a-102bb1bec947_925x739.png 848w, https://substackcdn.com/image/fetch/$s_!IwgS!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ff99c21-1666-4108-869a-102bb1bec947_925x739.png 1272w, https://substackcdn.com/image/fetch/$s_!IwgS!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ff99c21-1666-4108-869a-102bb1bec947_925x739.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!IwgS!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ff99c21-1666-4108-869a-102bb1bec947_925x739.png" width="925" height="739" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8ff99c21-1666-4108-869a-102bb1bec947_925x739.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:739,&quot;width&quot;:925,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:82151,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!IwgS!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ff99c21-1666-4108-869a-102bb1bec947_925x739.png 424w, https://substackcdn.com/image/fetch/$s_!IwgS!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ff99c21-1666-4108-869a-102bb1bec947_925x739.png 848w, https://substackcdn.com/image/fetch/$s_!IwgS!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ff99c21-1666-4108-869a-102bb1bec947_925x739.png 1272w, https://substackcdn.com/image/fetch/$s_!IwgS!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ff99c21-1666-4108-869a-102bb1bec947_925x739.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>As mentioned, the attention block takes as input a sequence of tokens represented as vectors. Suppose the input sequence is given by:</p><div class="latex-rendered" data-attrs="{&quot;persistentExpression&quot;:&quot;X = \\{x_1, x_2, \\dots, x_n\\},&quot;,&quot;id&quot;:&quot;TROFIYZJWD&quot;}" data-component-name="LatexBlockToDOM"></div><p>where each x_i is the vector representation (embedding) of a token.</p><p>Each input vector x_i is simultaneously projected into three different spaces using learned linear transformations. That is, for every token x_i, we compute:</p><div class="latex-rendered" data-attrs="{&quot;persistentExpression&quot;:&quot;q_i = x_i W^Q,\\quad k_i = x_i W^K,\\quad v_i = x_i W^V,&quot;,&quot;id&quot;:&quot;RTYBEAFKPI&quot;}" data-component-name="LatexBlockToDOM"></div><p>where:</p><ul><li><p>W^Q, W^K, and W^V are the weight matrices for the query, key, and value projections, respectively.</p></li><li><p>The sets of all query, key, and value vectors are often denoted as Q,  K, and V.</p></li></ul><p>For a given token x_i, we will compute a similarity score with every token x_j such that x_j comes before it in the sentence (or is the token itself). This is done by taking the dot product of the query vector for x_i with the key vector for k_j. This result is then divided by the square root of the key vector's dimension. Mathematically, this is given by:</p><div class="latex-rendered" data-attrs="{&quot;persistentExpression&quot;:&quot;\\alpha_{ij} = \\frac{q_i \\cdot k_j}{\\sqrt{d_k}}&quot;,&quot;id&quot;:&quot;GXXLDMYNFI&quot;}" data-component-name="LatexBlockToDOM"></div><p>This value is precisely the unnormalized measure of how much token i should attend to token j. In linear algebra, the dot product of two vectors is just a scaled version of the cosine of the angle between them. An angle of 0 degrees gives a cosine value of 1, while an angle of 180 degrees gives a cosine value of -1. Hence, the closer the two vectors get to pointing in the same direction, the closer their dot product gets to 1, and the closer the two vectors get to pointing in opposite directions, the closer their dot product gets to -1. Intuitively, then, cosine (and by extension the dot product) has very desirable properties to use as a similarity function in the attention mechanism.</p><p>Armed with these similarity scores, we now have a way of measuring how "similar" two tokens in our sequence are. However, in order to use them to produce new vector embeddings, we're going to want to rescale them. The dot product between the query and key vectors could end up being quite large, and using this value directly can cause large changes in the scale of the vector representation for a given token. Moreover, knowing the score of a particular token pair (let's say between tokens x_i and x_j) tells us nothing about how important that pair is - importance is always relative, and what if the score for x_i and x_k is bigger?</p><p>Given the above discussion, we know we need to introduce some function that will re-scale our scores in such a way that we do not radically change the magnitude of the token's vector representation and that we can quickly determine how "important" each token pair is. A convenient differentiable function that does just this is softmax. The equation to produce the softmax output for token x_i is given below:</p><div class="latex-rendered" data-attrs="{&quot;persistentExpression&quot;:&quot;\\tilde{\\alpha}_{ij} = \\frac{\\exp(\\alpha_{ij})}{\\sum_{j=1}^i \\exp(\\alpha_{ij})}&quot;,&quot;id&quot;:&quot;KBGRNDGLBV&quot;}" data-component-name="LatexBlockToDOM"></div><p>The softmax function will take our scores for all possible pairs made with x_i (i.e. all key vectors we multiplied by x_i's query vector) and squash them into the range of 0 to 1. Moreover, it will ensure that these values sum to 1. Hence, we can view these outputs, referred to as attention weights, as probabilities or percentages. It can be useful to think of the attention weight for the pair x_i and x_j as the percent of x_i's attention that should be paid to x_j. </p><p>Once we have these attention weights, we can use them to produce a new vector representation for token x_i. We do this by taking a weighted sum of the value vectors for each token, where the weight is the attention weight. As mentioned above, we can think of this attention weight as the percent of attention that x_i pays to each vector that precedes it in the sequence. Mathematically, this looks like:</p><div class="latex-rendered" data-attrs="{&quot;persistentExpression&quot;:&quot;\\tilde{x}_i = \\sum_{j=1}^i \\tilde{\\alpha}_{ij} v_j&quot;,&quot;id&quot;:&quot;WPLELAQFMS&quot;}" data-component-name="LatexBlockToDOM"></div><p>This weighted sum integrates information from the tokens that x_i &#8220;attends&#8221; to, based on the learned attention weights.</p><p>In multi-head attention, these operations happen in parallel multiple times over. That is, we will produce multiple instances of the query, key, and value vectors for each token in the sequence. We will then use those unique instances to produce distinct updated vector representations for each token. If we have k heads, then we will produce k updated vectors for token i:</p><div class="latex-rendered" data-attrs="{&quot;persistentExpression&quot;:&quot;\\tilde{x}_{i1}, \\tilde{x}_{i2}, \\ldots \\tilde{x}_{ik}&quot;,&quot;id&quot;:&quot;PRSVCGFTHD&quot;}" data-component-name="LatexBlockToDOM"></div><p>These vectors get concatenated together, forming a single vector to represent the updated token i:</p><div class="latex-rendered" data-attrs="{&quot;persistentExpression&quot;:&quot;\\tilde{x}_i = [\\tilde{x}_{i1}; \\tilde{x}_{i2}; \\ldots; \\tilde{x}_{ik}]&quot;,&quot;id&quot;:&quot;HWIMGBKDRV&quot;}" data-component-name="LatexBlockToDOM"></div><p>The intuition behind using multiple heads to create the final updated representation for token i is that each head can learn to capture different aspects of language. One might learn grammatical structure, while another might learn vocabulary related to the legal profession. By splitting responsibilities between the attention heads, each can learn unique, non-redundant information.</p><p>This vector will then be passed through a linear projection layer, producing the final output of the multi-head attention layer.</p><p>Let's walk through a toy example now to make things concrete.</p><h1>A Toy Example: "the dog barks"</h1><p>Suppose our sentence is "the dog barks", and our tokenizer splits it into three tokens: "the", "dog", and "barks". Initially, these tokens are embedded into vectors:</p><div class="latex-rendered" data-attrs="{&quot;persistentExpression&quot;:&quot;x_{\\text{the}},\\quad x_{\\text{dog}},\\quad x_{\\text{barks}}.&quot;,&quot;id&quot;:&quot;VPTYOASSTS&quot;}" data-component-name="LatexBlockToDOM"></div><p>When entering the first attention block, each of these vectors is projected into three new vectors:</p><div class="latex-rendered" data-attrs="{&quot;persistentExpression&quot;:&quot;\\begin{aligned}\n\n\\text{For \&quot;the\&quot;:} &amp;\\quad q_{\\text{the}} = x_{\\text{the}} W^Q,\\quad k_{\\text{the}} = x_{\\text{the}} W^K,\\quad v_{\\text{the}} = x_{\\text{the}} W^V, \\\\\n\n\\text{For \&quot;dog\&quot;:} &amp;\\quad q_{\\text{dog}} = x_{\\text{dog}} W^Q,\\quad k_{\\text{dog}} = x_{\\text{dog}} W^K,\\quad v_{\\text{dog}} = x_{\\text{dog}} W^V, \\\\\n\n\\text{For \&quot;barks\&quot;:} &amp;\\quad q_{\\text{barks}} = x_{\\text{barks}} W^Q,\\quad k_{\\text{barks}} = x_{\\text{barks}} W^K,\\quad v_{\\text{barks}} = x_{\\text{barks}} W^V.\n\n\\end{aligned}&quot;,&quot;id&quot;:&quot;OBBBHLQCJW&quot;}" data-component-name="LatexBlockToDOM"></div><p>Thus, we transform 3 input vectors into 9 new vectors (3 each for queries, keys, and values).</p><h2>Updating the "dog" Token</h2><p>Let&#8217;s focus on updating the token "dog". In our example, "dog" corresponds to the second token x_2. To update its representation, we use its query vector and compute dot-product scores with the key vectors of "the" and "dog" (i.e., the tokens that precede or are the token itself):</p><div class="latex-rendered" data-attrs="{&quot;persistentExpression&quot;:&quot;\\begin{aligned}\n\n\\alpha_{\\text{dog},\\text{the}} &amp;= \\frac{q_{\\text{dog}} \\cdot k_{\\text{the}}}{\\sqrt{d_k}}, \\\\\n\n\\alpha_{\\text{dog},\\text{dog}} &amp;= \\frac{q_{\\text{dog}} \\cdot k_{\\text{dog}}}{\\sqrt{d_k}},\n\n\\end{aligned}&quot;,&quot;id&quot;:&quot;OIAYCICNJN&quot;}" data-component-name="LatexBlockToDOM"></div><p>where d_k is the dimensionality of the key vectors. The division by the square root of d_k is used to normalize the scores.</p><p>These scores are then passed through a softmax function to obtain attention weights (or probabilities) that indicate how much attention "dog" should pay to itself and to "the":</p><div class="latex-rendered" data-attrs="{&quot;persistentExpression&quot;:&quot;\\begin{aligned}\n\n\\tilde{\\alpha}_{\\text{dog},\\text{the}} &amp;= \\frac{\\exp\\left(\\alpha_{\\text{dog},\\text{the}}\\right)}{\\exp\\left(\\alpha_{\\text{dog},\\text{the}}\\right) + \\exp\\left(\\alpha_{\\text{dog},\\text{dog}}\\right)}, \\\\\n\n\\tilde{\\alpha}_{\\text{dog},\\text{dog}} &amp;= \\frac{\\exp\\left(\\alpha_{\\text{dog},\\text{dog}}\\right)}{\\exp\\left(\\alpha_{\\text{dog},\\text{the}}\\right) + \\exp\\left(\\alpha_{\\text{dog},\\text{dog}}\\right)}.\n\n\\end{aligned}&quot;,&quot;id&quot;:&quot;YRFTFEVZYQ&quot;}" data-component-name="LatexBlockToDOM"></div><p>Softmax ensures that these attention weights are between 0 and 1, and that they all sum to 1. Hence, they are valid probabilities and, to make things easier, you can of what percent of its attention the word "dog" should pay to the word "the" or to itself.</p><p>With the attention probabilities computed, we update the original vector representation of "dog" by taking a weighted sum of the corresponding value vectors. In this case, we combine the value vector of "the" and the value vector of "dog":</p><div class="latex-rendered" data-attrs="{&quot;persistentExpression&quot;:&quot;\\tilde{x}_{\\text{dog}} = \\tilde{\\alpha}_{\\text{dog},\\text{the}} \\, v_{\\text{the}} + \\tilde{\\alpha}_{\\text{dog},\\text{dog}} \\, v_{\\text{dog}}&quot;,&quot;id&quot;:&quot;EVNVACWIHW&quot;}" data-component-name="LatexBlockToDOM"></div><p>This new vector is an updated representation that incorporates contextual information from the preceding token "the" as well as from "dog" itself.</p><h2>Recap</h2><p>To recap, these are the major steps for updating the "dog" vector in our example using causal self-attention:</p><p>1. <strong>Input Embedding:</strong> </p><p>   Each token is embedded into a vector x_i.</p><p>2. <strong>Linear Projections:</strong>  </p><p>   Each x_i is projected into query, key, and value vectors:  </p><div class="latex-rendered" data-attrs="{&quot;persistentExpression&quot;:&quot;   q_i = x_i W^Q,\\quad k_i = x_i W^K,\\quad v_i = x_i W^V.&quot;,&quot;id&quot;:&quot;NCHGMXGHNI&quot;}" data-component-name="LatexBlockToDOM"></div><p>3. <strong>Score Calculation (for causal attention):</strong></p><p>   For token "dog" (second token), calculate:</p><p>   </p><div class="latex-rendered" data-attrs="{&quot;persistentExpression&quot;:&quot;   \\alpha_{\\text{dog}, j} = \\frac{q_{\\text{dog}} \\cdot k_j}{\\sqrt{d_k}}, \\quad \\text{for } j \\in \\{\\text{\&quot;the\&quot;}, \\text{\&quot;dog\&quot;}\\}&quot;,&quot;id&quot;:&quot;NXASKXVPVR&quot;}" data-component-name="LatexBlockToDOM"></div><p>4. <strong>Softmax to Obtain Weights:</strong></p><p>   Convert scores to probabilities:</p><div class="latex-rendered" data-attrs="{&quot;persistentExpression&quot;:&quot;\\tilde{\\alpha}_{\\text{dog}, j} = \\frac{\\exp\\left(\\alpha_{\\text{dog}, j}\\right)}{\\sum_{j'} \\exp\\left(\\alpha_{\\text{dog}, j'}\\right)}&quot;,&quot;id&quot;:&quot;BPJVISXYAX&quot;}" data-component-name="LatexBlockToDOM"></div><p>5. <strong>Contextual Update:</strong>  </p><p>   Update "dog" by a weighted sum of the value vectors:</p><div class="latex-rendered" data-attrs="{&quot;persistentExpression&quot;:&quot;\\tilde{x}_{\\text{dog}} = \\sum_{j \\in \\{\\text{\&quot;the\&quot;}, \\text{\&quot;dog\&quot;}\\}} \\tilde{\\alpha}_{\\text{dog}, j} \\, v_j&quot;,&quot;id&quot;:&quot;JLFZHRNNAS&quot;}" data-component-name="LatexBlockToDOM"></div><p>In the full multi-head attention mechanism, this process is performed in parallel over multiple "heads" (with different learned projections), and the results are concatenated and transformed further to form the final output of the attention block.</p><h1>Key Takeaways</h1><p>Let's now summarize the key points of what we've learned:</p><ol><li><p>Attention is a neural network mechanism used to update vectors using the context from other vectors</p></li><li><p>Input vectors to an attention layer are replaced by 3 intermediate vectors: the query, key, and value vectors</p></li><li><p>The query and key vectors work together to produce similarity scores between pairs of vectors. If we are updating token i and want to know how much token j should influence our input, we multiply the query vector of token i by the key vector of token j.</p></li><li><p>The scores produced by the query and key vectors can be turned into probabilities using the softmax function. These probabilities are used to measure how much token i should consider the tokens that came before it in the sequence when updating its vector representation</p></li><li><p>The new vector representation for token i is produced by multiplying the softmax probabilities by the value vectors for each corresponding token. These are then summed together</p></li><li><p>The above process occurs independently across several parallel attention operations, called heads. At the end of the attention block, the new vector representations for token i coming from each head are concatenated together and passed through a linear projection layer.</p></li><li><p>The above process (steps 1-6) is performed in parallel for the full sequence of input tokens.</p></li></ol><p>If you keep these 7 key points in mind while reading Arxiv papers (or my future blog posts!), you'll have a strong understanding of what multi-head causal self-attention is doing, where it faces limitations, and whether or not a given architectural change actually addresses those limitations.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.chrishayduk.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Musings by Chris Hayduk! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Understanding DeepSeek Part I: DeepSeekMoE]]></title><description><![CDATA[Mixture of experts models with a twist]]></description><link>https://www.chrishayduk.com/p/understanding-deepseek-part-i-deepseekmoe</link><guid isPermaLink="false">https://www.chrishayduk.com/p/understanding-deepseek-part-i-deepseekmoe</guid><dc:creator><![CDATA[Chris Hayduk]]></dc:creator><pubDate>Thu, 30 Jan 2025 03:28:29 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!RrxT!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80afde5c-c788-4a82-b1d9-4779b07b4a11_826x465.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!RrxT!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80afde5c-c788-4a82-b1d9-4779b07b4a11_826x465.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!RrxT!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80afde5c-c788-4a82-b1d9-4779b07b4a11_826x465.jpeg 424w, https://substackcdn.com/image/fetch/$s_!RrxT!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80afde5c-c788-4a82-b1d9-4779b07b4a11_826x465.jpeg 848w, https://substackcdn.com/image/fetch/$s_!RrxT!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80afde5c-c788-4a82-b1d9-4779b07b4a11_826x465.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!RrxT!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80afde5c-c788-4a82-b1d9-4779b07b4a11_826x465.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!RrxT!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80afde5c-c788-4a82-b1d9-4779b07b4a11_826x465.jpeg" width="826" height="465" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/80afde5c-c788-4a82-b1d9-4779b07b4a11_826x465.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:465,&quot;width&quot;:826,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;What is DeepSeek: China's open-source AI research lab which rivals OpenAI |  World News - Business Standard&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="What is DeepSeek: China's open-source AI research lab which rivals OpenAI |  World News - Business Standard" title="What is DeepSeek: China's open-source AI research lab which rivals OpenAI |  World News - Business Standard" srcset="https://substackcdn.com/image/fetch/$s_!RrxT!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80afde5c-c788-4a82-b1d9-4779b07b4a11_826x465.jpeg 424w, https://substackcdn.com/image/fetch/$s_!RrxT!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80afde5c-c788-4a82-b1d9-4779b07b4a11_826x465.jpeg 848w, https://substackcdn.com/image/fetch/$s_!RrxT!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80afde5c-c788-4a82-b1d9-4779b07b4a11_826x465.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!RrxT!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80afde5c-c788-4a82-b1d9-4779b07b4a11_826x465.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h1>Series Introduction</h1><p>Recently, the announcement of DeepSeek-R1 shook the AI world, as an open source project managed to match the performance of OpenAI's state-of-the-art API, o1, within months of its release. The market reacted vehemently to this news, with Nvidia's stock dropping 18% in a single day. AI researchers, engineers, and commentators alike took to Twitter/X to share their thoughts on DeepSeek-R1's implications for the AI industry and the United States, with many asserting that the age of American AI had come and gone in a flash, with China now firmly taking the lead.</p><p>But were these takes correct?</p><p>In order to dissect the true implications for the world going forward, we first need to understand DeepSeek-R1 on a fundamental level - what it is, what it does, how it works, and what key innovations it introduced. This blog post series will aim to arm you with that knowledge.</p><p>To do this effectively, we are going to start at the beginning of DeepSeek's major papers and work our way forward in time, tracing out the researchers' reasoning and how they arrived at the final design for DeepSeek-R1. This final design included two key components: </p><ol><li><p>An efficient mixture of experts language model base </p></li><li><p>Reinforcement learning-tuned chain of thought capabilities</p></li></ol><p>In this blog series, we will explore two separate but related series of papers in order to deeply understand the two key components of DeepSeek-R1. First, we will trace the evolution of the mixture of experts architecture from DeepSeek-MOE to DeepSeek-V3, their newest state-of-the-art language model. We will then turn our attention to reinforcement learning-tuned chain of thought, beginning with the seminal DeepSeekMath paper and working our way forward to the current AI darling - DeepSeek-R1.</p><p>With this strong foundational knowledge of the theoretical underpinnings of DeepSeek-R1, we will be able to separate the hype from the noise. In light of what we've learned from these paper deep dives, this blog series will conclude with an analysis of the implications of DeepSeek-R1 from several perspectives:</p><ol><li><p>Technological progress</p></li><li><p>AI market dynamics</p></li><li><p>Geopolitical risks</p></li></ol><p>By the end of this series, you will have a clear, evidence-based understanding of DeepSeek-R1&#8212;what makes it powerful, where it stands relative to its competitors, and what its long-term impact might be. As the AI landscape continues to shift at an unprecedented pace, cutting through speculation and focusing on the fundamentals will be key to making sense of the road ahead. Let&#8217;s dive in.</p><div><hr></div><p><strong>Note: </strong>This post is part of the &#8220;Understanding DeepSeek&#8221; Series:</p><ol><li><p>[This article] Understanding DeepSeek Part I: DeepSeekMoE</p></li><li><p><a href="https://www.chrishayduk.com/p/understanding-deepseek-part-i-deepseek">Understanding DeepSeek Part II: DeepSeek-V2</a></p></li><li><p>[Upcoming] Understanding DeepSeek Part III: DeepSeekMath</p></li><li><p>[Upcoming] Understanding DeepSeek Part IV: DeepSeek-Prover-V1.5</p></li><li><p>[Upcoming] Understanding DeepSeek Part V: DeepSeek-V3</p></li><li><p>[Upcoming] Understanding DeepSeek Part VI: DeepSeek-R1</p></li><li><p>[Upcoming] Understanding DeepSeek Part VII: Implications for the AI Industry and the World</p></li></ol><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.chrishayduk.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Musings by Chris Hayduk! Subscribe to make sure you don&#8217;t miss any new posts in this series!</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><div><hr></div><h1>Paper Summary</h1><p>Mixture-of-experts (MoE) models are an extension of the standard transformer architecture in which a collection of expert modules (typically feed-forward networks) each learn to specialize in different aspects of the data. For a given token or input, only a subset of these specialized experts is activated, allowing the model to dynamically focus its computation on the most relevant components. This selective activation enables MoE models to achieve a high effective capacity&#8212;since many different specialists are available&#8212;while maintaining computational efficiency, because only a limited number of experts actually process each input. As a result, MoE approaches excel at capturing diverse patterns, efficiently scaling model size, and flexibly adapting to a wide variety of tasks.</p><p>Standard mixture-of-experts models, used prior to DeepSeekMoE, typically rely on selecting the top <em>K</em> experts (often 1 or 2) out of <em>N</em> possible experts for each token in a sequence. While this approach does reduce computational load&#8212;since only a small fraction of experts are activated&#8212;it also forces those few activated experts to capture <em>all</em> aspects of the token, including common linguistic structure that is often duplicated across experts. Consequently, an enormous portion of each expert&#8217;s capacity is spent memorizing redundant information, leaving less room for true specialization.</p><p>DeepSeekMoE improves upon the standard MoE architecture, solving this redundancy problem by:</p><p>1. <strong>Using a larger number of smaller experts (Fine-Grained Expert Segmentation)</strong></p><p>Instead of a few large experts, DeepSeek splits capacity into many more experts, each of which is smaller in dimensionality. The model then increases the number of selected experts by the same factor, creating a dramatically larger space of potential expert combinations. Despite this combinatorial explosion, the overall parameter count and per-token activated parameters remain <em>exactly the same</em> as in a conventional MoE setup&#8212;meaning we gain richer representational capacity without paying extra in total parameter count or computational cost.</p><p>2. <strong>Separating Experts into Shared and Routing Experts</strong></p><p>DeepSeek also partitions its experts into two sets. The shared experts, which are <em>always activated</em> for every token, learn the broad &#8220;common knowledge&#8221; required by all inputs (e.g., syntax, high-level semantics). The routing experts, by contrast, are only activated if they are relevant to a specific token, allowing them to focus on niche or domain-specific information. This further decreases redundancy and promotes parameter efficiency: shared experts handle language &#8220;fundamentals,&#8221; while routing experts handle specialization.</p><p>3. <strong>Load Balancing Through Additional Loss Terms</strong></p><p>Finally, DeepSeek addresses load balancing in two senses. It enforces a roughly equal usage of each active routing expert across tokens&#8212;ensuring no single expert is under- or over-utilized&#8212;and distributes the experts themselves across multiple GPUs to avoid hardware bottlenecks. Both of these aims are achieved by incorporating new balancing terms into the training objective.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!7dNn!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc305bcca-7705-40e3-8ddf-2a278c9e73a0_1086x664.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!7dNn!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc305bcca-7705-40e3-8ddf-2a278c9e73a0_1086x664.png 424w, https://substackcdn.com/image/fetch/$s_!7dNn!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc305bcca-7705-40e3-8ddf-2a278c9e73a0_1086x664.png 848w, https://substackcdn.com/image/fetch/$s_!7dNn!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc305bcca-7705-40e3-8ddf-2a278c9e73a0_1086x664.png 1272w, https://substackcdn.com/image/fetch/$s_!7dNn!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc305bcca-7705-40e3-8ddf-2a278c9e73a0_1086x664.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!7dNn!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc305bcca-7705-40e3-8ddf-2a278c9e73a0_1086x664.png" width="1086" height="664" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c305bcca-7705-40e3-8ddf-2a278c9e73a0_1086x664.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:664,&quot;width&quot;:1086,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:145250,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!7dNn!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc305bcca-7705-40e3-8ddf-2a278c9e73a0_1086x664.png 424w, https://substackcdn.com/image/fetch/$s_!7dNn!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc305bcca-7705-40e3-8ddf-2a278c9e73a0_1086x664.png 848w, https://substackcdn.com/image/fetch/$s_!7dNn!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc305bcca-7705-40e3-8ddf-2a278c9e73a0_1086x664.png 1272w, https://substackcdn.com/image/fetch/$s_!7dNn!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc305bcca-7705-40e3-8ddf-2a278c9e73a0_1086x664.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Taken together, these modifications produce a model that is both parameter-efficient and highly flexible. By boosting expert variety, removing needless duplication, and balancing the workload across experts and devices, DeepSeekMoE provides a substantially more effective way to leverage MoE architectures&#8212;achieving greater specialization and capacity without increasing the overall parameter footprint.</p><p>Let's dive in deeper into these three optimizations now and see how they alter the standard MoE transformer architecture.</p><h1>Standard Mixture of Experts Models</h1><p>In standard MoE architecture, expert layers will typically replace the feed-forward layer that occurs after self-attention. Experts can be thought of as a set of <em>N</em> feed-forward layers that are structurally identical to the original feed-forward layer. Only a subset of these <em>N</em> possible feed-forward networks will be activated for any individual token, with many prior MoE architectures selecting 1 or 2 of these <em>N</em> possible networks for a given token. </p><p>Whether or not a network is activated is determined by taking the dot product of the output of the attention layer for that token (i.e., the hidden vector for token i) with the centroid of the current expert. We then take the softmax of this value to force it into the range of 0 to 1. You can think of this like an attention score computed over the experts instead of the tokens - we want to see which expert aligns most closely with the current token under consideration. These scores are computed for each expert, and then the experts are ranked according to this score. The top <em>K</em> (usually 1 or 2) experts are selected based on this ranking, and the token embeddings are then passed to those feed-forward expert networks. </p><p>The output of these experts is added together alongside the initial hidden state for the token (i.e., the token vector prior to the application of the experts). This produces the final output for the given layer.</p><p>The major obstacle with this approach is the following: since most prior MoE models only selected the top 1 or 2 experts for each token, the selected expert(s) must capture <em>everything</em> about a given token, including redundant information such as language structure. This wastes a large amount of the model's capacity to learn useful information, forcing the weights of each expert to memorize redundant information that is already captured by the other experts.</p><h1>Fine-Grained Expert Segmentation</h1><p>One of DeepSeek's solutions to the redundancy problem is to <strong>make experts smaller but more numerous</strong>. That is, the DeepSeekMoE approach reduces the dimensionality of each individual expert's feed-forward network (and therefore its computational cost and representational capacity) by a factor of 1/m compared to the network's standard feed-forward networks. Correspondingly, it increases the number of total experts by a factor of m <em>and</em> the number of selected experts by the same factor of m. This results in the same number of parameters for the model on net, but allows for substantially more variety when selecting the experts to use for a specific token.</p><p>We can see this increased variety when examining the combinatorics of the expert space. Suppose our standard feed-forward network has hidden dimension 4096, and our standard mixture of experts model uses 8 of these experts in total, with 2 selected for any given token. This results in the following number of possible expert combinations for each token in the standard mixture of experts model:</p><div class="latex-rendered" data-attrs="{&quot;persistentExpression&quot;:&quot;{8 \\choose 2} = 28 \\; \\text{possible expert combinations}&quot;,&quot;id&quot;:&quot;SBDXFQQTLB&quot;}" data-component-name="LatexBlockToDOM"></div><p></p><p>Now, using the DeepSeekMoE architecture, suppose we have m = 8. That is, we are going to increase our number of experts by a factor of 8 (and reduce the hidden dimension by a factor of 1/8). This gives us a hidden dimension of 512 per expert, with 64 total experts and 16 experts selected for any given token. This results in the following number of possible expert combinations for each token in the DeepSeekMoE version of the model:</p><div class="latex-rendered" data-attrs="{&quot;persistentExpression&quot;:&quot;{64 \\choose 16} \\approx 489,000,000,000,000 \\; \\text{possible expert combinations}&quot;,&quot;id&quot;:&quot;HALZCSCSEQ&quot;}" data-component-name="LatexBlockToDOM"></div><p></p><p>That is, we go from 28 possible expert combinations to nearly 489 trillion possible expert combinations! This allows for <em>significantly</em> more specialization across experts and much more variety in knowledge application on a token-by-token basis. Astonishingly, even with this huge increase in variety, the number of tokens stays exactly the same! The number of total parameters in each model is given by:</p><div class="latex-rendered" data-attrs="{&quot;persistentExpression&quot;:&quot;\\begin{align*}\n\n\\text{Original MoE model parameters} &amp;= 8 \\text{ experts} * 4096 \\text{ parameters per expert} = 32,768\\\\\n\n\\text{DeepSeekMoE model parameters} &amp;= 64 \\text{ experts} * 512 \\text{ parameters per expert} = 32,768\n\n\\end{align*}&quot;,&quot;id&quot;:&quot;YLEXVROBCV&quot;}" data-component-name="LatexBlockToDOM"></div><p></p><p>Similarly, the number of parameters activated for any given token is exactly the same:</p><div class="latex-rendered" data-attrs="{&quot;persistentExpression&quot;:&quot;\\begin{align*}\n\n\\text{Original MoE activated parameters} &amp;= 2 \\text{ activated experts} * 4096 \\text{ parameters per expert}\\\\ &amp;= 8192\\\\\n\n\\text{DeepSeekMoE activated parameters} &amp;= 16 \\text{ activated experts} * 512 \\text{ parameters per expert}\\\\ &amp;= 8192\n\n\\end{align*}&quot;,&quot;id&quot;:&quot;DWYNCNGFFC&quot;}" data-component-name="LatexBlockToDOM"></div><p></p><p>Hence, we get basically a free lunch here - significantly higher representational capacity in our model with the same number of parameters used!</p><h1>Shared Experts</h1><p>Another approach DeepSeek took to avoid capturing redundancy in its experts is to segment the expert population into two groups: shared experts and routing experts.</p><p>Shared experts are <strong>always activated</strong>, regardless of the input token. This incentivizes these expert modules to capture common knowledge relevant to all queries (e.g., language semantics). By contrast, routing experts are only activated if the token is relevant to the expert, as described in the "Standard Mixture of Expert Models" section. </p><p>That is, the initial <em>mN</em> experts are split into two groups: <em>K_s</em> shared experts and <em>K_r = mN - K_s</em> routing experts. <em>All</em> of the <em>K_s</em> shared experts are activated for all tokens, while a subset of the <em>K_r</em> are selected for each token. Mathematically, this looks like the following:</p><div class="latex-rendered" data-attrs="{&quot;persistentExpression&quot;:&quot;\\begin{align}\n    \\mathbf{h}^l_t &amp;= \\sum_{i=1}^{K_s} \\text{FFN}_i(\\mathbf{u}^l_t) \n    + \\sum_{i=K_s+1}^{mN} \\left( g_{i,t} \\cdot \\text{FFN}_i(\\mathbf{u}_t^l) \\right) \n    + \\mathbf{u}_t^l, \\\\[8pt]\n    \\text{where} \\quad \n    \\mathbf{h}_t^l &amp;\\text{ is the hidden vector output for the } t\\text{-th token at the } l\\text{-th layer,} \\\\[5pt]\n    \\text{FFN}_i &amp;\\text{ is the feed-forward network representing the } i\\text{-th expert,} \\\\[5pt]\n    K_s &amp;\\text{ is the number of shared experts,} \\\\[5pt]\n    mN &amp;\\text{ is the total number of experts,} \\\\[5pt]\n    \\mathbf{u}_t^l &amp;\\text{ is the output of the attention mechanism for token } t \\text{ at layer } l, \\\\[5pt]\n    g_{i,t} &amp;= \\begin{cases} s_{i,t}, &amp; s_{i,t} \\in \\text{Top}_k\\left( \\{s_{j,t} \\mid 1 \\leq j \\leq mN\\}, mK \\right) \\\\[5pt] 0, &amp; \\text{otherwise} \\end{cases} \\\\[8pt]\ns_{i,t} &amp;= \\text{Softmax}_i \\left( \\mathbf{u}_t^l{}^\\top \\mathbf{e}_i^l \\right),\n\\end{align}&quot;,&quot;id&quot;:&quot;NOEKVJEXTI&quot;}" data-component-name="LatexBlockToDOM"></div><p></p><p>Hence, we can see that the hidden vector output of token t at layer L <em>always</em> uses all of the shared experts (denoted by the first summation in the equation) and <em>always</em> includes the residual (denoted by the last term). The middle term, representing the routing experts, includes a gating factor that controls which experts are turned on for any specific token. In particular, the gating factor is the output of a softmax if the expert ranked in the top <em>mK</em> experts. Otherwise, it is 0. As a result, not only do we eliminate most of the possible experts (thereby greatly reducing the number of active parameters), we also weight the final output based on how <em>close</em> each chosen routing expert is to the token. In other words, the more a chosen routing expert "knows" about a topic, the more heavily we weight its opinion.</p><p>This setup allows the routing experts to ignore the redundant information captured by the shared experts and instead focus on learning concepts and information that are relevant to their area of specialization. This promotes parameter efficiency in the model, as each marginal parameter added to the routing experts will be encouraged through the learning process to acquire information that is distinct from the existing parameters.</p><h1>Load Balancing</h1><p>Now that we have a better-designed MoE network with fine-grained experts and expert sharing, there still remains one major challenge to ensure the parameters are used maximally - we need to load balance requests across the available experts. Essentially, our goal is to force each token to attend to the outputs of the <em>mK</em> chosen routing experts roughly equally. This makes certain that, when we activate routing expert parameters to process a particular token, all of the activated parameters are contributing meaningfully to the output. As a result, we maximize the utilization of the MoE architecture.</p><p>In addition to load balancing across experts, we would like to load balance across devices. Experts are typically stored on many separate GPUs, since these models are too large to fit in the memory of a single GPU. Given this fact, we would like the chosen experts for a token to be evenly spread across devices, thus preventing overloading of any single GPU.</p><p>These two goals are achieved by DeepSeekMoE through introducing two new terms to the loss function.</p><h1>Results and Key Takeaways</h1><p>With the above optimizations, DeepSeek was able to mitigate many of the most challenging problems facing MoE models. Together, fine-grained segmentation, shared experts, and load balancing work to maximize the amount of unique, useful information stored in a given set of parameters. As a result, DeepSeekMoE is able to outperform models with <em>fewer</em> active parameters. Below, we can see that DeepSeekMoE outperformed LLaMA2 7B (a dense model that does <em>not</em> use any experts) across a number of benchmarks with fewer than half of the active parameters. </p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ZFgF!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ed31d09-a4d3-4b60-aa09-ba6ce8ae5a68_991x918.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ZFgF!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ed31d09-a4d3-4b60-aa09-ba6ce8ae5a68_991x918.png 424w, https://substackcdn.com/image/fetch/$s_!ZFgF!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ed31d09-a4d3-4b60-aa09-ba6ce8ae5a68_991x918.png 848w, https://substackcdn.com/image/fetch/$s_!ZFgF!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ed31d09-a4d3-4b60-aa09-ba6ce8ae5a68_991x918.png 1272w, https://substackcdn.com/image/fetch/$s_!ZFgF!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ed31d09-a4d3-4b60-aa09-ba6ce8ae5a68_991x918.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ZFgF!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ed31d09-a4d3-4b60-aa09-ba6ce8ae5a68_991x918.png" width="991" height="918" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/7ed31d09-a4d3-4b60-aa09-ba6ce8ae5a68_991x918.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:918,&quot;width&quot;:991,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:209291,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!ZFgF!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ed31d09-a4d3-4b60-aa09-ba6ce8ae5a68_991x918.png 424w, https://substackcdn.com/image/fetch/$s_!ZFgF!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ed31d09-a4d3-4b60-aa09-ba6ce8ae5a68_991x918.png 848w, https://substackcdn.com/image/fetch/$s_!ZFgF!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ed31d09-a4d3-4b60-aa09-ba6ce8ae5a68_991x918.png 1272w, https://substackcdn.com/image/fetch/$s_!ZFgF!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ed31d09-a4d3-4b60-aa09-ba6ce8ae5a68_991x918.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>When compared to another mixture of experts model, GShard, we see that DeepSeekMoE again outperforms it with the same total parameters and only half of the activated parameters.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!RZJF!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6101720-8db6-495a-99ad-d32ce5906a86_984x547.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!RZJF!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6101720-8db6-495a-99ad-d32ce5906a86_984x547.png 424w, https://substackcdn.com/image/fetch/$s_!RZJF!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6101720-8db6-495a-99ad-d32ce5906a86_984x547.png 848w, https://substackcdn.com/image/fetch/$s_!RZJF!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6101720-8db6-495a-99ad-d32ce5906a86_984x547.png 1272w, https://substackcdn.com/image/fetch/$s_!RZJF!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6101720-8db6-495a-99ad-d32ce5906a86_984x547.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!RZJF!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6101720-8db6-495a-99ad-d32ce5906a86_984x547.png" width="984" height="547" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b6101720-8db6-495a-99ad-d32ce5906a86_984x547.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:547,&quot;width&quot;:984,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:69102,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!RZJF!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6101720-8db6-495a-99ad-d32ce5906a86_984x547.png 424w, https://substackcdn.com/image/fetch/$s_!RZJF!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6101720-8db6-495a-99ad-d32ce5906a86_984x547.png 848w, https://substackcdn.com/image/fetch/$s_!RZJF!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6101720-8db6-495a-99ad-d32ce5906a86_984x547.png 1272w, https://substackcdn.com/image/fetch/$s_!RZJF!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6101720-8db6-495a-99ad-d32ce5906a86_984x547.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>In sum, DeepSeek's optimizations for the MoE architecture served to substantially expand the possibilities for local and edge inference. Since only a small percentage of the model's total parameters are active for any given token, during inference, the model's performance requirements are much closer to those of a small, weak model. However, its output quality matches that of a large, well-trained dense LLM. This innovation was critical for laying the groundwork towards DeepSeek-R1, ensuring that state-of-the-art base LLM performance would be possible for smaller models.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.chrishayduk.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Musings by Chris Hayduk! Subscribe to make sure you don&#8217;t miss any new posts in this series!</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Understanding Protein Language Models Part III: Structure Prediction without Multiple Sequence Alignment in ESMFold]]></title><description><![CDATA[How ESMFold and ESM3 replace explicit MSAs with encoder-only transformers]]></description><link>https://www.chrishayduk.com/p/understanding-protein-language-models-40b</link><guid isPermaLink="false">https://www.chrishayduk.com/p/understanding-protein-language-models-40b</guid><dc:creator><![CDATA[Chris Hayduk]]></dc:creator><pubDate>Wed, 22 Jan 2025 15:53:30 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!e9L5!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd14a546-d65b-4198-add1-c7e976554eb9_1277x330.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!e9L5!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd14a546-d65b-4198-add1-c7e976554eb9_1277x330.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!e9L5!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd14a546-d65b-4198-add1-c7e976554eb9_1277x330.png 424w, https://substackcdn.com/image/fetch/$s_!e9L5!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd14a546-d65b-4198-add1-c7e976554eb9_1277x330.png 848w, https://substackcdn.com/image/fetch/$s_!e9L5!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd14a546-d65b-4198-add1-c7e976554eb9_1277x330.png 1272w, https://substackcdn.com/image/fetch/$s_!e9L5!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd14a546-d65b-4198-add1-c7e976554eb9_1277x330.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!e9L5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd14a546-d65b-4198-add1-c7e976554eb9_1277x330.png" width="1277" height="330" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/dd14a546-d65b-4198-add1-c7e976554eb9_1277x330.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:330,&quot;width&quot;:1277,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;Meta's ESMfold: the rival of AlpahFold2 | by Salvatore Raieli | Medium&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Meta's ESMfold: the rival of AlpahFold2 | by Salvatore Raieli | Medium" title="Meta's ESMfold: the rival of AlpahFold2 | by Salvatore Raieli | Medium" srcset="https://substackcdn.com/image/fetch/$s_!e9L5!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd14a546-d65b-4198-add1-c7e976554eb9_1277x330.png 424w, https://substackcdn.com/image/fetch/$s_!e9L5!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd14a546-d65b-4198-add1-c7e976554eb9_1277x330.png 848w, https://substackcdn.com/image/fetch/$s_!e9L5!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd14a546-d65b-4198-add1-c7e976554eb9_1277x330.png 1272w, https://substackcdn.com/image/fetch/$s_!e9L5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd14a546-d65b-4198-add1-c7e976554eb9_1277x330.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><strong>Note: </strong>This post is part of the &#8220;Understanding Protein Language Models&#8221; Series:</p><ol><li><p><a href="https://www.chrishayduk.com/p/understanding-protein-language-models">Understanding Protein Language Models Part I: Multiple Sequence Alignment in AlphaFold2</a></p></li><li><p><a href="https://www.chrishayduk.com/p/understanding-protein-language-models-e1a">Understanding Protein Language Models Part II: Encoder-only Transformers as Continuous Fuzzy String Matching</a></p></li><li><p>[This article] Understanding Protein Language Models Part III: Structure Prediction without Multiple Sequence Alignment in ESMFold</p></li></ol><div><hr></div><p><strong>Overview of the Main Ideas</strong></p><ol><li><p><strong>AlphaFold2&#8217;s MSA:</strong> AlphaFold2 identifies evolutionarily related proteins to the target sequence and builds a multiple sequence alignment (MSA). In the Evoformer block, row-wise (within-sequence) and column-wise (across-sequences) attention on this MSA yields information about co-evolving residues. This MSA-based representation is then integrated into a pair representation matrix, ultimately helping AlphaFold2 predict the 3D structure.</p></li><li><p><strong>ESMFold&#8217;s Language Model Encoding:</strong> In ESMFold, the MSA step is replaced by a large protein language model (ESM-2) trained via a Masked Language Modeling (MLM) objective. As in standard large language models for text, the hidden layers of the encoder learn semantic and syntactic regularities&#8212;in this case, biochemical and structural patterns. The result is that ESMFold can leverage these learned encodings to identify motifs and co-evolving positions without explicitly performing genetic database searches or building large MSAs.</p></li><li><p><strong>Conceptual Motif Lookup:</strong> We can interpret ESM-2&#8217;s embeddings as performing a &#8220;continuous fuzzy lookup&#8221; within an implicit database of protein motifs. Because the language model was pretrained on massive amounts of protein data, it has effectively learned how residues co-occur&#8212;and thus co-evolve&#8212;within protein families. This internal representation replaces the explicit MSA step.</p></li></ol><p>Below, we will dive into how this replacement works in more detail, starting with a short recap of AlphaFold2&#8217;s MSA-based pipeline and then exploring how ESMFold (and ESM-2 as its core) sidesteps explicit alignment by using learned representations.</p><div><hr></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.chrishayduk.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Musings by Chris Hayduk! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><div><hr></div><h2>1. Revisiting AlphaFold2&#8217;s MSA-Based Approach</h2><h3>1.1 Gathering Evolutionary Information</h3><p>AlphaFold2 conducts genetic searches against databases such as MGnify, UniRef90, Uniclust30, and BFD to identify sequences that share evolutionary relationships with the target sequence. From these hits, it constructs an MSA:</p><div class="latex-rendered" data-attrs="{&quot;persistentExpression&quot;:&quot;\\text{MSA} \\;=\\; \\begin{pmatrix} s_{1,1} &amp; s_{1,2} &amp; \\dots &amp; s_{1,L} \\\\ s_{2,1} &amp; s_{2,2} &amp; \\dots &amp; s_{2,L} \\\\ \\vdots &amp; \\vdots &amp; \\ddots &amp; \\vdots \\\\ s_{S,1} &amp; s_{S,2} &amp; \\dots &amp; s_{S,L} \\end{pmatrix},&quot;,&quot;id&quot;:&quot;NLFVOYWKIH&quot;}" data-component-name="LatexBlockToDOM"></div><p>where L is the length of the target sequence, and S is the number of evolutionarily related sequences found. Here, s_{k,i} denotes the i-th residue of the k-th sequence in the alignment. By hypothesizing that residues co-evolve, the MSA is an external source of statistical correlations about which residues likely pair or contact each other in 3D space.</p><h3>1.2 Evoformer and Pair Representation</h3><p>In the AlphaFold2 pipeline:</p><ol><li><p><strong>MSA Representation</strong> <strong>M</strong>: A 3D tensor M,</p><div class="latex-rendered" data-attrs="{&quot;persistentExpression&quot;:&quot;\\mathbf{M} \\in \\mathbf{R}^{S \\times L \\times c}&quot;,&quot;id&quot;:&quot;TFKYKLSLZH&quot;}" data-component-name="LatexBlockToDOM"></div><p>where c is the dimensionality of each residue embedding.</p></li><li><p><strong>Pair Representation</strong> <strong>P</strong>: A 2D grid P,</p><div class="latex-rendered" data-attrs="{&quot;persistentExpression&quot;:&quot;\\mathbf{P} = \\mathbb{R}^{L \\times L \\times c_z}&quot;,&quot;id&quot;:&quot;HOLIQNNHIB&quot;}" data-component-name="LatexBlockToDOM"></div><p>where each P_{i,j} is a learned embedding representing the pairwise relationship between residue i and residue j in the target sequence.</p></li></ol><p>Inside the Evoformer block, row-wise and column-wise attention update the MSA representation:</p><ul><li><p><strong>Row-wise (Within-sequence) Attention</strong></p><div class="latex-rendered" data-attrs="{&quot;persistentExpression&quot;:&quot;\\text{AttnRow}(\\mathbf{M})_{k, i} = \\sum_{m=1}^{L} \\alpha_{i,m} \\, \\bigl(W^V \\mathbf{M}_{k,m}\\bigr)&quot;,&quot;id&quot;:&quot;FLMQJXBFVQ&quot;}" data-component-name="LatexBlockToDOM"></div><p></p><p>where &#945;_{i,m} are attention weights.</p></li><li><p><strong>Column-wise (Across-sequence) Attention</strong></p><div class="latex-rendered" data-attrs="{&quot;persistentExpression&quot;:&quot;\\text{AttnCol}(\\mathbf{M})_{k, i} = \\sum_{n=1}^{S} \\beta_{k,n} \\, \\bigl(W^V \\mathbf{M}_{n,i}\\bigr)&quot;,&quot;id&quot;:&quot;BCDSEKOEAZ&quot;}" data-component-name="LatexBlockToDOM"></div><p></p><p>where &#946;_{k,n} are attention weights.</p></li></ul><p>After these attention layers (plus MSA transitions via 2-layer MLP), AlphaFold2 computes an <strong>Outer Product Mean</strong> that integrates MSA embeddings into the pair representation:</p><div class="latex-rendered" data-attrs="{&quot;persistentExpression&quot;:&quot;\\mathbf{OPM}_{i,j} \\;=\\; \\Bigl(\\frac{1}{S}\\sum_{k=1}^S \\mathbf{u}_{k,i} \\otimes \\mathbf{u}_{k,j}\\Bigr) \\,W_{\\mathrm{proj}}&quot;,&quot;id&quot;:&quot;SMHJJISUOS&quot;}" data-component-name="LatexBlockToDOM"></div><p>where u_{k,i} is the final MSA embedding vector for residue i in sequence k. This OPM_{i,j} is then added (or concatenated and projected) into P_{i,j}, effectively injecting co-evolutionary signals gleaned from the MSA into the residue-pair representation.</p><div><hr></div><h2>2. ESMFold: Replacing MSA with Language Modeling</h2><h3>2.1 The Core Mechanism: Encoder-Only Transformer</h3><p>ESMFold (and its backbone ESM-2 model) is built around a large encoder-only transformer. It is trained with the <strong>masked language modeling</strong> objective, meaning it tries to reconstruct masked or hidden residues from context. This training strategy, originally popularized by BERT in natural language processing, has an important effect: it forces the model to encode in its weights the relevant &#8220;contexts&#8221; that predict each amino acid.</p><p>Mathematically, if x=(x_1,x_2,&#8230;,x_L) is the protein sequence and x_k is replaced by a special [MASK] token with some probability, the MLM training objective is</p><div class="latex-rendered" data-attrs="{&quot;persistentExpression&quot;:&quot;\\mathcal{L}_{\\mathrm{MLM}} = -\\sum_{k=1}^{L} \\log p_\\theta(x_k \\mid x_1, \\ldots, x_{k-1}, \\text{[MASK]}, x_{k+1}, \\ldots, x_L)&quot;,&quot;id&quot;:&quot;UFYLUMHRMO&quot;}" data-component-name="LatexBlockToDOM"></div><p></p><p>where p_&#952; is parameterized by the encoder transformer. Over billions of observed residues, the model internalizes the patterns of co-occurrence across diverse protein sequences.</p><h3>2.2 Implicit Motif Lookup</h3><p>Where AlphaFold2 uses explicit lookups in an MSA database (plus explicit attention across sequences), ESM-2&#8217;s learned embeddings do something analogous &#8220;in one shot.&#8221; After pretraining, the internal representation of each residue h_i (the hidden state at position i) captures average contexts encountered during training. In effect, for any position i,</p><ol><li><p>h_i has high similarity to h_j if residues x_i and x_j frequently appear in similar sequence contexts in the training set.</p></li><li><p>By extension, if an entire sequence <strong>x</strong> has patterns analogous to known motifs (e.g., an ATP-binding site pattern, a signal peptide motif, or secondary-structure fragments), then the embeddings reflect these patterns&#8212;allowing ESMFold to &#8220;retrieve&#8221; them without an explicit MSA.</p></li></ol><p>You can view this as a &#8220;continuous fuzzy matching&#8221; process, wherein the [KEY], [QUERY], and [VALUE] matrices of the transformer contain compressed representations of how residues co-occur. Rather than computing the dynamic-programming-based edit distances (or alignment) across a large external database, the model&#8217;s attention modules effectively do an alignment on-the-fly in a continuous, high-dimensional space.</p><h3>2.3 Integration into Folding</h3><p>ESMFold then appends a structure-prediction head on top of these ESM-2 embeddings, akin to how AlphaFold2 appends its structure module after the Evoformer. Even though ESMFold no longer has an explicit pair representation from an MSA, it still must estimate which residues interact or contact each other. In current ESMFold architectures:</p><ul><li><p>The final hidden states from the ESM-2 encoder are projected into a lower-dimensional representation that acts like a &#8220;pair embedding&#8221; for each (i,j).</p></li><li><p>A geometry module or a series of feed-forward layers further refines these embeddings to produce coordinates or distance/contact maps.</p></li></ul><p>In practice, ESMFold&#8217;s results are often on par with AlphaFold2 for many proteins, especially those with strong evolutionary constraints. For proteins with scant evolutionary data, ESMFold can sometimes do <em>better</em> than AlphaFold2, because it does not rely so heavily on a large MSA. On the flip side, certain proteins with well-studied deep MSAs can benefit from the explicit signals that AlphaFold2&#8217;s large MSA provides.</p><div><hr></div><h2>3. Mathematical Rationales for Replacing MSA</h2><h3>3.1 Complexity and Speed</h3><p>One major advantage of dropping MSAs is computational efficiency. MSA searches can be prohibitively expensive for large proteins or large sets of queries, requiring queries against massive databases (MGnify, UniRef, etc.) and heuristics to align thousands of sequences. In ESMFold:</p><ul><li><p><strong>No MSA Search:</strong> The model simply takes the query sequence and feeds it through the encoder in a single forward pass.</p></li><li><p><strong>Linear vs. Quadratic Complexity:</strong> A single Transformer forward pass for a sequence of length L has complexity O(L^2 d) where d is the dimension of embeddings, whereas building an MSA might involve searching and aligning thousands of sequences, each of length up to L.</p></li></ul><h3>3.2 Continuous Fuzzy Matching Perspective</h3><p>If we interpret the MSA as a form of nearest-neighbor search (looking for &#8220;neighboring&#8221; sequences in a large database), then the language model is effectively a learned data structure that has:</p><ul><li><p><strong>Compressed</strong> the manifold of known protein sequences into &#952; (the weights).</p></li><li><p><strong>Learned</strong> an attention-based mechanism to query that internal manifold for relevant contexts.</p></li></ul><p>In typical fuzzy string matching, one might compute edit distances between the query and every entry in the database. In the ESM-2 architecture, the attention mechanism:</p><div class="latex-rendered" data-attrs="{&quot;persistentExpression&quot;:&quot;\\text{Attn}(\\mathbf{Q},\\mathbf{K},\\mathbf{V}) = \\mathrm{softmax}\\Bigl(\\frac{\\mathbf{Q} \\mathbf{K}^T}{\\sqrt{d_k}}\\Bigr) \\mathbf{V}&quot;,&quot;id&quot;:&quot;JAWRVEHEKZ&quot;}" data-component-name="LatexBlockToDOM"></div><p>acts as a trainable similarity function to identify relevant contexts. The intangible advantage is that these contexts may mix and match partial motifs from multiple &#8220;virtual neighbors,&#8221; creating a new representation not limited to the top few explicit matches in a database.</p><h3>3.3 Co-evolutionary Signals Without Explicit Alignments</h3><p>A major reason MSA is so powerful is that it captures <em>co-evolving residues</em>&#8212;positions that change in correlated ways across evolutionary history. In a typical MSA-based approach, if residue i mutates from A to G, residue j might consistently switch from T to S. Over many sequences, one infers that i and j likely contact or interact structurally.</p><p>By training on a massive corpus, the language model sees countless such correlations in raw sequence form. The emergent embeddings reflect these patterns. Hence, the final hidden state h_i is (indirectly) sensitive to all correlated positions that have ever appeared near that residue in training. So even though ESMFold does not align the query sequence to a database, it has internalized an approximate version of that same statistical correlation from its pretrained weights.</p><div><hr></div><h2>4. Example: From MSA to Language Model&#8212;A Toy Mathematical Sketch</h2><p>Suppose we consider a short hypothetical protein sequence <strong>x</strong>=(M,K,L,L,P,V,L). In an MSA-based approach, you might gather 10,000 sequences from a database, building:</p><div class="latex-rendered" data-attrs="{&quot;persistentExpression&quot;:&quot;\\begin{pmatrix} M &amp; K &amp; L &amp; L &amp; P &amp; V &amp; L \\\\ M &amp; K &amp; L &amp; L &amp; T &amp; V &amp; L \\\\ M &amp; R &amp; L &amp; L &amp; P &amp; A &amp; L \\\\ \\vdots &amp; &amp; &amp; &amp; &amp; &amp; \\\\ \\text{(10,000 rows)} \\end{pmatrix}.&quot;,&quot;id&quot;:&quot;CPYPYAYXBH&quot;}" data-component-name="LatexBlockToDOM"></div><p>You then compute attention across these sequences (column-wise) and across residues (row-wise), deriving correlation maps.</p><p>In the ESM-2 approach, no explicit MSA is constructed. Instead, during training, the model saw thousands (or millions) of sequences resembling (M,K,L,L,P,V,L) or partial subsequences thereof. The MLM objective forced the model to fill in [MASK] tokens in contexts like _&#8201;K&#8201;L&#8201;_&#8201;P&#8201;V&#8201;_. Over many instances, it learned which next residues are probable. As a result, once we feed (M,K,L,L,P,V,L) into ESM-2, the hidden states reflect a &#8220;compressed MSA,&#8221; effectively picking up correlations that used to require explicit cross-sequence operations.</p><div><hr></div><h2>5. Implications and Future Directions</h2><ol><li><p><strong>Efficiency Gains:</strong> ESMFold runs <em>significantly faster</em> than AlphaFold2 when no large MSA is available, since it avoids the alignment process. For proteome-scale structure predictions, this is a game-changer.</p></li><li><p><strong>Handling Novel Proteins:</strong> If a target protein has few homologs in public databases, MSA-based models struggle. ESMFold is robust in these &#8220;low-homology&#8221; cases since it learned general protein grammar from the entire training corpus.</p></li><li><p><strong>Limited Interpretability:</strong> One downside is that MSA-based approaches produce an explicit record of hits and alignments, which can be biologically interpretable (e.g., which species and families contributed the signals). ESMFold&#8217;s learned embedding, while powerful, can be less transparent.</p></li><li><p><strong>Hybrid Approaches:</strong> Some emerging methods combine pre-trained embeddings with an MSA for the best of both worlds&#8212;particularly for proteins where deep MSAs exist.</p></li><li><p><strong>Scaling Laws and Emergent Behavior:</strong> As ESM models grow (ESM-2, ESM-3, etc.), they exhibit emergent behaviors akin to large language models in NLP. This suggests we may see further improvements in structure prediction, function annotation, and protein design.</p></li></ol><div><hr></div><h2>6. Conclusion</h2><p>AlphaFold2&#8217;s success showed how vital MSAs are in revealing <em>co-evolutionary signals</em>, which guide 3D structure inference. ESMFold&#8217;s fundamental insight is that you can <em>pre-learn</em> these signals at massive scale by treating protein sequences as &#8220;language.&#8221; Then, instead of collecting an MSA at inference time, the model effectively &#8220;queries&#8221; its internal knowledge of sequence co-occurrences, learned through the MLM objective.</p><p>In both approaches, the central idea is to approximate how residues covary. In AlphaFold2, that covariance emerges explicitly from a large MSA. In ESMFold, it is embedded implicitly in a high-dimensional transformer space. The advantage of the language-model approach is that it (1) eliminates the bottleneck of database searching/alignment, and (2) leverages far more global knowledge than just sequences that happen to align with the target protein.</p><p>Mathematically, we can view these approaches as two different ways to compute a &#8220;similarity function&#8221; over the manifold of protein sequences:</p><ul><li><p><strong>AlphaFold2 + MSA:</strong> An explicit alignment-based approach that organizes relevant sequences so the model can learn correlations.</p></li><li><p><strong>ESMFold + Transformer:</strong> A large-scale learned approach that stores correlation statistics in the weights, retrieving them through self-attention rather than explicit alignment.</p></li></ul><p>As these language models grow and become more accurate, their potential to replace, or at least augment, MSA-based pipelines will only increase, promising ever-faster and more versatile protein structure prediction.</p><p>In summary, ESMFold&#8217;s fundamental contribution is demonstrating how one can use a large, pretrained protein-language transformer to replicate (and in some cases surpass) the evolutionary context that an MSA provides. It is a step toward an era where generative models of protein sequence space might supersede explicit database lookups, enabling faster, more flexible, and equally accurate structure predictions&#8212;even for proteins with scarce evolutionary data.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.chrishayduk.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Musings by Chris Hayduk! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Understanding Protein Language Models Part II: Encoder-only Transformers as Continuous Fuzzy String Matching]]></title><description><![CDATA[How transformers learn from their input data]]></description><link>https://www.chrishayduk.com/p/understanding-protein-language-models-e1a</link><guid isPermaLink="false">https://www.chrishayduk.com/p/understanding-protein-language-models-e1a</guid><dc:creator><![CDATA[Chris Hayduk]]></dc:creator><pubDate>Wed, 15 Jan 2025 19:21:17 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!sMfG!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F97ec1cc1-2590-4cb8-947a-7e65144465a4_2026x1106.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!sMfG!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F97ec1cc1-2590-4cb8-947a-7e65144465a4_2026x1106.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!sMfG!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F97ec1cc1-2590-4cb8-947a-7e65144465a4_2026x1106.png 424w, https://substackcdn.com/image/fetch/$s_!sMfG!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F97ec1cc1-2590-4cb8-947a-7e65144465a4_2026x1106.png 848w, https://substackcdn.com/image/fetch/$s_!sMfG!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F97ec1cc1-2590-4cb8-947a-7e65144465a4_2026x1106.png 1272w, https://substackcdn.com/image/fetch/$s_!sMfG!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F97ec1cc1-2590-4cb8-947a-7e65144465a4_2026x1106.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!sMfG!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F97ec1cc1-2590-4cb8-947a-7e65144465a4_2026x1106.png" width="1456" height="795" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/97ec1cc1-2590-4cb8-947a-7e65144465a4_2026x1106.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:795,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;transformer - Why GPT uses decoder only architecture, when they can use  full encoder decoder architecture? - Artificial Intelligence Stack Exchange&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="transformer - Why GPT uses decoder only architecture, when they can use  full encoder decoder architecture? - Artificial Intelligence Stack Exchange" title="transformer - Why GPT uses decoder only architecture, when they can use  full encoder decoder architecture? - Artificial Intelligence Stack Exchange" srcset="https://substackcdn.com/image/fetch/$s_!sMfG!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F97ec1cc1-2590-4cb8-947a-7e65144465a4_2026x1106.png 424w, https://substackcdn.com/image/fetch/$s_!sMfG!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F97ec1cc1-2590-4cb8-947a-7e65144465a4_2026x1106.png 848w, https://substackcdn.com/image/fetch/$s_!sMfG!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F97ec1cc1-2590-4cb8-947a-7e65144465a4_2026x1106.png 1272w, https://substackcdn.com/image/fetch/$s_!sMfG!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F97ec1cc1-2590-4cb8-947a-7e65144465a4_2026x1106.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><strong>Note: </strong>This post is part of the &#8220;Understanding Protein Language Models&#8221; Series:</p><ol><li><p><a href="https://www.chrishayduk.com/p/understanding-protein-language-models">Understanding Protein Language Models Part I: Multiple Sequence Alignment in AlphaFold2</a></p></li><li><p>[This article] Understanding Protein Language Models Part II: Encoder-only Transformers as Continuous Fuzzy String Matching</p></li><li><p><a href="https://www.chrishayduk.com/p/understanding-protein-language-models-40b">Understanding Protein Language Models Part III: Structure Prediction without Multiple Sequence Alignment in ESMFold</a></p></li></ol><div><hr></div><p>In the quest to understand modern protein language models like ESM2 and ESM3, we often focus on their impressive empirical results while treating their internal mechanisms as a black box. This post attempts to build intuition about how encoder-only transformers work by drawing an analogy to a simpler, well-understood algorithm: fuzzy string matching. I argue that encoder-only transformers can be viewed as performing a kind of continuous fuzzy lookup against a compressed form of their training data, encoded in their weights and latent space representations.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.chrishayduk.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.chrishayduk.com/subscribe?"><span>Subscribe now</span></a></p><h2>The Central Analogy</h2><p>When working with large text corpora, we often need to find similar strings or patterns. The traditional approach employs fuzzy string matching: maintaining a database of all strings and computing edit distances to find matches. An alternative approach, which I argue is conceptually similar but mathematically more sophisticated, uses an encoder-only transformer to compress the patterns in the corpus into model weights, then uses attention mechanisms to find similarities.</p><p>Both approaches fundamentally solve the same problem - finding contextually appropriate matches - but do so in radically different ways. Understanding this connection helps demystify how encoder-only transformers work and suggests ways to improve them.</p><h2>How Encoder-Only Transformers Process Input</h2><p>To understand the analogy, we first need to build detailed intuition about how encoder-only transformers work. Unlike the full encoder-decoder architecture used in translation, encoder-only transformers take a sequence of tokens and return a sequence of the same length, where each output token is a refined representation incorporating contextual information.</p><p>The process begins in the embedding layer, where discrete tokens are converted into continuous vectors. Each input token is first converted to a one-hot encoding - a vector of zeros with a single one indicating the token's identity. This sparse vector is then multiplied by an embedding matrix to produce a dense vector representation. Mathematically, for a token x_i:</p><pre><code>one_hot = [0, 0, ..., 1, ..., 0] # 1 at position x_i 

embedding = one_hot @ W_emb # Matrix multiplication with embedding matrix</code></pre><p>To this embedding, we add a positional encoding vector that encodes information about where the token appears in the sequence. The original transformer paper used sinusoidal positional encodings:</p><pre><code>def get_positional_encoding(seq_len, d_model): 
   position = np.arange(seq_len)[:, np.newaxis] 
   div_term = np.exp(np.arange(0, d_model, 2) * -(np.log(10000.0) / d_model)) 
   pos_enc = np.zeros((seq_len, d_model)) 
   pos_enc[:, 0::2] = np.sin(position * div_term) 
   pos_enc[:, 1::2] = np.cos(position * div_term) 
   return pos_enc</code></pre><p>These positional encodings have an elegant property: the relative position of two tokens can be computed through linear combinations of their encodings.</p><p>The heart of the transformer architecture lies in its self-attention mechanism. For each position in the sequence, the model generates three vectors through learned linear transformations:</p><pre><code>Q = H @ W_Q # Query vectors 
K = H @ W_K # Key vectors 
V = H @ W_V # Value vectors</code></pre><p>where H is the matrix of hidden states. The attention scores are then computed as:</p><pre><code>attention_scores = softmax(Q @ K.T / sqrt(d_k)) 
output = attention_scores @ V</code></pre><p>This can be written more formally as:</p><div class="latex-rendered" data-attrs="{&quot;persistentExpression&quot;:&quot;\\text{Attention}(Q, K, V) = \\text{softmax}\\left(\\frac{QK^T}{\\sqrt{d_k}}\\right)V &quot;,&quot;id&quot;:&quot;AGZNKXYYMF&quot;}" data-component-name="LatexBlockToDOM"></div><p></p><p>This attention mechanism allows each position to gather information from all other positions, with the weights determined by learned compatibility scores. The scaling factor $\sqrt{d_k}$ prevents the dot products from growing too large in magnitude, which would push the softmax into regions of extremely small gradients.</p><p>The transformer employs multiple attention heads in parallel, each with its own set of query, key, and value projections:</p><div class="latex-rendered" data-attrs="{&quot;persistentExpression&quot;:&quot;\\text{MultiHead}(H) = \\text{Concat}(\\text{head}_1, \\ldots, \\text{head}_h)W^O&quot;,&quot;id&quot;:&quot;ZRQNCSGKPZ&quot;}" data-component-name="LatexBlockToDOM"></div><p>where each head is computed as:</p><div class="latex-rendered" data-attrs="{&quot;persistentExpression&quot;:&quot;\\text{head}_i = \\text{Attention}(HW^Q_i, HW^K_i, HW^V_i)&quot;,&quot;id&quot;:&quot;ICYATLGANP&quot;}" data-component-name="LatexBlockToDOM"></div><p>After attention, each position's representation goes through a two-layer feed-forward network:</p><div class="latex-rendered" data-attrs="{&quot;persistentExpression&quot;:&quot;\\text{FFN}(x) = \\text{max}(0, xW_1 + b_1)W_2 + b_2&quot;,&quot;id&quot;:&quot;WFWABMVVDQ&quot;}" data-component-name="LatexBlockToDOM"></div><h2>Traditional Fuzzy String Matching</h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!xxZW!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F57dde700-3ba6-49b6-bfcd-6f591b344001_547x335.gif" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!xxZW!,w_424,c_limit,f_webp,q_auto:good,fl_lossy/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F57dde700-3ba6-49b6-bfcd-6f591b344001_547x335.gif 424w, https://substackcdn.com/image/fetch/$s_!xxZW!,w_848,c_limit,f_webp,q_auto:good,fl_lossy/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F57dde700-3ba6-49b6-bfcd-6f591b344001_547x335.gif 848w, https://substackcdn.com/image/fetch/$s_!xxZW!,w_1272,c_limit,f_webp,q_auto:good,fl_lossy/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F57dde700-3ba6-49b6-bfcd-6f591b344001_547x335.gif 1272w, https://substackcdn.com/image/fetch/$s_!xxZW!,w_1456,c_limit,f_webp,q_auto:good,fl_lossy/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F57dde700-3ba6-49b6-bfcd-6f591b344001_547x335.gif 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!xxZW!,w_1456,c_limit,f_auto,q_auto:good,fl_lossy/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F57dde700-3ba6-49b6-bfcd-6f591b344001_547x335.gif" width="547" height="335" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/57dde700-3ba6-49b6-bfcd-6f591b344001_547x335.gif&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:335,&quot;width&quot;:547,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;Fuzzy String Matching with `stringdist` &#8211; Just R Things&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Fuzzy String Matching with `stringdist` &#8211; Just R Things" title="Fuzzy String Matching with `stringdist` &#8211; Just R Things" srcset="https://substackcdn.com/image/fetch/$s_!xxZW!,w_424,c_limit,f_auto,q_auto:good,fl_lossy/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F57dde700-3ba6-49b6-bfcd-6f591b344001_547x335.gif 424w, https://substackcdn.com/image/fetch/$s_!xxZW!,w_848,c_limit,f_auto,q_auto:good,fl_lossy/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F57dde700-3ba6-49b6-bfcd-6f591b344001_547x335.gif 848w, https://substackcdn.com/image/fetch/$s_!xxZW!,w_1272,c_limit,f_auto,q_auto:good,fl_lossy/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F57dde700-3ba6-49b6-bfcd-6f591b344001_547x335.gif 1272w, https://substackcdn.com/image/fetch/$s_!xxZW!,w_1456,c_limit,f_auto,q_auto:good,fl_lossy/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F57dde700-3ba6-49b6-bfcd-6f591b344001_547x335.gif 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>To appreciate the analogy, we need to understand how traditional fuzzy string matching works. Given an input string and a database of reference strings, fuzzy matching computes the edit distance between the input and each reference. The edit distance represents the minimum number of operations (insertions, deletions, or substitutions) needed to transform one string into another.</p><p>The core of fuzzy string matching is the computation of edit distance. For strings s and t, the dynamic programming recurrence is:</p><pre><code>def edit_distance(s, t): 
   m, n = len(s), len(t) 
   dp = np.zeros((m+1, n+1)) 

   # Initialize base cases 
   for i in range(m+1): 
      dp[i,0] = i 
   for j in range(n+1): 
      dp[0,j] = j 

   # Fill dp table 
   for i in range(1, m+1): 
      for j in range(1, n+1): 
         if s[i-1] == t[j-1]: 
            dp[i,j] = dp[i-1,j-1] 
         else: 
            dp[i,j] = 1 + min( 
               dp[i-1,j], # deletion 
               dp[i,j-1], # insertion 
               dp[i-1,j-1] # substitution 
            ) 
   
   return dp[m,n]</code></pre><p>Mathematically, the recurrence relation is:</p><div class="latex-rendered" data-attrs="{&quot;persistentExpression&quot;:&quot;D[i,j] = \\min \\begin{cases} D[i-1,j] + 1 &amp; \\text{deletion} \\\\ D[i,j-1] + 1 &amp; \\text{insertion} \\\\ D[i-1,j-1] + \\mathbb{1}_{s[i] \\neq t[j]} &amp; \\text{substitution} \\end{cases}&quot;,&quot;id&quot;:&quot;NQWAVECNJM&quot;}" data-component-name="LatexBlockToDOM"></div><p>The computation proceeds through dynamic programming, building a matrix where each cell represents the minimum number of operations needed to match a prefix of the input string to a prefix of the reference string. The final cell gives the total edit distance between the strings. By computing this distance for each reference string and sorting the results, we can find the closest matches in the database.</p><p>While conceptually simple and mathematically elegant, this approach becomes computationally expensive for large databases, requiring time proportional to both the string lengths and the size of the database. Various optimizations exist, such as trie structures and early pruning, but the fundamental challenge of scaling remains.</p><h2>The Transformer as Compressed Fuzzy Matching</h2><p>Here we arrive at the core insight: the encoder-only transformer effectively compresses the pattern-matching capabilities of fuzzy string matching into its weights. During training, the model learns to encode the essential patterns and relationships present in the training data into its parameters.</p><p>The embedding matrix learns to map tokens to a continuous space where similar tokens are close together. The attention weights learn which patterns of tokens commonly co-occur, while the feed-forward layers learn to combine these patterns into higher-level features. Each successive layer captures progressively more abstract relationships. This process is analogous to building an optimized index of the training data, but instead of storing exact strings, we store distributed representations of patterns and relationships.</p><p>The connection between transformers and fuzzy matching becomes clearer when we compare their similarity computations. In fuzzy matching, the similarity between strings s and t is:</p><div class="latex-rendered" data-attrs="{&quot;persistentExpression&quot;:&quot;\\text{similarity}(s,t) = -\\min_{\\text{operations}} \\sum_i \\text{cost}(\\text{op}_i)&quot;,&quot;id&quot;:&quot;ZDFABHWQEO&quot;}" data-component-name="LatexBlockToDOM"></div><p></p><p>In transformer attention, the similarity between vectors q and k is:</p><div class="latex-rendered" data-attrs="{&quot;persistentExpression&quot;:&quot;\\text{similarity}(q,k) = \\frac{q \\cdot k^T}{\\sqrt{d_k}} &quot;,&quot;id&quot;:&quot;HFJIELSMFL&quot;}" data-component-name="LatexBlockToDOM"></div><p></p><p>We can view the transformer's learned weights as parameterizing a continuous relaxation of edit distance. The attention mechanism implements this relaxed distance metric:</p><pre><code>def attention_similarity(query, key, value): 
   # query shape: [seq_len, d_k] 
   # key shape: [seq_len, d_k] 
   # value shape: [seq_len, d_v] 

   scores = query @ key.T / np.sqrt(query.shape[-1]) 
   attention_weights = softmax(scores) # [seq_len, seq_len] 
   return attention_weights @ value # [seq_len, d_v]</code></pre><p>Several observations support this compression view. The systematic improvement in model performance with increasing size suggests that larger models can store more detailed patterns from the training data. Analysis of attention patterns reveals that different heads learn interpretable relationships that match linguistic or domain structure. The organization of the embedding space shows meaningful clustering of similar tokens and preservation of analogical relationships.</p><p>When we run a sequence through the transformer, the process mirrors fuzzy matching but operates in a continuous space. The initial embedding maps tokens to vectors, analogous to preparing strings for comparison. The self-attention mechanism computes similarity scores between positions, playing a role similar to edit distance calculation, but using a learned, context-dependent metric. Multiple layers progressively refine these representations, like iteratively improving string alignment.</p><p>We can formalize this connection mathematically. In fuzzy matching, similarity is measured as the negative minimum cost of operations needed to transform one string into another. In transformer attention, similarity is measured through scaled dot products between query and key vectors. The transformer effectively learns a continuous approximation of edit distance that can capture more nuanced relationships.</p><h2>Concrete Example: Protein Sequences</h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!-iKO!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdc494fb8-639a-4a44-8cd0-ade19516ec06_600x536.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!-iKO!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdc494fb8-639a-4a44-8cd0-ade19516ec06_600x536.png 424w, https://substackcdn.com/image/fetch/$s_!-iKO!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdc494fb8-639a-4a44-8cd0-ade19516ec06_600x536.png 848w, https://substackcdn.com/image/fetch/$s_!-iKO!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdc494fb8-639a-4a44-8cd0-ade19516ec06_600x536.png 1272w, https://substackcdn.com/image/fetch/$s_!-iKO!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdc494fb8-639a-4a44-8cd0-ade19516ec06_600x536.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!-iKO!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdc494fb8-639a-4a44-8cd0-ade19516ec06_600x536.png" width="600" height="536" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/dc494fb8-639a-4a44-8cd0-ade19516ec06_600x536.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:536,&quot;width&quot;:600,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:74056,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!-iKO!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdc494fb8-639a-4a44-8cd0-ade19516ec06_600x536.png 424w, https://substackcdn.com/image/fetch/$s_!-iKO!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdc494fb8-639a-4a44-8cd0-ade19516ec06_600x536.png 848w, https://substackcdn.com/image/fetch/$s_!-iKO!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdc494fb8-639a-4a44-8cd0-ade19516ec06_600x536.png 1272w, https://substackcdn.com/image/fetch/$s_!-iKO!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdc494fb8-639a-4a44-8cd0-ade19516ec06_600x536.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Consider a concrete example from protein sequence analysis. In traditional fuzzy matching, we might have a query sequence "MKLLPVL" and search a database containing sequences like "MKLLTVL" (one substitution) or "MLKPVL" (two operations). Each comparison requires explicit computation of edit distances. This type of genetic database search is used by MSA-based models such as AlphaFold2 and AlphaFold3. The mechanics of this is described in the previous post in the series, <a href="https://www.chrishayduk.com/p/understanding-protein-language-models">Understanding Protein Language Models Part I: Multiple Sequence Alignment in AlphaFold2</a>.</p><p>The transformer approach is markedly different. After embedding the sequence into continuous vectors, self-attention finds similar patterns that have been compressed into the model's weights during training. The output reflects patterns observed during training, but crucially, the model can combine these patterns in novel ways. The transformer effectively "remembers" which amino acid substitutions are biochemically plausible in each context, without storing explicit sequences.</p><p>In the next post in this series, we will flesh out this section more and learn how protein language models allow us to replace the MSA step in AlphaFold, creating faster strucutre prediction models that generalize better to proteins that have few available related sequences.</p><h2>Conclusion</h2><p>Viewing encoder-only transformers as performing compressed fuzzy matching provides powerful intuition about their operation. Rather than seeing them as black boxes, we can understand them as learning to compress and query a vast database of patterns from their training data. This perspective suggests that improvements in transformer architecture may come from better compression techniques for storing training patterns, more efficient similarity computations, and explicit incorporation of string matching algorithms.</p><p>Future research might investigate how much pattern information is stored in different parts of the model, how different architectures affect compression quality, and whether we can design better compression mechanisms inspired by string algorithms. We might also explore the theoretical limits of this compression approach and its implications for model scaling.</p><p>The success of this architecture in domains like protein sequence modeling suggests that the ability to learn and compress domain-specific similarity metrics is a powerful paradigm. As we continue to develop these models, maintaining this conceptual connection to classical algorithms may help guide the way to more efficient and effective architectures.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.chrishayduk.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Musings by Chris Hayduk! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[OpenAI o3 and the Rise of the Intelligence Allocator]]></title><description><![CDATA[The implications of rapidly increasing inference costs]]></description><link>https://www.chrishayduk.com/p/openai-o3-and-the-rise-of-the-intelligence</link><guid isPermaLink="false">https://www.chrishayduk.com/p/openai-o3-and-the-rise-of-the-intelligence</guid><dc:creator><![CDATA[Chris Hayduk]]></dc:creator><pubDate>Fri, 20 Dec 2024 19:19:58 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!2iKn!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3188a284-6a08-423d-8f8c-964db47f75e6_1024x768.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!2iKn!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3188a284-6a08-423d-8f8c-964db47f75e6_1024x768.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!2iKn!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3188a284-6a08-423d-8f8c-964db47f75e6_1024x768.jpeg 424w, https://substackcdn.com/image/fetch/$s_!2iKn!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3188a284-6a08-423d-8f8c-964db47f75e6_1024x768.jpeg 848w, https://substackcdn.com/image/fetch/$s_!2iKn!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3188a284-6a08-423d-8f8c-964db47f75e6_1024x768.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!2iKn!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3188a284-6a08-423d-8f8c-964db47f75e6_1024x768.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!2iKn!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3188a284-6a08-423d-8f8c-964db47f75e6_1024x768.jpeg" width="1024" height="768" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3188a284-6a08-423d-8f8c-964db47f75e6_1024x768.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:768,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:176607,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!2iKn!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3188a284-6a08-423d-8f8c-964db47f75e6_1024x768.jpeg 424w, https://substackcdn.com/image/fetch/$s_!2iKn!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3188a284-6a08-423d-8f8c-964db47f75e6_1024x768.jpeg 848w, https://substackcdn.com/image/fetch/$s_!2iKn!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3188a284-6a08-423d-8f8c-964db47f75e6_1024x768.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!2iKn!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3188a284-6a08-423d-8f8c-964db47f75e6_1024x768.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>OpenAI's announcement of their o3 series of models represents a pivotal moment in AI development - but not for the reasons many might expect. While the headline achievement of 87% on ARC AGI is impressive, the more transformative aspect lies in the economics of the model's deployment.</p><p>Let's start with the raw numbers: A single inference task on o3 at its highest compute setting costs over $1,000 (see the figure below). This isn't $1,000 per evaluation set or per session - this is per individual task. To put this in perspective, that's roughly equivalent to 5-10 hours of skilled human labor cost, dedicated to solving a single problem. The model offers lower compute settings, but with corresponding decreases in capability. This creates a direct tradeoff between cost and intelligence that we haven't had to grapple with before.</p><p>This cost structure represents a sharp departure from the trend we've observed over the past two years. During that period, the cost of running general-purpose language models has approached zero, even as their capabilities have steadily improved. GPT-3.5 became GPT-4, yet inference costs remained relatively stable. GPT-4 then became GPT-4 Turbo and GPT-4o, maintaining intelligence while rapidly decreasing inference costs. This led to a proliferation of AI applications - we could afford to experiment freely, integrating AI into virtually every workflow to see what stuck.</p><p>The o3 series shatters this paradigm. When each inference costs more than a decent laptop, you can't simply "throw AI at the problem" anymore. Every use of high-compute o3 needs to be justified by the value it creates. This introduces what we might call the "inference allocation problem" - how do we determine which tasks are worth deploying our most powerful (and expensive) models on?</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!GwAu!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd1de8e6-34c6-43e0-bb18-4bbfb4b49874_1194x670.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!GwAu!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd1de8e6-34c6-43e0-bb18-4bbfb4b49874_1194x670.jpeg 424w, https://substackcdn.com/image/fetch/$s_!GwAu!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd1de8e6-34c6-43e0-bb18-4bbfb4b49874_1194x670.jpeg 848w, https://substackcdn.com/image/fetch/$s_!GwAu!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd1de8e6-34c6-43e0-bb18-4bbfb4b49874_1194x670.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!GwAu!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd1de8e6-34c6-43e0-bb18-4bbfb4b49874_1194x670.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!GwAu!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd1de8e6-34c6-43e0-bb18-4bbfb4b49874_1194x670.jpeg" width="1194" height="670" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/dd1de8e6-34c6-43e0-bb18-4bbfb4b49874_1194x670.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:670,&quot;width&quot;:1194,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:45229,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!GwAu!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd1de8e6-34c6-43e0-bb18-4bbfb4b49874_1194x670.jpeg 424w, https://substackcdn.com/image/fetch/$s_!GwAu!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd1de8e6-34c6-43e0-bb18-4bbfb4b49874_1194x670.jpeg 848w, https://substackcdn.com/image/fetch/$s_!GwAu!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd1de8e6-34c6-43e0-bb18-4bbfb4b49874_1194x670.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!GwAu!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd1de8e6-34c6-43e0-bb18-4bbfb4b49874_1194x670.jpeg 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.chrishayduk.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.chrishayduk.com/subscribe?"><span>Subscribe now</span></a></p><p>Consider a software development team using o3 for code analysis. Running the model at high compute to analyze a critical security vulnerability in a payment system might be easily justifiable. But what about using it to optimize a non-critical internal tool? Or to review routine pull requests? The team now needs to develop frameworks for making these decisions systematically.</p><p>This fundamentally transforms AI deployment into a capital allocation problem. Just as investment managers spread limited capital across opportunities to maximize returns, organizations must now optimize their allocation of inference compute to maximize value creation.</p><p>Consider a hypothetical AI budget of $1 million per month. Currently, this might support tens or hundreds of millions of GPT-4o inferences spread across hundreds of different use cases. With o3, the same budget only covers about 1,000 high-compute inferences. This scarcity forces us to think like capital allocators: Which thousand problems, if solved with our highest level of artificial intelligence, will generate the most value?</p><p>Beyond simply identifying high-value problems, intelligence allocators will need to understand the relationship between compute investment and value creation. Sometimes, a medium-compute inference at $100 might capture 80% of the potential value at 1/10th the cost. In other cases, the step-change in capability from high-compute might be worth the premium. Like any good investment decision, it requires understanding both the cost of capital and the expected returns.</p><p>In another parallel to traditional capital allocation, just as investors develop frameworks for evaluating investments across different sectors and risk levels, organizations will need frameworks for evaluating AI compute allocation across different use cases. These frameworks will need to consider factors like:</p><ul><li><p>The value delta between using high-compute versus lower-compute models</p></li><li><p>The cost of being wrong or suboptimal</p></li><li><p>The potential for value capture from improved accuracy</p></li><li><p>The frequency with which the task needs to be performed</p></li></ul><p>We might even see the emergence of "AI portfolio theory" - methods for optimizing the allocation of compute resources across different types of tasks to maximize expected return while managing risk. Some organizations might adopt a "barbell strategy" - using basic models for routine tasks while reserving expensive high-compute inferences for their most critical problems.</p><p>This shift in focus for AI engineers means that success looks more like developing the frameworks and metrics needed to make intelligence allocation decisions effectively, rather than focusing purely on technical implementations. The best AI engineers will be those who can think like capital allocators, understanding both the technical capabilities and the business value of different compute investments.</p><p>In this light, o3 represents the beginning of an era where artificial intelligence must be treated as a scarce resource requiring careful allocation. The organizations that thrive will be those that develop robust frameworks for deploying this resource where it can generate the highest returns.</p><p>The future of AI might look less like unlimited abundance and more like traditional capital markets, where success comes from making smart allocation decisions with limited resources. As models continue to become more powerful and computationally intensive, these allocation skills will only become more crucial.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.chrishayduk.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Musings by Chris Hayduk! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[On Algorithmic Moats and the Path to AGI]]></title><description><![CDATA[Google's path to winning the AI race]]></description><link>https://www.chrishayduk.com/p/on-algorithmic-moats-and-the-path</link><guid isPermaLink="false">https://www.chrishayduk.com/p/on-algorithmic-moats-and-the-path</guid><dc:creator><![CDATA[Chris Hayduk]]></dc:creator><pubDate>Thu, 19 Dec 2024 21:59:19 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!pLfc!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F065387d0-4f46-4d37-903f-6704a2a21ffa_1024x768.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!pLfc!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F065387d0-4f46-4d37-903f-6704a2a21ffa_1024x768.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!pLfc!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F065387d0-4f46-4d37-903f-6704a2a21ffa_1024x768.jpeg 424w, https://substackcdn.com/image/fetch/$s_!pLfc!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F065387d0-4f46-4d37-903f-6704a2a21ffa_1024x768.jpeg 848w, https://substackcdn.com/image/fetch/$s_!pLfc!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F065387d0-4f46-4d37-903f-6704a2a21ffa_1024x768.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!pLfc!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F065387d0-4f46-4d37-903f-6704a2a21ffa_1024x768.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!pLfc!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F065387d0-4f46-4d37-903f-6704a2a21ffa_1024x768.jpeg" width="1024" height="768" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/065387d0-4f46-4d37-903f-6704a2a21ffa_1024x768.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:768,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:248010,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!pLfc!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F065387d0-4f46-4d37-903f-6704a2a21ffa_1024x768.jpeg 424w, https://substackcdn.com/image/fetch/$s_!pLfc!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F065387d0-4f46-4d37-903f-6704a2a21ffa_1024x768.jpeg 848w, https://substackcdn.com/image/fetch/$s_!pLfc!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F065387d0-4f46-4d37-903f-6704a2a21ffa_1024x768.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!pLfc!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F065387d0-4f46-4d37-903f-6704a2a21ffa_1024x768.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><p>The past few weeks have provided a remarkable natural experiment in AI development dynamics. OpenAI releases what appears to be a breakthrough technology, and Google promptly demonstrates superior capabilities:</p><ul><li><p>OpenAI's Sora demonstrated remarkable text-to-video generation, only to be superseded by Google's Veo 2 with notably higher quality output</p></li><li><p>OpenAI's o1 introduced novel "thinking" capabilities, followed within weeks by Google's Gemini 2.0 Flash Thinking, implementing similar functionality</p></li><li><p>Gemini 2.0 has now surpassed both GPT-4 and Claude Sonnet across a broad range of benchmarks</p></li></ul><p>This pattern reveals something fundamental about the nature of competitive advantage in artificial intelligence. To understand why this Google dominance was inevitable, we need to examine a broader principle: the myth of algorithmic moats.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.chrishayduk.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.chrishayduk.com/subscribe?"><span>Subscribe now</span></a></p><h1>Algorithmic Moats</h1><p>It has frequently been said that part of Silicon Valley's success is the lack of non-compete clauses for employees. This allowed trade secrets to proliferate rapidly in the Bay Area, creating more efficient competition dynamics and allowing many engineers to learn from each other, rather than restricting learning and competitive advantages to a single firm.</p><p>However, I rarely see this same line of argument applied to business moats. If this holds true, then it implies that algorithms alone <em>cannot</em> provide a durable moat to a business. Employees can easily leave one company and take all of its hard-won knowledge to a competitor, allowing the competitor to catch up.</p><p>Consider a thought experiment: You discover a revolutionary new algorithm. How long can you maintain that advantage? In a world of mobile talent and reverse engineering, the half-life of algorithmic secrets approaches zero as their value approaches infinity.</p><p>This creates what we might call the algorithm diffusion principle: Any sufficiently valuable algorithm will spread through the industry at a rate proportional to its perceived importance. Silicon Valley's prohibition on non-compete clauses accelerates this process, creating an upper bound on how long any single player can maintain algorithmic superiority.</p><p>Hence, algorithms only provide moats insofar as they facilitate the construction of another type of moat. When we talk about algorithmic moats, we're really discussing two separate concepts: the technical implementation details that can be replicated, and the emergent properties that arise from being first to market with those implementations.</p><p>Consider Google's own history with PageRank. While revolutionary for its time, the core insight &#8211; that incoming links could be weighted by the importance of their source &#8211; was relatively straightforward to replicate once published. What made Google dominant wasn't PageRank itself, but rather the virtuous cycle it enabled: better search results &#8594; more users &#8594; more data &#8594; even better search results. The algorithm was merely the catalyst for building a data moat.</p><p>This pattern repeats across the technology landscape. Spotify's recommendation algorithms, while sophisticated, aren't what prevent users from switching to Apple Music or YouTube Music. Instead, it's the years of accumulated listening history, carefully curated playlists, and social sharing features that create switching costs. The algorithms enable these benefits, but they aren't the moat themselves.</p><h1>Moats on the Path to AGI</h1><p>The implications of the lack of direct algorithmic moats become clear when we consider AGI development as a function of three primary variables:</p><ol><li><p>Algorithmic innovation (A)</p></li><li><p>Computational resources (C)</p></li><li><p>Training data quality and quantity (D)</p></li></ol><p>We might express AGI capability as: AGI_capability = A * f(C,D)</p><p>Where f(C,D) represents the effective utilization of compute and data. The algorithm diffusion principle suggests that A will quickly equilibrate across major players. Therefore, the decisive factor becomes f(C,D).</p><p>This is where Google's position becomes overwhelming. Consider their structural advantages:</p><p>Data Supremacy:</p><ul><li><p>Google Search: The world's most comprehensive map of human knowledge and intent</p></li><li><p>YouTube: The largest repository of human audio-visual communication</p></li><li><p>Google Books/Scholar: A near-complete corpus of formal human knowledge</p></li><li><p>Android/Gmail: Vast behavioral and communication datasets</p></li></ul><p>Compute Dominance:</p><ul><li><p>Custom TPU architecture optimized for AI workloads</p></li><li><p>Vertical integration from silicon to software</p></li><li><p>World-class data center infrastructure</p></li><li><p>Decades of distributed systems optimization</p></li></ul><p>These advantages compound non-linearly. Having twice the data and twice the compute doesn't yield four times the capability &#8211; it might yield eight times or more due to emergent properties in large-scale systems.</p><p>The recent pattern of Google rapidly matching and exceeding OpenAI's innovations perfectly illustrates this dynamic. When OpenAI develops a new technique, Google can quickly replicate it (algorithm diffusion) and then apply it with vastly superior resources, achieving better results almost immediately.</p><p>This creates what game theorists would call a dominant strategy for Google: Wait for algorithmic innovations, replicate them with superior resources, and achieve better results than the original inventors. The math becomes almost deterministic.</p><p>One might object that breakthrough algorithms could create discontinuous advantages that trump resource differences. However, the observed scaling laws in neural networks suggest otherwise. The smooth power-law relationships we've seen indicate that resource advantages compound predictably rather than being disrupted by algorithmic breakthroughs.</p><p>In retrospect, the tech industry's focus on OpenAI and other startups represents a failure to reason from first principles. In a world where algorithmic innovations cannot be contained, the player with overwhelming advantages in compute and data will inevitably emerge victorious. Google's position isn't just strong &#8211; it's strategically dominant in a game-theoretic sense.</p><p>The universal rule of algorithmic diffusion suggests a surprising corollary: The most effective strategy for other players might not be to compete directly with Google, but rather to focus on specialized domains where Google's general-purpose advantages are less relevant. This could, ironically, lead to a more specialized and diverse AI ecosystem than many currently predict.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.chrishayduk.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Musings by Chris Hayduk! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Rejuvenating the Political System]]></title><description><![CDATA[How do we get the US government unstuck?]]></description><link>https://www.chrishayduk.com/p/rejuvinating-the-political-system</link><guid isPermaLink="false">https://www.chrishayduk.com/p/rejuvinating-the-political-system</guid><dc:creator><![CDATA[Chris Hayduk]]></dc:creator><pubDate>Mon, 22 Jul 2024 23:39:38 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!y82n!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fac188ae5-df90-466a-928e-0d204fbfb2e1_1024x698.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>At this point, there&#8217;s fairly broad bipartisan agreement that the United States government has stagnated. Our politicians are too old, our government institutions less effective, our spending less efficient than in the halcyon days of Big Government dating back to the 1940-1970 period of US history. But why has the government ossified? To answer that question, let&#8217;s start by looking at one system that has managed to continually renew itself and improve since its widespread appearance roughly 200 years ago &#8212; the free market economy.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.chrishayduk.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.chrishayduk.com/subscribe?"><span>Subscribe now</span></a></p><h2>Renewal and Markets</h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!y82n!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fac188ae5-df90-466a-928e-0d204fbfb2e1_1024x698.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!y82n!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fac188ae5-df90-466a-928e-0d204fbfb2e1_1024x698.png 424w, https://substackcdn.com/image/fetch/$s_!y82n!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fac188ae5-df90-466a-928e-0d204fbfb2e1_1024x698.png 848w, https://substackcdn.com/image/fetch/$s_!y82n!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fac188ae5-df90-466a-928e-0d204fbfb2e1_1024x698.png 1272w, https://substackcdn.com/image/fetch/$s_!y82n!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fac188ae5-df90-466a-928e-0d204fbfb2e1_1024x698.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!y82n!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fac188ae5-df90-466a-928e-0d204fbfb2e1_1024x698.png" width="1024" height="698" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/ac188ae5-df90-466a-928e-0d204fbfb2e1_1024x698.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:698,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;Capitalism is the greatest anti-poverty program the world has ever known&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Capitalism is the greatest anti-poverty program the world has ever known" title="Capitalism is the greatest anti-poverty program the world has ever known" srcset="https://substackcdn.com/image/fetch/$s_!y82n!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fac188ae5-df90-466a-928e-0d204fbfb2e1_1024x698.png 424w, https://substackcdn.com/image/fetch/$s_!y82n!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fac188ae5-df90-466a-928e-0d204fbfb2e1_1024x698.png 848w, https://substackcdn.com/image/fetch/$s_!y82n!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fac188ae5-df90-466a-928e-0d204fbfb2e1_1024x698.png 1272w, https://substackcdn.com/image/fetch/$s_!y82n!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fac188ae5-df90-466a-928e-0d204fbfb2e1_1024x698.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>For most of recorded human history, progress was nearly too small to measure along nearly all relevant axes: life expectancy, infant mortality, GDP per capita, etc. Progress along each of these vectors was measured on the order of centuries or millennia, and any hard-won progress was extremely fragile, with events such as the fall of the Western Roman Empire taking centuries to recover from in areas such as Britain. However, beginning in the late 1700s, a confluence of technological, sociological, and ideological innovations led to capital P &#8220;Progress&#8221; &#8212; the sort of rapid, hockey stick growth that we&#8217;ve come to expect as a natural part of our world. One of these key innovations, perhaps the most important of them all, was the introduction of capitalism and the free market economy. </p><p>Capitalism and free markets together unleashed several key forces that ignited global growth: competition, distributed information processing, and incentive alignment. I&#8217;ll outline briefly the importance of each of these below:</p><ul><li><p><strong>Competition. </strong>Competition is the driving force behind innovation and efficiency in capitalist economies. When companies vie for market share, they are compelled to improve their products, lower prices, and increase efficiency to attract consumers. This constant pressure to outperform rivals fosters an environment where only the most efficient and innovative businesses thrive. Competition not only benefits consumers with better and more affordable products but also pushes companies to continually innovate, ensuring that progress never stagnates.</p></li><li><p><strong>Distributed Information Processing.</strong> Information about consumer preferences, resource availability, and technological possibilities is widely distributed among individuals and businesses. This decentralized information processing allows markets to quickly adapt to changes. Entrepreneurs and businesses can respond to signals from the market, such as shifts in consumer demand or new technological advancements, enabling a dynamic and responsive economic system.</p></li><li><p><strong>Incentive Alignment. </strong>Capitalism aligns incentives by tying the success of individuals to the success of the companies they own or work for. When management and workers have a stake in the ownership of their companies, their incentives align closely with the economic growth and dynamism of the broader economy. Managers and employees who hold equity in their firms are directly rewarded for the company&#8217;s performance, encouraging them to work harder, innovate, and make decisions that enhance the company&#8217;s value. This ownership structure creates a powerful alignment of interests that drives productivity and fosters a culture of continuous improvement.</p></li></ul><p>These three factors have endowed capitalist free market economies with a remarkable capacity to enrich society through a built-in mechanism for rejuvenation. While individual companies may not endure indefinitely, the market system itself thrives and evolves, welcoming new entrants and jettisoning outdated competitors. </p><p>The cycle of rejuvenation in free markets follows these steps:</p><ol><li><p><strong>Innovation Emergence</strong>: A new technology, methodology, or market becomes available, often driven by advancements in science and technology or shifts in consumer behavior.</p></li><li><p><strong>Entrepreneurial Entry</strong>: Entrepreneurs and new companies seize the opportunity to exploit this new angle, creating products or services that meet emerging demands.</p></li><li><p><strong>Incumbent Inertia</strong>: Established companies, often characterized by bureaucratic inertia and resistance to change, struggle to adapt to the new paradigm.</p></li><li><p><strong>Competitive Displacement</strong>: The new entrants, leveraging their innovative approaches and exploiting the incumbents' weaknesses, begin to outcompete and displace these established firms.</p></li><li><p><strong>Decline of the Old Guard</strong>: Incumbent companies shrink or exit the market, unable to maintain their previous dominance.</p></li><li><p><strong>Cycle Continuation</strong>: The high-growth new companies eventually become the new incumbents, setting the stage for the next wave of innovation and disruption.</p></li></ol><p>This cycle ensures that, while no single company survives for eternity, <em>the system itself</em> survives and grows more efficient and prosperous. And, crucially, <strong>virtually everything about the companies within the system can change</strong>. They can have different governance structures, different incentive structures, leverage different technologies, target different market segments, and more. While the system itself remains in place, its constituent parts can change to the point of being unrecognizable.  This allows our economy to respond dynamically to rapidly changing technological and environmental factors and to leverage these changes for continuous improvement &#8212; that is, in the parlance of Nassim Nicholas Taleb, to be <a href="https://fs.blog/antifragile-a-definition/">Antifragile</a>.</p><h2>The Ossified Nation-State</h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!_bGK!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6160c1eb-9598-49f6-a3ea-ec0bd80bca26_1136x852.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!_bGK!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6160c1eb-9598-49f6-a3ea-ec0bd80bca26_1136x852.jpeg 424w, https://substackcdn.com/image/fetch/$s_!_bGK!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6160c1eb-9598-49f6-a3ea-ec0bd80bca26_1136x852.jpeg 848w, https://substackcdn.com/image/fetch/$s_!_bGK!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6160c1eb-9598-49f6-a3ea-ec0bd80bca26_1136x852.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!_bGK!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6160c1eb-9598-49f6-a3ea-ec0bd80bca26_1136x852.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!_bGK!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6160c1eb-9598-49f6-a3ea-ec0bd80bca26_1136x852.jpeg" width="1136" height="852" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/6160c1eb-9598-49f6-a3ea-ec0bd80bca26_1136x852.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:852,&quot;width&quot;:1136,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;Congress Today: Oldest in American History - Business Insider&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Congress Today: Oldest in American History - Business Insider" title="Congress Today: Oldest in American History - Business Insider" srcset="https://substackcdn.com/image/fetch/$s_!_bGK!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6160c1eb-9598-49f6-a3ea-ec0bd80bca26_1136x852.jpeg 424w, https://substackcdn.com/image/fetch/$s_!_bGK!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6160c1eb-9598-49f6-a3ea-ec0bd80bca26_1136x852.jpeg 848w, https://substackcdn.com/image/fetch/$s_!_bGK!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6160c1eb-9598-49f6-a3ea-ec0bd80bca26_1136x852.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!_bGK!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6160c1eb-9598-49f6-a3ea-ec0bd80bca26_1136x852.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>In stark contrast to the above, <strong>there is no equivalent process for rejuvenation and change in nation-states</strong>. Governments are often enshrined in constitutions that either remain largely unchanged (as in the United States) or undergo only superficial modifications while retaining the same overarching structure (as seen in many Latin American countries). The only mechanism for large-scale governmental change appears to be a total collapse of the existing order, which can come through: </p><ul><li><p><strong>Violent Revolution</strong>: Internal uprisings that overthrow the existing order.</p></li><li><p><strong>External Conquest</strong>: Invasions and conquests that impose new governmental structures.</p></li><li><p><strong>Post-War Reorganization</strong>: Major wars resulting in governments' dissolution and reformation, such as Germany after World War I.</p></li></ul><p>Historically, the structure of democratic republics was able to partially fill the role of free market economies in the context of nation-states. Democratic Republics such as the United States allowed voters to periodically replace presidents, cabinets, and policy agendas through regularly scheduled elections, avoiding the violent regime changes mentioned above. The combination of these regularly scheduled elections with de facto term limits on the executive branch of government created a system where new leaders, new ideas, and new policy structures were continually brought in to renew the government. However, several contemporary factors are undermining the Democratic Republic as a rejuvenating system:</p><ol><li><p><strong>Aging Politicians</strong>: Politicians are getting older and staying in office longer. Increased lifespans, coupled with the absence of term limits for Congress, contribute to this trend. Additionally, mass media and a larger, less informed voting populace make name recognition a crucial determinant of electoral success, favoring long-term incumbents.</p></li><li><p><strong>Influence of the Supreme Court</strong>: The Supreme Court plays a more significant role than in the past, as partisan interest groups frequently reach gridlock. By elevating many decisions to the Supreme Court, the nation&#8217;s decision-making power becomes more and more centralized in a group of justices who receive lifetime appointments, limiting the capacity for new views to be brought in.</p></li><li><p><strong>Persistent Bureaucracy</strong>: The bureaucratic apparatus surrounding the government has increased in size and does not change with each administration, so the personnel and ideologies making up the government are more constant from administration to administration.</p></li></ol><p>All of these factors together make the US government increasingly brittle &#8212; it is less able to change in response to rapidly changing environments, and so is likely to break. Think of the difference between a wood plank and water. Water changes in response to even tiny perturbations, but the benefit is that it will not "break" from a large impact. By contrast, a wood plank will resist most changes until one change that is large enough to overcome its resistance snaps the plank in half. Similarly, the increased rigidity of the US government has now made it harder to bend, making it more likely to break in response to extreme stress and environmental, economic, or sociological change.</p><h2>How do we move forward?</h2><p>To improve the situation, we need to ensure a much higher turnover in the government. This includes personnel, policies, legislation, and ideologies. We need to ensure that the government can respond rapidly to changing conditions, which can only occur by allowing the governance structure in place to shift without the strong <em>status quo</em> bias we see today. There are a few factors that I think need to change to allow this to happen:</p><ol><li><p><strong>Radically reducing the size and scope of government agencies with unelected officials. </strong>Bureaucratic agencies since WW2 have been able to run roughshod over the country, interpreting and enforcing laws as they see fit and being largely impervious to public scrutiny. The opacity of the organizations and the inability of voters to remove officials from them through elections has allowed them to ossify, making the government apparatus significantly more stagnant from administration to administration. The <a href="https://www.scotusblog.com/2024/06/supreme-court-strikes-down-chevron-curtailing-power-of-federal-agencies/">Supreme Court&#8217;s ruling in </a><em><a href="https://www.scotusblog.com/2024/06/supreme-court-strikes-down-chevron-curtailing-power-of-federal-agencies/">Loper Bright Enterprises v. Raimondo</a></em><a href="https://www.scotusblog.com/2024/06/supreme-court-strikes-down-chevron-curtailing-power-of-federal-agencies/"> and </a><em><a href="https://www.scotusblog.com/2024/06/supreme-court-strikes-down-chevron-curtailing-power-of-federal-agencies/">Relentless, Inc. v. Department of Commerce</a> </em>is a step in the right direction in this case, limiting the ability of agencies to actively interpret the law without much oversight (which can lead to ambiguous and shape-shifting regulatory implementations, creating huge strains on businesses). </p></li><li><p><strong>Setting expiry dates on all government legislation. </strong>Laws existing in perpetuity create strong inertia in the government &#8212; laws that already exist will likely continue to exist because keeping them requires no action, whereas repealing them requires some action. This results in ever-increasing regulations, tax codes, and legal codes, which sap dynamism from the economy and provide significant market opportunities to purely extractive companies and institutions. Both individuals and corporations would benefit from simpler tax codes and reduced regulatory burdens.</p></li><li><p><strong>Imposing term limits on Congress and the Supreme Court. </strong>Term limits would ensure a regular influx of new ideas and reduce the likelihood of long-term incumbency leading to stagnation. In particular, poor choices for appointments to the Supreme Court are disastrous, as lifetime appointments can last for several decades before the spot is turned over to a new justice. By making appointments temporary, we can limit the negative impact of bad appointments while also ensuring that new generations of justices regularly filter into the court.</p></li><li><p><strong>Implementing cognitive testing for all elected officials. </strong>This one is self-explanatory. We should not have to repeat the embarrassments of this national election cycle, where each party has competed to prove that the other&#8217;s candidate is the more geriatric and clueless. We as a nation should have the confidence (and frankly, the dignity) to be able to assume that our leaders are cognitively capable of handling the job.</p></li></ol><p>The goal of the above measures is to create a system that fosters innovation and dynamism in the government by allowing it to <em>change</em>. Each measure aims to remove bloat from the current system, increase competition for positions among elected officials, and reduce the status quo bias that is so prevalent in politics today. There is certainly substantially more that needs to be done above and beyond these suggestions, but it does present a starting point for a government that isn&#8217;t stuck doing a cheap parody of FDR&#8217;s administration.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.chrishayduk.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Musings by Chris Hayduk! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item></channel></rss>