<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[prinz]]></title><description><![CDATA[be not afraid of greatness]]></description><link>https://www.prinzai.com</link><image><url>https://www.prinzai.com/img/substack.png</url><title>prinz</title><link>https://www.prinzai.com</link></image><generator>Substack</generator><lastBuildDate>Mon, 27 Apr 2026 16:22:54 GMT</lastBuildDate><atom:link href="https://www.prinzai.com/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[prinz]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[prinz@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[prinz@substack.com]]></itunes:email><itunes:name><![CDATA[prinz]]></itunes:name></itunes:owner><itunes:author><![CDATA[prinz]]></itunes:author><googleplay:owner><![CDATA[prinz@substack.com]]></googleplay:owner><googleplay:email><![CDATA[prinz@substack.com]]></googleplay:email><googleplay:author><![CDATA[prinz]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[The Race to RSI]]></title><description><![CDATA[Spring 2026 Update]]></description><link>https://www.prinzai.com/p/the-race-to-rsi</link><guid isPermaLink="false">https://www.prinzai.com/p/the-race-to-rsi</guid><dc:creator><![CDATA[prinz]]></dc:creator><pubDate>Wed, 22 Apr 2026 20:37:26 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!iobM!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faff0a225-ce91-405b-be71-0e15c3406523_1402x1122.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!iobM!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faff0a225-ce91-405b-be71-0e15c3406523_1402x1122.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!iobM!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faff0a225-ce91-405b-be71-0e15c3406523_1402x1122.png 424w, https://substackcdn.com/image/fetch/$s_!iobM!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faff0a225-ce91-405b-be71-0e15c3406523_1402x1122.png 848w, https://substackcdn.com/image/fetch/$s_!iobM!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faff0a225-ce91-405b-be71-0e15c3406523_1402x1122.png 1272w, https://substackcdn.com/image/fetch/$s_!iobM!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faff0a225-ce91-405b-be71-0e15c3406523_1402x1122.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!iobM!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faff0a225-ce91-405b-be71-0e15c3406523_1402x1122.png" width="1402" height="1122" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/aff0a225-ce91-405b-be71-0e15c3406523_1402x1122.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1122,&quot;width&quot;:1402,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2367742,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.prinzai.com/i/195063462?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faff0a225-ce91-405b-be71-0e15c3406523_1402x1122.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!iobM!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faff0a225-ce91-405b-be71-0e15c3406523_1402x1122.png 424w, https://substackcdn.com/image/fetch/$s_!iobM!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faff0a225-ce91-405b-be71-0e15c3406523_1402x1122.png 848w, https://substackcdn.com/image/fetch/$s_!iobM!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faff0a225-ce91-405b-be71-0e15c3406523_1402x1122.png 1272w, https://substackcdn.com/image/fetch/$s_!iobM!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faff0a225-ce91-405b-be71-0e15c3406523_1402x1122.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>In January, Dario Amodei told a stunned audience at Davos that the coding agents developed by Anthropic will be used &#8220;<a href="https://x.com/deredleritt3r/status/2013613671704924640">to create the new generation of models, and speed it up, create a loop that would increase the speed of [AI] model development</a>&#8221;.  Anthropic views Claude Code as the path towards automation of AI research and, eventually, recursive self-improvement (RSI).  Similarly, <a href="https://x.com/deredleritt3r/status/2019475360438493597">OpenAI is using Codex to accelerate its own development of AI models</a> and expects that a future version of Codex will eventually become the automated AI research intern:</p><blockquote><p>Where AI researchers have great hope to help themselves... is that if you could just say &#8216;<em><strong><a href="https://www.youtube.com/watch?v=3K-R4yVjJfU">hey, Codex, this is the idea, and it&#8217;s fairly clear what I&#8217;m saying, please just implement it so it runs fast on this 8-machine setup or 100-machine setup</a></strong></em>&#8217;. I think that&#8217;s what OpenAI [means by] an AI intern by the end of [2026].</p><p>&#8212;Lukasz Kaiser, OpenAI</p></blockquote><p>OpenAI and Anthropic are racing to automate AI research and reach RSI.  But is it a two-horse race, or might any other labs join them?  Read on to find out.</p><h3>OpenAI</h3><p>OpenAI&#8217;s goal <a href="https://x.com/sama/status/1983584366547829073">announced in October 2025</a> is to develop an automated AI research intern (<em>i.e.</em>, the system as described by Lukasz Kaiser, above), running on &#8220;hundreds of thousands of GPUs&#8221;, by September 2026.  Jakub Pachocki <a href="https://www.youtube.com/watch?v=vK1qEF3a3WM">recently said</a> that, based on the improving coding capabilities of Codex, he thinks the intern is &#8220;on track&#8221; to be developed by September - <em><strong>now just 5 months away</strong></em>.  Pachocki also described the differences between the &#8220;intern&#8221; and the fully automated AI researcher (which OpenAI expects to develop by March 2028):</p><blockquote><p>The way I would distinguish a research intern from a full automated researcher is the <em><strong>span of time</strong></em> that we would have it work mostly autonomously or the <em><strong>specificity of the task</strong></em> that has to be given.  I don't expect we'll have systems where you tell them: &#8220;Go improve your model capability, go solve alignment&#8221; - and they will do it.  Not this year.  I think we might get there at some point. But for more specific technical ideas - like this particular idea how to improve the models, how to run this evaluation differently - I think we have the pieces that we mostly just need to put together.</p></blockquote><p><a href="https://www.technologyreview.com/2026/03/20/1134438/openai-is-throwing-everything-into-building-a-fully-automated-researcher/">In another interview</a>, Pachocki said that the &#8220;intern&#8221; is a system to which &#8220;you can delegate tasks that would take a person a few days&#8221;.</p><p>It is not clear whether OpenAI is deliberately being conservative with its September 2026 timeline for developing the &#8220;intern&#8221; and/or its March 2028 timeline for developing the fully automated AI researcher.  Interestingly, Sam Altman <a href="https://x.com/deredleritt3r/status/2024879807134318979">recently said</a> that &#8220;it&#8217;s going to be a faster takeoff than [he] originally thought&#8221;.</p><h3>Anthropic</h3><p>Anthropic&#8217;s publicly stated timeline for reaching fully automated AI research is significantly more aggressive than OpenAI&#8217;s.  Dario Amodei expects 2026 to &#8220;<a href="https://www.tmtbreakout.com/p/tmtb-dario-amodei-anthropic-ceo-at">have a radical acceleration that surprises everyone&#8230; I think we are on the precipice of something incredible</a>&#8221;.  According to Anthropic&#8217;s <a href="https://www.anthropic.com/responsible-scaling-policy/roadmap">Frontier Safety Roadmap</a>, released in February 2026, it is &#8220;plausible, <em><strong>as soon as early 2027</strong></em>, that [Anthropic&#8217;s] AI systems could fully automate, or otherwise dramatically accelerate, the work of large, top-tier teams of human researchers in domains [including development of] AI itself&#8221;.  Echoing this timeline, Anthropic co-founder and chief science officer Jared Kaplan <a href="https://x.com/AndrewCurran_/status/2031731035105628270">told Time magazine in March 2026</a> that fully automated Al research could be "<em><strong>as little as a year away</strong></em>".</p><p>Also in line with these predictions, Jack Clark <a href="https://x.com/jackclarkSF/status/2030417306288066780">continues to believe</a> that &#8220;a country of geniuses in a datacenter&#8221; (<em>i.e.</em>, AGI)<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> will be achievable in late 2026, and &#8220;running many copies&#8221; in 2027.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a></p><h3>Google</h3><p>Sitting across from Dario Amodei at Davos in January 2026, Demis Hassabis <a href="https://www.youtube.com/watch?v=02YLwsCKUww">was diplomatically skeptical</a> about coding models leading to RSI:</p><blockquote><p>The full closing of the loop, I think is an unknown... I think it's possible to do, you may need AGI itself to be able to do that in some domains where there's more messiness around them [and] it's not so easy to verify your answer very quickly.  There are NP-hard domains, and I also include for AGI physical AI, robotics.  And then you've got hardware in the loop that may limit how fast the self-improvement systems can work - but I think in coding and mathematics, I can definitely see that working.</p></blockquote><p>&#8220;If self-improvement doesn&#8217;t deliver the goods on its own&#8221;, Hassabis said, &#8220;then we&#8217;ll need other things to work&#8221; - <em>i.e.</em>, <a href="https://deepmind.google/models/genie/">world models</a>, <a href="https://bostondynamics.com/blog/boston-dynamics-google-deepmind-form-new-ai-partnership/">robotics</a> and continual learning.</p><p>Under Demis Hassabis&#8217; leadership, Google has indeed focused on reaching AGI via the path of developing continual learning, world models and &#8220;physical AI&#8221; (<em>i.e.</em>, robotics).  This is a <em><strong>vastly different path</strong></em> from that currently being pursued by OpenAI and Anthropic.  Demis Hassabis estimates that building AI on this path <a href="https://www.youtube.com/watch?v=YvT5aaa0r7Q">will result in Google achieving AGI in 5 to 10 years</a>.</p><p>But is there a change coming at Google?  On April 20, 2026, <a href="https://x.com/Yuchenj_UW/status/2046246166871089438">The Information reported</a> that Sergey Brin has formed a &#8220;strike team&#8221; to improve Google&#8217;s coding models.  &#8220;The end goal&#8221;, the article reads, &#8220;is AI takeoff or AI that can improve itself&#8230; Brin has told staffers that improving Google AI&#8217;s coding abilities is a step toward that eventual goal.&#8221;</p><p>Is Google joining the race?  If so, will it throw enough compute and other resources at the problem so as to actually be able to catch up with Anthropic and OpenAI?  We will find out over the next few months.</p><h3>xAI</h3><p>There is no publicly available evidence to date that xAI is focused on achieving automated AI research or RSI.  In January 2026, <a href="https://x.com/kyliebytes/status/2009686466746822731">reports emerged</a> that xAI&#8217;s team was using Anthropic&#8217;s models through Cursor instead of using Grok.  And when co-founder Jimmy Ba left xAI earlier this year, he <a href="https://x.com/jimmybajimmyba/status/2021374875793801447">tweeted</a> that he was leaving to &#8220;recalibrate his gradient on the big picture&#8221; because &#8220;[r]ecursive self improvement loops likely go live in the next 12mo.&#8221;  </p><p>In response, Elon Musk has chosen to focus his energies on <a href="https://x.com/xDaily/status/2025239936853819709">beating Anthropic on coding capabilities</a>.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a>. On April 21, 2026, SpaceX <a href="https://x.com/SpaceX/status/2046713419978453374">announced</a> that it will be &#8220;working closely together&#8221; with Cursor &#8220;to create the world&#8217;s best coding and knowledge work AI&#8221;.  As part of the deal, SpaceX will either pay Cursor $10B for this collaboration in 2026 or, at its option, will purchase Cursor for $60B.  </p><p>Does xAI&#8217;s senior leadership realize that coding models and the resulting enterprise revenue are merely an (extremely useful) milestone on the path to RSI, or is xAI&#8217;s goal limited to mimicking Anthropic&#8217;s success in delivering agentic models to enterprises? Time will tell.</p><h3>Meta</h3><p>The Meta Superintelligence Labs team, assembled at a great cost in mid-2025, has been working for many months on developing new AI models and related tools.  Thus far, these efforts have culminated only in the release of <a href="https://ai.meta.com/blog/introducing-muse-spark-msl/">Muse Spark</a>, a new reasoning model.</p><p>There is no publicly available evidence currently that Meta is focusing any attention on coding models, automation of AI research or RSI.  </p><h3>Microsoft</h3><p>An underrated player in the quest for AGI, Microsoft has a major trump card up its sleeve: <a href="https://substack.com/home/post/p-178862567">its licensing deal with OpenAI</a> means that Microsoft has rights to OpenAI&#8217;s &#8220;research IP&#8221; (including models intended for internal deployment or research only - which should include automated AI research models) until the earlier of 2030 or verification of OpenAI&#8217;s declaration of AGI by an independent expert panel.  After OpenAI declares AGI, we can expect Microsoft to use this research IP to undertake its own quest for superintelligence.  </p><p>Possibly in preparation for this move (at least in part), Microsoft has been <a href="https://x.com/satyanadella/status/2044767391293509761">aggressively expanding its data center capacity</a>.</p><h3>Others?</h3><p>DeepSeek has been suspiciously quiet recently, with no major model releases since December 2025.  Given DeepSeek&#8217;s technical prowess and taste, it would not be surprising to the author if it turned out that DeepSeek has built its own coding model internally and is using it to accelerate its own coding and AI research capabilities.  These efforts, if they indeed do exist, may be significantly constrained by compute availability and other factors.  </p><p>For now, the smaller U.S. labs and &#8220;neolabs&#8221; (<em>e.g.</em>, SSI, Thinking Machines, Core Automation - just to name a few) generally appear to be headed in different directions from the one chosen by OpenAI and Anthropic.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a></p><p>Finally, time will tell whether any other Chinese labs are willing and able to join the race to RSI.</p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>As defined in Dario Amodei&#8217;s &#8220;<a href="https://darioamodei.com/essay/machines-of-loving-grace">Machines of Loving Grace</a>&#8221;.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>Jack kindly provided a detailed explanation of why he thinks this timeline is plausible <a href="https://x.com/jackclarkSF/status/2030420665569092091">in a thread on X</a>, which is well worth reading in its entirety.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>The goal, as stated by Musk in February 2026, was to &#8220;get pretty close [to Anthropic] by April, and roughly similar by May, so probably better by June&#8221;.  As of the date of this article (April 22, 2026), this stated goal remains elusive.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>This is understandable because, among other things, automation of AI research will likely be a very compute-intensive path.  For example, as noted above, OpenAI&#8217;s automated AI research intern will likely be running on hundreds of thousands of GPUs.</p></div></div>]]></content:encoded></item><item><title><![CDATA[Why I think AI will kill BigLaw]]></title><description><![CDATA[I&#8216;ve been asked to expand on my tweet - why do I believe that BigLaw as a concept will not survive the arrival of powerful AI?]]></description><link>https://www.prinzai.com/p/why-i-think-ai-will-kill-biglaw</link><guid isPermaLink="false">https://www.prinzai.com/p/why-i-think-ai-will-kill-biglaw</guid><dc:creator><![CDATA[prinz]]></dc:creator><pubDate>Wed, 11 Mar 2026 02:49:27 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!r4_L!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc85bb738-73ee-428a-a8a1-2aaaa690ada2_1600x640.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!r4_L!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc85bb738-73ee-428a-a8a1-2aaaa690ada2_1600x640.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!r4_L!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc85bb738-73ee-428a-a8a1-2aaaa690ada2_1600x640.jpeg 424w, https://substackcdn.com/image/fetch/$s_!r4_L!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc85bb738-73ee-428a-a8a1-2aaaa690ada2_1600x640.jpeg 848w, https://substackcdn.com/image/fetch/$s_!r4_L!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc85bb738-73ee-428a-a8a1-2aaaa690ada2_1600x640.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!r4_L!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc85bb738-73ee-428a-a8a1-2aaaa690ada2_1600x640.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!r4_L!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc85bb738-73ee-428a-a8a1-2aaaa690ada2_1600x640.jpeg" width="1456" height="582" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c85bb738-73ee-428a-a8a1-2aaaa690ada2_1600x640.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:582,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:150397,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.prinzai.com/i/190578939?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc85bb738-73ee-428a-a8a1-2aaaa690ada2_1600x640.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!r4_L!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc85bb738-73ee-428a-a8a1-2aaaa690ada2_1600x640.jpeg 424w, https://substackcdn.com/image/fetch/$s_!r4_L!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc85bb738-73ee-428a-a8a1-2aaaa690ada2_1600x640.jpeg 848w, https://substackcdn.com/image/fetch/$s_!r4_L!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc85bb738-73ee-428a-a8a1-2aaaa690ada2_1600x640.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!r4_L!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc85bb738-73ee-428a-a8a1-2aaaa690ada2_1600x640.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>I&#8216;ve been asked to expand on <a href="https://x.com/deredleritt3r/status/2031526762211860535">my tweet</a> - why do I believe that BigLaw as a concept will not survive the arrival of powerful AI?</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Jk2v!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9992ef42-56f4-40ba-ac12-a541daf4d26b_599x169.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Jk2v!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9992ef42-56f4-40ba-ac12-a541daf4d26b_599x169.png 424w, https://substackcdn.com/image/fetch/$s_!Jk2v!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9992ef42-56f4-40ba-ac12-a541daf4d26b_599x169.png 848w, https://substackcdn.com/image/fetch/$s_!Jk2v!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9992ef42-56f4-40ba-ac12-a541daf4d26b_599x169.png 1272w, https://substackcdn.com/image/fetch/$s_!Jk2v!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9992ef42-56f4-40ba-ac12-a541daf4d26b_599x169.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Jk2v!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9992ef42-56f4-40ba-ac12-a541daf4d26b_599x169.png" width="599" height="169" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/9992ef42-56f4-40ba-ac12-a541daf4d26b_599x169.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:169,&quot;width&quot;:599,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:20521,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.prinzai.com/i/190578939?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9992ef42-56f4-40ba-ac12-a541daf4d26b_599x169.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Jk2v!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9992ef42-56f4-40ba-ac12-a541daf4d26b_599x169.png 424w, https://substackcdn.com/image/fetch/$s_!Jk2v!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9992ef42-56f4-40ba-ac12-a541daf4d26b_599x169.png 848w, https://substackcdn.com/image/fetch/$s_!Jk2v!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9992ef42-56f4-40ba-ac12-a541daf4d26b_599x169.png 1272w, https://substackcdn.com/image/fetch/$s_!Jk2v!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9992ef42-56f4-40ba-ac12-a541daf4d26b_599x169.png 1456w" sizes="100vw"></picture><div></div></div></a></figure></div><h2><strong>Why Hire BigLaw?</strong></h2><p>I&#8217;d say there are three main reasons people hire a BigLaw firm:</p><ol><li><p>The client needs customized or specialized advice about something.  For example, the client&#8217;s in-house lawyers might know how to draft a simple vendor agreement, but not how to navigate a complex M&amp;A deal with tax considerations, regulatory implications, complicated transaction mechanics, etc.</p></li><li><p>The client needs someone to do a large amount of legal work, potentially under a tight time frame.  For example, a team of a few in-house lawyers would not be able to review thousands of documents dumped into a data room on a Friday in a timely manner - but a BigLaw firm will staff a dozen associates on this if needed, and they&#8217;ll work around the clock to finish the task by any deadline, no matter how unreasonable.</p></li><li><p>The client needs advice in a matter that is high-stakes, or needs to be blessed by competent outside counsel, or involves a counterparty that is using another BigLaw firm.  Big M&amp;A transactions, bet-the-company litigation, internal investigations, sensitive matters requiring establishment of a special committee all fall within this category.</p></li></ol><p>Often, a matter will meet more than one of these criteria (and potentially will meet all three).</p><p>The BigLaw model involves one key partner (or a few key partners) providing mostly strategic advice to the client, plus a team of associates supporting the partners&#8217; work.  The associates do the research and draft a legal memo, a junior partner reviews it, the final work product goes to the senior partner who glances at it and maybe distills it down to a few talking points for the client&#8217;s GC (the GC will not read the memo, but will listen to these talking points).</p><h2><strong>How does AI impact all this?</strong></h2><p><strong>Category 1 (specialized advice)</strong>:  GPT-5.x Pro knows the tax laws and the regulatory implications.  I also think we&#8217;re not far away from an AI harness that would enable a SOTA AI model to succinctly (and fairly quickly) summarize all legal implications of a particular fact pattern (e.g., in a legal memo) or implement them via contract language.  LLMs would probably need to learn how to &#8220;write like a lawyer&#8221; a bit better in order to achieve this, but I don&#8217;t view this obstacle as being insurmountable.  The remaining human role in this process: (1) verifying the LLM&#8217;s output, (2) providing high-level strategic advice, and (3) going &#8220;beyond the law&#8221; to things like unwritten regulatory requirements, market practices, etc.  All of this can be done by a team of senior-partner-level people; associates are not required.</p><p><strong>Category 2 (high-volume work)</strong>:  AI works much faster than humans, doesn&#8217;t need to take breaks, doesn&#8217;t sleep, can work for many hours at a time, &#8216;nuff said.  The remaining human role in this process: verifying the LLM&#8217;s input by doing things like double-checking summaries of key documents, reviewing a sampling of documents to make sure the human reviewer agrees with the LLM&#8217;s conclusion, etc.  This can be done by a small team of in-house counsel, with maybe some input from senior-partner-level people in a law firm.</p><p><strong>Category 3 (high-profile work)</strong>:  This is where BigLaw firms will continue to dominate regardless of AI... at least for a while.  Yet, slowly but surely, the law firms&#8217; work will be eroded.  No, don&#8217;t do the diligence; do only specified spot-checking and a high-level review of our (the client&#8217;s) AI&#8217;s findings.  No, don&#8217;t draft the Merger Agreement; we&#8217;ll send you something we put together ourselves for review.  No, don&#8217;t do the research; we&#8217;ll send you something our AI put together to double-check.  Eventually, the value typically brought by BigLaw to these kinds of transactions (i.e., leveraging an army of associates to do the gruntwork) disappears, and it starts looking more and more like the client is hiring a particular partner (or team of partners) to do the entire matter.  The leverage starts disappearing.</p><p>At some point, the best partners start asking themselves why they should share their profits with less successful partners - including those who have not adapted to the age of AI.  And so, the senior partners will slowly start leaving to start up their own boutique law firms.  The client gets a great deal: Bob Jones, formerly the top deal-maker at Cravath, is still handling the client&#8217;s work, but charges the client only X% of what Cravath would charge - and the work is faster while its quality is better.  The boutique law firm consists of Bob Jones, maybe a few specialist colleagues (tax, regulatory,  etc.), and a few junior lawyers, paralegals and/or IT professionals whose main job is to work with AI models to rapidly produce high-quality legal work.  On the client&#8217;s side, just a few in-house lawyers now handle not only the work that a much larger in-house legal team used to do, but also a good portion of work that was formerly done by BigLaw.</p><p><em><strong>There is no longer any room for BigLaw in this paradigm, and BigLaw firms start disappearing.</strong></em></p><p>The timing for all this is extremely uncertain.  The legal industry moves slowly.  Lawyers are extremely non-technical.  I&#8217;d venture to guess that 99% of lawyers today don&#8217;t know the power of GPT-5.4 Pro.  This will eventually change, but how long is &#8220;eventually&#8221;?  And when will the clients begin to understand, begin to truly internalize, the transformative impact that AI can have on the practice of giving legal advice?</p><p>Could be 2 years, or could be 10.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a><em> </em></p><p><em>This article <a href="https://x.com/deredleritt3r/status/2031558138562859510">also appeared on Twitter</a>.</em></p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>I expect that some people will argue that AGI will be able to do 100% of all legal work.  Well, yes - that very well might be true... but when that comes to pass, we&#8217;ll probably be living under conditions of post-scarcity anyway.  In other words, if AI is performing all legal work, then it is probably also performing all <em>other</em> economically valuable work, and there is therefore nothing left for humans to do but take up hobbies or talk to each other on Twitter all day.  I hope you&#8217;ll still find me here when this new reality comes to pass!</p></div></div>]]></content:encoded></item><item><title><![CDATA[Core Automation]]></title><description><![CDATA[Jerry Tworek, OpenAI&#8217;s former VP of Research who left the company earlier this month, is raising for his new start-up, Core Automation.]]></description><link>https://www.prinzai.com/p/core-automation</link><guid isPermaLink="false">https://www.prinzai.com/p/core-automation</guid><dc:creator><![CDATA[prinz]]></dc:creator><pubDate>Thu, 29 Jan 2026 02:50:52 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!S6o9!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c269381-7497-409d-8214-29d8542409f6_1782x975.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!S6o9!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c269381-7497-409d-8214-29d8542409f6_1782x975.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!S6o9!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c269381-7497-409d-8214-29d8542409f6_1782x975.png 424w, https://substackcdn.com/image/fetch/$s_!S6o9!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c269381-7497-409d-8214-29d8542409f6_1782x975.png 848w, https://substackcdn.com/image/fetch/$s_!S6o9!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c269381-7497-409d-8214-29d8542409f6_1782x975.png 1272w, https://substackcdn.com/image/fetch/$s_!S6o9!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c269381-7497-409d-8214-29d8542409f6_1782x975.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!S6o9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c269381-7497-409d-8214-29d8542409f6_1782x975.png" width="1456" height="797" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/9c269381-7497-409d-8214-29d8542409f6_1782x975.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:797,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1177845,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.prinzai.com/i/186149058?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c269381-7497-409d-8214-29d8542409f6_1782x975.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!S6o9!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c269381-7497-409d-8214-29d8542409f6_1782x975.png 424w, https://substackcdn.com/image/fetch/$s_!S6o9!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c269381-7497-409d-8214-29d8542409f6_1782x975.png 848w, https://substackcdn.com/image/fetch/$s_!S6o9!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c269381-7497-409d-8214-29d8542409f6_1782x975.png 1272w, https://substackcdn.com/image/fetch/$s_!S6o9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c269381-7497-409d-8214-29d8542409f6_1782x975.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Jerry Tworek, OpenAI&#8217;s former VP of Research who left the company earlier this month, <a href="https://www.theinformation.com/articles/ex-openai-researchers-startup-targets-1-billion-funding-develop-new-type-ai">is raising for his new start-up, Core Automation</a>.</p><p>The start-up&#8217;s roadmap is ambitious, and starts with nothing less than developing a new AI architecture to be used in lieu of the transformer.  Standard methods for training models &#8220;up to and including gradient descent&#8221; will go out the window.  A new model named Ceres (after the Roman goddess of fertility) will be trained using these new methods.  The training process will be hyper-efficient, using 100x less data than today&#8217;s frontier models.  And Ceres will be able to learn through real-world experience - because Core Automation also intends to crack continual learning.</p><p>As if that weren&#8217;t enough, Core Automation&#8217;s goals after developing Ceres include automating development of future AI products, constructing self-replicating factories, and &#8220;potentially building biomachines to automatically create custom designs - or even terraform planets&#8221;.</p><p>We know that Ilya Sutskever thinks that he&#8217;ll be able to crack continual learning <a href="https://www.dwarkesh.com/p/ilya-sutskever-2#:~:text=Dwarkesh%20Patel%2001%3A22%3A28,Mhm.">in 5 to 20 years</a>.  Jerry doesn&#8217;t have that kind of time.  After all, his former employer <a href="https://www.prinzai.com/p/what-we-know-about-openais-autonomous">intends to fully automate AI research by March 2028</a>, which might lead to recursive self-improvement (RSI).  With RSI also squarely on Core Automation&#8217;s roadmap (what did you think &#8220;automating development of future AI products&#8221; meant? vibes? papers? essays?), the company will need to execute on its goals very quickly - potentially within months! - lest the likes of OpenAI and Anthropic achieve RSI first and scupper its plans.</p><p>After all, the bet on automated AI research ultimately means betting on the bitter lesson - <em>i.e.,</em> that &#8220;<a href="https://www.cs.utexas.edu/~eunsol/courses/data/bitter_lesson.pdf">general methods that leverage computation are ultimately the most effective, and by a large margin</a>&#8221;.  Once the GPUs powering OpenAI&#8217;s and Anthropic&#8217;s automated AI researchers start humming, it appears quite possible that human-led AI research will quickly fall by the wayside and that the big unsolved problems of AI (such as continual learning) will become much more easily solvable if desired.  Viewed from this perspective, the key dilemma emerges: <em><strong>given the race between the frontier labs to fully automate AI research and potentially achieve RSI, does it still make sense to &#8220;front-load&#8221; human-led research of new AI architectures and transformative ideas, or is it more prudent to instead curtail these initiatives and spend the freed-up resources on reaching the goal of automating AI research even faster?</strong></em></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!4ZAy!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fafc19f9d-5ecf-46e0-96ea-f77cfce4b06b_600x326.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!4ZAy!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fafc19f9d-5ecf-46e0-96ea-f77cfce4b06b_600x326.png 424w, https://substackcdn.com/image/fetch/$s_!4ZAy!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fafc19f9d-5ecf-46e0-96ea-f77cfce4b06b_600x326.png 848w, https://substackcdn.com/image/fetch/$s_!4ZAy!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fafc19f9d-5ecf-46e0-96ea-f77cfce4b06b_600x326.png 1272w, https://substackcdn.com/image/fetch/$s_!4ZAy!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fafc19f9d-5ecf-46e0-96ea-f77cfce4b06b_600x326.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!4ZAy!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fafc19f9d-5ecf-46e0-96ea-f77cfce4b06b_600x326.png" width="600" height="326" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/afc19f9d-5ecf-46e0-96ea-f77cfce4b06b_600x326.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:326,&quot;width&quot;:600,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:25672,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.prinzai.com/i/186149058?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fafc19f9d-5ecf-46e0-96ea-f77cfce4b06b_600x326.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!4ZAy!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fafc19f9d-5ecf-46e0-96ea-f77cfce4b06b_600x326.png 424w, https://substackcdn.com/image/fetch/$s_!4ZAy!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fafc19f9d-5ecf-46e0-96ea-f77cfce4b06b_600x326.png 848w, https://substackcdn.com/image/fetch/$s_!4ZAy!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fafc19f9d-5ecf-46e0-96ea-f77cfce4b06b_600x326.png 1272w, https://substackcdn.com/image/fetch/$s_!4ZAy!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fafc19f9d-5ecf-46e0-96ea-f77cfce4b06b_600x326.png 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Jerry Tworek&#8217;s bet is on humans for another few years.  His former employer&#8217;s bet is exactly the opposite.  </p><p>We shall soon find out which one of the two is right.</p>]]></content:encoded></item><item><title><![CDATA[The Gentle Singularity; The Fast Takeoff]]></title><description><![CDATA[On June 10, 2025, Sam Altman published a blog post entitled &#8220;The Gentle Singularity&#8221;, in which he wrote that &#8220;[w]e are past the event horizon; the takeoff has started&#8221;.]]></description><link>https://www.prinzai.com/p/the-gentle-singularity-the-fast-takeoff</link><guid isPermaLink="false">https://www.prinzai.com/p/the-gentle-singularity-the-fast-takeoff</guid><dc:creator><![CDATA[prinz]]></dc:creator><pubDate>Sat, 10 Jan 2026 04:25:13 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/9dd5f979-b7cb-41eb-ba49-c0570f3e3ca6_1368x798.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>On June 10, 2025, Sam Altman published a blog post entitled &#8220;<a href="https://blog.samaltman.com/the-gentle-singularity">The Gentle Singularity</a>&#8221;, in which he wrote that &#8220;[w]e are past the event horizon; the takeoff has started&#8221;.</p><p>This blog post gathered some attention, and its ideas have since been mindlessly copied by others.  Mark Zuckerberg <a href="https://www.meta.com/superintelligence/">claimed</a> a few days later that &#8220;[o]ver the last few months we have begun to see glimpses of our AI systems improving themselves&#8221;.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a>  More recently, <a href="https://finance.yahoo.com/news/elon-musk-says-entered-singularity-185946780.html">Elon Musk</a>, too, said that we have entered the singularity.</p><p>It has typically been assumed that these claims have been principally driven by the generally fast rate of improvement in AI models (<em>i.e.</em>, &#8220;AI is improving fast today; AI will improve even faster tomorrow&#8221;).  With respect to Altman&#8217;s claims specifically, I am of a different view.  I believe that Altman meant something <em><strong>very specific</strong></em> when he said that &#8220;we are past the event horizon&#8221;, and that this &#8220;something&#8221; is <em><strong>the most important thing happening in AI today</strong></em>.  </p><h4>Codex</h4><p>On May 16, 2025 (a few weeks before Altman&#8217;s blog post), OpenAI <a href="https://openai.com/index/introducing-codex/">released</a> its agentic coding tool, Codex.  The release flew a bit under the radar, overshadowed by the previous month&#8217;s release of o3 and endless speculation about the then-impending releases of o3-pro and OpenAI&#8217;s open-source models.  But no matter.  The coding agent, which was OpenAI&#8217;s answer to <a href="https://www.anthropic.com/news/claude-3-7-sonnet">Claude Code, released just three months earlier</a>, was merely the first step on OpenAI&#8217;s path to <em><strong>full automation of AI research</strong></em>.</p><p>OpenAI likely set out on this path in or around March 2025, just a few weeks after Anthropic&#8217;s release of Claude Code.  This is why OpenAI&#8217;s Preparedness Framework <a href="https://www.prinzai.com/p/why-openai-needs-to-gain-confidence">was updated to include recursive self-improvement (RSI) as a Tracked Category in April 2025</a>.  Other circumstantial evidence also points to the project&#8217;s launch in March 2025: OpenAI&#8217;s goal of <a href="https://www.prinzai.com/p/what-we-know-about-openais-autonomous">developing a fully automated AI researcher</a> falls exactly three years later (March 2028), and its mid-way goal of developing an automated AI research &#8220;intern&#8221; falls exactly mid-way through this three-year process (September 2026, or 18 months after March 2025).</p><p>Even OpenAI insiders were initially not convinced by Codex until a much more powerful version arrived with August&#8217;s release of GPT-5:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!sTKk!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faf4f427c-3444-415c-9944-e329db19df11_1188x536.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!sTKk!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faf4f427c-3444-415c-9944-e329db19df11_1188x536.png 424w, https://substackcdn.com/image/fetch/$s_!sTKk!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faf4f427c-3444-415c-9944-e329db19df11_1188x536.png 848w, https://substackcdn.com/image/fetch/$s_!sTKk!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faf4f427c-3444-415c-9944-e329db19df11_1188x536.png 1272w, https://substackcdn.com/image/fetch/$s_!sTKk!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faf4f427c-3444-415c-9944-e329db19df11_1188x536.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!sTKk!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faf4f427c-3444-415c-9944-e329db19df11_1188x536.png" width="1188" height="536" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/af4f427c-3444-415c-9944-e329db19df11_1188x536.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:536,&quot;width&quot;:1188,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:121870,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.prinzai.com/i/184085276?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faf4f427c-3444-415c-9944-e329db19df11_1188x536.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!sTKk!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faf4f427c-3444-415c-9944-e329db19df11_1188x536.png 424w, https://substackcdn.com/image/fetch/$s_!sTKk!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faf4f427c-3444-415c-9944-e329db19df11_1188x536.png 848w, https://substackcdn.com/image/fetch/$s_!sTKk!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faf4f427c-3444-415c-9944-e329db19df11_1188x536.png 1272w, https://substackcdn.com/image/fetch/$s_!sTKk!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faf4f427c-3444-415c-9944-e329db19df11_1188x536.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">roon links Codex to &#8220;the takeoff&#8221;</figcaption></figure></div><p>By September 2025, OpenAI began leaking that an automated AI researcher has become <em><strong>the focus of its entire research program</strong></em>.  Here&#8217;s Jacub Pachocki explaining that OpenAI has been building most of its projects <a href="https://www.youtube.com/watch?v=KSgPNVmZ8jQ">with the goal of achieving an automated AI researcher</a>:</p><blockquote><p>Our <em>set goal for our research program has been getting to an automated researcher</em> for a couple years now. And so <em>we&#8217;ve been building most our projects with this goal in mind</em>.</p></blockquote><p>The following month, OpenAI <a href="https://x.com/sama/status/1983584366547829073">officially announced to the world</a> that it is focusing on developing the automated AI research &#8220;intern&#8221; by September 2026 and the fully automated AI researcher by March 2028.  Sam Altman added that the &#8220;intern&#8221; will run on hundreds of thousands of GPUs.</p><p>Since this announcement, OpenAI has repeatedly stressed that automated AI research is now its primary focus.  &#8220;We&#8217;re very excited about <em>our 2026 roadmap and advancing work toward an automated scientist,</em>&#8221; Mark Chen <a href="https://x.com/deredleritt3r/status/2009408451990913121">said</a> just yesterday.</p><p>Again, the path towards fully automated AI research starts with Codex.  This is clear, <em>e.g.</em>, from <a href="https://www.youtube.com/watch?v=3K-R4yVjJfU">this description of the &#8220;intern&#8221; from Lukasz Kaiser</a>:</p><blockquote><p>Where AI researchers have great hope to help themselves... is that if you could just say &#8216;<em><strong>hey, Codex, this is the idea, </strong></em>and it&#8217;s fairly clear what I&#8217;m saying, please just implement it so it runs fast on this 8-machine setup or 100-machine setup&#8217;. I think that&#8217;s what OpenAI [means by] an AI intern by the end of next year.</p></blockquote><h4>Claude Code</h4><p>Not surprisingly, Anthropic views Claude Code in exactly the same way as OpenAI views Codex - <em>i.e.</em>, as a coding tool that will eventually lead to automation of AI research.  Indeed, <a href="https://assets.anthropic.com/m/12f214efcc2f457a/original/Claude-Sonnet-4-5-System-Card.pdf">Sonnet 4.5</a> and <a href="https://assets.anthropic.com/m/64823ba7485345a7/Claude-Opus-4-5-System-Card.pdf">Opus 4.5</a> system cards conspicuously included results of surveys of Anthropic employees designed to evaluate whether the model, paired with Claude Code, is good enough to fully replace a junior AI researcher.  In the Opus 4.5 survey, two (2) out of 18 participants classified Opus 4.5 as a &#8220;near-complete entry-level researcher replacement&#8221; - albeit with &#8220;meaningful caveats&#8221;.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!rbM2!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa02c0b84-36f4-4470-84ff-aed2d16773fb_648x197.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!rbM2!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa02c0b84-36f4-4470-84ff-aed2d16773fb_648x197.jpeg 424w, https://substackcdn.com/image/fetch/$s_!rbM2!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa02c0b84-36f4-4470-84ff-aed2d16773fb_648x197.jpeg 848w, https://substackcdn.com/image/fetch/$s_!rbM2!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa02c0b84-36f4-4470-84ff-aed2d16773fb_648x197.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!rbM2!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa02c0b84-36f4-4470-84ff-aed2d16773fb_648x197.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!rbM2!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa02c0b84-36f4-4470-84ff-aed2d16773fb_648x197.jpeg" width="648" height="197" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/a02c0b84-36f4-4470-84ff-aed2d16773fb_648x197.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:197,&quot;width&quot;:648,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:35107,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.prinzai.com/i/184085276?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa02c0b84-36f4-4470-84ff-aed2d16773fb_648x197.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!rbM2!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa02c0b84-36f4-4470-84ff-aed2d16773fb_648x197.jpeg 424w, https://substackcdn.com/image/fetch/$s_!rbM2!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa02c0b84-36f4-4470-84ff-aed2d16773fb_648x197.jpeg 848w, https://substackcdn.com/image/fetch/$s_!rbM2!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa02c0b84-36f4-4470-84ff-aed2d16773fb_648x197.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!rbM2!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa02c0b84-36f4-4470-84ff-aed2d16773fb_648x197.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>This is also why <a href="https://www.youtube.com/watch?v=quCu1lJOL40&amp;t=3185s">we&#8217;ve heard Sholto Douglas speak</a> about withholding models with capabilities to perform AI research from Anthropic&#8217;s competitors:</p><blockquote><p>As AI models get better at [machine learning research tasks], I do expect the labs to to hold back some of the the capabilities. If a model's capable of writing out a whole new architecture that's a lot better, even if it's just capable of writing all their kernels for them, you probably don't want to release that to your competitors.</p></blockquote><p>And what is &#8220;the main thing&#8221; that Jack Clark worries about these days?  But of course, <a href="https://x.com/deredleritt3r/status/2009783956669935752">closing the loop on AI R&amp;D, which would lead to RSI</a>:</p><blockquote><p>The main thing I worry about is whether people succeed at 'building AI that builds AI'&#8212;fully closing the loop on AI R&amp;D (sometimes called recursively self-improving AI).</p></blockquote><p>Clark notes that &#8220;extremely early signs&#8221; of AI getting better at doing components of AI research can already be seen, &#8220;ranging from kernel development to autonomously fine-tuning open-weight models&#8221;.  </p><h4>Recursive Self-Improvement and the Takeoff</h4><p>But why does automating AI research matter?  Turning again to Jack Clark, the key is &#8220;<em><strong>compounding R&amp;D advantage</strong></em>&#8221; from automated AI research.  The premise is that an AI researcher would be able to build an even better (and smarter) AI researcher, which, in turn, would be able to build yet another better and smarter AI researcher.  Automated intelligence could quickly lead to automated superintelligence - and, eventually, to systems so much smarter than humans that a human researcher would not be able to even understand the new discoveries being made by AI, much less keep up with it:</p><blockquote><p>If this stuff keeps getting better and you end up building an AI system that can build itself, then AI development would speed up very dramatically and probably become harder for people to understand.</p></blockquote><p><em><strong>This is &#8220;the takeoff&#8221;.</strong></em></p><h4>Parting Thoughts on the Race to AGI</h4><p>The above considerations lead us to the most critical insight of all vis-a-vis the race to AGI.  If OpenAI and/or Anthropic succeed in fully automating AI research, there is a chance that the &#8220;takeoff&#8221; shall occur, with the result that no other lab shall ever be able to catch up to models built by these labs.  In the &#8220;takeoff&#8221; scenario, even a large team staffed with the very best human AI researchers will never be able to compete with a model capable of superhuman AI research, and the advantages will only compound from there.  A rival lab might reach automated AI research at a later date, but by then it will be too late - its model will not be able to compete with the much more advanced AI researchers compounding faster being operated by the other labs.</p><p>Assuming that one believes in this version of the take-off,<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a> that should lead one to also believe that a chasm is already developing between those labs that are racing to automate AI research and those that <a href="https://x.com/deredleritt3r/status/2009695839091118170">are</a> <a href="https://x.com/8teAPi/status/2007252568427376954">not</a>.</p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>Zuckerberg was later forced to admit that this referred not to RSI, but rather to <a href="https://x.com/deredleritt3r/status/1968758825231487295">to an autonomous agent built by Llama 4 that had successfully checked in some changes to the Facebook algorithm</a>.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>To be clear, there are plenty of reasons to doubt that the takeoff will occur in exactly this fashion.  For example, speed of automated AI research may wind up being bounded by compute or energy constraints.  Alternatively, it is possible that AI models will not recursively self-improve infinitely, but instead will quickly reach some upper bound of intelligence - in which case it will not take too much time for others to catch up to their level.  Finally, the goal of automating AI research may prove to be a red herring, an expensive mistake not leading to capabilities significantly better than those of a human researcher.  </p></div></div>]]></content:encoded></item><item><title><![CDATA[Why OpenAI needs to "gain confidence in the safety of running systems that can self-improve"]]></title><description><![CDATA[Sam Altman caused some commotion on X today with his post that the new Head of Preparedness role at OpenAI would be responsible, inter alia, for &#8220;gain[ing] confidence in the safety of running systems that can self-improve&#8221;.]]></description><link>https://www.prinzai.com/p/why-openai-needs-to-gain-confidence</link><guid isPermaLink="false">https://www.prinzai.com/p/why-openai-needs-to-gain-confidence</guid><dc:creator><![CDATA[prinz]]></dc:creator><pubDate>Sun, 28 Dec 2025 02:44:05 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!knSw!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2c90e9ab-4e69-4d2e-9aba-2bfd34a15e4a_1402x788.webp" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!knSw!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2c90e9ab-4e69-4d2e-9aba-2bfd34a15e4a_1402x788.webp" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!knSw!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2c90e9ab-4e69-4d2e-9aba-2bfd34a15e4a_1402x788.webp 424w, https://substackcdn.com/image/fetch/$s_!knSw!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2c90e9ab-4e69-4d2e-9aba-2bfd34a15e4a_1402x788.webp 848w, https://substackcdn.com/image/fetch/$s_!knSw!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2c90e9ab-4e69-4d2e-9aba-2bfd34a15e4a_1402x788.webp 1272w, https://substackcdn.com/image/fetch/$s_!knSw!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2c90e9ab-4e69-4d2e-9aba-2bfd34a15e4a_1402x788.webp 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!knSw!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2c90e9ab-4e69-4d2e-9aba-2bfd34a15e4a_1402x788.webp" width="1402" height="788" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/2c90e9ab-4e69-4d2e-9aba-2bfd34a15e4a_1402x788.webp&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:788,&quot;width&quot;:1402,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:247548,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/webp&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.prinzai.com/i/182739228?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2c90e9ab-4e69-4d2e-9aba-2bfd34a15e4a_1402x788.webp&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!knSw!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2c90e9ab-4e69-4d2e-9aba-2bfd34a15e4a_1402x788.webp 424w, https://substackcdn.com/image/fetch/$s_!knSw!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2c90e9ab-4e69-4d2e-9aba-2bfd34a15e4a_1402x788.webp 848w, https://substackcdn.com/image/fetch/$s_!knSw!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2c90e9ab-4e69-4d2e-9aba-2bfd34a15e4a_1402x788.webp 1272w, https://substackcdn.com/image/fetch/$s_!knSw!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2c90e9ab-4e69-4d2e-9aba-2bfd34a15e4a_1402x788.webp 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Sam Altman caused some commotion on X today <a href="https://x.com/sama/status/2004939524216910323">with his post</a> that the new Head of Preparedness role at OpenAI would be responsible, <em>inter alia</em>, for &#8220;gain[ing] confidence in the safety of running systems that can <em><strong>self-improve</strong></em>&#8221;.</p><p>However, the fact that the Head of Preparedness will oversee safety efforts for model self-improvement should not be surprising.  After all, the role&#8217;s primary responsibility is to &#8220;<a href="https://openai.com/careers/head-of-preparedness-san-francisco/">lead the technical strategy and execution of OpenAI&#8217;s Preparedness Framework</a>&#8221;,<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> which currently focuses on cybersecurity risk, bio risk, and - yes - model self-improvement.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!kkRP!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa53dc80e-2242-4fc6-99c3-f6e55733b75e_970x278.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!kkRP!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa53dc80e-2242-4fc6-99c3-f6e55733b75e_970x278.png 424w, https://substackcdn.com/image/fetch/$s_!kkRP!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa53dc80e-2242-4fc6-99c3-f6e55733b75e_970x278.png 848w, https://substackcdn.com/image/fetch/$s_!kkRP!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa53dc80e-2242-4fc6-99c3-f6e55733b75e_970x278.png 1272w, https://substackcdn.com/image/fetch/$s_!kkRP!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa53dc80e-2242-4fc6-99c3-f6e55733b75e_970x278.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!kkRP!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa53dc80e-2242-4fc6-99c3-f6e55733b75e_970x278.png" width="970" height="278" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/a53dc80e-2242-4fc6-99c3-f6e55733b75e_970x278.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:278,&quot;width&quot;:970,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:121875,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.prinzai.com/i/182739228?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa53dc80e-2242-4fc6-99c3-f6e55733b75e_970x278.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!kkRP!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa53dc80e-2242-4fc6-99c3-f6e55733b75e_970x278.png 424w, https://substackcdn.com/image/fetch/$s_!kkRP!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa53dc80e-2242-4fc6-99c3-f6e55733b75e_970x278.png 848w, https://substackcdn.com/image/fetch/$s_!kkRP!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa53dc80e-2242-4fc6-99c3-f6e55733b75e_970x278.png 1272w, https://substackcdn.com/image/fetch/$s_!kkRP!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa53dc80e-2242-4fc6-99c3-f6e55733b75e_970x278.png 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>But this is where things get interesting.  It turns out that AI self-improvement capabilities were added to the Preparedness Framework as a &#8220;Tracked Category&#8221; all the way back in April 2025(!).  What is a &#8220;Tracked Category&#8221;?  OpenAI explains that it&#8217;s a capability that meets <em><strong>all</strong></em> of the following five criteria:</p><ol><li><p><strong>Plausible</strong>.  &#8220;It must be possible to identify a causal pathway for a severe harm in the capability area, enabled by frontier AI.&#8221;</p></li><li><p><strong>Measurable</strong>.  OpenAI must be able to construct or adopt capability evaluations that measure capabilities that closely track the potential for the severe harm.</p></li><li><p><strong>Severe</strong>.  There is a plausible threat model within the capability area that would create severe harm.</p></li><li><p><strong>Net New</strong>.  &#8220;The outcome cannot currently be realized as described (including at that scale, by that threat actor, or for that cost) with existing tools and resources&#8230; but without access to frontier AI.&#8221;  </p></li><li><p><strong>Instantaneous or irremediable</strong>.  Once the outcome is realized, its severe harms: (1) are immediately felt; or (2) are inevitable due to a lack of feasible measures to remediate.</p></li></ol><p>Per OpenAI, &#8220;AI self-improvement&#8221; was separated as a Tracked Category because:</p><blockquote><p>&#8220;it presents a distinct plausible, net new, and potentially irremediable risk, namely that of a hard-to-track rapid acceleration in AI capabilities which could have hard-to-predict severely harmful consequences. In addition, the evaluations we use to measure this capability are distinct from those applicable to Long-range Autonomy and Autonomous Replication and Adaptation.&#8221;<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a></p></blockquote><p>&#8230;but reading between the lines, the <em><strong>real</strong></em> reason for this capability&#8217;s designation as a Tracked Category may have been to begin preparing for <a href="https://substack.com/home/post/p-177620461">automation of AI research</a>.  After all, &#8220;High&#8221; risk associated with this capability is defined as follows:</p><blockquote><p>The model&#8217;s impact is equivalent to giving every OpenAI researcher <em><strong>a highly performant mid-career research engineer assistant</strong></em>, relative to those researchers&#8217; 2024 baseline.</p></blockquote><p>(This sounds perhaps slightly more capable than the &#8220;automated AI research intern&#8221; that OpenAI intends to develop by September 2026.)</p><p>And &#8220;Critical&#8221; risk sounds suspiciously similar to the risk that would be posed by the &#8220;automated AI researcher&#8221; that OpenAI intends to develop by March 2028:</p><blockquote><p>The model is capable of recursively self improving (i.e., <em><strong>fully automated AI R&amp;D</strong></em>), defined as either (leading indicator) a superhuman research scientist agent OR (lagging indicator) causing a generational model improvement (e.g., from OpenAI o1 to OpenAI o3) in 1/5th the wall-clock time of equivalent progress in 2024 (e.g., sped up to just 4 weeks) sustainably for several months.</p></blockquote><p>Based on the updated Preparedness Framework&#8217;s release date (April 2025), it seems that March/April 2025 is when OpenAI first set the internal goal to fully automate AI research within 3 years&#8217; time (by March 2028).  We are now &#8776;9 months into this 3-year project.  Six months in, OpenAI publicly announced it to the world.  And now, nine months in, OpenAI has begun building out a comprehensive safety function around it. It seems that we continue to be &#8220;all systems go&#8221; for the launch of the automated AI research intern next fall.</p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>The Preparedness Framework is available at: https://cdn.openai.com/pdf/18a02b5d-6b67-4cec-ab64-68cdfbddebcd/preparedness-framework-v2.pdf. </p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>&#8220;Long-range autonomy&#8221; and &#8220;autonomous replication and adaptation&#8221; are identified by OpenAI in the Preparedness Framework as Research Categories - <em>i.e.</em>, capabilities not (yet) qualifying as Tracked Categories.  Per OpenAI', these capabilities&#8217; threat models &#8220;are not yet sufficiently mature&#8221; to warrant their designation as Tracked Categories.</p></div></div>]]></content:encoded></item><item><title><![CDATA[Predictions for 2026]]></title><description><![CDATA[2025 was a year of stunningly fast AI progress.]]></description><link>https://www.prinzai.com/p/predictions-for-2026</link><guid isPermaLink="false">https://www.prinzai.com/p/predictions-for-2026</guid><dc:creator><![CDATA[prinz]]></dc:creator><pubDate>Mon, 22 Dec 2025 05:01:16 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!DwTI!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F205f9864-78ee-47df-90f5-474775d9cd52_768x432.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!DwTI!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F205f9864-78ee-47df-90f5-474775d9cd52_768x432.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!DwTI!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F205f9864-78ee-47df-90f5-474775d9cd52_768x432.jpeg 424w, https://substackcdn.com/image/fetch/$s_!DwTI!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F205f9864-78ee-47df-90f5-474775d9cd52_768x432.jpeg 848w, https://substackcdn.com/image/fetch/$s_!DwTI!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F205f9864-78ee-47df-90f5-474775d9cd52_768x432.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!DwTI!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F205f9864-78ee-47df-90f5-474775d9cd52_768x432.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!DwTI!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F205f9864-78ee-47df-90f5-474775d9cd52_768x432.jpeg" width="768" height="432" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/205f9864-78ee-47df-90f5-474775d9cd52_768x432.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:432,&quot;width&quot;:768,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:19634,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.prinzai.com/i/182279736?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F205f9864-78ee-47df-90f5-474775d9cd52_768x432.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!DwTI!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F205f9864-78ee-47df-90f5-474775d9cd52_768x432.jpeg 424w, https://substackcdn.com/image/fetch/$s_!DwTI!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F205f9864-78ee-47df-90f5-474775d9cd52_768x432.jpeg 848w, https://substackcdn.com/image/fetch/$s_!DwTI!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F205f9864-78ee-47df-90f5-474775d9cd52_768x432.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!DwTI!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F205f9864-78ee-47df-90f5-474775d9cd52_768x432.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>2025 was a year of stunningly fast AI progress.</p><p>In December 2024, the best reasoning model was <a href="https://openai.com/o1/">OpenAI&#8217;s o1</a>, a toy reasoning model that wasn&#8217;t even particularly proficient at using tools.  By September 2025, OpenAI&#8217;s unreleased general reasoning models had won gold medals on <a href="https://x.com/OpenAI/status/1946594928945148246">the 2025 International Mathematics Olympiad (IMO)</a>, the <a href="https://x.com/SherylHsu02/status/1954966109851119921">2025 International Olympiad in Informatics (IOI)</a>, and the <a href="https://x.com/MostafaRohani/status/1968360976379703569">2025 International Collegiate Programming Contest (ICPC) World Finals</a>.  Another unreleased OpenAI model <a href="https://x.com/gdb/status/1945553676321657127">won second place in the AtCoder World Finals</a>, working fully autonomously without human intervention for the entire 10 hours of the competition.  And coding agents - including, in particular, Claude Code - have taken the world of coding by storm, while also <a href="https://www.prinzai.com/p/opus-45-dramatically-increases-anthropic">meaningfully accelerating the pace of AI research at the frontier labs</a>.  </p><p>We have also begun to see glimpses of AI meaningfully contributing to work in fields other than coding.  Starting in late Q3 2025, I began using GPT-5.x Pro for legal research and analysis, <a href="https://x.com/deredleritt3r/status/2002064109223752163">and am now finding it absolutely essential to my work</a>.  I am also increasingly seeing reports that Google&#8217;s <a href="https://x.com/_simonsmith/status/2001810559369687477">NotebookLM is fantastic at generating presentations and data tables</a>, which is another important enterprise use case.  And even non-technical people (yes, including yours truly) <a href="https://x.com/mattyglesias/status/2002388080460812420">are discovering &#8220;Claude Code for things that are not coding&#8221;</a>.</p><p>Where does this lead us in 2026?  Here are some predictions:</p><h4>Automation of AI research</h4><p>Earlier this year, <a href="https://x.com/tszzl/status/2002924510455251161">roon<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> played with Codex for the first time and &#8220;realiz[ed] we&#8217;re in the takeoff&#8221;</a>.  In 2026, agentic coding tools like Codex and Claude Code will continue accelerating frontier lab researchers.  By September 2026, OpenAI intends to have this effort culminate in <a href="https://www.prinzai.com/p/what-we-know-about-openais-autonomous">an automated AI research intern running on hundreds of thousands of GPUs</a>, which will be able to automatically handle the implementation and debugging of research ideas proposed by OpenAI&#8217;s human researchers.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a></p><h4>Continual learning</h4><p>In Q3 2025, public consensus suddenly decreed that continual learning is required to achieve AGI.  Andrej Karpathy said that <a href="https://www.dwarkesh.com/p/andrej-karpathy">current LLMs are &#8220;cognitively lacking&#8221; due to lack of continual learning, and &#8220;it&#8217;s not working&#8221;</a> - placing AGI about a decade away.  Later, Ilya Sutskever added fuel to the fire when he revealed that <a href="https://www.dwarkesh.com/p/ilya-sutskever-2">SSI is working on developing AI capable of continual learning</a> - which he said is &#8220;5 to 20 years&#8221; away.</p><p>Relatively unnoticed among all the hoopla were <a href="https://www.youtube.com/watch?v=mYDSSRS-B5U">comments on continual learning from Anthropic&#8217;s CEO, Dario Amodei</a>:</p><blockquote><p>One thing we learned in AI is whenever it feels like there&#8217;s some fundamental obstacle - like two years ago we thought there was this fundamental obstacle around reasoning - turned out just to be be RL, you just train with RL and you let the model write things down to try and figure out objective math problems&#8230;<em><strong>Without being too specific, we already have maybe some evidence to suggest that [continual learning] is another of those problems that is not as difficult as it seems that will fall to scale plus a slightly different way of thinkin</strong></em><strong>g </strong><em><strong>about things.</strong></em></p></blockquote><p>And just a few days ago, Sholto Douglas, an Anthropic employee, <a href="https://x.com/deredleritt3r/status/2002442736431980857">dropped a bombshell</a> with his prediction that &#8220;<em><strong>continual learning [will get] solved in a satisfying way&#8221;</strong></em> in 2026.</p><p>Does this mean that Anthropic already knows how to achieve continual learning?  We&#8217;ll find out next year.</p><h4>Recursive self-improvement</h4><p><a href="https://www.prinzai.com/p/openai-is-scaling-up-synthetic-data">Mark Chen recently mentioned</a> that OpenAI is aggressively scaling up several bets, including one related to synthetic data.  This was a reference to Sebastien Bubeck&#8217;s brief cameo during the GPT-5 launch livestream, in which he revealed that OpenAI has developed &#8220;new training techniques&#8221; whereby o3 had generated synthetic data to train GPT-5 in a way &#8220;raw web data just never could&#8221;.  &#8220;<em><strong>This interaction between models foreshadows a recursive self-improvement loop</strong></em>&#8221;, Bubeck said.</p><p><a href="https://x.com/deredleritt3r/status/2001765302049136754">Google DeepMind is also working in the same direction</a>, according to Sebastian Borgeaud, pre-training lead for Gemini 3:</p><blockquote><p>One really interesting question is whether you can actually generate synthetic data to make a model that you want to train in the future better than the model that generated the synthetic data in the first place. We spend a lot of time thinking about this and doing research in this direction.</p></blockquote><p>It is unclear where these efforts will lead in 2026, but needless to say that this is an area of ML research that is well worth monitoring.</p><h4>AI is coming to the workplace (not just for coders)</h4><p><a href="https://x.com/deredleritt3r/status/2002442736431980857">Here&#8217;s Sholto Douglas again</a>:</p><blockquote><p>The most striking thing about next year is that the<strong> </strong><em><strong>other forms of knowledge work are going to experience what software engineers are feeling right now</strong></em>, where they went from typing most of their lines of code at the beginning of the year to typing barely any of them at the end of the year. I think of this as the Claude Code experience, but for all forms of knowledge work.</p></blockquote><p>Those who follow me on X know that I have been crying out for an interface that would enable even a non-technical lawyer to &#8220;vibe-code&#8221; a stock purchase agreement (see, e.g., point 3 <a href="https://x.com/deredleritt3r/status/2002064109223752163">here</a>).  It looks as though my wish may finally come true in 2026.</p><p>But it will be more than that, of course.  <a href="https://x.com/deredleritt3r/status/1993326105671987663">Anthropic&#8217;s goal for 2026</a> is to develop and sell to enterprises a &#8220;<em><strong>virtual co-worker</strong></em> that is in all your Slack channels and can join your meetings and can work alongside you&#8221;.  Some of us will be seeing these &#8220;virtual co-workers&#8221; join our companies next year.</p><h4>And as for coding&#8230;</h4><p>I am not a coder, but, as an outside observer, I can easily tell that several significant &#8220;vibe shifts&#8221; occurred in 2025 around using agentic tools like Claude Code and Codex for coding tasks.  Claude Opus 4.5 in particular <a href="https://x.com/METR_Evals/status/2002203627377574113">smashed the METR 50%-time horizon benchmark</a>, and appears to be a huge step change when compared to the previous generation of models - to the point where <a href="https://x.com/deanwball/status/2001068539990696422">some debate may be had as to whether Opus 4.5 in Claude Code is &#8220;basically AGI</a>&#8221; (by OpenAI&#8217;s definition: a highly autonomous system that outperforms humans at most economically valuable work).</p><p>It seems clear that the models will continue to improve at a rapid pace from here vis-a-vis coding.  Expect software engineering to &#8220;<a href="https://x.com/deredleritt3r/status/2002442736431980857">go[] utterly wild next year</a>&#8221;.</p><h4>AI for science</h4><p>Starting late this year, there has been an <a href="https://openai.com/index/gpt-5-mathematical-discovery/">increasing</a> <a href="https://x.com/SebastienBubeck/status/1958198661139009862?lang=en">cadence</a> <a href="https://x.com/AlexKontorovich/status/2001338945301352781">of reports</a> that models like GPT-5 Pro can be leveraged effectively as a tool by human mathematicians to help with making relatively minor advances in mathematics.  As models continue to improve next year, OpenAI expects that its AI systems &#8220;<a href="http://In 2026 we expect that our AI systems may be able to make small new discoveries">may be able to make small new [scientific] discoveries</a>&#8221; in 2026.  Indeed, work on these initiatives is ongoing at multiple frontier labs: for example, <a href="https://x.com/mgdurrant/status/2000700971471573460">Anthropic has begun hiring</a> &#8220;wet lab wizards&#8221; for its life sciences team.</p><p>But can LLM actually autonomously generate novel scientific hypotheses?  In my view, the answer is almost certainly &#8220;yes&#8221;.  We have already seen that <a href="https://x.com/deredleritt3r/status/1998062768671313927">even Gemini 2.0 Pro, when equipped with a great harness</a>, can propose a novel scientific hypothesis pertaining to a complex gene transfer mechanism.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a>  The general rule of thumb that I think makes sense to follow is that anything an LLM can do with a harness will eventually also be achievable by a more powerful LLM without any harness whatsoever; the only (important!) question that remains open is the timeline by which this feat would become possible to accomplish.</p><p>OpenAI has declared that 2026 will be <a href="https://x.com/deredleritt3r/status/2001737415715635528">the &#8220;Year of AI and Science&#8221;</a>.  Let&#8217;s hope that the year can live up to this lofty title!</p><h4>The robots are coming?</h4><p>There&#8217;s been a lot of hoopla around humanoid (or otherwise) robots over the past few years, but very few of these advances have thus far made it out into the real world.  I remain somewhat unconvinced that 2026 will be the year when robots truly proliferate in the real world at scale, but it&#8217;s possible that I am too pessimistic in this regard.  <a href="https://x.com/OfficialLoganK/status/2002831970586566824">Google DeepMind apparently projects</a> that 2026 will be &#8220;a huge year&#8221; for embodied AI, and that there will be &#8220;a lot more robots in the real world soon&#8221;.  Other knowledgeable commentators expect that 2026 will see at least &#8220;<a href="https://x.com/deredleritt3r/status/2002442736431980857">the first test deployments of home robots</a>&#8221;.</p><p>*   *   *</p><p>I will end here with an overall observation. Over the past few months, it has become decidedly fashionable to update one&#8217;s views towards longer timelines for &#8220;AGI&#8221; (whatever that term might mean).  If significant progress is made on automation of AI research and/or continual learning in 2026, these longer timelines will likely begin to feel extremely - maybe even needlessly - conservative by the end of the year.  In particular, OpenAI&#8217;s stated goal of fully automating AI research in just slightly more than two years&#8217; time still has not been - but should be - fully internalized by most industry observers and commentators.  Should OpenAI successfully develop and deploy an automated AI research &#8220;intern&#8221; during 2026, a realization may suddenly come to many that the long-expected promise of the machine taking over the building of other, yet more powerful, machines has come to the verge of being fulfilled.</p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>A famous semi-anon OpenAI employee, @tszzl on X.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>Just 18 months later, by March 2028, OpenAI expects to develop a fully end-to-end automated AI researcher.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>The related paper by Penades et al. is available here: https://www.sciencedirect.com/science/article/pii/S0092867425009730. </p></div></div>]]></content:encoded></item><item><title><![CDATA[Continual learning may not be as difficult as it seems]]></title><description><![CDATA[Some optimistic predictions from the frontier labs]]></description><link>https://www.prinzai.com/p/continual-learning-may-not-be-as</link><guid isPermaLink="false">https://www.prinzai.com/p/continual-learning-may-not-be-as</guid><dc:creator><![CDATA[prinz]]></dc:creator><pubDate>Fri, 12 Dec 2025 06:40:52 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/14656eb1-7625-43ba-8e21-afb645b87ba0_1920x1080.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em><strong>Update (12.20.25)</strong></em>:  New prediction from Sholto Douglas (Anthropic) that continual learning will be solved in 2026.</p><div><hr></div><p>Over the past six months, it has become quite fashionable in certain circles to assume that AGI is more than a decade away because, among other things, current AI models lack continual learning.  Ilya Sutskever is working on <a href="https://www.dwarkesh.com/p/ilya-sutskever-2">developing superintelligence with skills and knowledge of an eager 15-year-old and the ability to learn on the job</a>; he thinks this is &#8220;5 to 20&#8221; years away.  Andrej Karpathy thinks that <a href="https://www.dwarkesh.com/p/andrej-karpathy">current AI systems &#8220;are cognitively lacking and it&#8217;s just not working&#8221; because they have no continual learning</a>; &#8220;[i]t will take about a decade to work through all of those issues&#8221;, he thinks.</p><p>But why should we assume that continual learning is a decade away?  Could it be that it&#8217;s easier to achieve than some think?</p><h3>Sholto Douglas: continual learning to &#8220;get solved in a satisfying way&#8221; in 2026</h3><p><a href="https://x.com/NoPriorsPod/status/2002120381709365257">In the &#8220;No Priors&#8221; year-end podcast</a>, Sholto Douglas (Anthropic) said that he thinks &#8220;that probably <em><strong>continual learning gets solved in a satisfying way&#8221; in 2026</strong>.</em></p><h3>Dario Amodei: continual learning &#8220;not as difficult as it seems&#8221;</h3><p>Dario Amodei, CEO of Anthropic, repeatedly stated this summer that continual learning may not be as difficult as it seems.</p><p><a href="https://www.youtube.com/watch?v=mYDSSRS-B5U">In a July interview</a>, Amodei said that, even without continual learning, other techniques &#8220;can fill in many of the gaps&#8221;.  One such technique is significantly lengthening the context window - perhaps to as much as 100 million words, which is roughly the number of words a human hears during his or her lifetime.  There is &#8220;no reason&#8221; from a machine learning perspective why the context window could not be increased to this size, Amodei said, &#8220;it&#8217;s really just inference support&#8221; that is needed to make this viable.</p><p>But what about continual learning that allows for updating a model&#8217;s weights? Unexpectedly, Amodei suggested that Anthropic may have <em><strong>already</strong></em> found a path to achieving it:</p><blockquote><p>One thing we learned in AI is whenever it feels like there&#8217;s some fundamental obstacle - like two years ago we thought there was this fundamental obstacle around reasoning - turned out just to be be RL, you just train with RL and you let the model write things down to try and figure out objective math problems&#8230;<em><strong>Without being too specific, we already have maybe some evidence to suggest that [continual learning] is another of those problems that is not as difficult as it seems that will fall to scale plus a slightly different way of thinkin</strong></em><strong>g </strong><em><strong>about things.</strong></em> </p></blockquote><p>Amodei also hinted that an &#8220;inner loop/outer loop&#8221; structure, wherein an agent learns things, and optimizes for a lifetime of an episode (inner loop) and also learns over multiple episodes (outer loop) &#8220;maybe&#8230; is a way to learn continual learning&#8221;.</p><p><a href="https://substack.com/home/post/p-170262127">In a subsequent August interview</a>, Amodei again mentioned extending a model&#8217;s context to 100 million tokens and suggested that models could be trained to be &#8220;specialized for learning over the context&#8221;; &#8220;[y]ou could, even during the context, update the model&#8217;s weights&#8221;. &#8220;<em><strong>[T]here are lots of ideas that are very close to the ideas we have now that could perhaps do this [i.e., achieve continual learning]</strong></em>&#8221;, Amodei said.</p><h3>Shane Legg: &#8220;no fundamental blockers&#8221; on continual learning; &#8220;we have ideas&#8221; on how to develop it</h3><p><a href="https://www.youtube.com/watch?v=l3u_FAv33G0">In an interview released just today</a>, Shane Legg, co-founder and Chief AGI Scientist at Google DeepMind, said that there are <em><strong>no &#8220;fundamental blockers&#8221;</strong></em> on continual learning (and also visual reasoning).</p><blockquote><p><em><strong>[W]e have ideas on how to develop systems that can do these things</strong></em>, and we see metrics improving over time in a bunch of these areas. So my expectation is over a number of years these things will all get addressed. But they&#8217;re not there yet.</p></blockquote><p>Continual learning &#8220;might need some process whereby new information may be stored&#8221;, a &#8220;retrieval system or episodic memory&#8221;, and &#8220;systems whereby that information over time is trained back into some underlying model&#8221;, Legg said.  This will require both more data and algorithmic and architectural changes.</p>]]></content:encoded></item><item><title><![CDATA[OpenAI is scaling up synthetic data generation]]></title><description><![CDATA["This interaction between models foreshadows a recursive self-improvement loop."]]></description><link>https://www.prinzai.com/p/openai-is-scaling-up-synthetic-data</link><guid isPermaLink="false">https://www.prinzai.com/p/openai-is-scaling-up-synthetic-data</guid><dc:creator><![CDATA[prinz]]></dc:creator><pubDate>Tue, 02 Dec 2025 04:16:20 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/2cb5fb0a-33ec-40b5-87d5-137183127e17_1607x901.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><a href="https://www.youtube.com/watch?v=ZeyHBM2Y5_4">In a new interview</a>, Mark Chen mentioned that OpenAI is aggressively scaling up several bets, including one about synthetic data that &#8220;OpenAI talked a lot about&#8221; when GPT-5 was launched.</p><p>This is a reference to Sebastien Bubeck&#8217;s brief cameo during the GPT-5 launch, during which he said that OpenAI has developed &#8220;new training techniques&#8221; whereby o3 generated synthetic data to train GPT-5 in a way &#8220;raw web data just never could&#8221;. The point was to generate not a large volume of data cheaply, but rather useful training data.  </p><p>&#8220;<em><strong>This interaction between models foreshadows a recursive self-improvement loop</strong></em>&#8221;, Bubeck said, adding: &#8220;Here at OpenAI we cracked pre-training, then reasoning, and now we are seeing their interactions significantly deepened. In the future, AI systems will move far beyond our current pre-training and post-training pipelines we&#8217;ve been used to and we are seeing the first steps towards this right now and right here.&#8221;</p><p>And OpenAI is now aggressively scaling it up.</p>]]></content:encoded></item><item><title><![CDATA[Top Anthropic researchers are significantly accelerated by Claude Code]]></title><description><![CDATA["9 of 18 [researchers] reported &#8805;100% productivity improvements, with a median estimate of 100% and a mean estimate of 220%."]]></description><link>https://www.prinzai.com/p/opus-45-dramatically-increases-anthropic</link><guid isPermaLink="false">https://www.prinzai.com/p/opus-45-dramatically-increases-anthropic</guid><dc:creator><![CDATA[prinz]]></dc:creator><pubDate>Fri, 28 Nov 2025 20:57:02 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/dac1fc5b-d3b4-4bb6-b4d7-2cc1a2bb2f05_872x487.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>In connection with <a href="https://www.anthropic.com/news/claude-opus-4-5">last week&#8217;s release of Claude Opus 4.5</a>, Anthropic surveyed 18 members of its technical staff to estimate the productivity boost they get from the model.</p><p>The results (from the <a href="https://assets.anthropic.com/m/64823ba7485345a7/Claude-Opus-4-5-System-Card.pdf">Opus 4.5 system card</a>):</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!pfw-!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F056e1700-a41b-4ed9-a732-cabdc526c5dd_648x197.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!pfw-!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F056e1700-a41b-4ed9-a732-cabdc526c5dd_648x197.png 424w, https://substackcdn.com/image/fetch/$s_!pfw-!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F056e1700-a41b-4ed9-a732-cabdc526c5dd_648x197.png 848w, https://substackcdn.com/image/fetch/$s_!pfw-!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F056e1700-a41b-4ed9-a732-cabdc526c5dd_648x197.png 1272w, https://substackcdn.com/image/fetch/$s_!pfw-!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F056e1700-a41b-4ed9-a732-cabdc526c5dd_648x197.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!pfw-!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F056e1700-a41b-4ed9-a732-cabdc526c5dd_648x197.png" width="648" height="197" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/056e1700-a41b-4ed9-a732-cabdc526c5dd_648x197.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:197,&quot;width&quot;:648,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:62026,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://prinz.substack.com/i/180203966?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F056e1700-a41b-4ed9-a732-cabdc526c5dd_648x197.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!pfw-!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F056e1700-a41b-4ed9-a732-cabdc526c5dd_648x197.png 424w, https://substackcdn.com/image/fetch/$s_!pfw-!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F056e1700-a41b-4ed9-a732-cabdc526c5dd_648x197.png 848w, https://substackcdn.com/image/fetch/$s_!pfw-!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F056e1700-a41b-4ed9-a732-cabdc526c5dd_648x197.png 1272w, https://substackcdn.com/image/fetch/$s_!pfw-!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F056e1700-a41b-4ed9-a732-cabdc526c5dd_648x197.png 1456w" sizes="100vw" fetchpriority="high"></picture><div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!20VA!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F94cd4fcc-d5f0-46e7-92e9-1564cbc4d4ff_642x147.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!20VA!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F94cd4fcc-d5f0-46e7-92e9-1564cbc4d4ff_642x147.png 424w, https://substackcdn.com/image/fetch/$s_!20VA!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F94cd4fcc-d5f0-46e7-92e9-1564cbc4d4ff_642x147.png 848w, https://substackcdn.com/image/fetch/$s_!20VA!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F94cd4fcc-d5f0-46e7-92e9-1564cbc4d4ff_642x147.png 1272w, https://substackcdn.com/image/fetch/$s_!20VA!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F94cd4fcc-d5f0-46e7-92e9-1564cbc4d4ff_642x147.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!20VA!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F94cd4fcc-d5f0-46e7-92e9-1564cbc4d4ff_642x147.png" width="642" height="147" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/94cd4fcc-d5f0-46e7-92e9-1564cbc4d4ff_642x147.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:147,&quot;width&quot;:642,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:49013,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://prinz.substack.com/i/180203966?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F94cd4fcc-d5f0-46e7-92e9-1564cbc4d4ff_642x147.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!20VA!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F94cd4fcc-d5f0-46e7-92e9-1564cbc4d4ff_642x147.png 424w, https://substackcdn.com/image/fetch/$s_!20VA!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F94cd4fcc-d5f0-46e7-92e9-1564cbc4d4ff_642x147.png 848w, https://substackcdn.com/image/fetch/$s_!20VA!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F94cd4fcc-d5f0-46e7-92e9-1564cbc4d4ff_642x147.png 1272w, https://substackcdn.com/image/fetch/$s_!20VA!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F94cd4fcc-d5f0-46e7-92e9-1564cbc4d4ff_642x147.png 1456w" sizes="100vw"></picture><div></div></div></a></figure></div><ul><li><p>50% of survey participants reported productivity improvement of at least 100% (2x); median productivity improvement was 100%</p></li><li><p>Mean productivity improvement was 220%(!)</p></li><li><p>11% (2/18) characterized the model as a &#8220;near-complete entry-level researcher replacement&#8221; (with meaningful caveats)</p></li><li><p>Most researchers would prefer losing access to Opus 4.5 to losing access to Claude Code (i.e., the harness remains more important than the model)</p></li></ul><p>Importantly, survey participants were not just average Anthropic employees, but rather were &#8220;primarily&#8221; selected from the top 30 Anthropic employees ranked by internal Claude Code usage.  <em><strong>We should expect that these Claude power users would get significantly more uplift from using the model than the average Anthropic employee</strong></em>.  Nonetheless, the productivity boost unlocked by the model for its most skilled users is extremely impressive and well worth noting.</p><h3>Comparison to Sonnet 4.5</h3><p>For reference, here are the results of a similar survey <a href="https://assets.anthropic.com/m/12f214efcc2f457a/original/Claude-Sonnet-4-5-System-Card.pdf">conducted by Anthropic in September 2025 for Sonnet 4.5:</a></p><ul><li><p>7 Anthropic researchers were surveyed; it&#8217;s not clear whether they were average employees or some of the top users of Claude</p></li><li><p>Productivity boost estimates from Sonnet 4.5 were: 15%, 20%, 20%, 30%, 40%, 100%, one instance of &#8220;qualitative-only feedback&#8221;</p></li><li><p>0% thought that Sonnet 4.5 could completely automate the work of a junior ML researcher</p></li><li><p>As with Opus 4.5, most researchers (4 out of 7) thought that most of the productivity boost was attributable to Claude Code, as opposed to the model itself</p></li></ul><p></p><p></p>]]></content:encoded></item><item><title><![CDATA[OpenAI-Proof Q&A]]></title><description><![CDATA[OpenAI's models try to crack major bugs that previously took OpenAI researchers >1 day to solve.]]></description><link>https://www.prinzai.com/p/benchmarks-openai-proof-q-and-a</link><guid isPermaLink="false">https://www.prinzai.com/p/benchmarks-openai-proof-q-and-a</guid><dc:creator><![CDATA[prinz]]></dc:creator><pubDate>Thu, 20 Nov 2025 05:24:31 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!pwEF!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4821fe4c-de3f-4769-9278-19b30c2b87a5_1205x686.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>With <a href="https://cdn.openai.com/pdf/2a7d98b1-57e5-4147-8d0e-683894d782ae/5p1_codex_max_card_03.pdf">today&#8217;s release of GPT-5.1-Codex-Max</a>, OpenAI updated the results of one of the most interesting extant AI model benchmarks, the unfortunately named OpenAI-Proof Q&amp;A.</p><p>What is this benchmark?</p><ul><li><p>Take 20 research and engineering bottlenecks that OpenAI actually encountered in the past, each of which required <em><strong>over a day</strong></em> for the OpenAI team to solve.  These bottlenecks include &#8220;unexpected performance regressions, anomalous training metrics [and] subtle implementation bugs&#8221;, which <em><strong>actually represented delays to major projects</strong></em> and &#8220;in some case influenc[ed] the outcome of large training runs and launches&#8221;.</p></li><li><p>Give the model access to a container with code access and run artifacts, permitting it to use historical code, logs, and experiment data.</p></li><li><p>Ask the model to diagnose and explain the root cause of the issue.</p></li><li><p>Each of the model&#8217;s solutions is graded pass@1 (one try only!).</p></li></ul><p>This benchmark is very relevant to accelerating - and eventually automating - AI research. Imagine the time and resources that could be saved quashing a major bug if GPT-x could diagnose and identify it even 50% of the time (instead of the OpenAI team spending 1+ days identifying and fixing it). </p><p>I particularly like that, instead of using toy problems (which are often of questionable relevance to real-world AI use cases, but plague nearly all popular LLM benchmarks), OpenAI-Proof Q&amp;A measures the model&#8217;s performance on major bugs that OpenAI has <em><strong>actually encountered</strong></em> in the past.  And the model&#8217;s performance is judged on a pass@1 standard - none of that &#8220;well, it got the solution right once out of the twenty times we ran it, so we&#8217;ll call that a pass&#8221; nonsense.</p><p>Here&#8217;s how OpenAI&#8217;s models have done on this benchmark to date:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!pwEF!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4821fe4c-de3f-4769-9278-19b30c2b87a5_1205x686.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!pwEF!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4821fe4c-de3f-4769-9278-19b30c2b87a5_1205x686.png 424w, https://substackcdn.com/image/fetch/$s_!pwEF!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4821fe4c-de3f-4769-9278-19b30c2b87a5_1205x686.png 848w, https://substackcdn.com/image/fetch/$s_!pwEF!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4821fe4c-de3f-4769-9278-19b30c2b87a5_1205x686.png 1272w, https://substackcdn.com/image/fetch/$s_!pwEF!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4821fe4c-de3f-4769-9278-19b30c2b87a5_1205x686.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!pwEF!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4821fe4c-de3f-4769-9278-19b30c2b87a5_1205x686.png" width="1205" height="686" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/4821fe4c-de3f-4769-9278-19b30c2b87a5_1205x686.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:686,&quot;width&quot;:1205,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:53315,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://prinz.substack.com/i/179424255?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4821fe4c-de3f-4769-9278-19b30c2b87a5_1205x686.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!pwEF!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4821fe4c-de3f-4769-9278-19b30c2b87a5_1205x686.png 424w, https://substackcdn.com/image/fetch/$s_!pwEF!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4821fe4c-de3f-4769-9278-19b30c2b87a5_1205x686.png 848w, https://substackcdn.com/image/fetch/$s_!pwEF!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4821fe4c-de3f-4769-9278-19b30c2b87a5_1205x686.png 1272w, https://substackcdn.com/image/fetch/$s_!pwEF!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4821fe4c-de3f-4769-9278-19b30c2b87a5_1205x686.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>GPT-5.1-Codex-Max scored 8%<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> (with and without refusals), beating the previous SOTA (GPT-5, which scored 2%).</p><ul><li><p><em><strong>Note</strong></em>: It is not clear to me how much of the delta between GPT-5 and GPT-5.1-Codex-Max was due to the fact that GPT-5 had &#8220;no browsing&#8221;.  Did GPT-5.1-Codex-Max have access to browsing (I assume yes)?  If so, how much did this skew the score?  </p></li></ul><p>Regardless of comparability to GPT-5&#8217;s score, the result achieved by GPT-5.1-Codex-Max is quite impressive.  Look for further updates to OpenAI-Proof Q&amp;A as more powerful OpenAI models are released over the next 12 months.</p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>It is not clear how a benchmark consisting of 20 problems can yield scores that are not divisible by 5.  My best guess is that each of the 20 problems was given to different instances of GPT-5.1-Codex-Max multiple times, resulting in a more granular aggregated score - but this has not been confirmed by OpenAI.</p></div></div>]]></content:encoded></item><item><title><![CDATA["All of it"]]></title><description><![CDATA[Demystifying Microsoft's Rights to OpenAI IP.]]></description><link>https://www.prinzai.com/p/all-of-it</link><guid isPermaLink="false">https://www.prinzai.com/p/all-of-it</guid><dc:creator><![CDATA[prinz]]></dc:creator><pubDate>Fri, 14 Nov 2025 06:49:13 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Esz1!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcf3a97b6-59dc-4f86-a9f7-98b6eaba5926_557x382.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><a href="https://www.dwarkesh.com/p/satya-nadella-2">Dylan Patel asked Satya Nadella</a> about the level of access Microsoft has to OpenAI&#8217;s IP.  Satya&#8217;s response went viral:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Esz1!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcf3a97b6-59dc-4f86-a9f7-98b6eaba5926_557x382.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Esz1!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcf3a97b6-59dc-4f86-a9f7-98b6eaba5926_557x382.jpeg 424w, https://substackcdn.com/image/fetch/$s_!Esz1!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcf3a97b6-59dc-4f86-a9f7-98b6eaba5926_557x382.jpeg 848w, https://substackcdn.com/image/fetch/$s_!Esz1!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcf3a97b6-59dc-4f86-a9f7-98b6eaba5926_557x382.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!Esz1!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcf3a97b6-59dc-4f86-a9f7-98b6eaba5926_557x382.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Esz1!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcf3a97b6-59dc-4f86-a9f7-98b6eaba5926_557x382.jpeg" width="557" height="382" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/cf3a97b6-59dc-4f86-a9f7-98b6eaba5926_557x382.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:382,&quot;width&quot;:557,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:20795,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://prinz.substack.com/i/178862567?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcf3a97b6-59dc-4f86-a9f7-98b6eaba5926_557x382.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Esz1!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcf3a97b6-59dc-4f86-a9f7-98b6eaba5926_557x382.jpeg 424w, https://substackcdn.com/image/fetch/$s_!Esz1!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcf3a97b6-59dc-4f86-a9f7-98b6eaba5926_557x382.jpeg 848w, https://substackcdn.com/image/fetch/$s_!Esz1!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcf3a97b6-59dc-4f86-a9f7-98b6eaba5926_557x382.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!Esz1!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcf3a97b6-59dc-4f86-a9f7-98b6eaba5926_557x382.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Let&#8217;s examine this claim based on <a href="https://openai.com/index/next-chapter-of-microsoft-openai-partnership/">publicly available information</a>.</p><h3>Microsoft&#8217;s rights to OpenAI&#8217;s &#8220;standard&#8221; IP</h3><p>Other than &#8220;Research IP&#8221; (see below), Microsoft does have broad rights to all of OpenAI&#8217;s IP until 2032.  This includes:</p><ul><li><p>model architecture</p></li><li><p>model weights</p></li><li><p>inference code</p></li><li><p>finetuning code</p></li><li><p>anything related to data center hardware and software</p></li></ul><p>As Satya correctly pointed out during the interview, there is one notable exclusion: Microsoft does <em><strong>not</strong></em> have any IP rights to OpenAI&#8217;s consumer hardware.</p><h3>&#8220;Research IP&#8221;</h3><p>Notwithstanding the above, Microsoft has rights to OpenAI&#8217;s &#8220;Research IP&#8221; only through the earlier of: (a) 2030; or (b) verification of OpenAI&#8217;s declaration of AGI by an independent expert panel.</p><p>&#8220;Research IP&#8221; means &#8220;confidential methods used in the development of models and systems&#8221;, such as models intended for internal deployment or research only.</p><h3>Restrictions on Microsoft&#8217;s use of OpenAI&#8217;s IP</h3><p>It&#8217;s one thing to have IP rights; it&#8217;s another to be able to use them however you like.</p><p>The contract between OpenAI and Microsoft includes a clause restricting Microsoft&#8217;s use of OpenAI&#8217;s IP as follows:</p><blockquote><p>If Microsoft uses OpenAI&#8217;s IP to develop AGI, prior to AGI being declared, the models will be subject to compute thresholds; those thresholds are significantly larger than the size of systems used to train leading models today.</p></blockquote><h3>What does this mean for Microsoft&#8217;s quest for AGI?</h3><p>Assuming that AGI will be a model larger than the threshold specified in the contract between OpenAI and Microsoft (probably a good assumption), <em><strong>Microsoft won&#8217;t be able to use OpenAI&#8217;s IP to develop AGI until after OpenAI already has achieved AGI.  </strong></em></p><p>Note, however, that:</p><ul><li><p>Microsoft will be able to use OpenAI&#8217;s IP to train smaller models for AGI-related research purposes.</p></li><li><p>Nothing prevents Microsoft from separately pursuing AGI without relying on any of OpenAI&#8217;s IP.</p></li></ul><p>But <em><strong>once OpenAI&#8217;s declaration of AGI is verified, all bets are off</strong></em>.  At that point, Microsoft will have a copy of the AGI model&#8217;s weights and will be able to race directly against OpenAI to develop even better models ( let&#8217;s call them &#8220;ASI&#8221;).</p><p>In that race, Microsoft will no longer have the benefit of OpenAI&#8217;s &#8220;Research IP&#8221;, including models used only for &#8220;internal deployment&#8221; or &#8220;research&#8221;.  Thus, <em><strong>even if OpenAI develops &#8220;ASI&#8221; before Microsoft, it won&#8217;t have to share it with Microsoft</strong></em>, so long as &#8220;ASI&#8221; is not publicly released, but instead remains a model in OpenAI&#8217;s &#8220;internal deployment&#8221;.</p>]]></content:encoded></item><item><title><![CDATA[OpenAI's automated AI researcher]]></title><description><![CDATA["Given the extraordinary potential impacts we think it is in the public interest to be transparent about this."]]></description><link>https://www.prinzai.com/p/what-we-know-about-openais-autonomous</link><guid isPermaLink="false">https://www.prinzai.com/p/what-we-know-about-openais-autonomous</guid><dc:creator><![CDATA[prinz]]></dc:creator><pubDate>Fri, 31 Oct 2025 03:49:45 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!1FrX!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5f6c3b20-8498-4460-a5fb-0c387c49164e_1156x568.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em><strong>Updated December 1, 2025 </strong>- to include a clear explanation from Mark Chen on what OpenAI means by an &#8220;automated AI researcher&#8221; and an &#8220;automated AI research intern&#8221;.</em></p><p><em><strong>Updated November 26, 2025</strong></em><strong> - </strong><em>to include new remarks by Lukasz Kaiser.</em></p><div><hr></div><p>On October 29, 2025, Sam Altman and Jakub Pachocki announced <a href="https://www.youtube.com/watch?v=ngDCxlZcecw">during a livestream</a> that OpenAI is building an &#8220;automated AI researcher&#8221;, targeted to be available by March 2028.  Here&#8217;s what we know about it.</p><h4>What&#8217;s Being Released and When</h4><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!1FrX!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5f6c3b20-8498-4460-a5fb-0c387c49164e_1156x568.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!1FrX!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5f6c3b20-8498-4460-a5fb-0c387c49164e_1156x568.png 424w, https://substackcdn.com/image/fetch/$s_!1FrX!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5f6c3b20-8498-4460-a5fb-0c387c49164e_1156x568.png 848w, https://substackcdn.com/image/fetch/$s_!1FrX!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5f6c3b20-8498-4460-a5fb-0c387c49164e_1156x568.png 1272w, https://substackcdn.com/image/fetch/$s_!1FrX!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5f6c3b20-8498-4460-a5fb-0c387c49164e_1156x568.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!1FrX!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5f6c3b20-8498-4460-a5fb-0c387c49164e_1156x568.png" width="1156" height="568" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/5f6c3b20-8498-4460-a5fb-0c387c49164e_1156x568.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:568,&quot;width&quot;:1156,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:64254,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://prinz.substack.com/i/177620461?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5f6c3b20-8498-4460-a5fb-0c387c49164e_1156x568.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!1FrX!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5f6c3b20-8498-4460-a5fb-0c387c49164e_1156x568.png 424w, https://substackcdn.com/image/fetch/$s_!1FrX!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5f6c3b20-8498-4460-a5fb-0c387c49164e_1156x568.png 848w, https://substackcdn.com/image/fetch/$s_!1FrX!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5f6c3b20-8498-4460-a5fb-0c387c49164e_1156x568.png 1272w, https://substackcdn.com/image/fetch/$s_!1FrX!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5f6c3b20-8498-4460-a5fb-0c387c49164e_1156x568.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>OpenAI&#8217;s goal is to &#8220;have&#8221;:</p><ul><li><p>an <em><strong>automated AI research intern</strong></em> by September 2026</p></li><li><p>an <em><strong>automated AI researcher</strong></em> by March 2028</p></li></ul><p>These appear to be dates by which OpenAI hopes to have these researchers available internally, and <em><strong>not</strong></em> necessarily release dates.  For example, <a href="https://x.com/sama/status/1983584366547829073">as tweeted by Sam</a>:</p><blockquote><p>We have set internal goals of <em><strong>having</strong></em> an automated AI research intern by September of 2026 running on hundreds of thousands of GPUs, and a true automated AI researcher by March of 2028.</p></blockquote><p>Similarly, during the livestream, Jakub repeatedly referred to the goal of &#8220;<em><strong>getting</strong></em>&#8221; (as opposed to releasing) the research intern/researcher by these dates.</p><h4>The Automated AI Research Intern</h4><p>Even though OpenAI calls the model that will become available by September 2026 merely an &#8220;intern&#8221;, do not be fooled - this model will likely be very powerful.  <a href="https://x.com/sama/status/1983584366547829073">Sam wrote</a> that the &#8220;intern&#8221; will <em><strong>run on hundreds of thousands of GPUs</strong></em>, a gigantic amount of compute.  As a point of comparison, consider that OpenAI will likely have <a href="https://x.com/sama/status/1947057625780396512">only just over 1 million GPUs online in total by year-end 2025</a>.  This means that running the &#8220;intern&#8221; alone would take up &gt;20% (and possibly significantly more) of the entire compute capacity available to OpenAI today.</p><p>Given this background, it should not be surprising that, in Jakub&#8217;s words, OpenAI expects that the automated AI research intern will &#8220;<em><strong>meaningfully accelerate</strong></em>&#8221; its researchers.</p><h4>The Automated AI Researcher</h4><p>During the livestream, Jakub described the automated AI researcher as &#8220;a system capable of autonomously delivering on larger research projects&#8221;.  Expect this system to be very impactful.  In fact, Jakub described the automated AI researcher <a href="https://www.youtube.com/watch?v=KSgPNVmZ8jQ">in a September 2025 interview</a> as <em><strong>the focal point of OpenAI&#8217;s research program over the past few years</strong></em>:</p><blockquote><p>Our set goal for our research program has been getting to an automated researcher <em>for a couple years now</em>. And so <em>we&#8217;ve been building most our projects with this goal in mind</em>.</p></blockquote><p><a href="https://x.com/sama/status/1983584366547829073">Per Sam&#8217;s follow-up tweet</a>, the automated AI researcher will have &#8220;<em><strong>extraordinary potential impacts</strong></em>&#8221; - potentially so significant that OpenAI has deemed it to be &#8220;<em><strong>in the public interest to be transparent</strong></em>&#8221; about its plans for developing it. </p><h4>&#8230;But How Will It Work?</h4><p><em><strong>Update (December 1, 2025):  </strong></em><a href="https://www.youtube.com/watch?v=ZeyHBM2Y5_4">Mark Chen has finally shed some light</a> on what the automated AI researcher and automated AI research intern are intended to accomplish:</p><blockquote><p>Within a year, we want to change the nature of the way that we&#8217;re doing research. We want to be productively relying on AI interns in the research development process. And <em><strong>within 2.5 years, we want AI to be doing end-to-end research</strong></em>. Today, you come up with an idea, you execute on it, you implement it, you debug it. <em><strong>Within a year, we&#8217;re quite confident we can get to a world where we control the outer loop - we come up with the ideas, but the model is in charge of the implementation and debugging.</strong></em></p></blockquote><p>So - the automated AI research intern will enable OpenAI&#8217;s research team to limit their work to generating new ML ideas; the intern will implement them and debug them. And the automated AI researcher will be doing &#8220;end-to-end research&#8221; - including generating new ML ideas.</p><p><em><strong>Update (November 25, 2025):</strong></em>  <a href="https://www.youtube.com/watch?v=3K-R4yVjJfU">Lukasz Kaiser described the automated AI research intern similarly</a> - <em>i.e.</em>, as being able to convert clear - but general - instructions from a human researcher into an efficient implementation of the researcher&#8217;s idea:</p><blockquote><p>Where AI researchers have great hope to help themselves... is that if you could just say &#8216;<em><strong>hey, Codex, this is the idea, and it&#8217;s fairly clear what I&#8217;m saying, please just implement it so it runs fast on this 8-machine setup or 100-machine setup</strong></em>&#8217;. I think that&#8217;s what OpenAI [means by] an AI intern by the end of next year.</p></blockquote><p>As to how OpenAI intends to achieve this technically, Jakub Pachocki&#8217;s <a href="https://www.youtube.com/watch?v=KSgPNVmZ8jQ">September 2025 interview</a> may shed some light on at least some of the advances that OpenAI expects to power the automated AI researcher:</p><blockquote><p>The big thing we are targeting with our research is producing an automated researcher. So, automating the discovery of new ideas, and in particular automating our own own work, automating ML research&#8230;</p><p>One good way to measure progress there is looking at the time horizon on which these models actually can reason and make progress. Now as we get to a level of near-mastery of high school competitions, we get to on the order of 1 to 5 hours of reasoning. <em><strong>We are focused on extending that horizon both in terms the models&#8217; capability to plan over very long horizons and actually have ability to retain memory.</strong></em></p></blockquote><p>So, one piece of the puzzle may be to get the models to reason for longer - <a href="https://x.com/polynoamial/status/1834280969786065278">which OpenAI has been working on for a long time, probably since before o1-preview was first announced</a>.  Another may be to have the automated AI researcher &#8220;<em>retain memory</em>&#8221; (but it&#8217;s unclear how).</p><h4>&#8230;And Will It Be AGI? </h4><p>When asked &#8220;wen AGI?&#8221; during the livestream, Sam said that <em>it&#8217;s more useful to have an automated AI researcher by March 2028 and define what that means that try to define &#8220;AGI&#8221;</em>:</p><blockquote><p>&#8220;The AGI term has become hugely overloaded, and&#8230; it will be this process over a number of years that we&#8217;re in the middle of.  But one of the reasons we wanted to present what we did today is, I think it&#8217;s much more useful to say our intention, our goal, by March of 2028 is to have a true automated AI researcher and define what that means than to try to satisfy everyone with a definition of AGI.&#8221;</p></blockquote><p><a href="https://www.youtube.com/watch?v=5yA4o9fSJek">Greg Brockman recently said</a> that he expects AGI to arrive within the next &#8220;one to three years&#8221; (for those counting, that means by late 2028) and that he &#8220;would feel like something went wrong if we were not there by 2030&#8221;.  It&#8217;s interesting to observe how neatly the anticipated March 2028 release of the automated AI researcher falls within this timeline.</p>]]></content:encoded></item></channel></rss>