<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[AI Pathways]]></title><description><![CDATA[Musings on the trajectories of AI development and policy]]></description><link>https://www.pathwaysai.org</link><image><url>https://substackcdn.com/image/fetch/$s_!OC13!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fafc6c13c-0c42-4728-9911-49af59c17369_400x400.png</url><title>AI Pathways</title><link>https://www.pathwaysai.org</link></image><generator>Substack</generator><lastBuildDate>Wed, 06 May 2026 10:31:15 GMT</lastBuildDate><atom:link href="https://www.pathwaysai.org/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Herbie Bradley]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[herbiebradley@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[herbiebradley@substack.com]]></itunes:email><itunes:name><![CDATA[Herbie Bradley]]></itunes:name></itunes:owner><itunes:author><![CDATA[Herbie Bradley]]></itunes:author><googleplay:owner><![CDATA[herbiebradley@substack.com]]></googleplay:owner><googleplay:email><![CDATA[herbiebradley@substack.com]]></googleplay:email><googleplay:author><![CDATA[Herbie Bradley]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[Glimpses of AI Progress]]></title><description><![CDATA[Mental models for fast times]]></description><link>https://www.pathwaysai.org/p/glimpses-of-ai-progess</link><guid isPermaLink="false">https://www.pathwaysai.org/p/glimpses-of-ai-progess</guid><dc:creator><![CDATA[Herbie Bradley]]></dc:creator><pubDate>Sun, 16 Mar 2025 22:44:13 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!8pQJ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1af19aa-b849-4333-9cff-82663ca601e0_1440x750.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>In AI, <strong>2025 is a year for strategic clarity</strong>. We can finally part the fog of war and glimpse where this technology is headed, how companies &amp; governments are likely to react to it, and how you, dear reader, should think about its effects on your life and work. </p><p>In this essay I explain some of my guesses for where AI development is going in 2025, using a dense set of heuristics and mental models that I hope are particularly useful for those who work in AI policy.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.pathwaysai.org/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading AI Pathways! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>I argue for several framings that are key to predicting future developments:</p><ul><li><p><strong>AI will eventually diffuse widely. </strong><a href="https://x.com/EpochAIResearch/status/1900264630473417006">Costs are falling extremely rapidly</a>, and distillation and compression are powerful. It will simply not be very expensive, in 2030 $, to make AGI.</p></li><li><p>Reasoning model progress is driven by repeated iteration in domains with verifiable <em>or close to verifiable </em>reward. The number of domains can be expanded over time to achieve wide coverage. This is <em>complemented </em>by periodic scaling up of pretraining.</p></li><li><p><em><a href="https://zhengdongwang.com/2024/12/29/2024-letter.html">&#8220;The model does the eval&#8221;</a></em>&#8212;models are good at things which we can build evals for, and poor at things for which we struggle. However, building evals becomes easier with more capable models.</p></li><li><p><strong>Agents are </strong>key to economic effects, not chatbots, but they are expensive to run! On current trends, there will be a &#8220;capabilities overhang&#8221; for a few years, with significantly more demand for compute than can be met globally.</p></li><li><p>An easy way to measure the abilities of AI agents is their <strong>time horizon</strong>&#8212;the length of time they can act autonomously and reliably. The reliable time horizon of frontier AI agents will be significantly ahead of others, but given enough time even open source agents will be able to act as a &#8220;drop in remote worker&#8221;.</p></li><li><p>The <strong>&#8220;automated researcher&#8221;</strong> is possible, and we should expect them to significantly accelerate progress over the next few years.</p></li></ul><div class="pullquote"><p>&#8220;How did you go bankrupt?&#8221; Bill asked. </p><p>&#8220;Two ways,&#8221; Mike said. &#8220;Gradually and then suddenly.&#8221;</p><p><strong>Ernest Hemingway, The Sun Also Rises</strong></p></div><h1>Tick-tock?</h1><p>In the midst of a chaotic information environment, I think it&#8217;s worth stepping back to reflect upon how far we&#8217;ve come and consider the current drivers of AI progress. Throughout 2023 and into late 2024, we have seen the steady development of language models into AI chatbots, as they become more and more integrated into our daily lives. We use them as search tools, as writing assistants, but also increasingly for advice, counsel, and as a <em>brain partner</em>.</p><p>We are now moving into a new, even more rapid phase of AI development, focused on autonomous AI agents and reasoning models. The term &#8220;agent&#8221; is overhyped, but the core concept is still valuable: an autonomous system that can take actions over many steps to achieve its goals, without needing direct supervision. In today&#8217;s context, this increasingly means actions on the internet or in a coding environment.</p><p>The main engine of this new rush in progress is the breakthrough, years in the making, of the ability to get AI models to <em>think</em> reliably for longer periods of time, reasoning their way to a better answer. We first saw true inference-time scaling with OpenAI&#8217;s <a href="https://openai.com/index/introducing-openai-o1-preview/">o1 release</a>, although researchers had been trying to make variations work for years prior. <a href="https://arxiv.org/pdf/2501.12948">DeepSeek&#8217;s r1 paper</a> lifts the veil from the core research idea: using reinforcement learning on LLMs, with rewards on the outcome of tasks with verifiably correct answers, will automatically teach the models to error correct, generate hypotheses, check their work, and reason towards a final answer. <strong><a href="https://situational-awareness.ai/from-gpt-4-to-agi/#Unhobbling">&#8220;Unhobbling&#8221;</a> has arrived in earnest.</strong></p><p>A real piece of magic here is that the model&#8217;s learned reasoning heuristics&#8212;the ability to consider an idea, then backtrack and reconsider upon gaining further information&#8212;generalize outside of the domains it was trained on. You can ask DeepSeek&#8217;s r1 to write you a story, and its chain of thought will follow similar patterns, despite clearly not being trained with RL to write stories.</p><p>This generalization shows a tantalizing path forward to further, self-reinforcing, advances in AI&#8217;s capabilities. Of course, to some extent these will still depend on the ability to verify the correctness of an answer, and get reward for the RL training process. Many commentators have pointed out that this may significantly limit the potential of reasoning models in the short term, because only a relatively narrow range of tasks have exact <em>and cheap</em> verification.</p><p>But very often, <strong>verification need not be exact</strong>! There are a surprising number of ways to obtain proxy rewards with a combination of <a href="https://verdict.haizelabs.com/">specialized models</a>, heuristics, and <a href="https://dspy.ai/">format specifications</a>, even in relatively open-ended domains like <a href="https://openai.com/index/harvey/">legal writing</a> or <a href="https://docs.anthropic.com/en/docs/agents-and-tools/computer-use">navigating</a> <a href="https://openai.com/index/introducing-operator/">web pages</a>. These are domains with <strong>pseudo-verifiers</strong>. For many web browsing or enterprise software navigation tasks, synthetic environments and rollouts can be generated which are sufficiently realistic (consider that a huge portion of white collar work simply occurs in Google Docs, Office, Gmail/Outlook, and Slack). Researchers must be cautious, of course, to not <a href="http://sohl-dickstein.github.io/2022/11/06/strong-Goodhart.html">Goodhart</a> or over-optimize models for these proxy rewards, causing worse performance in the real world. And using a pure LLM reward model as a verifier without any <em>grounding</em> will <a href="https://arxiv.org/abs/2407.21787">usually fail</a> to scale. But often, <a href="https://arxiv.org/abs/2206.05802">the implicit gap</a> between the difficulty of verification and the difficulty of generation will be sufficient to progress in important domains.</p><p>In hindsight, perhaps it&#8217;s not surprising that this works so well. Previous attempts to get models to reason their way through complex problems often depended on supervising the correctness of each step of reasoning&#8212;<a href="https://www.stephendiehl.com/posts/process_reward/">process-based supervision</a>. But especially for large and complex tasks, why should we expect models to follow the same reasoning process as humans? AI models have <em>non-smooth skill distributions</em>: in contrast to humans, we cannot reliably predict how capable an AI model will be on closely related tasks. This property is improving as models get better and more robust, but there are still, for example, many types of problems for which GPT-4o is worse than a nominally weaker competitor, or where performance varies drastically based on minor prompt differences. As a result, the model designer should not try and force the models down a reasoning path that is most natural for humans. Instead, <a href="https://x.com/brandonwilson/status/1883914771726315810">let the model figure out for itself</a> the best path to solve the problem: give them a goal and let them run. <em><a href="https://www.youtube.com/watch?v=CM_DP7pkJQk">The models just wanna learn.</a></em></p><p>In parallel to advances in reasoning (and post-training), we also expect improvements via scaling up pretraining. <a href="https://epochai.substack.com/p/ai-progress-is-about-to-speed-up">Grok 3 and particularly GPT-4.5</a> are the early signs of this&#8212;GPT-4.5 is clearly a much bigger model, but it hasn&#8217;t received significant RL fine tuning for reasoning, so is most clearly comparable to GPT-4 to see the effects of scale. And what do we see? The public benchmarks are relatively underwhelming (mostly due to saturation), but the model topped the Chatbot arena and qualitatively seems to be an extremely good model, with some users reporting that they prefer it for common coding questions.</p><p>Overall, development seems to be trending towards a &#8220;<a href="https://en.wikipedia.org/wiki/Tick%E2%80%93tock_model">tick-tock</a>&#8221; model, in which pre-training scale-ups every few years are complemented by increasingly fast progress in continually finetuning the models using RL across a spread of verifiable or pseudo-verifiable domains.</p><p>Significant efficiency improvements are also being driven by the power of <strong>distillation</strong>. As researchers train larger and more capable models with each generation, they can use them to generate data and reward signals to train smaller, more compressed, cheaper models which retain much of the original capability. The abilities of the small Gemma 3 models would have shocked the AI researchers of 2021. The fact that such strong capabilities are possible in so few weights is a powerful hint from the universe that many components of human intelligence are just not that complex, and that this technology is ultimately destined to proliferate cheaply.</p><p>These factors drive the widespread, albeit lagged, <strong>capability diffusion </strong>of AI into smaller and cheaper models over time, including those created by non-frontier AI labs. Because larger models tend to be more capable, and yet more expensive to run, AI labs are increasingly incentivized to follow a scheme of &#8220;train large strong model, distill into smaller cheap model&#8221;, as seen with Gemini 2.0 and o3-mini. Wide deployment then ends up being limited largely by the effectiveness of distillation and compression.</p><p>So far, we have seen distillation (and other algorithmic advances) improve over time, to the point that small models can continually increase in performance despite not changing in size. For economically useful tasks, models at the frontier are still markedly more valuable than their smaller, less capable cousins. But most tasks which are highly valuable have a fixed &#8220;difficulty&#8221;&#8212;at some point, due to distillation and compute improvements, models of a fixed low $/token price will be capable of automating the work of e.g. a junior investment bank analyst, whilst the most capable models will still be struggling to complete much higher complexity tasks. This dynamic will likely continue until the vast majority of economically valuable tasks can be cheaply automated, even if it takes some time to convert the models from expensive to cheap&#8212;although many difficulties remain, especially around tasks which are harder to verify or build good environments for.</p><p>Finally, in the bigger picture, all the progress we have seen so far is a result of the effort put into building better and better <strong>evaluations</strong> for capabilities. Anything we can robustly quantify, we can turn into a benchmark&#8212;and then <em><strong><a href="https://zhengdongwang.com/2024/12/29/2024-letter.html">the model does the eval.</a></strong> </em>In the long term, this means that we should expect AI models to do well on almost anything we can specify and quantify, and poorly in things we struggle to define rigorously, like top 0.1% fiction writing ability.</p><h1>The Automated Researcher Dream</h1><p>A huge driver of progress, still underestimated, is the advent of AI agents capable of AI research: <a href="https://epoch.ai/blog/interviewing-ai-researchers-on-automation-of-ai-rnd">automated researchers</a>. The frontier AI labs are racing fast towards end-to-end AI R&amp;D agents. They know that once this capability is achieved, progress could be made with much greater velocity, potentially in a self-reinforcing loop&#8212;newly discovered algorithmic advances themselves helping to bring the next breakthroughs closer. The best AI researchers in the world are rushing headlong into a grand project to automate their own jobs. Indeed, there are rumors that within Anthropic, some researchers have raised concerns to management that their jobs could be at risk.</p><p>Of course, AI safety advocates have many concerns over this: it may be harder to supervise in some way&#8212;how could we tell whether our automated researcher is capable of robustly evaluating the latest, more capable AI model? This problem, known as <em><a href="https://arxiv.org/abs/2211.03540">scalable oversight</a></em>, has long been a reason for some to argue against building superhuman AI systems. But I&#8217;m more optimistic: the advent of inference-time compute scaling implies that the scalable oversight problem is potentially solvable by simply providing more compute to the supervisor or evaluator model.</p><p>However, automated research has <a href="https://inferencemagazine.substack.com/i/155018281/ai-research-will-be-automatable-but-the-practical-details-will-matter-a-lot">more difficulties than meets the eye</a>. First, most AI lab research teams are compute bottlenecked for their experiments, and are limited to some GPU allocation handed down from on high. Researchers are strongly encouraged to use all of their compute, and typically have far more useful experiment ideas than they can carry out, even if they somehow coded them up instantly. To achieve significantly better utilization of a compute budget and thus faster research, automated researchers need to have better ideas than the average frontier AI lab researcher: quite unlikely, at least in the short term.</p><p>That leaves the other parts of the research loop, most prominently <em>engineering</em>: coding up new research ideas and developing faster or more efficient infrastructure. Significant parts of frontier AI lab workflows are bottlenecked on simple implementation speed and testable software improvements (e.g., <a href="https://metr.org/blog/2025-02-14-measuring-automated-kernel-engineering/">CUDA kernels</a>), and this looks much more tractable for the first automated researchers to tackle. I believe that automated researchers are likely to provide a large boost here, but it may be effectively a one-time boost to overall research speed without improvements in the other parts of the research loop.</p><p>Overall then, the incentives are strong for AI labs to race towards capable automated researchers. Attaining this capability means that labs are less talent constrained: it enables a pure conversion of compute into intellectual effort, and implies that even labs which struggle to attract talent are guaranteed at least some baseline of research ability, provided they can obtain a good enough starting model. </p><p>However, the biggest unanswered question is still that we don&#8217;t know how fast the automated researcher feedback loops will run. Will we get a modest &#8220;one and done&#8221; bump in the short-term from automating engineering, or will the self-reinforcing feedback loop be so strong that we unlock significantly more capable automated researchers in short order? To some extent this is a fundamental question about the difficulty of AI research itself, and whether the marginal research secret (or &#8220;micro-Nobel&#8221;) becomes easier or harder to obtain over time, if you are continually improving at research ability. Regardless, we will likely find out the answer soon: I currently believe leading AI labs are on track to have the first fully automated researcher prototypes <strong>by the end of 2025.</strong></p><p>An AI research ecosystem with a dramatically larger capability differential between labs only ~6 months apart in progress may have interestingly destabilising effects. As new capabilities appear even quicker, new implications for economic growth and national security are unlocked with even less time for companies and governments to react. Currently, the speed of automated research is set to be closely guarded by AI labs&#8212;I think that reporting some statistics about this to say, the Office of Science and Technology Policy in the U.S. government could help improve decision making significantly in the future.</p><h1>Agents and their Effects</h1><p>I&#8217;ve been playing with Deep Research and Operator for several weeks now, and I&#8217;m convinced: these systems<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> are our first glimpse at the agents of the future. They have an unprecedented degree of coherence and reliability over long time horizons, despite many rough edges. It&#8217;s often captivating to see these models generate tens or even hundreds of thousands of tokens of reasoning before giving their final answer.</p><p>However, there are still many issues to overcome as this new form-factor develops. Agents operating within a browser have key limitations on a technical side: reliability (which has noticeably increased, driven by dataset improvements), visual understanding (for operator-like agents), planning capability for high level tasks, and familiarity with key common apps and websites. This last one is grounds for optimism about the speed of economically-relevant development: since vast quantities of office work takes place essentially within a handful of apps (Google Suite, Office, GMail, Slack, etc), it should be possible to vastly increase the reliability of agents in these specific apps by creating handmade training environments for each.</p><p>There is a framework for describing progress here which I quite like, called <em><strong><a href="https://www.lesswrong.com/posts/BoA3agdkAzL6HQtQP/clarifying-and-predicting-agi">t-AGI</a></strong></em>. If we take AGI to mean the &#8220;<a href="https://blog.samaltman.com/three-observations">drop-in remote worker</a>&#8221; imagined by many <a href="https://darioamodei.com/machines-of-loving-grace">AI lab leaders</a>, the idea of t-AGI is that we should, before full AGI, expect to have systems capable of acting like a drop-in remote worker for time-bounded tasks that would take an expert human, say, 30 minutes. Then, upon further development, the AI system will become capable of 2 hour tasks, and so on. We should also consider the probability of success: the system may be capable of the average 30 minute task with an 80% chance of success, and we can plot the progress over time at a fixed success rate.</p><p>But this is too low-resolution, since we know that <em><strong>AI capabilities are spiky</strong></em>! Models which can reliably code features that would take an expert software engineer 2 hours may struggle terribly to do things that a junior consultant or physicist can do in 10 minutes. Of course, there is some amount of general capability transfer, and in many ways AI progress for the past few years has had the effect of making things less spiky. But overall, I currently prefer to refer to t-AI, not t-AGI, and describe things in specific but broad domains.</p><p>We already see this t-AI dynamic with agents to some degree, especially in areas easier to evaluate like software engineering and <a href="https://metr.org/AI_R_D_Evaluation_Report.pdf">AI research</a> (as <a href="https://metr.github.io/autonomy-evals-guide/openai-o1-preview-report/">shown by METR</a>). I currently think we have AIs averaging 80% reliability at a broad sweep of 10 minute tasks, though the time horizon is probably closer to 30 minutes or longer for specific domains like software engineering and compiling basic research reports from the web&#8212;and longer still for 50% reliability. I expect these time horizons to extend to ~8-12h by mid 2026, and a full day by 2027.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!8pQJ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1af19aa-b849-4333-9cff-82663ca601e0_1440x750.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!8pQJ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1af19aa-b849-4333-9cff-82663ca601e0_1440x750.png 424w, https://substackcdn.com/image/fetch/$s_!8pQJ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1af19aa-b849-4333-9cff-82663ca601e0_1440x750.png 848w, https://substackcdn.com/image/fetch/$s_!8pQJ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1af19aa-b849-4333-9cff-82663ca601e0_1440x750.png 1272w, https://substackcdn.com/image/fetch/$s_!8pQJ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1af19aa-b849-4333-9cff-82663ca601e0_1440x750.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!8pQJ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1af19aa-b849-4333-9cff-82663ca601e0_1440x750.png" width="1440" height="750" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b1af19aa-b849-4333-9cff-82663ca601e0_1440x750.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:750,&quot;width&quot;:1440,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!8pQJ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1af19aa-b849-4333-9cff-82663ca601e0_1440x750.png 424w, https://substackcdn.com/image/fetch/$s_!8pQJ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1af19aa-b849-4333-9cff-82663ca601e0_1440x750.png 848w, https://substackcdn.com/image/fetch/$s_!8pQJ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1af19aa-b849-4333-9cff-82663ca601e0_1440x750.png 1272w, https://substackcdn.com/image/fetch/$s_!8pQJ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1af19aa-b849-4333-9cff-82663ca601e0_1440x750.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Research by METR, benchmarking agent performance against the time taken by humans.</figcaption></figure></div><p>We also see it with Operator and Anthropic&#8217;s computer-use agents, which can often reliably complete web browsing tasks of short duration (e.g., 1-10 minutes) but which struggle to continue the more actions are required. I expect this increase in <em>t</em>, for many domains, to be one of the biggest factors underlying the development of economically useful models. Deep Research is currently a leading indicator of this&#8212;the model is capable of searching over the web up to ~30 minutes in duration, before compiling a report about its findings.</p><p>Operator inherently delegates authority to humans for some actions which are particularly important or sensitive. This functions as an implicit "decision threshold", a quality that is partially social and cultural, which we can expect to rise over time as the system proves its reliability &amp; trustworthiness. Eventually, this threshold may be high enough for contracting or managing workers, making major purchases, sending critical emails, and more. The <a href="https://www.dwarkeshpatel.com/p/ai-firm">autonomous corporation</a> beckons?</p><h1>Compute Bottlenecks</h1><p>Now, suppose an AI company develops a 60m-AI agent for many tasks common in the professional workplace, including software engineering, producing, analyzing, and reviewing reports &amp; slide decks using publicly available information, drafting or editing articles or papers, and so on&#8212;tasks chosen because they appear to be amongst the easiest to automate with current technology. This AI agent would be immediately extremely valuable to many businesses&#8212;but <strong>how many businesses would actually be able to run it?</strong> Do we have enough compute for widespread economic benefits? Indeed, labs have shown some indications that reasoning agents are strongly compute constrained: even paying for OpenAI&#8217;s pro $200 subscription only gives you 120 Deep Research queries per month.</p><p>Assuming that this 60m-AI agent might be based on a model of similar size to DeepSeek&#8217;s r1 (or indeed larger), then it requires at least a full set of 8xGH200 GPUs to run. This is the size of NVIDIA&#8217;s latest AI server, costing $300k. Some basic <a href="https://x.com/evanjconrad/status/1881937662787117438">napkin math</a> shows that OpenAI&#8217;s announced Stargate project could sustain running at least ~3.1 million of these agents 24/7 by 2029 for a cost of $500b, (albeit not counting batching, so a very loose lower bound). </p><p>In 2025, the approximate existing stock of NVIDIA H100s in the US could sustain, on similar assumptions, at least ~125k agents. If we assume distillation and model compression improves on trend, this capability could eventually fit onto a single chip, making the numbers ~25m and ~1m respectively (let me know in the comments if you have better numbers!). The excellent <span class="mention-wrap" data-attrs="{&quot;name&quot;:&quot;Rohit Krishnan&quot;,&quot;id&quot;:12282408,&quot;type&quot;:&quot;user&quot;,&quot;url&quot;:null,&quot;photo_url&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/0aa4c22d-4b25-4bec-9587-3ec4d4dcce01_2228x2228.jpeg&quot;,&quot;uuid&quot;:&quot;5a02d403-efb3-43ab-9d07-5f959359ee85&quot;}" data-component-name="MentionToDOM"></span> <a href="https://www.strangeloopcanon.com/p/what-would-a-world-with-agi-look">provides some similar numbers</a> for the long-term estimate (on a single chip) via a different route.</p><p><strong>These are not huge numbers, </strong>especially if models grow to be capable of automating large portions of work in many professional sectors. Similar calculations may be driving some of Sam Altman&#8217;s desire to build Stargate. Of course, there are many other factors in inference economics which could both alleviate or increase this potential bottleneck. For example, distillation will continue to be strong (enabling older GPUs to be used at scale), inference runs well on older GPUs or alternative chip providers, other algorithmic factors like batch sizes, parallelism, and token speed will likely improve, and adoption is set to be slow throughout many sectors even after the technology exists. On the other hand,if pretrained model size needs to be scaled significantly to achieve high reliability, the inherent inefficiency of serving large models (see GPT-4.5 token costs) may delay large-scale use of effective agents for some time. We also have yet to explore how useful it will be to run many agents in parallel for a given task&#8212;perhaps the best configuration consists of a swarm of agents completing different subtasks? And what new forms of work may be enabled by the existence of these agents?</p><h1>Conclusion</h1><p>Overall, I expect <strong>AI agents to be very compute bottlenecked</strong> in the next couple of years, largely because of the sheer speed of progress creating an &#8220;economic overhang&#8221; of sorts. By the time we have 24h AI agents for many common tasks&#8212;perhaps mid to late 2026&#8212;most data center projects currently planned &amp; approved will be nearing completion. Other infrastructure is not so speedy, and things like energy and power transmission construction have significantly longer lead times. At some point, progress (and importantly, economic adoption) may slow down due to fundamental constraints on the availability &amp; cost of chips or energy.</p><p>A large compute supply bottleneck will create new (temporary) political and economic dilemmas for both AI labs and governments. Should compute be preferentially allocated to companies, academic researchers, and other groups, to satisfy the interests of both the public and AI developers? As we move towards capable autonomous scientific researchers, can governments use their compute to accelerate key public interest research projects in many domains of science?</p><p>Longer term however, there is plenty of light at the end of the tunnel, driven by the combination of <a href="https://epoch.ai/data-insights/nvidia-chip-production">increased AI chip production</a> and efficiency improvements for a fixed level of capabilities. Eventually, I expect t-AGI for most common workplace tasks<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a> to be at a level of capability distillable into fairly cheap models, which will greatly decrease the compute requirements and potentially unlock t-AI on consumer devices.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a> <strong>Ultimately, I believe the cost of intelligence will tend towards zero.</strong></p><p>This is the first of two big picture essays bringing together my views on AI progress and where we are headed. The second part, coming soon, discusses the implications of this AI progress for the <strong>economy and geopolitics</strong>. I will argue that in the context of the US-China competition, AI should be viewed as a form of raw <strong>economic advantage</strong>, with compute as the key lever that the US can use to secure this advantage.</p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>See: Gemini Deep Research, Operator, OpenAI Deep Research, Manus, etc.</p><p></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>Estimates compiled from dozens of conversations with lab researchers as well as soon-to-be-released research by METR</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>Note that this is specifically for the current, 2025 distribution of tasks in the workplace. I expect this distribution to change rapidly if automation is rapid&#8212;making AGI a target that continually moves further away, by some definitions.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>Apple is set to be a big winner of this dynamic as long as they can <a href="https://daringfireball.net/2025/03/something_is_rotten_in_the_state_of_cupertino">rescue Apple Intelligence</a>.</p><p></p></div></div>]]></content:encoded></item><item><title><![CDATA[AI Strategy for a New American President]]></title><description><![CDATA[What can we expect for the next few years of U.S. AI policy?]]></description><link>https://www.pathwaysai.org/p/ai-strategy-for-a-new-american-president</link><guid isPermaLink="false">https://www.pathwaysai.org/p/ai-strategy-for-a-new-american-president</guid><dc:creator><![CDATA[Herbie Bradley]]></dc:creator><pubDate>Sun, 05 Jan 2025 02:21:36 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!OC13!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fafc6c13c-0c42-4728-9911-49af59c17369_400x400.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>Welcome to <strong>AI Pathways</strong></em>, <em>a new tech blog that aims to see the future of AI first and make it more publicly legible. In 2025, AI progress is faster than ever, and yet public awareness of where we are headed significantly lags behind the private insider consensus. The possibility space narrows, and it feels like the right time to articulate a unifying view of the likely <strong>pathways </strong>for this transformative technology, looking at both the latest trends in technical AI research and the latest developments in AI policy. However, I will endeavour not to take a purely predictive viewpoint&#8212;<a href="https://michaelnotebook.com/optimism/index.html#fnref35">we are active participants in this future</a> and it is important to see ourselves as capable of shaping it with our imagination.</em></p><p>Two months ago, the political board in the U.S. was flipped, and many assumptions in the world of policy that felt like constants are no more. We&#8217;re in a New Year, and in this, the first piece of <em><strong>AI Pathways</strong></em>, I aim to answer the question: what can we expect from the next government&#8217;s AI policy? And what might this imply for the future development and deployment of the world&#8217;s most advanced AI systems?</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.pathwaysai.org/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading AI Pathways! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>My main message is this: <strong>the strategic landscape has dramatically shifted. </strong>After spending the last two months chatting to AI researchers and tech policy folk around Washington, D.C., San Francisco, New York, London, and Brussels, I'm convinced that almost everyone outside of D.C. is strongly <em><strong>underrating</strong></em> the implications of the election. Many exciting new opportunities have opened up for both AI research and tech policy, while some Biden-era policy paradigms (e.g., pre-deployment evaluations and risk assessment as a dominant form of governance for frontier AI) are on their way out.</p><p>I&#8217;ll briefly list the main takeaways here, before going into specifics. The main policy motifs I see from the new administration are:</p><ol><li><p>An increased focus on <strong><a href="https://alexw.substack.com/p/war">China as a technological adversary</a></strong> and effort to maintain American advantage in AI. This is a core motivation behind many potential policy actions, including boosting domestic energy production, onshoring of more semiconductor manufacturing, export controls on chips, and building out AI applications for national security.</p></li><li><p>One of the most underappreciated effects in AI policy of the U.S.&#8211;China dynamic is greater interest in using AI for national security and military applications. There are two sides to this coin:</p><ol><li><p>building defense-specific AI applications, which will likely lead to the government conceiving of AI as increasingly a defense technology;</p></li><li><p>securing existing frontier AI systems and IP from <a href="https://www.justice.gov/opa/pr/chinese-national-residing-california-arrested-theft-artificial-intelligence-related-trade">foreign corporate espionage</a> or other attacks. As long as the capabilities gap with China is maintained (an open question given DeepSeek&#8217;s impressive progress), there will be strong incentives for tighter security partnerships between the U.S. government and the AI labs.</p></li></ol></li><li><p>A potential conflict in the new admin is between those favoring tighter government control over the security of frontier AI model weights &amp; IP, motivated by viewing AI through the U.S.&#8211;China competition lens, and those arguing in favor of minimal security restrictions &amp; boosting the open-source AI sector to make adoption throughout the economy easier. Senior figures in the new admin&#8217;s tech policy, including David Sacks and Sriram Krishnan, are strong advocates for open-source, and there are good arguments in favor of their views given the likely inevitable diffusion of frontier AI capabilities over time. I strongly believe there is a viable way to unify these two perspectives&#8212;the securitization view and the open-source view&#8212;and hope to articulate a path forward in a future piece.</p></li></ol><p>A key concept here is that as capabilities grow, we should increasingly expect governments to view AI as a <strong>strongly dual-use technology</strong>. Up until now, AI has been an almost entirely <em><a href="https://ora.ox.ac.uk/objects/uuid:ea3c7cb8-2464-45f1-a47c-c7b568f27665/files/maebbb4b2123a17bb923478bec1812d92#page=234">civilian-first</a></em> technology, but government interest in it seems likely to grow along with capabilities that provide more possible applications to defense.</p><p>This leads to some guesses for Trump&#8217;s AI-relevant policy moves:</p><ul><li><p>Moves to boost AI datacenter growth, by <a href="https://ifp.org/future-of-ai-compute/#challenges-to-building-in-america">unblocking permitting</a>, or <a href="https://manhattan.institute/article/a-playbook-for-ai-policy">funding and incentivizing</a> the construction of new energy supply and transmission capacity. U.S. energy demand forecasts for the next 5 years <a href="https://www.csis.org/analysis/strategic-perspectives-us-electric-demand-growth">have increased rapidly</a>, largely due to AI datacenters but also because of greater than expected growth in manufacturing. The electricity grid is underprepared for this, and only speedy action can prevent economic growth being bottlenecked by power.</p></li><li><p>Scaling up of initiatives within the federal government and the defense sector to use frontier AI in national security or military applications. We see steps towards this outside government in the burgeoning <a href="https://archive.is/si8qE">Palantir-Anduril defense tech consortium</a>, OpenAI&#8217;s partnership with Anduril, and Palantir&#8217;s partnership with Anthropic &amp; AWS. I expect <strong>defense-specific finetunes</strong> of leading LLMs to be ultimately sold as products to the government, which strongly motivates a need to secure these model weights from cyber attacks. We may see <strong>export controls on model weights </strong>to prevent some models being deployed outside of U.S. datacenters.</p></li><li><p>Testing and evaluation work within the federal government is likely to move to more national security focused agencies and teams. For example, USAISI may become a much smaller share of this work, and could be moved out of NIST into a more relevant place like the Department of Energy (DOE).</p></li><li><p>Trump&#8217;s administration seems likely to maintain and tighten export controls on semiconductors, to slow down Chinese AI efforts. Their effectiveness will depend on how well the multiple different agencies working on export controls, including BIS in Commerce, can actually be coordinated and run efficiently. DOGE may play a part in the latter effort.</p></li><li><p>Deeper partnerships between frontier AI labs and the U.S. Intelligence Community to improve the cyber and info-security of lab development &amp; deployment. This may involve a so-called public-private partnership (PPP) between e.g., the DOD/DOE and one or more leading AI companies, including potential government security experts embedded in the labs. Due to Elon&#8217;s involvement, x.ai is an obvious lab to watch closely for potential partnerships with the government.</p></li><li><p>A greater unknown is how to deal with the patchwork of currently-proposed state-level AI bills. The more insane the patchwork appears (<a href="https://www.hyperdimensional.co/p/texas-plows-ahead">*cough* Texas *cough*</a>), the greater the incentive for the federal government to pre-empt states by passing some federal level AI bill that overrides them. This might be a light-touch AI bill (to avoid slowing down developers, particularly open-source) aimed at reducing uncertainty for businesses deploying AI systems and boosting economic growth. But it is currently unclear how high this will be on the priority list for Congress, especially given the slim Republican majority.</p></li></ul><p>In summary: if you&#8217;re working on AI policy or on technical research which you think will be useful for policy (particularly on anything in the bucket sometimes called <a href="https://arxiv.org/abs/2407.14981">technical AI governance</a>), then you should seriously consider the <strong>vibe shift</strong> and its implications. Some people like to make <a href="https://laurajung.substack.com/p/062-my-ins-and-outs-for-2025">In/Out lists</a> for the New Year, so my guess is that in terms of usefulness for AI policy, we&#8217;re looking at <strong>In:</strong> work on AI security, ensuring robust &amp; reliable AI for defense applications, and forecasting the likely economic impacts of AI agents; <strong>Out:</strong> AI bias evals, pre-deployment third-party safety evals, safety cases, adversarial robustness and jailbreaking, risk assessment frameworks, and mechanistic interpretability. I will expand on these takes in a separate piece.</p><h2>Background</h2><p>There are already <a href="https://time.com/7174210/what-donald-trump-win-means-for-ai/">several</a> <a href="https://arstechnica.com/ai/2024/11/trump-victory-signals-major-shakeup-for-us-ai-regulations/">good</a> <a href="https://techcrunch.com/2024/11/06/what-trumps-victory-could-mean-for-ai-regulation/">pieces</a> you can read to get a sense of Trump&#8217;s existing statements on AI, as well as hints from those working with the Trump campaign. The main explicit policy action we know of is the existing commitment to repeal Biden&#8217;s AI Executive Order. This seems likely to be replaced with a Trump AI Executive Order at some point.</p><p>The <a href="https://www.project2025.org/policy/">Project 2025 policy agenda</a> is also a useful an indicator of interest from a group connected to the transition team, and contains a variety of AI-relevant suggestions (findable by searching &#8220;AI&#8221; or similar). However, Project 2025 should be viewed more like a list of potential directions, considering the document was drawn up many months ago now.</p><h3>Elon Musk&#8212;A Wildcard</h3><p>Of course, much also depends on Elon Musk&#8212;how influential will he be in the new administration, and what will his preferred AI approach be? Will he maintain a good relationship with Trump for the entire term? </p><p>This is currently very uncertain, since we both don&#8217;t know what work Elon will spend most of his time on (for the short-term it is clearly DOGE), and Elon&#8217;s own preferences for AI policy are unclear, although he has taken a strong interest in its development for many years.</p><p>Many in Silicon Valley AI circles hope that Elon&#8217;s deep familiarity with the AI industry and influence in the Trump administration will help shape the next government&#8217;s policy. So far, this seems likely to play out, potentially to the benefit of x.ai.</p><p>More broadly there are, as we have seen with the recent immigration discourse, strong tensions between the populist right side of the Republican party, and the &#8220;tech right&#8221; (encompassing Elon, Silicon Valley Republicans &amp; libertarians, e/accs, etc). Whether Elon maintains influence depends to a large degree on how these political tensions shake out.</p><h2>The U.S. AI Safety Institute</h2><p>The fate of the U.S. AI Safety Institute (USAISI)&#8212;a small team of technical experts &amp; policy staff within NIST in the Department of Commerce&#8212;is a question on the minds of many in AI policy this month. Trump has promised to repeal the Biden AI Executive Order, part of which concerns pre-deployment safety evaluations of frontier AI systems (USAISI&#8217;s main activity). The later Biden National Security Memorandum also directs USAISI to act as the central point within the government for frontier AI work &amp; engagement with the labs.</p><p>USAISI has attracted attention from <a href="https://www.commerce.senate.gov/2024/12/cruz-calls-out-potentially-illegal-foreign-influence-on-u-s-ai-policy">influential anti-regulation Republicans</a> in 2024 for its collaborations with UKAISI and other non-governmental AI safety organizations. This particularly includes the November International Network of AISIs event in San Francisco, with a side event on <a href="https://www.aisi.gov.uk/work/conference-on-frontier-ai-safety-frameworks">safety frameworks</a> (e.g., regulatory mechanisms) partly run by UKAISI. When I speak to observers in AI policy, many express concern that USAISI simply hedged insufficiently for a Trump, and is now trying to re-work its agenda and focus to be closer to the expected desires of the new administration. A narrowing of <a href="https://www.commerce.gov/news/press-releases/2024/11/us-ai-safety-institute-establishes-new-us-government-taskforce">focus onto AI for national security</a> seems like the right path for them&#8212;they have recruited some valuable technical experts, whose abilities will be a useful resource for other government teams seeking to use frontier AI.</p><p>As a result, a full winding down of USAISI seems unlikely. Although they are small (their technical team is a handful of people, the rest are policy staff), they have accumulated a number of well-known experts at the working level with strong track records. However, both hiring new talent and retaining existing experts will likely be even more challenging than previously. USAISI are known in the community to have a very restrictive conflict of interest policy that nixed at least one potential ex-industry-lab hire (which generally does not bode well for government&#8217;s ability to hire top AI talent).</p><p>Going forward, USAISI seems likely to still serve a useful function as an advisory body for various parts of the federal government with a stake in frontier AI, as well as a convenient point of engagement with the AI labs for <a href="https://www.nist.gov/news-events/news/2024/11/us-ai-safety-institute-establishes-new-us-government-taskforce-collaborate">some parts of pre-deployment testing</a>.</p><p>Another rumored possibility, and perhaps a way to alleviate hiring inefficiencies, is that USAISI may be moved outside of Commerce. In many ways USAISI&#8217;s place in DOC is a historical accident born of Gina Raimondo&#8217;s interest in AI&#8212;given that much of USAISI&#8217;s testing work and coordination has been on national security risks, the Department of Energy is a potential natural home. The DOE has a much larger budget, significant computing expertise within the national labs, the existing <a href="https://www.energy.gov/articles/doe-and-commerce-department-sign-memorandum-understanding-advance-safe-secure-and">DOE-DOC testing collaboration</a>, and is <a href="https://openai.com/index/openai-and-los-alamos-national-laboratory-work-together/">separately running tests</a> with several leading AI labs. Either the DOE or DOD would fit nicely as a home for USAISI, given the increasing interest in the <a href="https://www.nist.gov/news-events/news/2024/11/us-ai-safety-institute-establishes-new-us-government-taskforce-collaborate">testing and use of frontier AI for national security</a> applications &amp; risks.</p><h2>International Collaboration</h2><p>What then, for international collaboration on AI? Under Biden, the international AI governance landscape saw a great proliferation of mostly symbolic fora, agreements, dialogues, summits, policy frameworks, and commitments. Examples include the <a href="https://www.soumu.go.jp/hiroshimaaiprocess/en/index.html">Hiroshima Process</a> (via the G7), the <a href="https://www.oecd.org/en/topics/sub-issues/ai-principles.html">OECD AI Principles</a>, the <a href="https://www.un.org/en/ai-advisory-body">UN AI Advisory Body &amp; report</a>, the AI Safety Summits in <a href="https://www.gov.uk/government/topical-events/ai-safety-summit-2023">the UK</a> and Seoul, and the <a href="https://www.nist.gov/news-events/news/2024/11/fact-sheet-us-department-commerce-us-department-state-launch-international">International Network of AISIs</a>.</p><p>Some of the most significant for frontier AI were the AI Safety Summits (although I am biased, having had a small hand in organizing the first), which were able to secure commitments from the leading companies &amp; many countries to collaborate on testing for frontier AI systems. Despite appearing vague and high-level, these kind of major international commitments can serve as a form of social pressure, since they can be cited in any engagement between countries and companies as a motivator for why the frontier AI labs should do, for example, joint safety testing.</p><p>This works fine if your goal is purely imposing political cost on companies that do not engage with government interest in AI deployments. However, there is nothing binding here&#8212;these commitments do not either force companies to take action or <em>offer additional options</em> (e.g., additional types of deployment mode or form factor), they simply raise the political cost for a company to follow a path that a government may disagree with, a cost that may be worth the price in a more high-stakes situation.</p><p>From an America First perspective, the motivation to engage significantly in internationalizing frontier AI governance looks quite weak given that all the main companies involved are U.S.-based, and so I do not expect as much engagement on AI evaluation &amp; testing as under Biden (whose administration was more ideologically motivated to pursue that project).</p><h2>The International Network of AISIs</h2><p>The <a href="https://www.commerce.gov/news/fact-sheets/2024/11/fact-sheet-us-department-commerce-us-department-state-launch-international">International Network of AISIs</a> is an informal coordination mechanism between the governments of several countries to collaborate on the testing and evaluation of frontier AI systems. I say informal, because it is fundamentally bound by Memoranda of Understanding (MoUs) and other forms of non-hard law agreement.</p><p>The Network was preceded by <a href="https://www.commerce.gov/news/press-releases/2024/04/us-and-uk-announce-partnership-science-ai-safety">the MoU between the U.S. and U.K. AISIs</a>, who collaborated on joint safety for Anthropic&#8217;s Claude 3.5 Sonnet, and <a href="https://www.nist.gov/news-events/news/2024/12/pre-deployment-evaluation-openais-o1-model">OpenAI&#8217;s o1 model</a>. These two AISIs were the first and are, as far as I know, the largest. As a demo of testing amongst a larger group, this pair recently ran a joint safety testing project on <a href="https://www.nist.gov/system/files/documents/2024/11/21/Improving%20International%20Testing%20of%20Foundation%20Models-%20%20%20A%20Pilot%20Testing%20Exercise%20from%20the%20International%20Network%20of%20AI%20Safety%20Institutes.pdf">Meta&#8217;s Llama-3 405B</a> together with Singapore AISI.</p><p>The Network&#8217;s activities are likely to be significantly affected by USAISI&#8217;s fate, the Network&#8217;s Chair. This is because structurally, AI labs based in the U.S. have a strong incentive to work primarily or only with the U.S. federal government on safety testing of their models&#8212;the U.S. government is the most natural partner here, especially when it comes to testing on national security risks like <a href="https://en.wikipedia.org/wiki/CBRN_defense">CBRN</a>. The Network acts as a mechanism to transfer this incentive, via U.S. AISI, to other countries who have the capability to do safety testing, setting a precedent towards internationalization of this testing.</p><p>However, this mechanism only happens via the scope of testing USAISI conducts; if USAISI loses their mandate from the Biden AI EO to act as <em>the</em> central point for safety testing of frontier AI within the U.S. government, we should expect more testing to be done in other parts of the federal government, outside the scope of the International Network. This is doubly true since the Biden White House acted as a forcing function for this centralization to happen, often proactively nudging parts of the government interested in frontier AI to collaborate with U.S. AISI.</p><p>Therefore, it is likely we&#8217;ll see less international collaboration for safety testing on frontier AI models. This has pros and cons: in some sense the Network can be viewed as an attempt to reduce centralization of governance over frontier AI, but on the other hand it may also bring increased security risks or cause confusion by adding too many additional actors with different definitions of safety.</p><p>But will national AI Safety Institutes other than the U.S. or U.K. be relevant to the global trajectory of AI? This move to internationalize evaluation-based AI governance faces a fundamental problem: the United States is the base for of all of the world&#8217;s leading frontier AI labs, except for (arguably) DeepSeek in China. In regulation and governance, other jurisdictions rely purely on the desire of foreign tech companies to make profit from deploying there (e.g., the so-called Brussels effect, significantly overhyped by the EU Commission).</p><p>Almost none of the other countries in the network have so far built sizeable safety testing teams staffed with technical experts, or shown intention to do so. Singapore AISI, for instance, is primarily focused on encouraging more safety research in neglected directions via academia and <a href="https://aiverifyfoundation.sg/">building evaluation tools</a> for the ecosystem. Other countries (Canada, Australia, France, Japan) have announced or begun setting up their own AISI, yet these look more like a gesture towards AI&#8217;s importance from their respective governments rather than efforts likely to lead to highly productive research organizations. And does <a href="https://www.kenyanews.go.ke/ambassador-thigo-leads-kenyas-participation-in-historic-ai-safety-network-launch-in-the-us/">Kenya</a> really need an AI Safety Institute? To what extent is pushing an international consensus on things like best practices for misuse risk evaluations actually useful, given the rapidly shifting state of evaluations research?</p><p>Finally, it&#8217;s worth noting the inclusion of the EU AI Office in the International Network, which is in many ways the &#8220;odd one out&#8221;. Unlike the other AISIs, the Office is a regulatory body within the EU Commission, tasked with implementing the EU AI Act. This has given NIST a delicate line to navigate, since the Network thus ties NIST to a foreign regulator, even if it is for informal collaboration purposes.</p><h2>Securitization</h2><p>The securitization of AI is increasingly a hot topic in the tech policy world, which I define as work to improve the cyber and information security of frontier AI training and deployment&#8212;particularly model weights and IP. The last year has seen numerous public calls to improve this security, including Leopold&#8217;s <a href="https://situational-awareness.ai/">Situational Awareness</a> essay, and the RAND <a href="https://www.rand.org/pubs/research_reports/RRA2849-1.html">securing model weights report</a>. The primary motivation for this goal is to maintain the U.S. lead in AI capabilities by ensuring that other actors, including nation-state adversaries, cannot hack or otherwise obtain frontier AI capabilities, under the assumption that these capabilities will become increasingly important for national security.</p><p>As many observers have noted, the effectiveness of this strategy is much reduced when both China and open-source are close behind the frontier AI capabilities of &#8220;closed&#8221; industry labs like DeepMind, and where that capabilities gap may credibly narrow. Nonetheless, my personal intuition is that securitization is still very valuable to pursue, because it provides much greater optionality for long-term U.S. AI strategy. Many China hawks with connections to the Trump administration are <a href="https://therepublicjournal.com/journal/11-elements-of-american-ai-supremacy/">also advocates for this view</a>. </p><p>Relatedly, we see a strong desire to apply frontier AI systems for national security and military applications, including moves like <a href="https://www.anduril.com/article/anduril-partners-with-openai-to-advance-u-s-artificial-intelligence-leadership-and-protect-u-s/">OpenAI&#8217;s partnership with Anduril</a>, <a href="https://investors.palantir.com/news-details/2024/Anthropic-and-Palantir-Partner-to-Bring-Claude-AI-Models-to-AWS-for-U.S.-Government-Intelligence-and-Defense-Operations/">Palantir&#8217;s partnership with Anthropic &amp; AWS</a>, and OpenAI&#8217;s <a href="https://openai.com/index/openai-appoints-retired-us-army-general/">board appointment of an ex-NSA general</a>. Indeed, fine-tunes of leading frontier LLMs for defense applications are a likely way that frontier AI labs end up increasing their security in the short-term. The U.S. government&#8217;s interest in securitization of frontier AI is likely strongly correlated with the utility of these defense applications&#8212;and given the rapid progress in AI development towards autonomous agents, it is plausible that some applications could soon become a significant national security capability.</p><p>This overall direction also coincides with Silicon Valley&#8217;s increasing recognition, in the making since the Ukraine invasion, of the necessity of tech&#8217;s involvement in maintaining American security and military leadership, and the stakes of the geopolitical tensions with China. In stark contrast to the anti-military vibe of several years ago, defense tech is finally cool in SF&#8212;I attended a recent Palantir party where guests cheered at the host&#8217;s description of the &#8220;Stanford to crypto to AI to defense tech&#8221; pipeline.</p><p>Early moves by the government to encourage this security may include creating stronger partnerships between the national security world and the frontier labs. For example, this could include embedding government cyber and info security experts into labs to beef up their measures. </p><p>The DOE has the potential to play a role here, given that it contains the National Labs, which posesses several AI compute clusters (both classified and unclassified) as well as significant high-performance computing expertise. If high-security AI training or deployment clusters are needed to support national security applications, the DOE is a natural home for their construction, in partnership with industry labs.</p><h2>Energy, Semiconductors, and Compute</h2><p>For context, there are already <a href="https://manhattan.institute/article/a-playbook-for-ai-policy">several</a> <a href="https://ifp.org/compute-in-america/">excellent</a> <a href="https://ifp.org/future-of-ai-compute/">policy</a> proposals arguing the case for serious investment in energy for U.S. datacenters, and in the compute build-out itself. In this, the new administration is likely to support similar high level industrial strategy themes to the Biden administration: building more energy supply, onshoring more chip manufacturing, and building more compute capacity. President Trump&#8217;s tariff agenda may become relevant here, since he floated the idea of putting tariffs on chips from Taiwan on the Joe Rogan podcast.</p><p>Even if AI capabilities stagnate, we still expect energy demand to <a href="https://www.csis.org/analysis/strategic-perspectives-us-electric-demand-growth">increase drastically over the next 5-10 years</a>&#8212;driven by demand from batteries, semiconductors, and other manufacturing being onshored. Adding AI on top, the potential of increasingly capable AI agents to create vast deployments throughout the economy, using many times more compute than now, means that the U.S. is in danger of severely underbuilding its power supply and energy transmission infrastructure. Much recent industrial policy from both Democrats and Republicans recognizes this and is advocating for huge investment in energy. </p><p>The Trump administration seems likely to continue this trend. One indicator is the proposed appointment of Jacob Helburg to undersecretary of State for economic growth, energy, and the environment. Helburg is a prominent Silicon Valley Republican with deep connections to the AI industry, who has <a href="https://therepublicjournal.com/journal/11-elements-of-american-ai-supremacy/">recently advocated for</a> facilitating energy investment via permitting reform, including oil, gas, and nuclear, as well as reshoring manufacturing of all elements in the AI supply chain.</p><p>The CHIPS Act, which subsidizes domestic chip manufacturing, has received criticism from some Republicans, primarily due to provisions around <a href="https://fortune.com/2023/03/02/biden-chips-woke-republican-senators-mitt-romney-child-care-union-labor/">union labor</a> and climate research. However, its core idea of boosting U.S. chip production still receives bipartisan support. Given the urgency and looming supply bottlenecks, further legislation to incentivize private-sector investment in energy, semiconductors, or AI datacenters could be on the table for the new administration.</p><h4>A Note on DOGE</h4><p>DOGE&#8217;s focus is likely not particularly relevant to AI policy, though it may have some impact depending on which teams and offices it is directed to focus on by Elon and Vivek. To briefly summarize, the latest <a href="https://www.business-standard.com/world-news/elon-musk-taps-trump-s-ex-aide-tech-execs-for-staffing-doge-initiative-124121900097_1.html">rumor about DOGE</a> is that it:</p><ul><li><p>is likely to operate analogously to Palantir: a talented team of software engineers, many of whom will be &#8220;forward-deployed&#8221; and embedded inside government agencies. These embedded SWEs will be able to see how things work, solve problems, and build tools within the bureaucracy, but also have rapid lines of communication to the White House, in case executive action can help fulfil DOGE&#8217;s mission;</p></li><li><p>will primarily focus on efficency, both by slimming down staffing at key government offices, but also by building efficient technology to let government employees be more productive.</p></li></ul><p>Most of the parts of the government DOGE could focus on to achieve the greatest financial savings are not that relevant to AI, but one suggestion might be this: dedicate some DOGE staff to the Bureau of Industry and Security (BIS) in the Department of Commerce. The BIS handles (a part of) the implementation and enforcement of export controls on semiconductors, and its function can plausibly become significantly more effective with DOGE assistance.</p><div><hr></div><p>If you have any thoughts or disagreements with this piece, do post in the comments section or on Twitter, and let me know what you think!  You can also reach out at mail [at] herbiebradley.com for a chat.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.pathwaysai.org/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading AI Pathways! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item></channel></rss>