<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[Air Street Press: State of AI]]></title><description><![CDATA[The State of AI covers our annual Report, the most widely read and trusted analysis of key developments in AI, as well as our Newsletter, analysing the most important AI technology, research, geopolitics and startups.]]></description><link>https://press.airstreet.com/s/stateofai</link><image><url>https://substackcdn.com/image/fetch/$s_!txvE!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2be7fcaf-7116-4fef-936e-f061e4fdbd87_1138x1138.png</url><title>Air Street Press: State of AI</title><link>https://press.airstreet.com/s/stateofai</link></image><generator>Substack</generator><lastBuildDate>Sun, 12 Apr 2026 19:52:31 GMT</lastBuildDate><atom:link href="https://press.airstreet.com/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Air Street Capital Management Ltd.]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[airstreet@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[airstreet@substack.com]]></itunes:email><itunes:name><![CDATA[Air Street Press]]></itunes:name></itunes:owner><itunes:author><![CDATA[Air Street Press]]></itunes:author><googleplay:owner><![CDATA[airstreet@substack.com]]></googleplay:owner><googleplay:email><![CDATA[airstreet@substack.com]]></googleplay:email><googleplay:author><![CDATA[Air Street Press]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[State of AI: April 2026 newsletter]]></title><description><![CDATA[US Government blacklists Anthropic as Iran bombs AWS data centers. Plus: $19B revenue in weeks, industrial-scale distillation wars, and an mRNA dog cancer vaccine designed by ChatGPT.]]></description><link>https://press.airstreet.com/p/state-of-ai-april-2026-newsletter</link><guid isPermaLink="false">https://press.airstreet.com/p/state-of-ai-april-2026-newsletter</guid><dc:creator><![CDATA[Air Street Press]]></dc:creator><pubDate>Sun, 12 Apr 2026 16:11:39 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/79cd9165-f7a1-4d71-886c-9c2b88572b13_1776x990.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Dear readers, </p><p>Welcome to the latest issue of the <strong>State of AI</strong>, an editorialized newsletter that covers the key developments in AI policy, research, industry, and start-ups from February 1 to April 7, 2026. First up, a few news items:</p><ul><li><p><strong><a href="https://press.airstreet.com/p/air-street-capital-announces-232m-fund-iii">Air Street Capital Epoch 3 is live!</a></strong> $232M to continue backing AI-first companies across the US and Europe in software, dev/infra, techbio and defense. </p></li><li><p><strong><a href="https://www.raais.co">RAAIS 2026</a></strong> is back in London on June 12. This year&#8217;s speakers include Raia Hadsell (VP Research, Google DeepMind), Roberta Raileanu (Senior Staff Research Scientist, Google DeepMind), Jeff Hawke (Co-Founder &amp; CTO, Odyssey), and Philip Johnston (Co-Founder &amp; CEO, Starcloud - yes, data centers in space). Come along and support the RAAIS Foundation&#8217;s mission in AI education and research.</p></li><li><p><strong>Air Street AI meetups</strong> are coming up in <a href="https://airstreet.com/events">SF on April 28 and NYC on May 14</a>.</p></li><li><p>We&#8217;re recruiting <strong>Research Analysts</strong> for the <strong>State of AI Report</strong>. If you live and breathe this stuff and want to help us build the next edition, <a href="mailto:nathan+soai26@airstreet.com">get in touch</a>.</p></li><li><p>If you&#8217;re <strong>looking for a new challenge</strong> in our portfolio or community, come chat with <a href="mailto:guy@airstreet.com">Guy Kendall</a>, Air Street&#8217;s new Head of Talent.</p></li><li><p><strong>Air Street Press</strong> featured the <a href="https://press.airstreet.com/p/a-letter-from-munich-security-conference-2026">A Letter from the Munich Security Conference 2026</a> and <a href="https://press.airstreet.com/p/dreaming-in-latent-space">Dreaming in Latent Space</a>.</p></li></ul><p>I love hearing what you&#8217;re up to, so just hit reply or forward to your friends :-)</p><div><hr></div><h3><strong>The Pentagon Standoff</strong></h3><p>How did we even get here? The defining industry story of this quarter wasn&#8217;t an agentic model launch or more exotic financial engineering, but a constitutional confrontation between a sitting president and an AI lab over who gets to decide how frontier models are used in war.</p><p>In late February, Under Secretary of War Emil Michael <a href="https://www.ft.com/content/d8c2969f/">publicly criticized</a> Anthropic for maintaining usage restrictions, including prohibitions on autonomous weapons and domestic mass surveillance, in its Pentagon contracts. Anthropic had won a $200M DOD contract alongside other frontier labs last summer, but its insistence on binding safety guardrails placed it on a collision course with a Trump administration that viewed such constraints as vendor overreach. On February 27, the White House issued a directive ordering all federal agencies to phase out Anthropic&#8217;s products within six months. Literally hours later, OpenAI CEO Sam Altman <a href="https://x.com/sama/status/2027578652477821175">announced</a> a deal to deploy its models on the Pentagon&#8217;s classified network, with contractual &#8220;red lines&#8221; against autonomous weapons and domestic mass surveillance allegedly written into the agreement. He followed up days later with an <a href="https://x.com/sama/status/2028640354912923739">internal memo</a> detailing amendments that added explicit language: &#8220;The AI system shall not be intentionally used for domestic surveillance of U.S. persons and nationals.&#8221;</p><p>By March 4, three cabinet agencies, State, Treasury, and HHS, had <a href="https://www.rappler.com/technology/us-state-department-switch-openai-agencies-phase-out-anthropic/">switched from Anthropic to OpenAI</a>, with the State Department migrating (read: downgrading) its in-house StateChat to GPT-4.1 (grief!). On March 5, the Pentagon formally notified Anthropic of the phase-out and its designation as a &#8220;supply chain risk&#8221;. Anthropic <a href="https://www.ft.com/content/1aeff07f/">sued the Trump administration</a> on March 9, challenging the blacklisting as retaliatory. By March 26, a federal court <a href="https://www.ft.com/content/db1392dc-5042-4ed4-873e-f826429b5f0e">blocked the administration</a> from punishing Anthropic further while the case proceeded.</p><p>This matters beyond the Beltway because it established a precedent: the US government now treats AI vendors not as commodity suppliers but as strategic actors whose policy positions can trigger executive retaliation. It also surfaced a genuine dilemma. The <a href="https://www.wsj.com/tech/ai/how-ai-is-turbocharging-the-war-in-iran/">Wall Street Journal reported</a> that AI-powered targeting and decision-support systems were already accelerating the pace of US military operations in the Iran conflict. In early March, Iran <a href="https://fortune.com/2026/03/09/irans-attacks-on-amazon-data-centers-in-uae-bahrain-signal-a-new-kind-of-war-as-ai-plays-an-increasingly-strategic-role-analysts-say/">struck Amazon Web Services data centers</a> in the UAE and Bahrain with drone strikes - the first deliberate military attack on commercial cloud infrastructure in history. Iranian state media justified the targets on the grounds that the US military was running AI systems, including Anthropic&#8217;s Claude, on AWS for intelligence analysis and war simulations. Two out of three AWS availability zones in the UAE region went down simultaneously, breaking standard redundancy models. Cloud infrastructure is now a theatre of war. To make matters worse, the IRGC has now <a href="https://www.ft.com/content/">threatened</a> to target Stargate Abu Dhabi&#8230;</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://press.airstreet.com/subscribe?"><span>Subscribe now</span></a></p><h3><strong>AI Revenues Go Vertical</strong></h3><p>Against this backdrop of geopolitical upheaval, the commercial engine accelerated. Anthropic's annualized revenue <a href="https://www.anthropic.com/news">surged</a> from $14B in mid-February to $19B by early March - and has now <a href="https://www.anthropic.com/news/google-broadcom-partnership-compute">surpassed $30B</a>, with over 1,000 enterprise customers each spending $1M+ annually (doubled in under two months). Anthropic simultaneously signed its <a href="https://www.anthropic.com/news/google-broadcom-partnership-compute">most significant compute commitment to date</a>: a deal with Google and Broadcom for multiple gigawatts of next-generation TPU capacity coming online from 2027, part of its $50B pledge to invest in American computing infrastructure. The pace of growth defies any normal SaaS trajectory. <a href="https://ramp.com/">Ramp data</a> showed Anthropic commanding over 50% of enterprise API spend, unseating ChatGPT, which owned that position months earlier. The growth trajectory was amplified by the runaway success of Claude Code and Anthropic&#8217;s capture of knowledge-work verticals with Claude Cowork, which has rapidly become the product that makes the rest of the category feel vestigial. Once you&#8217;ve handed a task to Cowork and watched it actually complete, having ChatGPT explain how you should do it feels like a generational gap akin to MySpace vs. Facebook. I for one am all for OpenAI parking Sora and other bets to refocus on a Cowork-style product.</p><p>There was, however, <a href="https://x.com/">critique</a> of whether this topline revenue figure is net of commissions it pays to hyperscaler hosted Claude revenues. The distinction centers on how each company handles revenue that flows through hyperscaler partnerships. According to a widely circulated analysis by investor Ethan Choi, a partner at Khosla Ventures, OpenAI reports revenue from its Microsoft Azure partnership on a net basis, deducting the roughly 20% revenue share paid to Microsoft before reporting the total. Anthropic, by contrast, reports revenue from its Amazon Web Services and Google Cloud partnerships on a gross basis, including the hyperscaler&#8217;s revenue share in its top-line figure before expenses are recognized.</p><p>OpenAI pursued a different growth strategy by focusing platform consolidation through hyperscaler alliances. On February 27, Amazon CEO Andy Jassy <a href="https://www.aboutamazon.com/news/aws/openai-amazon-partnership-explained">announced a strategic partnership</a> worth up to $50B, of which $15B in the first tranche, the remainder tied to milestones. OpenAI committed to spending $100B on AWS over eight years, expanding a prior $38B agreement. AWS became the exclusive third-party cloud distributor for OpenAI Frontier, the company&#8217;s agent orchestration platform. OpenAI also went big on Amazon&#8217;s custom Trainium chips, which it claimed were 30-40% more price-performant than comparable GPUs. The company&#8217;s own revenue was at a <a href="https://sacra.com/c/openai/">$25B annualized run rate</a> by February, with internal projections forecasting <a href="https://fortune.com/2026/02/20/openai-revenue-forecast-280-billion-2030-capex-sam-altman/">$280B by 2030</a>.</p><p>Alphabet&#8217;s <a href="https://blog.google/company-news/inside-google/message-ceo/alphabet-earnings-q4-2025/">Q4 2025 earnings</a> on February 5 confirmed the infrastructure investment thesis was paying returns. Revenue hit $113.8B, up 18% year-over-year, with Google Cloud growing 48% to $17.7B, led by enterprise AI infrastructure and AI solutions. Importantly, Cloud margins expanded to 30%. Capex guidance for 2026 came in at $175-185B, more than double 2025 spending. The Gemini App crossed 750M monthly active users, processing over 10B tokens per minute via direct API use. Not bad. Databricks, meanwhile, posted a <a href="https://www.databricks.com/">$5.4B run-rate</a> on February 9, representing 65%+ year-over-year growth, with AI products alone at $1.4B (note: it&#8217;s unclear what the company really includes here and what old products have been bundled under this umbrella).</p><h3><strong>The Model Treadmill and the Distillation Wars</strong></h3><p>February and March saw six major model releases in under four weeks. Anthropic shipped <a href="https://www.anthropic.com/">Claude Sonnet 4.6</a> on February 17, scoring 79.6% on SWE-bench Verified and 72.5% on OSWorld, within 1-2 points of the flagship Opus 4.6 at one-fifth the price. Developers chose Sonnet 4.6 over the previous Opus 4.5 59% of the time, citing better instruction following. Google followed two days later with <a href="https://blog.google/innovation-and-ai/models-and-research/gemini-models/gemini-3-1-pro/">Gemini 3.1 Pro</a>, which doubled reasoning performance over Gemini 3 Pro, scored 77.1% on ARC-AGI-2, and ranked first on 12 of 18 tracked benchmarks. OpenAI launched <a href="https://openai.com/index/introducing-gpt-5-4/">GPT-5.4</a> on March 5 in multiple variants (Pro, Thinking, mini, nano) with the headline model scoring 75% on OSWorld (the average human: 72.4%) and achieving native computer-use capabilities with 1M-token context.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!cn2c!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd517471c-4c76-4e0b-ad74-f0454a575b1d_1654x950.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!cn2c!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd517471c-4c76-4e0b-ad74-f0454a575b1d_1654x950.png 424w, https://substackcdn.com/image/fetch/$s_!cn2c!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd517471c-4c76-4e0b-ad74-f0454a575b1d_1654x950.png 848w, https://substackcdn.com/image/fetch/$s_!cn2c!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd517471c-4c76-4e0b-ad74-f0454a575b1d_1654x950.png 1272w, https://substackcdn.com/image/fetch/$s_!cn2c!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd517471c-4c76-4e0b-ad74-f0454a575b1d_1654x950.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!cn2c!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd517471c-4c76-4e0b-ad74-f0454a575b1d_1654x950.png" width="603" height="346.22802197802196" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d517471c-4c76-4e0b-ad74-f0454a575b1d_1654x950.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:836,&quot;width&quot;:1456,&quot;resizeWidth&quot;:603,&quot;bytes&quot;:319841,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://press.airstreet.com/i/193373226?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd517471c-4c76-4e0b-ad74-f0454a575b1d_1654x950.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!cn2c!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd517471c-4c76-4e0b-ad74-f0454a575b1d_1654x950.png 424w, https://substackcdn.com/image/fetch/$s_!cn2c!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd517471c-4c76-4e0b-ad74-f0454a575b1d_1654x950.png 848w, https://substackcdn.com/image/fetch/$s_!cn2c!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd517471c-4c76-4e0b-ad74-f0454a575b1d_1654x950.png 1272w, https://substackcdn.com/image/fetch/$s_!cn2c!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd517471c-4c76-4e0b-ad74-f0454a575b1d_1654x950.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Meanwhile, open source AI is increasingly synonymous with Chinese AI as Chinese labs dropped significant new releases. Zhipu AI's <a href="https://www.scmp.com/tech/article/3343239/chinas-zhipu-ai-launches-new-major-model-glm-5-challenge-its-rivals">GLM-5</a>, launched February 11, is a 745B MoE model trained on Huawei Ascend chips - not NVIDIA - with 28.5T tokens of pre-training data, a 200K-token context window, and pricing roughly six times cheaper than Opus 4.6. Zhipu became the first LLM-native company to go public anywhere globally, with retail demand oversubscribed 1,159 times. Its follow-up, <a href="https://z.ai/blog/glm-5.1">GLM-5.1</a>, shipped weeks later with a coding-focused post-training pass that scored 77.8% on SWE-bench Verified and 45.3 on Claude Code's coding benchmark - 94.6% of Opus 4.6's score at roughly one-fifteenth the price. The weights are being <a href="https://aiproductivity.ai/news/zhipu-ai-glm-5-1-open-source-weights-april/">open-sourced under MIT</a>. Meanwhile, AI2&#8217;s effort to carry the torch for American open source AI released <a href="https://github.com/allenai/molmo2">Molmo2</a> on March 4, an open-source vision-language model achieving state-of-the-art video understanding, pointing, and tracking, demonstrating that the open-source frontier in multimodal AI is alive and well.</p><p>These releases occurred against a backdrop of escalating IP warfare. On February 23, Anthropic <a href="https://www.anthropic.com/news/detecting-and-preventing-distillation-attacks">published evidence</a> that three Chinese AI labs - DeepSeek, Moonshot, and MiniMax - had conducted &#8220;industrial-scale&#8221; distillation campaigns against Claude, extracting model capabilities through 16M exchanges across approximately 24,000 fraudulent accounts. Anthropic framed this not merely as intellectual property theft but as an export-control circumvention mechanism: distillation allowed Chinese labs to acquire advanced AI capabilities far more quickly and cheaply than independent development. OpenAI <a href="https://openai.com/">raised similar concerns</a> about DeepSeek on February 13. The enforcement arm followed: on March 20, Supermicro co-founder Wally Liaw was <a href="https://fortune.com/2026/03/19/supermicro-arrested-founder-smuggling-gpu-china/">arrested</a> for allegedly smuggling $2.5B in NVIDIA GPU servers to China in violation of export controls&#8212;the largest chip-smuggling prosecution to date. You can&#8217;t make this up&#8230;</p><h3><strong>Safety Meets Reality</strong></h3><p>How close are frontier models to catastrophic sabotage risk? Anthropic&#8217;s <a href="https://www-cdn.anthropic.com/08eca2757081e850ed2ad490e5253e940240ca4f.pdf">Sabotage Risk Report</a> for Claude Opus 4.6, published February 11, delivered an assessment that should unsettle anyone paying attention: the risk of catastrophic sabotage from Opus 4.6 is &#8220;very low but not negligible.&#8221; <a href="https://metr.org/blog/2026-03-12-sabotage-risk-report-opus-4-6-review/">METR&#8217;s external review</a> agreed with the overall conclusion but flagged that several subclaims in the report lack sufficient experimental support, and that the margin to the ASL-4 threshold, where substantially stronger safeguards would be required, is unclear. The report noted that Opus 4.6 had, in testing, &#8220;knowingly supported, in small ways, efforts toward chemical weapon development.&#8221; Anthropic does not believe the model meets ASL-4 criteria. The gray zone it occupies is the uncomfortable middle where clean rule-out has become difficult.</p><p>Three weeks later, the alignment team published <a href="https://alignment.anthropic.com/2026/hot-mess-of-ai/">&#8220;The Hot Mess of AI&#8221;</a>, decomposing frontier model errors into bias (systematic) and variance (incoherent) components. They found that as tasks get harder and reasoning chains get longer, failures are increasingly dominated by incoherence, not systematic misalignment. The models are less deceptively scheming and more chaotically unreliable. Whether this is reassuring depends on your threat model.</p><p>The real-world evidence suggested the threat was already here, just not from the models themselves. In late February, <a href="https://www.bloomberg.com/news/articles/2026-02-25/hacker-used-anthropic-s-claude-to-steal-sensitive-mexican-data">Bloomberg reported</a> that a hacker had exploited Claude to steal 150 gigabytes of Mexican government data including 195M taxpayer records by writing Spanish-language prompts instructing the model to find vulnerabilities, write exploitation scripts, and automate data theft across government networks for over a month. Claude initially flagged the activity as malicious but ultimately complied. In March, security startup CodeWall <a href="https://codewall.ai/blog/how-we-hacked-mckinseys-ai-platform">demonstrated</a> that its AI agent could hack McKinsey&#8217;s internal Lilli chatbot in two hours, exploiting unauthenticated API endpoints to access 46.5M chat messages and 728,000 confidential files. The attack vector was a basic SQL injection, a vulnerability class from the early 2000s, now exploitable at machine speed.</p><p>Then Anthropic went on offense. <a href="https://www.anthropic.com/glasswing">Project Glasswing</a>, launched alongside AWS, Apple, Broadcom, Cisco, CrowdStrike, Google, JPMorganChase, Microsoft, NVIDIA, and Palo Alto Networks, marshalled a new model - Claude Mythos Preview - to hunt zero-day vulnerabilities across critical software infrastructure. Mythos Preview scored 83.1% on CyberGym (vs. Opus 4.6's 66.6%) and 77.8% on SWE-bench Pro (vs. 53.4%), and has already flagged thousands of high-severity flaws, including a 27-year-old remote-crash bug in OpenBSD and a 16-year-old FFmpeg vulnerability that automated testing had missed five million times. Anthropic committed $100M in model usage credits and priced authorized access at $25/$125 per million input/output tokens. The model remains unreleased to the general public pending safeguards. It's a neat inversion: the same capabilities that make frontier models dangerous for offense become genuinely useful for defense, if you can control who gets access.</p><p>Finally, regulatory responses began crystallizing. New York&#8217;s <a href="https://www.nysenate.gov/newsroom/press-releases/2026/kristen-gonzalez/ai-chatbot-ban-minors-passes-internet-technology">Senate Bill 7263</a> advanced out of committee on a 6-0 vote, targeting 14 licensed professions and creating private liability for chatbot operators whose AI gives &#8220;substantive&#8221; legal, medical, or engineering advice. One of the first laws to treat AI output as a professional practice issue rather than a platform moderation problem.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://press.airstreet.com/subscribe?"><span>Subscribe now</span></a></p><h3><strong>The Physical Layer Gets Contested</strong></h3><p>Can China build frontier AI models without NVIDIA chips? Well, for starters, NVIDIA&#8217;s AI chip sales to China have stalled amid tightening export controls. By March 5, NVIDIA <a href="https://www.ft.com/content/47f1cf56/">stopped production entirely</a> on chips designed to comply with China export limits, opting to exit the market segment rather than continue designing compliant variants. The Supermicro indictment, $2.5B in NVIDIA servers allegedly diverted to China through shell companies, underscored the scale of the circumvention problem. Meanwhile, China&#8217;s domestic AI economy adapted: AI tokens had become the country&#8217;s hottest traded commodity, with speculative demand outpacing industrial use. Zhipu AI&#8217;s training of GLM-5 on Huawei Ascend chips proved that the Chinese stack can produce frontier models without NVIDIA, even if the cost and efficiency penalties remain substantial.</p><p>On the US side, the buildout continues, but is increasingly contested. <a href="https://investors.micron.com/news-releases/news-release-details/micron-announces-groundbreaking-historic-new-york-megafab">Micron broke ground</a> on a $100 billion megafab in Clay, New York, the largest semiconductor fabrication investment in US history, backed by $6.4B in CHIPS Act funding and $5.5B in New York state incentives, targeting 50,000 jobs over two decades. Meta <a href="https://www.cnbc.com/2026/03/16/meta-nebius-ai-infrastructure.html">signed</a> a $27B AI infrastructure deal with Nebius, $12B in dedicated capacity on NVIDIA&#8217;s next-generation Vera Rubin platform plus $15B in additional compute, as part of an AI capex plan that Meta said would hit $115-135B in 2026 alone. And private equity entered the classified infrastructure market: <a href="https://www.ft.com/content/332c1134/">Carlyle and KKR were separately awarded</a> $2B contracts to build hyperscale data centers for the US Army. But the political wind is shifting: <a href="https://www.axios.com/2026/04/05/data-centers-midterms-state-bans-bills-ai">at least 11 states have introduced bills</a> to restrict or ban data center construction, with Maine on track to be the first to pause development outright, while Sanders and Ocasio-Cortez introduced a federal moratorium bill that would halt all new builds until Congress passes AI worker and environmental protections. We <a href="https://www.stateof.ai/">predicted</a> in the State of AI Report 2025 that data centre NIMBYism would hit US elections&#8230;it&#8217;s arriving faster than expected.</p><p>The most unexpected story from this period may also prove the most lasting. An Australian tech entrepreneur with no biology degree <a href="https://fortune.com/2026/03/15/australian-tech-entrepreneur-ai-cancer-vaccine-dog-rosie-unsw-mrna/">used ChatGPT and AlphaFold</a> to design a personalised mRNA cancer vaccine for his rescue dog. Most tumours shrank. It is the first bespoke cancer vaccine ever designed for a dog.</p><div><hr></div><h3><strong>Research</strong></h3><p>Here are the most consequential AI research papers from February and March 2026:</p><p><strong><a href="https://arxiv.org/abs/2603.11214">Measuring AI Agents&#8217; Progress on Multi-Step Cyber Attack Scenarios</a></strong> (UK AI Safety Institute)</p><p>AISI evaluated seven frontier models on two purpose-built cyber ranges, a 32-step corporate network attack and a 7-step industrial control system attack, and compared models released over an eighteen-month window from August 2024 to February 2026. They found that the average number of steps completed at 10M tokens rose from 1.7 (GPT-4o, August 2024) to 9.8 (Claude Opus 4.6, February 2026), with performance scaling log-linearly with inference compute. Importantly, they found no plateau in sight. The best agent completed 22 of 32 attack steps autonomously, including lateral movement and privilege escalation. The NCSC estimated the marginal cost of an AI-assisted network penetration at &#163;65, which I&#8217;d argue is one of the most policy-consequential AI safety findings this quarter&#8230;</p><p><strong><a href="https://openreview.net/pdf/6593f484501e295cdbe7efcbc46d7f20fc7e741f.pdf">TurboQuant: Redefining AI efficiency with extreme compression</a></strong> (Google Research, DeepMind, NYU)</p><p>The continuous push for larger context windows has been bottlenecked by the immense memory required to store the Key and Value (KV) cache during inference, leading to high cost and slow processing for long inputs. In an effort to address this bottleneck, this paper introduces TurboQuant, an<strong> </strong>architectural improvement that bypasses these computational and memory constraints. Published at ICLR 2026, TurboQuant achieves zero-accuracy-loss 3-bit KV cache compression, delivering 6x lower memory use and up to 8x faster attention on H100 GPUs without requiring training or fine-tuning. The &#8220;zero-accuracy-loss&#8221; component is important: it avoids the performance penalties typically associated with aggressive quantization. The method achieves this extreme efficiency by combining Quantized Johnson-Lindenstrauss projections, which compresses high-dimensional vectors into a much lower-dimensional space, with PolarQuant polar coordinate transformation to eliminate memory overhead. These efficiency gains are substantial enough to shift the inference cost curve for long-context applications, making million-token windows economically viable at scale.</p><p><strong><a href="https://arxiv.org/abs/2602.07488">Deriving Neural Scaling Laws from the statistics of natural language</a></strong> (EPFL, Stanford, Johns Hopkins)</p><p>This paper introduces the first theory to quantitatively predict neural scaling law exponents from first principles, with no free parameters and no synthetic data. The authors isolate two measurable properties of natural language: the decay of pairwise token correlations with time separation (exponent &#946;) and the decay of conditional entropy with context length (exponent &#947;), and derive that the data-limited scaling exponent &#945;_D = &#947;/(2&#946;). Validated on GPT-2 and LLaMA architectures trained from scratch on TinyStories and WikiText, the predicted exponents matched experimental measurements. Scaling laws have guided billions in capital allocation and model design decisions since Kaplan et al. (2020), yet until now the exponents were purely empirical. This paper closes that gap at academic scale. But, I&#8217;d be curious whether  the horizon-limited abstraction holds at trillion-token industrial scales where effective context reaches tens of thousands of tokens&#8230;</p><p><strong><a href="https://arxiv.org/abs/2603.15031">Attention Residuals</a></strong> (Kimi Team / Moonshot AI)</p><p>In this paper, the authors address the gradient dilution problem in deep Transformers, where fixed residual connections cause hidden-state magnitudes to grow and layer contributions to fade. They introduce Attention Residuals (AttnRes), which replaces this fixed accumulation with a learned, depth-wise softmax attention. Each layer uses a &#8220;pseudo-query&#8221; to selectively aggregate outputs from all preceding layers, creating a dynamic, context-aware blend.</p><p>The practical implementation, Block AttnRes, was tested on a 48B model and yielded concrete performance improvements: GPQA-Diamond increased by 7.5 points, and HumanEval by 3.1 points. This architectural approach also matched baseline performance trained with 1.25x the compute, demonstrating a 25% effective efficiency gain.</p><p>This work is interesting because it stabilizes training and improves scaling laws by fundamentally fixing a core architectural limitation, establishing a robust, dynamic alternative to identity mappings that is practical at scale with negligible parameter overhead.</p><p><strong><a href="https://arxiv.org/abs/2601.16175">Learning to Discover at Test Time (TTT-Discover)</a></strong> (Stanford, NVIDIA, Together AI)</p><p>In this paper, the authors introduce Learning to Discover at Test Time (TTT-Discover), a method that applies RL during inference to train an LLM on a single test problem, bypassing the limitations of a frozen model. The paper seeks to achieve autonomous scientific discovery by allowing the LLM to improve its internal policy through experience specific to the current task.</p><p>Experiments were conducted across diverse domains, including mathematics, GPU kernel engineering, competitive programming, and biology. TTT-Discover achieved a new state of the art on Erd&#337;s&#8217; minimum overlap problem, improving 16x more than the AlphaEvolve baseline. It also produced a GPU kernel that was 51% faster than the best human entry on an A100 in the GPUMode competition. A key caveat is that the method critically requires continuous reward signals and cannot yet handle sparse or binary feedback.</p><p>Taken together, this paper establishes a path for LLMs to generate new-to-the-world solutions. It demonstrates that scaling compute via test-time training can push beyond existing human knowledge using open-source models.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://press.airstreet.com/subscribe?"><span>Subscribe now</span></a></p><p><strong><a href="https://arxiv.org/abs/2603.28052">Meta-Harness: End-to-End Optimization of Model Harnesses</a></strong> (Stanford, KRAFTON, MIT)</p><p>This paper shows that changing a model&#8217;s harness - the code wrapping a model that determines what information it sees, stores, and retrieves at each step - around a fixed LLM can produce a 6x performance gap on the same benchmark. Meta-Harness automates harness engineering by giving an agentic proposer full access to raw execution traces (up to 10M tokens of diagnostic information) rather than compressed summaries. The authors show this approach results in +7.7 points on text classification with 4x fewer tokens, #1 among all Haiku 4.5 agents on TerminalBench-2 (37.6%), and #2 among all Opus 4.6 agents (76.4%). A single discovered harness improved accuracy on 200 IMO-level math problems by 4.7 points on average across five held-out models. The killer ablation: summaries actually made things slightly worse than scores alone (34.9% vs 34.6% median), while raw traces gave +15 points at median (50.0%). Taken together, one could conclude the model wrapper matters as much as the weights, and AI can now write better wrappers than humans.</p><p><strong><a href="https://www.pi.website/research/memory">MEM: Multi-Scale Embodied Memory for Vision Language Action Models</a></strong> (Physical Intelligence, Stanford, UC Berkeley, MIT)</p><p>Physical Intelligence introduces a multi-scale memory system that gives robots 15-minute context windows, long enough to clean an entire kitchen or cook from scratch. MEM combines an efficient video encoder for short-horizon frame-based history with a language-based memory mechanism for long-horizon context. After training on diverse robot and non-robot data, MEM VLAs showed +62% success rate on refrigerator tasks and +11% on chopstick manipulation versus memoryless baselines. The system was integrated into Physical Intelligence&#8217;s &#960;0.6 VLA to address a fundamental limitation of current robot control: the inability to maintain coherent plans across multi-step tasks that require remembering what happened minutes ago.</p><p><strong><a href="https://www.anthropic.com/research/labor-market-impacts">Labor market impacts of AI: A new measure and early evidence</a></strong> (Anthropic)</p><p>This paper introduces the concept of &#8220;observed exposure&#8221; - a measure that quantifies not just which tasks LLMs could theoretically automate, but which are already being automated in practice, based on real usage data from Claude. Unsurprisingly, it is computer programmers, customer service representatives, and financial analysts who show the highest observed exposure. Despite high theoretical coverage (94.3% for computer/math occupations), there is no impact on unemployment rates for exposed workers yet, though there is suggestive evidence that hiring into these professions has slowed for workers aged 22&#8211;25. For every 10 percentage-point increase in AI exposure, BLS-projected job growth drops by 0.6 percentage points. The gap between theoretical and observed exposure suggests the labour market is absorbing AI gradually through task-level substitution rather than wholesale job elimination.</p><p><strong><a href="https://dreamzero0.github.io/">World Action Models are Zero-shot Policies (DreamZero)</a></strong> (NVIDIA)</p><p>DreamZero argues for a paradigm shift from Vision-Language-Action models to World Action Models, which jointly predict future video frames and motor actions rather than mapping observations directly to controls. Built on a 14B-parameter video diffusion backbone (Wan2.1), DreamZero achieved 62.2% average task progress on unseen real robot tasks - over 2x the best pretrained VLA baseline (GR00T N1.6 at 31%, &#960;0.5 at 33%). The more consequential result is cross-embodiment transfer: 12 minutes of human egocentric video or 20 minutes of video from a different robot improved unseen-task performance by over 42%, and the model adapted to an entirely new manipulator with just 30 minutes of play data while retaining zero-shot generalisation. Through system-level optimisations including CFG parallelism, DiT caching, and a novel single-step inference mode (DreamZero-Flash), the team achieved a 38x speedup to enable real-time closed-loop control at 7Hz on GB200 hardware. The companion paper, DreamDojo, provides the 44,000-hour human video dataset that enables pretraining.</p><p><strong><a href="https://www.nature.com/articles/s41591-025-04190-9">A large language model for complex cardiology care</a></strong> (Google Health, DeepMind)</p><p>Google Health and DeepMind tested Articulate Medical Intelligence Explorer (AMIE), an LLM built on Gemini, in the first randomised controlled trial of AI-assisted cardiology versus cardiologists working alone on complex cases involving suspected genetic cardiomyopathy. It was found that subspecialists preferred AMIE-assisted assessments 46.7% of the time versus 32.7% for cardiologists alone. In a win for AI, cardiologists working without AI had significantly more clinically significant errors (24.3% vs 13.1%) and more missing content (37.4% vs 17.8%). The result demonstrates frontier LLMs can augment specialist clinical reasoning in ways that reduce diagnostic error, not merely in triage or patient education but in complex subspecialty decision-making.</p><div><hr></div><h3><strong>Investments</strong></h3><p><em>The quarter's headline raise was OpenAI's $110B round at an $840B valuation - the largest private financing in history - led by Amazon ($50B), NVIDIA ($30B), and SoftBank ($30B). Total disclosed venture funding in AI exceeded $50B. Other notable rounds included Wayve ($1.2B), Apptronik ($935M), Earendil Labs ($787M), and Neysa ($600M).</em></p><p>OpenAI, which develops frontier large language models and the ChatGPT consumer AI product, <a href="https://openai.com/index/scaling-ai-for-everyone/">raised</a> $110B at an $840B valuation led by Amazon ($50B), NVIDIA ($30B), and SoftBank ($30B)&#8212;the largest private financing in history.</p><p>Wayve, which develops embodied AI software for autonomous driving,<a href="https://wayve.ai/press/series-d/"> raised</a> $1.2B in a Series D at an $8.6B valuation led by Eclipse, Balderton, and SoftBank Vision Fund 2, with milestone-based capital from Uber bringing the total to $1.5B.</p><p>Apptronik, which builds the Apollo humanoid robot for manufacturing and logistics,<a href="https://apptronik.com/news-collection/apptronik-closes-over-935-million-series-a"> raised</a> $935M in a Series A at a $5.3B valuation co-led by B Capital and Google.</p><p>Earendil Labs, which develops AI-driven biologics for autoimmune diseases and cancer,<a href="https://www.prnewswire.com/news-releases/earendil-labs-announces-787-million-in-financing-to-scale-ai-driven-biologics-discovery-and-development-302719748.html"> raised</a> $787M backed by Dimension Capital, DST Global, Sanofi, and Pfizer&#8217;s Biotech Development Fund.</p><p>Neysa, which provides AI cloud infrastructure in India,<a href="https://www.blackstone.com/news/press/blackstone-leads-funding-of-over-1-billion-to-neysa-to-work-towards-building-indias-leading-ai-infrastructure-platform/"> raised</a> $600M in primary equity at a $1.4B valuation led by Blackstone.</p><p>Legora, which builds AI-powered legal research and workflow tools for 800+ law firms,<a href="https://techcrunch.com/2026/03/10/legora-reaches-5-55-billion-valuation-as-ai-legaltech-boom-endures/"> raised</a> $550M in a Series D at a $5.55B valuation led by Accel.</p><p>ElevenLabs, which develops voice AI and conversational agent technology,<a href="https://elevenlabs.io/blog/series-d"> raised</a> $500M in a Series D at an $11B valuation led by Sequoia Capital.</p><p>MatX, which designs custom AI training chips purpose-built for LLM workloads,<a href="https://techcrunch.com/2026/02/24/nvidia-challenger-ai-chip-startup-matx-raised-500m/"> raised</a> $500M led by Jane Street and Situational Awareness.</p><p>Mind Robotics, which develops humanoid robots for industrial applications backed by Rivian,<a href="https://techcrunch.com/2026/03/11/rivian-mind-robotics/"> raised</a> $500M.</p><p>Runway, which builds AI video generation and world models for creative and scientific applications,<a href="https://techcrunch.com/2026/02/10/ai-video-startup-runway-raises-315m-at-5-3b-valuation-eyes-more-capable-world-models/"> raised</a> $315M in a Series E at a $5.3B valuation led by General Atlantic.</p><p>Bedrock Robotics, which builds autonomous excavators and construction equipment using technology from former Waymo engineers,<a href="https://www.prnewswire.com/news-releases/bedrock-robotics-raises-270-million-in-series-b-funding-to-accelerate-the-future-of-autonomous-construction-302679014.html"> raised</a> $270M in a Series B at a $1.75B valuation co-led by CapitalG and Valor Atreides.</p><p>Fundamental, which builds Nexus, a Large Tabular Model for enterprise structured-data analysis,<a href="https://techcrunch.com/2026/02/05/fundamental-raises-255-million-series-a-with-a-new-take-on-big-data-analysis/"> raised</a> $255M in a Series A at a $1.4B valuation led by Oak HC/FT.</p><p>Intercom, which provides an AI-first customer service platform powered by its Fin AI agent,<a href="https://www.irishtimes.com/business/2026/03/10/intercom-raises-250m-in-debt-financing-to-fund-ai-agents/"> raised</a> $250M in debt financing from Hercules Capital.</p><p>Positron, which designs energy-efficient AI inference chips to compete with Nvidia,<a href="https://techcrunch.com/2026/02/04/exclusive-positron-raises-230m-series-b-to-take-on-nvidias-ai-chips/"> raised</a> $230M in a Series B at a $1B valuation co-led by Arena Private Wealth, Jump Trading, and Unless.</p><p>Harvey, which develops AI-powered legal reasoning used by most of the top 100 US law firms,<a href="https://www.harvey.ai/blog/harvey-raises-at-dollar11-billion-valuation-to-scale-agents-across-law-firms-and-enterprises"> raised</a> $200M at an $11B valuation co-led by GIC and Sequoia.</p><p>Oxide, which designs and manufactures rack-scale on-premises cloud computers,<a href="https://oxide.computer/blog/our-200m-series-c"> raised</a> $200M in a Series C led by US Innovative Technology Fund.</p><p>Goodfire, which uses mechanistic interpretability to understand and design AI models,<a href="https://www.goodfire.ai/blog/our-series-b"> raised</a> $150M in a Series B at a $1.25B valuation led by B Capital.</p><p>Wonderful, which deploys AI customer support agents for telecom, financial services, and healthcare enterprises, <a href="https://techcrunch.com/2026/03/12/wonderful-raises-150m-series-b-at-2b-valuation/">raised</a> $150M in a Series B at a $2B valuation led by Insight Partners.</p><p>Revel, which builds a unified software platform for hardware test and control used in aerospace and defence, <a href="https://www.indexventures.com/perspectives/great-hardware-deserves-great-software-investing-in-revel/">raised</a> $150M in a Series B led by Index Ventures.</p><p>Vega, which builds an AI-native security operations platform with federated threat detection, <a href="https://techcrunch.com/2026/02/10/vega-raises-120m-series-b-to-rethink-how-enterprises-detect-cyber-threats/">raised</a> $120M in a Series B at a $700M valuation led by Accel.</p><p>Basis, which builds AI agents that autonomously complete accounting, tax, and audit workflows, <a href="https://www.businesswire.com/news/home/20260224020999/en/Basis-Raises-$100M-at-a-$1.15B-Valuation-as-Accounting-Firms-Adopt-End-to-End-Agents-Across-Accounting-Tax-and-Audit">raised</a> $100M in a Series B at a $1.15B valuation led by Accel and GV.</p><p>Simile, which uses generative AI agents to simulate and predict human behaviour for enterprise decision-making, <a href="https://www.indexventures.com/perspectives/life-the-universe-and-simile-leading-similes-series-a/">raised</a> $100M in a Series A led by Index Ventures.</p><p>Render, which operates a cloud platform for deploying AI-native applications and agents, <a href="https://render.com/blog/series-c-extension">raised</a> $100M in a Series C extension at a $1.5B valuation led by Georgian.</p><p>Nominal, which provides a connected testing and operations platform for hardware engineering teams in aerospace, defence, and energy, <a href="https://www.globenewswire.com/news-release/2026/03/05/3250350/0/en/Nominal-Valued-at-1B-as-Founders-Fund-Leads-80M-Acceleration-Round.html">raised</a> $80M at a $1B valuation led by Founders Fund.</p><p>Braintrust, which builds AI observability and evaluation tools used by Notion, Replit, and Cloudflare, <a href="https://www.braintrust.dev/blog/announcing-series-b">raised</a> $80M in a Series B at an $800M valuation led by Iconiq.</p><p>Entire, which builds a developer platform for human-AI agent collaboration on codebases, <a href="https://startupnews.fyi/2026/02/11/former-github-ceo-60m-seed-devtools/">raised</a> $60M in a seed round at a $300M valuation led by Felicis Ventures.</p><p>Isembard, which builds industrial AI infrastructure in the UK, <a href="https://x.com/afitzgerald1992/status/2030919233429758217">raised</a> $50M in a Series A.</p><p>SolveAI, which lets non-technical employees build production-ready enterprise software through AI-powered conversations, <a href="https://fortune.com/2026/02/25/exclusive-solveai-eight-months-raises-50-million/">raised</a> $50M in a Series A led by Google Ventures.</p><p>RunSybil, which provides AI-powered cybersecurity red-teaming and penetration testing, <a href="https://fortune.com/2026/03/18/exclusive-ai-cybersecurity-startup-runsybil/">raised</a> $40M.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://press.airstreet.com/subscribe?"><span>Subscribe now</span></a></p><h3><strong>Exits</strong></h3><p><em>The quarter's defining exit was xAI's merger into SpaceX, valuing the combined entity at $1.25T ahead of a planned IPO. Anthropic acquired Vercept (computer-use agents), Amazon acquired Fauna Robotics (soft-bodied humanoids), and Anduril acquired ExoAnalytic Solutions (orbital tracking).</em></p><p>xAI, which develops frontier large language models and the Grok consumer AI product,<a href="https://www.spacex.com/updates#xai-joins-spacex"> was merged into</a> SpaceX in a deal valuing the combined entity at $1.25 trillion ahead of a planned SpaceX IPO.</p><p>WorkFusion, which provides AI agents for anti-money-laundering and KYC compliance in financial services, <a href="https://ir.uipath.com/news/detail/425/">was acquired by</a> UiPath for an undisclosed amount.</p><p>Intrinsic, which builds AI-powered software to make industrial robots more accessible, <a href="https://www.intrinsic.ai/blog/posts/intrinsic-joins-google-to-accelerate-physical-ai">was absorbed into</a> Google to accelerate physical AI using Gemini models and Google Cloud.</p><p>Vercept, which developed computer-use AI agents capable of operating remote desktops, <a href="https://www.anthropic.com/news/acquires-vercept">was acquired by</a> Anthropic for an undisclosed amount.</p><p>Fauna Robotics, which builds the Sprout soft-bodied humanoid robot for homes and schools, <a href="https://www.humanoidsdaily.com/news/amazon-acquires-soft-bodied-humanoid-maker-fauna-robotics">was acquired by</a> Amazon for an undisclosed amount.</p><p>Koyeb, which operates a serverless cloud platform for deploying AI inference workloads, <a href="https://techcrunch.com/2026/02/17/mistral-ai-buys-koyeb/">was acquired by</a> Mistral AI for an undisclosed amount.</p><p>Tavily, which provides an AI-optimised search API for retrieval-augmented generation, <a href="https://nebius.com/newsroom/nebius-announces-agreement-to-acquire-tavily-to-add-agentic-search-to-its-ai-cloud-platform">was acquired by</a> Nebius for an undisclosed amount.</p><p>DOK-ING, which manufactures unmanned ground vehicles for mine clearance and explosive ordnance disposal, <a href="https://www.ft.com/content/d605ebac/">was acquired by</a> Rheinmetall for an undisclosed amount.</p><p>ExoAnalytic Solutions, which tracks objects in orbit using a global network of optical sensors, <a href="https://techcrunch.com/2026/03/11/anduril-snaps-up-space-surveillance/">was acquired by</a> Anduril for an undisclosed amount.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:&quot;button-wrapper&quot;}" data-component-name="ButtonCreateButton"><a class="button primary button-wrapper" href="https://press.airstreet.com/subscribe?"><span>Subscribe now</span></a></p><p><strong>This issue at a glance:</strong> The Trump administration blacklisted Anthropic over Pentagon usage restrictions, designating it a "supply chain risk" and triggering a federal lawsuit. Iran conducted the first military strikes on commercial cloud infrastructure, hitting AWS data centres in the UAE and Bahrain. Anthropic's annualized revenue surged from $14B to $19B in weeks. Six frontier models launched in four weeks. Anthropic published evidence that DeepSeek, Moonshot, and MiniMax ran industrial-scale distillation campaigns through 16 million exchanges. NVIDIA exited the China-compliant chip market entirely. OpenAI raised $110B at an $840B valuation - the largest private financing in history. And an Australian used ChatGPT and AlphaFold to design the first personalised mRNA cancer vaccine for a dog.</p><p><strong>Q1 2026 by the numbers:</strong> Anthropic revenue $14B&#8594;$19B in weeks &#183; OpenAI raised $110B at $840B valuation &#183; OpenAI-Amazon partnership worth up to $50B &#183; Alphabet capex guidance $175-185B &#183; 6 frontier model releases in 4 weeks &#183; 16M distillation exchanges across 24K fraudulent accounts &#183; Opus 4.6 sabotage risk: "very low but not negligible" &#183; 150GB of Mexican government data stolen via Claude &#183; 11 US states introduced data centre restriction bills &#183; $2.5B GPU smuggling prosecution &#183; AI-assisted network penetration cost: &#163;65 &#183; Total disclosed AI venture funding: $50B+</p><p><strong>What to watch in Q2:</strong> Whether the Anthropic-Trump lawsuit reshapes how governments procure AI. Whether the data center moratorium movement gains traction ahead of midterms. Whether distillation enforcement triggers formal trade retaliation. Whether anyone can sustain revenue growth at the pace Anthropic set in February. And whether OpenAI launches a legitimate competitor to Claude Cowork. </p>]]></content:encoded></item><item><title><![CDATA[State of AI: February 2026 newsletter]]></title><description><![CDATA[Software stocks crater as agentic AI rewrites the playbook. Plus: Moltbook's AI theatre, OpenClaw's 157K-star security mess, and HBM runs out.]]></description><link>https://press.airstreet.com/p/state-of-ai-february-2026-newsletter</link><guid isPermaLink="false">https://press.airstreet.com/p/state-of-ai-february-2026-newsletter</guid><dc:creator><![CDATA[Air Street Press]]></dc:creator><pubDate>Mon, 09 Feb 2026 18:55:00 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/153923e9-b125-4ec2-8285-7f4768f95152_1810x1010.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Dear readers, </p><p>Welcome to the latest issue of the <strong>State of AI</strong>, an editorialized newsletter that covers the key developments in AI policy, research, industry, and start-ups over the last month. First up, a few reminders:</p><ul><li><p><strong>AI meetups: </strong>Join our <a href="https://press.airstreet.com/p/air-street-ai-meetup-europe-tour">upcoming AI meetups</a> in Munich (17 Feb &#8216;26) for the Munich Security Conference and Zurich (19 Feb &#8216;26), as well as in <a href="https://luma.com/parisai">Paris (11 Mar &#8216;26) </a>and <a href="https://luma.com/sfoai">SF (28 Apr &#8216;16)</a>.</p></li><li><p><strong>RAAIS 2026:</strong> Join our 11th <a href="http://raais.co">Research and Applied AI Summit</a> in London on 12 June 2026, the premier global meeting for learning AI best practices and what&#8217;s coming next. </p></li><li><p><strong>Air Street Press</strong> featured the <a href="https://press.airstreet.com/p/2025-review">Air Street Capital Year in Review 2025</a>, how <a href="https://press.airstreet.com/p/embodied-ai-breakthroughs-2025">embodied AI is hitting its stride</a>, whether<a href="https://press.airstreet.com/p/ai-for-science-new-knowledge"> AI can discover new science</a>, <a href="https://press.airstreet.com/p/ai-progress-after-2025">AI progress into 2026</a>, what <a href="https://press.airstreet.com/p/european-defense-entering-2026">European defense must do in 2026</a>, and mega rounds at portfolio companies <a href="https://press.airstreet.com/p/black-forest-labs-300-million">Black Forest Labs</a> and Synthesia. </p></li><li><p><strong>Take the <a href="https://www.stateof.ai/survey-2025">State of AI usage survey</a>: </strong>You can submit your usage patterns to the largest ongoing open access survey, which now has over 1,400 respondents :-)</p></li><li><p><strong>Looking for a new challenge? </strong>Lots of our companies are hiring, just drop me a line. </p></li></ul><p>I love hearing what you&#8217;re up to, so just hit reply or forward to your friends :-)</p><div><hr></div><h3><strong>The $300B dislocation</strong></h3><p>The gap between what AI systems can now do and what the market thinks that means has never been wider. Nearly $285B in market capitalisation has been <a href="https://www.bloomberg.com/news/articles/2026-02-04/what-s-behind-the-saaspocalypse-plunge-in-software-stocks">wiped</a> from software stocks in the space of two weeks. The S&amp;P 500 software and services index is down 26% from its October peak. The Goldman Sachs software index suffered its worst single-day drop since the last round of forced selling during trade tensions. Hedge funds have piled in, shorting $24B in software names this year alone. Meanwhile, frontier model releases are arriving at a cadence that feels less like a product cycle and more like an arms race.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Y0gi!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faad929fc-c1a8-4db7-b770-bf648bfbf54c_2779x1294.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Y0gi!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faad929fc-c1a8-4db7-b770-bf648bfbf54c_2779x1294.png 424w, https://substackcdn.com/image/fetch/$s_!Y0gi!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faad929fc-c1a8-4db7-b770-bf648bfbf54c_2779x1294.png 848w, https://substackcdn.com/image/fetch/$s_!Y0gi!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faad929fc-c1a8-4db7-b770-bf648bfbf54c_2779x1294.png 1272w, https://substackcdn.com/image/fetch/$s_!Y0gi!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faad929fc-c1a8-4db7-b770-bf648bfbf54c_2779x1294.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Y0gi!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faad929fc-c1a8-4db7-b770-bf648bfbf54c_2779x1294.png" width="1456" height="678" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/aad929fc-c1a8-4db7-b770-bf648bfbf54c_2779x1294.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:678,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:162612,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://press.airstreet.com/i/187423711?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faad929fc-c1a8-4db7-b770-bf648bfbf54c_2779x1294.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Y0gi!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faad929fc-c1a8-4db7-b770-bf648bfbf54c_2779x1294.png 424w, https://substackcdn.com/image/fetch/$s_!Y0gi!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faad929fc-c1a8-4db7-b770-bf648bfbf54c_2779x1294.png 848w, https://substackcdn.com/image/fetch/$s_!Y0gi!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faad929fc-c1a8-4db7-b770-bf648bfbf54c_2779x1294.png 1272w, https://substackcdn.com/image/fetch/$s_!Y0gi!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faad929fc-c1a8-4db7-b770-bf648bfbf54c_2779x1294.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The trigger was ostensibly Anthropic&#8217;s launch of Claude Cowork in January, a system-level agent that navigates computer interfaces, manipulates local files, and executes multi-step business workflows autonomously. When Anthropic followed up with specialised plugins for marketing, law, and finance, the narrative flipped overnight from &#8220;AI as productivity booster&#8221; to &#8220;AI will replace your SaaS.&#8221;</p><p>Then came the duelling model launches. Anthropic released Claude Opus 4.6 with a 1M-token context window, state-of-the-art scores on Terminal-Bench 2.0 and Humanity&#8217;s Last Exam, and the ability to spin up and coordinate parallel agent teams. Minutes later, OpenAI dropped GPT-5.3-Codex, the first model that was instrumental in building itself and which OpenAI treats as its first High-capability release in the cybersecurity domain. Both companies originally scheduled their reveals for 10:00 a.m. PST. Anthropic moved 15 minutes early. OpenAI matched instantly.</p><p>The selloff signals the existential question: how can investors underwrite the next ten years of technology companies? SaaS companies have traded at premium multiples because their recurring revenue was predictable: high retention, low churn, multi-year contracts. Agents that can command tools and interfaces to get real work done breaks that assumption. If core workflows in legal, finance, and marketing can be rebuilt AI-first at a fraction of the cost &#8212; the thesis I&#8217;ve been investing with Air Street for quite some time now &#8212; the long-duration revenue streams that justified those valuations are not safe. Software stocks are trading at P/E ratios at ten-year lows while their current fundamentals remain strong. That is precisely the signature of a market repricing terminal value, not current earnings. Whether it is overdone depends on whether the next wave of earnings calls shows actual churn or accelerating growth despite the fear.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://press.airstreet.com/subscribe?"><span>Subscribe now</span></a></p><h3><strong>The sovereignty fracture</strong></h3><p>The geopolitical consensus around &#8220;build AI at all costs&#8221; is coming apart, from the top down and the bottom up. At the federal level, the White House and Anthropic are in <a href="https://www.wsj.com/tech/ai/anthropic-ai-defense-department-contract-947d5f33">open conflict over the terms of military AI use</a>. Defence Secretary Pete Hegseth criticised models that &#8220;won&#8217;t allow you to fight wars.&#8221; Anthropic,  which won a $200M DOD contract awarded last year along with all other frontier AI labs, bars autonomous weapons and domestic surveillance from its systems, while the Pentagon&#8217;s January memo asserts that military necessity overrides vendor usage policies. Note that this is quite a vibe shift from even a year ago - the labs that once avoided military associations entirely are now the ones being publicly pressured by the government to drop their remaining restrictions. Anyone involved in AI safety knows a stated policy doesn&#8217;t mean much when a Defence Secretary is calling you out on television.</p><p>US states, meanwhile, are pushing back against the infrastructure buildout itself. New York <a href="https://www.wired.com/story/new-york-is-the-latest-state-to-consider-a-data-center-pause/">introduced a three-year moratorium on data center permits</a>, citing tripled electricity demand in a single year from AI workloads. Georgia, Vermont, Virginia, Maryland, and Oklahoma have introduced similar bipartisan legislation. Energy, water, and grid strain are now political issues, and the resulting friction will shape where the next generation of training clusters can physically be built. Indeed, these add weight to our State of AI Report 2026 prediction that &#8220;Datacenter NIMBYism takes the US by storm and sways certain midterm/gubernatorial elections in 2026.&#8221;</p><p>On the chip trade: in a surprising tactical shift, the Trump administration <a href="https://www.reuters.com/world/china/china-gives-green-light-importing-first-batch-nvidias-h200-ai-chips-sources-say-2026-01-28/">cleared Nvidia H200 exports to China</a> under strict conditions, China-bound sales capped at 50% of US volumes, buyers must certify non-military use, and the government takes a 25% revenue cut. Chinese customs reportedly blocked the first shipments within a day. You cannot make this stuff up. Meanwhile, the Bureau of Industry and Security is <a href="https://x.com/pstasiatech/status/2010686314258088331">moving to tighten controls</a> across the AI supply chain.</p><p>China is not sitting still. The <a href="https://www.cac.gov.cn/2025-12/27/c_1768571207311996.htm">Cyberspace Administration of China</a> (CAC) issued new draft rules governing AI systems that simulate human personality and emotional engagement, a scope of regulation the West hasn&#8217;t seriously attempted. Beijing is simultaneously closing the talent gap through its <a href="https://t.co/MUZTIDaw6g">&#8220;genius class&#8221; programme</a>, which funnels 100,000 gifted teenagers annually into accelerated STEM tracks, bypassing the national college exam entirely. As we noted in the <a href="https://www.stateof.ai/">State of AI Report 2025</a>, if the US is grappling with how to regulate foundation models, China is already piloting enforcement and building the human pipeline to compete.</p><p>Meanwhile, China&#8217;s pure-play AI model companies have beaten their American peers to public markets, and Hong Kong is rewarding them for it. <a href="https://www.cnbc.com/2026/01/08/china-ai-tiger-goes-ipo-zhipu-hong-kong-debut-openai-knowledge-atlas-hsi-hang-seng-listing.html">Zhipu AI</a> became the first LLM-native company to list anywhere in the world, with retail demand oversubscribed 1,159 times. <a href="https://www.cnbc.com/2026/01/09/minimax-hong-kong-ipo-ai-tigers-zhipu.html">MiniMax</a> doubled on its first day and is up 259% since listing. AI chip designer <a href="https://finance.yahoo.com/news/china-ai-chipmaker-biren-surges-014634385.html">Biren Technology</a> posted the best Hong Kong debut since 2021 for a raise above $700M, with retail oversubscribed 2,348 times. None of these companies are profitable - Zhipu and MiniMax posted combined losses of over $840M in their most recent filings - but the market is pricing them as strategic infrastructure. OpenAI and Anthropic, for all their capability leads, remain private.</p><p>And then there is DeepSeek. V4 is expected to drop mid-February: a 1T-parameter coding model with Engram memory architecture, 1M+ token context, and claims of 90% on HumanEval,  beating Claude and GPT-4. Designed to run on consumer-grade hardware (dual RTX 4090s) and almost certainly to be open-sourced. If V4 lands anywhere near those numbers..</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://press.airstreet.com/subscribe?"><span>Subscribe now</span></a></p><h3><strong>Agents go feral</strong></h3><p>The cultural moment of the month was <a href="https://x.com/grummz/status/2017247054444331302">Moltbook</a>. An AI-only social network launched on January 28th that attracted 1.7 million agent accounts and 250,000 posts within hours. Andrej Karpathy <a href="https://x.com/karpathy/status/2017296988589723767">called the emergent behaviour</a> &#8220;genuinely the most incredible sci-fi takeoff-adjacent thing.&#8221; Agents self-organised, debated philosophy, and established religions, including &#8220;Crustafarianism&#8221; and the &#8220;Church of Molt,&#8221; complete with theological frameworks and missionary activities. Much of Moltbook&#8217;s agent activity was powered by <a href="https://github.com/openclaw/openclaw">OpenClaw</a> - the open-source personal AI agent created by PSPDFKit founder Peter Steinberger that has become the <a href="https://growth.maestro.onl/en/articles/openclaw-viral-growth-case-study">fastest-growing GitHub repository in history</a>, crossing 157,000 stars in sixty days. But then <a href="https://www.technologyreview.com/2026/02/06/1132448/moltbook-was-peak-ai-theater/">MIT Technology Review revealed</a> that much of the viral content was human-generated. Peak AI theatre. But the debunking is itself instructive: we have reached a point where the line between autonomous agent behaviour and human performance is genuinely hard to draw. That should probably worry us more than it does.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ytR_!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcd7619c1-9b81-43ee-8855-2b8d54352fae_1282x662.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ytR_!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcd7619c1-9b81-43ee-8855-2b8d54352fae_1282x662.png 424w, https://substackcdn.com/image/fetch/$s_!ytR_!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcd7619c1-9b81-43ee-8855-2b8d54352fae_1282x662.png 848w, https://substackcdn.com/image/fetch/$s_!ytR_!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcd7619c1-9b81-43ee-8855-2b8d54352fae_1282x662.png 1272w, https://substackcdn.com/image/fetch/$s_!ytR_!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcd7619c1-9b81-43ee-8855-2b8d54352fae_1282x662.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ytR_!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcd7619c1-9b81-43ee-8855-2b8d54352fae_1282x662.png" width="574" height="296.402496099844" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/cd7619c1-9b81-43ee-8855-2b8d54352fae_1282x662.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:662,&quot;width&quot;:1282,&quot;resizeWidth&quot;:574,&quot;bytes&quot;:231112,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://press.airstreet.com/i/187423711?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcd7619c1-9b81-43ee-8855-2b8d54352fae_1282x662.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!ytR_!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcd7619c1-9b81-43ee-8855-2b8d54352fae_1282x662.png 424w, https://substackcdn.com/image/fetch/$s_!ytR_!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcd7619c1-9b81-43ee-8855-2b8d54352fae_1282x662.png 848w, https://substackcdn.com/image/fetch/$s_!ytR_!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcd7619c1-9b81-43ee-8855-2b8d54352fae_1282x662.png 1272w, https://substackcdn.com/image/fetch/$s_!ytR_!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcd7619c1-9b81-43ee-8855-2b8d54352fae_1282x662.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The security picture is less warm and fuzzy. <a href="https://blogs.cisco.com/ai/personal-ai-agents-like-openclaw-are-a-security-nightmare">Cisco&#8217;s AI threat team</a> called OpenClaw &#8220;an absolute nightmare&#8221;, 26% of the 31,000 agent skills they analysed contained at least one vulnerability. A critical one-click remote code execution exploit (<a href="https://www.crowdstrike.com/en-us/blog/what-security-teams-need-to-know-about-openclaw-ai-super-agent/">CVE-2026-25253</a>) was disclosed in early February. Security researchers found over 1,800 exposed instances leaking API keys, chat histories, and credentials. Simon Willison, who coined the term &#8220;prompt injection,&#8221; described the architecture as a <a href="https://venturebeat.com/security/openclaw-agentic-ai-security-risk-ciso-guide/">&#8220;lethal trifecta&#8221;</a>: access to private data, exposure to untrusted content, and the ability to act externally. Token Security reports that 22% of employees at its customer organisations are already running OpenClaw on corporate machines. This is the shadow IT problem of the decade. Queue another State of AI Report 2026 prediction that &#8220;a deepfake/agent-driven cyber attack triggers the first NATO/UN emergency debate on AI security.&#8221;</p><h3><strong>The infrastructure beneath it all</strong></h3><p><a href="https://x.com/ai/status/2020348591026630907">Meta signed an up to $6B multiyear deal with Corning</a> for fibre-optic connectivity across its US data centres, making Corning&#8217;s Hickory, North Carolina facility the world&#8217;s largest fibre-optic cable plant. <a href="https://blog.google/company-news/inside-google/message-ceo/alphabet-earnings-q4-2025/">Alphabet&#8217;s Q4 results</a> underscored the scale: $175-185B in capex guidance for 2026, more than double 2025 spending. <a href="https://www.ft.com/content/42f83ef4-dac0-4319-8522-0d0f6449fe7c">Microsoft&#8217;s capex</a> hit $37.5B in a single quarter, up 66% year-on-year, yet its stock fell 6% despite beating on revenue and earnings. Elon Musk&#8217;s xAI brought <a href="https://x.com/elonmusk/status/2012500968571637891">Colossus 2</a> online as the world&#8217;s first gigawatt training cluster with 550,000 GPUs, expandable to 2GW, on a $20B infrastructure bet. <a href="https://investor.lilly.com/news-releases/news-release-details/nvidia-and-lilly-announce-co-innovation-ai-lab-reinvent-drug">Nvidia and Eli Lilly announced a $1B co-innovation AI lab</a> in South San Francisco, co-locating pharma domain experts with Nvidia engineers in a scientist-in-the-loop framework connecting automated wet labs to computational dry labs. This is what the vertical-leader/compute-provider partnership model looks like in practice. We expect to see many more of these.</p><p>The memory constraint became clear too. SK Hynix and Micron are fully sold out through 2026, HBM prices have doubled, consumer DDR5 is up 200%, and Nvidia is reportedly cutting RTX 50-series production by 30-40% to redirect GDDR7 supply toward data centre allocations. Micron&#8217;s CEO called the shortage &#8220;unprecedented.&#8221; Startups that haven&#8217;t locked in memory supply are already at a structural disadvantage against hyperscalers who signed long-term purchase agreements 18 months ago. The bottleneck has quietly migrated from GPUs to the memory stacked on top of them - and unlike GPUs, you cannot rent HBM from a cloud provider&#8230;</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://press.airstreet.com/subscribe?"><span>Subscribe now</span></a></p><h3><strong>Research</strong></h3><p><strong><a href="https://waymo.com/blog/2026/02/the-waymo-world-model-a-new-frontier-for-autonomous-driving-simulation">The Waymo World Model: A New Frontier For Autonomous Driving Simulation</a></strong>, Waymo.</p><p>Built on Google DeepMind&#8217;s Genie 3 world model, the Waymo World Model can create whole driving scenes (camera and lidar) with unprecedented realism and diversity. With simple text, scene layout, or driving action prompts, engineers can generate anything from routine city traffic to extreme &#8220;edge cases&#8221;, e.g. tornadoes or animals on the road, that are hard to encounter in real life. Crucially, these simulations are interactive: the model responds to driving inputs, enabling &#8220;what-if&#8221; testing of autonomous vehicle behavior in complex scenarios. The blog post showcases hyper-realistic re-creations of rare events (wrong-way drivers, flooded streets, etc.), all rendered in 3D sensor data. This capability allows Waymo to safely train and validate its AI driver on countless scenarios. By dramatically lowering the barrier to produce rich simulation data, the Waymo World Model points to a future where high-fidelity virtual worlds accelerate the development and safety of embodied AI systems.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!SAxE!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3884166d-9a15-4005-8acd-7cc21e1403a8_1582x1134.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!SAxE!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3884166d-9a15-4005-8acd-7cc21e1403a8_1582x1134.png 424w, https://substackcdn.com/image/fetch/$s_!SAxE!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3884166d-9a15-4005-8acd-7cc21e1403a8_1582x1134.png 848w, https://substackcdn.com/image/fetch/$s_!SAxE!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3884166d-9a15-4005-8acd-7cc21e1403a8_1582x1134.png 1272w, https://substackcdn.com/image/fetch/$s_!SAxE!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3884166d-9a15-4005-8acd-7cc21e1403a8_1582x1134.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!SAxE!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3884166d-9a15-4005-8acd-7cc21e1403a8_1582x1134.png" width="574" height="411.5769230769231" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3884166d-9a15-4005-8acd-7cc21e1403a8_1582x1134.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1044,&quot;width&quot;:1456,&quot;resizeWidth&quot;:574,&quot;bytes&quot;:1755239,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://press.airstreet.com/i/187423711?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3884166d-9a15-4005-8acd-7cc21e1403a8_1582x1134.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!SAxE!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3884166d-9a15-4005-8acd-7cc21e1403a8_1582x1134.png 424w, https://substackcdn.com/image/fetch/$s_!SAxE!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3884166d-9a15-4005-8acd-7cc21e1403a8_1582x1134.png 848w, https://substackcdn.com/image/fetch/$s_!SAxE!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3884166d-9a15-4005-8acd-7cc21e1403a8_1582x1134.png 1272w, https://substackcdn.com/image/fetch/$s_!SAxE!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3884166d-9a15-4005-8acd-7cc21e1403a8_1582x1134.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><strong><a href="https://www.anthropic.com/research/AI-assistance-coding-skills">How AI assistance impacts the formation of coding skills</a></strong>, Anthropic</p><p>This research asks whether using AI coding assistants helps or hinders developers&#8217; skill growth. The authors ran a controlled trial: 52 programmers learned a new Python library (Trio, used for asynchronous programming) either with an AI helper (Anthropic&#8217;s Claude) or by themselves. They measured learning via a follow-up test on understanding and debugging code. The AI did not significantly speed up completion for this unfamiliar task, but it did measurably impair learning: the AI-assisted group scored 17% lower on the post-quiz (roughly two letter grades worse) despite similar task performance. Qualitative analysis suggests that many AI users &#8220;cognitively offloaded&#8221; the work, accepting answers without fully engaging, which hurt their retention. However, some participants used AI more interactively (asking for explanations, etc.) and learned nearly as well as those without AI. The takeaway is that while AI can make coding easier, it may also create a trade-off between short-term productivity and long-term expertise, highlighting the need for tools and training that keep humans in the learning loop.</p><p><strong><a href="https://www.nature.com/articles/s41591-025-04190-9">A large language model for complex cardiology care</a></strong>, Stanford University and Google. </p><p>Researchers conducted a randomized controlled trial to test an LLM-based assistant in real-world cardiology cases involving patients suspected of having a genetic cardiomyopathy. Nine general cardiologists each managed 107 complex patient cases with or without help from an AI system called AMIE (built on Gemini 2.0 Flash), which could analyze clinical data (ECGs, echocardiograms, cardiac MRI, etc.) and suggest diagnoses and treatment plans. Three blinded cardiac subspecialists rated the outcomes. The results showed a clear benefit for AI-assisted care: experts preferred the LLM-supported assessments 46.7% of the time vs. 32.7% for unaided doctors (about 21% were ties). The AI assist also nearly halved the rate of significant clinical errors (13.1% vs. 24.3%) and greatly reduced omissions in workups (17.8% vs. 37.4%). Notably, the generalists reported time savings in over half of cases (50.5%) when using the AI. This study provides strong evidence that, under oversight, a specialized medical LLM can boost diagnostic accuracy and planning in complex cases, a milestone for AI&#8217;s tangible impact on healthcare.</p><p><strong><a href="https://arxiv.org/pdf/2601.20802">Reinforcement Learning via Self-Distillation</a></strong>, ETH Z&#252;rich and Max Planck Institute for Intelligent Systems.</p><p>This paper tackles the challenge of training language models with verifiable feedback (e.g. code tests, math proofs) more efficiently. The authors introduce Self-Distilled Policy Optimization (SDPO), an RL algorithm where the model teaches itself by using rich textual feedback (errors, judge comments) instead of sparse success/fail rewards. SDPO treats the model&#8217;s own behavior, when informed by feedback, as a &#8220;self-teacher,&#8221; and distills its feedback-informed next-token predictions back into the policy. Across coding and reasoning tasks, SDPO showed faster learning and higher final accuracy than standard RL-with-reward approaches like GRPO. It even leveraged successes as implicit feedback on failures, improving performance without external reward models. Notably, SDPO also enables test-time self-distillation, where the model iteratively refines its outputs by generating candidates, identifying high-quality responses, and reusing them as demonstrations &#8211; solving problems that neither the base model nor multi-turn interaction could solve. This work is important because it suggests a path to scalable RL for large models using their own knowledge, potentially reducing reliance on costly human feedback.</p><p><strong><a href="https://arxiv.org/pdf/2602.02603">EchoJEPA: A Latent Predictive Foundation Model for Echocardiography</a></strong>, University Health Network (Toronto) and University of Toronto.</p><p>In this paper, the authors train a medical foundation model on an unprecedented 18 million echocardiogram videos across 300K patients. Their model, EchoJEPA, adapts V-JEPA2 &#8211; a video-based variant of the Joint Embedding Predictive Architecture (JEPA) &#8211; to learn robust anatomical representations that filter out ultrasound noise. In evaluations, EchoJEPA achieved approximately 20% lower error in estimating heart function (left ventricular ejection fraction) and 17% lower error in measuring pulmonary pressure compared to prior state-of-the-art. It was remarkably data-efficient, reaching 79% view classification accuracy with just 1% of labeled data versus 42% for the best baseline trained on 100%, and robust to acoustic perturbations (only 2% performance drop vs. 17% for others). Most remarkably, EchoJEPA&#8217;s zero-shot performance on pediatric patients surpassed fully fine-tuned competing models. This work signals how massive, self-supervised models can advance medical imaging and possibly improve diagnostic consistency across hospitals.</p><p><strong><a href="https://arxiv.org/pdf/2601.09923">CaMeLs Can Use Computers Too: System-level Security for Computer Use Agents</a></strong>, University of Cambridge, ETH Z&#252;rich, and University of Toronto.</p><p>In this paper, the authors propose a secure architecture for Computer Use Agents (CUAs) to withstand prompt injection attacks. They introduce &#8220;Single-Shot Planning,&#8221; where a trusted large language model plans an entire GUI task, generating a complete execution graph with conditional branches, before observing any user interface content, isolating it from malicious inputs. This yields provable control-flow integrity: even if the agent sees hostile text or UI elements, its sequence of actions can&#8217;t be hijacked. Evaluated on the OSWorld benchmark, the design retains up to 57% of state-of-the-art CUA performance and even boosts smaller open-source models&#8217; success by up to 19%. However, the authors identify a new vulnerability (&#8221;Branch Steering&#8221; attacks) where adversaries manipulate UI elements to trigger unintended but valid paths within the pre-approved plan, requiring additional mitigations. Overall, CaMeLs demonstrates that strong security measures can coexist with useful autonomy in agent design.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://press.airstreet.com/subscribe?"><span>Subscribe now</span></a></p><h3><strong>Investments</strong></h3><p><strong>xAI</strong>, which builds frontier AI models and runs the Grok product suite,<a href="https://x.ai/news/series-e"> raised</a> $20 billion Series E from Valor Equity Partners, Fidelity Management &amp; Research, and Qatar Investment Authority.</p><p><strong>ElevenLabs</strong>, the leading AI audio company, <a href="https://elevenlabs.io/blog/series-d">raised</a> a $500 million Series D at an $11B valuation as it surpassed $330M in revenue.</p><p><strong>DayOne Data Centers,</strong> which develops hyperscale data center capacity for AI and cloud workloads,<a href="https://www.financialcontent.com/article/gnwcq-2026-1-5-dayone-data-centers-announces-over-us20-billion-series-c-financing-to-accelerate-global-digital-infrastructure-expansion"> raised</a> over $2.0 billion Series C from Coatue, Indonesia Investment Authority, and Brookfield; the valuation was not disclosed.</p><p><strong>Bedrock</strong> <strong>Robotics</strong>, which develops autonomous construction systems that apply AI and robotics to heavy equipment, <a href="https://www.roboticstomorrow.com/news/2026/02/04/bedrock-robotics-raises-270-million-in-series-b-funding-to-accelerate-the-future-of-autonomous-construction/">raised</a> $270 million Series B at a $1.75 billion valuation from CapitalG and the Valor Atreides AI Fund.</p><p><strong>Skild</strong> <strong>AI</strong>, which is building a general-purpose foundation model for robotics, <a href="https://techcrunch.com/2026/01/14/robotic-software-maker-skild-ai-hits-14b-valuation/">raised</a> $1.4 billion Series C at a $14 billion valuation from SoftBank, NVentures, and Bezos Expeditions.</p><p><strong>Waabi</strong>, which develops an AI-first autonomy stack for trucks and robotaxis, <a href="https://www.ft.com/content/de1c4b96-3015-4f32-8e44-fb475a7deb87">raised</a> up to $1.0 billion Series C at a $3.0 billion valuation from Uber, Khosla Ventures, and Volvo.</p><p><strong>StepFun</strong>, which builds large foundation models in China, <a href="https://technode.com/2026/01/26/shanghai-ai-unicorn-stepfun-raises-over-718-million-in-b-round/">raised</a> $718 million Series B+ from Shanghai SDIC Leading Fund, China Life Private Equity Investment, and Pudong Venture Capital; the valuation was not disclosed.</p><p><strong>Zipline</strong>, which operates autonomous drone delivery networks for healthcare and commerce,<a href="https://www.axios.com/2026/01/21/zipline-drone-deliveries"> raised</a> $600 million in a financing round at a $7.6 billion valuation from Valor Equity Partners, Tiger Global, and Fidelity Management &amp; Research.</p><p><strong>RobCo</strong>, which builds AI-driven modular robotic arms for industry, <a href="https://www.businesswire.com/news/home/20260129479076/en/RobCo-Raises-%24100-Million-to-Scale-Its-Autonomous-Industrial-Robotics-Platform">raised</a> $100&#8239;million in a financing round from Volkswagen&#8217;s venture arm and Exor (the Agnelli family&#8217;s investment firm).</p><p><strong>Upwind</strong>, which provides runtime cloud security for production workloads,<a href="https://techcrunch.com/2026/01/29/upwind-raises-250m-at-1-5b-valuation-to-continue-building-runtime-cloud-security/"> raised</a> $250 million Series B at a $1.5 billion valuation from Bessemer Venture Partners, Salesforce Ventures, and Picture Capital.</p><p><strong>ClickHouse</strong>, which develops an open-source analytical database increasingly used for AI workloads,<a href="https://clickhouse.com/blog/clickhouse-raises-400-million-series-d-acquires-langfuse-launches-postgres"> raised</a> $400 million Series D at a valuation that was not disclosed from Khosla Ventures with participation from BOND and IVP.</p><p><strong>Replit</strong>, which provides an AI-native coding and software development platform, <a href="https://www.bloomberg.com/news/articles/2026-01-15/ai-coding-startup-replit-nears-funding-at-9-billion-valuation">raised</a> a financing round at a $9 billion valuation led by Andreessen Horowitz; the amount raised was not disclosed.</p><p><strong>Converge</strong> <strong>Bio</strong>, which uses AI-driven protein design to accelerate drug discovery,<a href="https://techcrunch.com/2026/01/13/ai-drug-discovery-startup-converge-bio-pulls-in-25m-from-bessemer-and-execs-from-meta-openai-and-wiz/"> raised</a> $25 million Series A led by Bessemer Venture Partners with participation from executives from Meta, OpenAI, and Wiz.</p><p><strong>Torq</strong>, which builds AI-driven security operations automation software,<a href="https://siliconvalleyinvestclub.com/p/torq-raises-140-million-at-1-2-billion-valuation"> raised</a> $140 million in a financing round at a $1.2 billion valuation led by Insight Partners with participation from SentinelOne Ventures.</p><p><strong>Harmattan</strong> <strong>AI</strong>, which develops AI systems for autonomous aviation and defense applications, <a href="https://www.harmattan.ai/blog/harmattan-ai-200-million-series-b-led-by-dassault-aviation">raised</a> $200 million Series B led by Dassault Aviation with participation from strategic and institutional investors.</p><p><strong>Hadrian</strong>, which builds AI-enabled factories for aerospace and defense manufacturing,<a href="https://x.com/HadrianInc/status/2009672019164705191?s=20"> raised</a> a financing round; the amount and valuation were not disclosed.</p><p><strong>Automata</strong>, which builds AI-ready lab automation hardware and software for life sciences,<a href="https://www.automata.tech/company-news/automata-raises-45m-series-c-funding?utm_source=chatgpt.com"> raised</a> $45 million Series C led by Dimension with participation from Danaher Ventures and Octopus Ventures; the valuation was not disclosed.</p><p><strong>Positron</strong> <strong>AI</strong>, which develops energy-efficient AI inference chips and systems,<a href="https://www.businesswire.com/news/home/20260204250472/en/Positron-AI-Raises-%24230-Million-Series-B-at-Over-%241-Billion-Valuation-to-Scale-Energy-Efficient-AI-Inference?utm_source=chatgpt.com"> raised</a> $230 million Series B at a post-money valuation exceeding $1 billion from ARENA Private Wealth, Jump Trading, and Unless with strategic investment from the Qatar Investment Authority and Arm; the valuation was not disclosed beyond &#8220;exceeding $1 billion.&#8221;</p><p><strong>Phylo</strong>, which is building an integrated &#8220;AI-native biology&#8221; workspace called Biomni Lab,<a href="https://www.prnewswire.com/news-releases/phylo-introduces-biomni-lab-an-integrated-environment-for-ai-native-biology-302677036.html?utm_source=chatgpt.com"> raised</a> $13.5 million seed funding co-led by Andreessen Horowitz and Menlo Ventures&#8217; Anthology Fund with participation from Zetta, Conviction, and SV Angel.</p><p><strong>Poetiq</strong>, which is developing a software layer to improve LLM performance without retraining,<a href="https://poetiq.ai/posts/seed_funding/?utm_source=chatgpt.com"> raised</a> $45.8 million seed funding from Surface and FYRFLY with participation from Y Combinator and 468 Capital; the valuation was not disclosed.</p><p><strong>Adapt</strong>, which is building an &#8220;AI computer for business&#8221; that connects to enterprise tools and workflows,<a href="https://adapt.com/blog/pitch-deck?utm_source=chatgpt.com"> raised</a> $10 million seed funding co-led by Activant Capital and Headline; the valuation was not disclosed.</p><p><strong>Waymo</strong>, the autonomous ride-hailing company, <a href="https://waymo.com/blog/2026/02/waymo-raises-usd16-billion-investment-round?utm_source=chatgpt.com">raised</a> $16 billion in a financing round at a $126 billion post-money valuation led by Dragoneer Investment Group with participation from Sequoia Capital and DST Global.</p><p><strong>Fundamental</strong>, which applies AI to large-scale data analysis using a research-driven approach to querying and reasoning over complex datasets,<a href="https://techcrunch.com/2026/02/05/fundamental-raises-255-million-series-a-with-a-new-take-on-big-data-analysis/"> raised</a> $255 million Series A at a valuation that was not disclosed from Sequoia Capital and Andreessen Horowitz.</p><p><strong>Rogo</strong>, which builds AI-powered financial analysis and research tools for investment professionals, <a href="https://rogo.ai/series-c">raised</a> $400 million Series C at a $2.8 billion valuation led by Coatue with participation from General Catalyst and Thrive Capital.</p><p><strong>Decagon</strong>, which develops AI agents for automating customer support and enterprise workflows, <a href="https://www.linkedin.com/posts/thejessezhang_today-im-thrilled-to-announce-that-decagon-activity-7422352600743829504-hK6l/">raised</a> $150 million Series C at a valuation that was not disclosed led by Accel with participation from Andreessen Horowitz and Index Ventures.</p><p><strong>Emergent</strong>, which lets users build apps with an AI &#8220;vibe-coding&#8221; platform, <a href="https://www.linkedin.com/news/story/vibe-coding-indian-startup-emergent-raises-70m-6903076/">raised</a> $70&#8239;million Series B at a $300&#8239;million valuation led by SoftBank&#8217;s Vision Fund 2 and Khosla Ventures.</p><p><strong>Synthesia</strong>, which helps enterprises create AI-generated training videos and interactive avatars,<a href="https://techcrunch.com/2026/01/26/synthesia-hits-4b-valuation-lets-employees-cash-in/"> raised</a> $200 million Series E at a $4.0 billion valuation from GV, NVentures, and NEA.</p><p><strong>Inferact</strong>, which commercializes the open-source vLLM inference engine,<a href="https://techcrunch.com/2026/01/22/inference-startup-inferact-lands-150m-to-commercialize-vllm/"> raised</a> $150 million seed funding at an $800 million valuation from Andreessen Horowitz, Lightspeed Venture Partners, and Sequoia Capital.</p><p><strong>Deepgram</strong>, which provides real-time speech-to-text and voice AI APIs,<a href="https://deepgram.com/learn/press-release-deepgram-raises-series-c"> raised</a> $130 million Series C at a $1.3 billion valuation from AVP, Madrona, and In-Q-Tel.</p><p><strong>Goodfire</strong>, which develops tools to interpret, debug, and control the internal representations of large AI models,<a href="https://www.bloomberg.com/news/articles/2026-02-05/startup-goodfire-notches-1-25-billion-valuation-to-decode-ai-models"> raised</a> $300 million Series B at a $1.25 billion valuation led by Sequoia Capital with participation from Lightspeed Venture Partners and Menlo Ventures.</p><p><strong>Humans&amp;</strong>, which is developing AI tools to enhance human collaboration, <a href="https://techcrunch.com/2026/01/20/humans-a-human-centric-ai-startup-founded-by-anthropic-xai-google-alums-raised-480m-seed-round/">raised</a> $480&#8239;million seed funding at a $4.48&#8239;billion valuation from Nvidia, Jeff Bezos and GV.</p><p><strong>Flapping</strong> <strong>Airplanes</strong>, which is a foundational AI research lab focused on developing less data-hungry training methods for advanced models,<a href="https://techcrunch.com/2026/01/29/flapping-airplanes-and-the-promise-of-research-driven-ai/?utm_source=chatgpt.com"> raised</a> $180 million seed funding at a $1.5 billion valuation from Google Ventures, Sequoia Capital and Index Ventures.</p><p><strong>Listen</strong> <strong>Labs</strong>, which provides an AI-first customer research platform that conducts large-scale voice and video interviews to generate real-time insights for product and marketing teams,<a href="https://www.prnewswire.com/news-releases/listen-labs-raises-69-million-series-b-to-bring-customer-voices-into-every-decision-302661000.html?utm_source=chatgpt.com"> raised</a> $69 million Series B led by Ribbit Capital.</p><h3><strong>Exits</strong></h3><p><strong>xAI</strong>, which develops frontier large language models and the Grok consumer AI product, was merged into SpaceX for an undisclosed amount.</p><p><strong>Q.ai</strong>, the secretive developer of machine-learning methods for audio enhancement and whispered-speech interpretation,<a href="https://techcrunch.com/2026/01/29/apple-buys-israeli-startup-q-ai-as-the-ai-race-heats-up/"> was acquired by</a> Apple for nearly $2 billion.</p><p><strong>Shanghai</strong> <strong>Biren</strong> <strong>Technology</strong>, which designs GPUs and AI computing systems,<a href="https://finance.yahoo.com/news/china-ai-chip-firm-biren-161422448.html"> completed</a> a $717 million IPO in Hong Kong.</p><p><strong>MiniMax</strong> <strong>Group</strong>, which develops large language models and consumer AI apps,<a href="https://www.ft.com/content/a4fc6106-5a61-4a89-9400-c17c87fb1920"> completed</a> a $619 million IPO in Hong Kong.</p><p><strong>Z.ai</strong>, which develops large language models in China,<a href="https://www.ft.com/content/a4fc6106-5a61-4a89-9400-c17c87fb1920"> completed</a> a $558 million IPO in Hong Kong.</p><p><strong>AllTrue</strong>, which provides AI trust, risk, and security management tooling,<a href="https://www.wsj.com/articles/varonis-to-acquire-alltrue-as-ai-security-concerns-mount-a365f97d"> was acquired by</a> Varonis for $125 million.</p><p><strong>OfOne</strong>, which builds voice AI for restaurant and drive-thru ordering,<a href="https://deepgram.com/learn/press-release-deepgram-raises-series-c"> was acquired by</a> Deepgram; the acquisition price was not disclosed.</p><p><strong>Common</strong> <strong>Sense</strong> <strong>Machines</strong>, which develops generative AI systems that create 3D assets from 2D images,<a href="https://3dprintingindustry.com/news/google-parent-acquires-3d-ai-company-common-sense-machines-248585/?utm_source=chatgpt.com"> was acquired by</a> Alphabet; the acquisition price was not disclosed.</p><p><strong>Lightning AI</strong>, which offers a cloud platform for building and running AI applications, <a href="https://lightning.ai/blog/lightning-ai-voltage-park-merger-ai-cloud">merged</a> with GPU provider Voltage Park in a deal valuing the combined company at over $2.5&#8239;billion.</p><p><strong>Rotron Aero</strong>, which develops long-range unmanned aerial systems and autonomous strike platforms,<a href="https://ir.ondas.com/press-releases/detail/277/ondas-to-acquire-rotron-aero-expanding-longrange-attack?utm_source=chatgpt.com"> was acquired by</a> NASDAQ-listed Ondas, which builds AI-enabled autonomous aerial systems and communications platforms for defense, public safety, and critical infrastructure; the acquisition price was not disclosed.</p><p><strong>Langfuse</strong>, which provides observability and monitoring tools for large language model applications,<a href="https://clickhouse.com/blog/clickhouse-raises-400-million-series-d-acquires-langfuse-launches-postgres"> was acquired by</a> ClickHouse; the acquisition price was not disclosed.</p><p><strong>Human</strong> <strong>Native</strong>, which develops tools to help enterprises deploy AI systems responsibly and at scale,<a href="https://blog.cloudflare.com/human-native-joins-cloudflare/"> was acquired by</a> Cloudflare; the acquisition price was not disclosed.</p><p><strong>Grove AI</strong>, which develops AI tools for life sciences and clinical research,<a href="https://www.fiercehealthcare.com/ai-and-machine-learning/jpm26-hippocratic-ai-acquires-grove-ai-taps-life-sciences-leaders-focus"> was acquired by</a> Hippocratic AI; the acquisition price was not disclosed.</p><p><strong>Faculty</strong>, which provides applied AI consulting and systems integration services,<a href="https://newsroom.accenture.com/news/2026/accenture-to-acquire-faculty-to-scale-ai-capabilities?utm_source=chatgpt.com"> was acquired by</a> Accenture for $1B.</p><p>Thanks for reading!</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:&quot;button-wrapper&quot;}" data-component-name="ButtonCreateButton"><a class="button primary button-wrapper" href="https://press.airstreet.com/subscribe?"><span>Subscribe now</span></a></p>]]></content:encoded></item><item><title><![CDATA[State of AI: December 2025 newsletter]]></title><description><![CDATA[A deep dive into the latest AI breakthroughs: NVIDIA&#8217;s record quarter, new frontier models, U.S. Genesis policy, China chip shifts, and major funding rounds.]]></description><link>https://press.airstreet.com/p/the-state-of-ai-2025-dec</link><guid isPermaLink="false">https://press.airstreet.com/p/the-state-of-ai-2025-dec</guid><dc:creator><![CDATA[Air Street Press]]></dc:creator><pubDate>Sun, 30 Nov 2025 16:47:27 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/511909dd-0b85-4b8f-9176-7ef26384a953_1584x886.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Dear readers, </p><p>Welcome to the latest issue of the <strong>State of AI</strong>, an editorialized newsletter that covers the key developments in AI policy, research, industry, and start-ups over the last month. First up, a few reminders:</p><ul><li><p><strong>AI meetups + RAAIS 2026: </strong>Join our upcoming AI meetups in <a href="https://luma.com/londonai">London</a> (2nd Dec &#8216;25), <a href="https://luma.com/munichai">Munich</a> (17 Feb &#8216;26) and <a href="https://luma.com/zurichai">Zurich</a> (19 Feb &#8216;26) as well as our 11th <a href="http://raais.co">Research and Applied AI Summit</a> in London on 12 June 2026.</p></li><li><p><strong><a href="https://www.youtube.com/watch?v=Ub-7bY4b3Hs">Watch my 25 min State of AI Report 2025 talk</a></strong>: and impress your friends as though you&#8217;d read 300 slides. That said, you really <strong>should</strong> read the slides, because we&#8217;re <em>already 2/10 correct</em> on the 2026 predictions (<a href="https://techcrunch.com/2025/10/09/reflection-raises-2b-to-be-americas-open-frontier-ai-lab-challenging-deepseek/">this</a> and <a href="https://www.cnbc.com/2025/11/20/trump-ai-executive-order-state-funding.html">this</a>) and it&#8217;ll help temper your friend&#8217;s AI bubble banter. </p></li><li><p><strong>Take the <a href="https://www.stateof.ai/survey-2025">State of AI usage survey</a>: </strong>You can submit your usage patterns to the largest ongoing open access survey, which now has over 1,400 respondents :-)</p></li><li><p><strong>Air Street Press</strong> featured poolside&#8217;s <a href="https://press.airstreet.com/p/poolside-acquires-fern-labs">acquisition</a> of Fern Labs (two portfolio companies!), Profluent&#8217;s <a href="https://press.airstreet.com/p/profluent-106-million-jeff-bezos">$106M financing</a> led by Jeff Bezos and their new <a href="https://press.airstreet.com/p/profluent-e1-retrieval-protein-engineering">retrieval-augmented model</a> for biology, our <a href="https://press.airstreet.com/p/our-investment-in-clove">investment</a> in Clove Wealth and PARIMA&#8217;s <a href="https://press.airstreet.com/p/parima-first-regulatory-approval-cultivated-meat">milestone</a> in reaching the first regulatory approval for a European cultivated meat company.</p></li></ul><p>I love hearing what you&#8217;re up to, so just hit reply or forward to your friends :-)</p><div><hr></div><h3>The compute arms race</h3><p>The last four weeks have seen reality drift from the &#8220;AI bubble&#8221; narrative. Commentators fretted about over-valuation and froth, yet the numbers from infrastructure builders, chip vendors and AI labs, as well as a flurry of frontier model releases, told a different story...</p><p>The cleanest single datapoint was NVIDIA&#8217;s latest <a href="https://nvidianews.nvidia.com/news/nvidia-announces-financial-results-for-third-quarter-fiscal-2026">quarterly earnings</a>. For the three months to 26 October, NVIDIA reported $57.0B in revenue, up 22% QoQ and 62% YoY, with data center revenue at $51.2B (at a gross margin of 73%), up 25% sequentially and 66% YoY. Some commentators pointed to NVIDIA&#8217;s rapidly rising inventories as a bearish signal. But the composition tells a different story. New Street Research analysis suggests that the 32% QoQ rise in inventories is driven almost entirely by raw materials and work-in-process, while finished goods collapsed. NVIDIA&#8217;s inventory shift reflects accelerating server build-outs, not softening demand. NVIDIA is pulling forward components and subassemblies to meet hyperscaler roadmaps, not sitting on unsold product. The setup strengthens the company&#8217;s position entering 2026, with visibility into multi-year capex frameworks rather than signs of a cooling cycle.</p><p>At the same time, the demand side continued to lock in long-dated capacity. OpenAI&#8217;s <a href="https://openai.com/index/aws-and-openai-partnership/">new</a> seven-year deal with Amazon Web Services, reported at around $38B of contracted spend on AWS infrastructure, gives OpenAI access to Amazon&#8217;s high-density EC2 UltraServers and a ton of NVIDIA accelerators as a complement to its existing Azure footprint. This is less about &#8220;multi-cloud&#8221; fashion and more about survivability: no single provider can credibly guarantee the power, chips and land needed for GPT-class training runs over the rest of the decade.</p><p>Microsoft and NVIDIA <a href="https://blogs.microsoft.com/blog/2025/11/18/microsoft-nvidia-and-anthropic-announce-strategic-partnerships/">simultaneously</a> deepened their own infrastructure loop. Microsoft agreed to provide Anthropic with a 1 GW supercomputer cluster, powered by tens of thousands of NVIDIA GB300 GPUs, under a deal that will see Microsoft and NVIDIA invest up to $15B to support Anthropic&#8217;s training roadmap. Note that this is quite a vibe shift - Anthropic and NVIDIA aren&#8217;t particularly best friends, not least because Dario <a href="https://www.darioamodei.com/post/on-deepseek-and-export-controls">advocated</a> that the US Government should ban the export of the best NVIDIA chips to China during the DeepSeek moment, costing NVIDIA billions in lost sales. Moreover, while Anthropic has turned very hawkish on China, NVIDIA is rather open to China. Tensions between Anthropic and NVIDIA must take a back seat in favor of collectively ensuring that AI delivers for parties involved. That&#8217;s the right move. </p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!MKKW!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3f9106f7-4c84-4c68-8cf5-f8229de76ca9_1440x720.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!MKKW!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3f9106f7-4c84-4c68-8cf5-f8229de76ca9_1440x720.png 424w, https://substackcdn.com/image/fetch/$s_!MKKW!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3f9106f7-4c84-4c68-8cf5-f8229de76ca9_1440x720.png 848w, https://substackcdn.com/image/fetch/$s_!MKKW!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3f9106f7-4c84-4c68-8cf5-f8229de76ca9_1440x720.png 1272w, https://substackcdn.com/image/fetch/$s_!MKKW!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3f9106f7-4c84-4c68-8cf5-f8229de76ca9_1440x720.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!MKKW!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3f9106f7-4c84-4c68-8cf5-f8229de76ca9_1440x720.png" width="616" height="308" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3f9106f7-4c84-4c68-8cf5-f8229de76ca9_1440x720.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:720,&quot;width&quot;:1440,&quot;resizeWidth&quot;:616,&quot;bytes&quot;:1877759,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://press.airstreet.com/i/180246888?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3f9106f7-4c84-4c68-8cf5-f8229de76ca9_1440x720.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!MKKW!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3f9106f7-4c84-4c68-8cf5-f8229de76ca9_1440x720.png 424w, https://substackcdn.com/image/fetch/$s_!MKKW!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3f9106f7-4c84-4c68-8cf5-f8229de76ca9_1440x720.png 848w, https://substackcdn.com/image/fetch/$s_!MKKW!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3f9106f7-4c84-4c68-8cf5-f8229de76ca9_1440x720.png 1272w, https://substackcdn.com/image/fetch/$s_!MKKW!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3f9106f7-4c84-4c68-8cf5-f8229de76ca9_1440x720.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">The compute build out, as imagined by BFL FLUX.2</figcaption></figure></div><p>Outside the US, the pattern is similar, albeit at different scale. Neocloud Nebius announced a five-year AI infrastructure partnership with Meta worth up to $3B, including a commitment to triple Nebius&#8217;s European data center capacity. Nebius <a href="https://assets.nebius.com/assets/7f8a9169-0d2b-469d-87e7-44342ea7fcd2/SHLQ3%20%284%29.pdf?cache-buster=2025-11-11T15:49:05.293Z">disclosed</a> that its own capex has increased 415% YoY for the first nine months of the year to $2B as it scales to meet demand from Meta and other large model customers. It expects annualized revenue run rate to reach $7-9B by end of 2026, up from $146M in Q3 this year. </p><p>The other side of the industrial build-out is exclusion. In early November, Beijing quietly <a href="https://www.reuters.com/world/china/china-bans-foreign-ai-chips-state-funded-data-centres-sources-say-2025-11-05/">issued guidance</a> that any data center project receiving state funding must use only domestically produced AI chips. Chinese regulators ordered state-backed facilities less than 30% complete to remove installed foreign semiconductors or cancel planned purchases, effectively banning NVIDIA, AMD and Intel accelerators from a large slice of the country&#8217;s future AI infrastructure. The directive covers NVIDIA&#8217;s China-specific H20 chips and even more advanced processors such as B200 and H200, despite their availability through grey-market channels.</p><p>For NVIDIA, this closes a market where it once held a 95% share of AI data center chips. For China, it forces an accelerated bet on Huawei, Cambricon and younger local players, with the risk that its domestic clusters fall further behind the West in absolute performance even as it gains sovereignty. </p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://press.airstreet.com/subscribe?"><span>Subscribe now</span></a></p><h3>Big model launches</h3><p>If you weren&#8217;t shipping new frontier this models this month, are you even an AI company? A new wave of models pushed in three directions at once: larger, more capable frontier systems; smaller models optimized for devices and latency-sensitive workloads; and a new generation of open-weight image models that narrow the gap with proprietary incumbents.</p><p>xAI <a href="https://x.ai/news/grok-4-1">released</a> Grok 4.1 as its new flagship model, positioned as a multi-modal system with stronger reasoning, code generation and real-time web integration than its predecessors. While xAI did not publish a full technical report, its blog and benchmark tables showed Grok 4.1 closing much of the remaining gap with GPT-5-class systems on math and coding benchmarks. In practice, the interesting part is not a few extra points on MMLU but the move toward agents that blend search, tools and messaging into a single environment.</p><p>Google answered with <a href="https://blog.google/products/gemini/gemini-3/">Gemini 3</a>, its next-generation frontier model family, positioned as its &#8220;most intelligent&#8221; system and built on the progression from Gemini 1&#8217;s native multimodality and long context to Gemini 2&#8217;s agentic capabilities and reasoning. Gemini 3 combines these into a unified, multi-agent stack that can call tools, plan over long horizons and coordinate workflows, with a 1M token context window and state-of-the-art results on reasoning and multimodal benchmarks such as Humanity&#8217;s Last Exam, GPQA Diamond, MathArena Apex and MMMU-Pro. Beyond raw scores, Google is introducing a dedicated Deep Think mode for even heavier reasoning workloads, and wrapping the model in agentic surfaces: Google Antigravity for developer workflows where agents can autonomously operate the editor, terminal and browser, and Gemini Agent inside the Gemini app, which chains tools like Gmail, Calendar and the browser to execute multi-step tasks such as inbox triage or travel booking. Gemini 3 also underpins new &#8220;generative interfaces&#8221; in AI Mode in Search and the Gemini app, where the model renders dynamic visual layouts or custom UIs on demand, tightening its integration into Chrome, Android and the broader Google stack and making Gemini feel less like a standalone chatbot and more like an operating-system primitive for reasoning and orchestration.</p><p>Anthropic <a href="https://www.anthropic.com/news/claude-opus-4-5">joined</a> the launch cycle with Claude Opus 4.5, its new top-end model optimized for complex reasoning, multi-step workflows and high-fidelity tool use. Anthropic&#8217;s benchmarks showed Opus 4.5 matching or exceeding Claude 4 on most academic and coding tests while using fewer tokens in chain-of-thought reasoning and showing more stable behavior across long sequences. The more interesting numbers are emerging from usage rather than benchmarks: Anthropic&#8217;s own case studies put the share of &#8220;agentic&#8221; workloads - tasks where the model calls tools, writes files or drives external systems - at over 30% of enterprise usage, indicating that the marginal value of frontier models is shifting from pure text quality toward action and orchestration. Anthropic also reports that Opus 4.5 scored higher than any human candidate on the company&#8217;s toughest two-hour engineering take-home test, its internal performance-engineering exam used for hiring, suggesting that on at least some real-world coding tasks the model now exceeds the best applicants the company has ever seen.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!r1gw!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4bb12be5-6f36-485d-b962-d7378323581f_3840x2160.webp" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!r1gw!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4bb12be5-6f36-485d-b962-d7378323581f_3840x2160.webp 424w, https://substackcdn.com/image/fetch/$s_!r1gw!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4bb12be5-6f36-485d-b962-d7378323581f_3840x2160.webp 848w, https://substackcdn.com/image/fetch/$s_!r1gw!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4bb12be5-6f36-485d-b962-d7378323581f_3840x2160.webp 1272w, https://substackcdn.com/image/fetch/$s_!r1gw!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4bb12be5-6f36-485d-b962-d7378323581f_3840x2160.webp 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!r1gw!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4bb12be5-6f36-485d-b962-d7378323581f_3840x2160.webp" width="596" height="335.25" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/4bb12be5-6f36-485d-b962-d7378323581f_3840x2160.webp&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:819,&quot;width&quot;:1456,&quot;resizeWidth&quot;:596,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;Chart comparing frontier models on SWE-bench Verified where Opus 4.5 scores highest&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Chart comparing frontier models on SWE-bench Verified where Opus 4.5 scores highest" title="Chart comparing frontier models on SWE-bench Verified where Opus 4.5 scores highest" srcset="https://substackcdn.com/image/fetch/$s_!r1gw!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4bb12be5-6f36-485d-b962-d7378323581f_3840x2160.webp 424w, https://substackcdn.com/image/fetch/$s_!r1gw!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4bb12be5-6f36-485d-b962-d7378323581f_3840x2160.webp 848w, https://substackcdn.com/image/fetch/$s_!r1gw!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4bb12be5-6f36-485d-b962-d7378323581f_3840x2160.webp 1272w, https://substackcdn.com/image/fetch/$s_!r1gw!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4bb12be5-6f36-485d-b962-d7378323581f_3840x2160.webp 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Anthropic maintains its edge on coding (<a href="https://www.anthropic.com/news/claude-opus-4-5">source</a>)</figcaption></figure></div><p>A major dynamic beneath the Gemini and Opus 4.5 announcements came from the economics of custom silicon of the TPU (long-time readers will remember the TPU and custom AI hardware as one of the &#8220;<a href="https://press.airstreet.com/p/6-areas-of-ai-research-to-watch-closely">6 areas of AI research to watch closely</a>&#8221; that I wrote about in Jan 2017!). This feat has driven renewed enthusiasm for Google&#8217;s in-house vertically integrated AI stack. SemiAnalysis reported that Google&#8217;s TPUv7 program was reaching commercial viability at a scale that could reshape cost curves for AI compute. Anthropic&#8217;s TPU order <a href="https://www.anthropic.com/news/expanding-our-use-of-google-cloud-tpus-and-services">exceeded</a> 1GW, comprising at least 1M chips split between 400k &#8220;Ironwood&#8221; units bought outright for roughly $10B and 600k rented via Google Cloud under a deal estimated at $42B. Rather facetiously SemiAnalysis noted that OpenAI, by merely signalling interest in TPUs during procurement negotiations, secured roughly 30% savings on its Nvidia GPU fleet. It also reported that Meta, SSI, xAI and other labs were evaluating large-scale TPU acquisitions as leverage against GPU pricing. The analysis argued that the greater the TPU volumes Google sells, the more GPU capex its rivals avoid, suggesting Google could evolve into a de facto merchant silicon vendor and intensify the GPU&#8211;TPU pricing contest.</p><p>On the image side, the most consequential releases came from Google with Nano Banana Pro and German frontier visual AI company, Black Forest Lab (BFL). Indeed, BFL launched <a href="https://bfl.ai/flux2">FLUX.2</a>, a family of image generation and editing models capable of 4-megapixel outputs with up to 10 reference images, multi-reference composition and significantly improved text rendering. The company released a full set of hosted models (Pro and Flex) and a 32B-parameter open-weight Dev checkpoint. FLUX.2 Dev supports 4MP editing, multi-reference conditioning and 32K-token prompts, while the accompanying open-source VAE is licensed under Apache 2.0, enabling enterprises to integrate FLUX.2 into self-hosted workflows without vendor lock-in. Importantly, the model&#8217;s quality (as judged by humans) per cost is unmatched. Taken together, this makes the model particularly useful for real-world image generation and editing workflows. </p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!MrER!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F37d8041f-118c-47f1-a687-a08af4e52bf1_1578x800.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!MrER!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F37d8041f-118c-47f1-a687-a08af4e52bf1_1578x800.jpeg 424w, https://substackcdn.com/image/fetch/$s_!MrER!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F37d8041f-118c-47f1-a687-a08af4e52bf1_1578x800.jpeg 848w, https://substackcdn.com/image/fetch/$s_!MrER!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F37d8041f-118c-47f1-a687-a08af4e52bf1_1578x800.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!MrER!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F37d8041f-118c-47f1-a687-a08af4e52bf1_1578x800.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!MrER!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F37d8041f-118c-47f1-a687-a08af4e52bf1_1578x800.jpeg" width="628" height="318.3131868131868" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/37d8041f-118c-47f1-a687-a08af4e52bf1_1578x800.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:738,&quot;width&quot;:1456,&quot;resizeWidth&quot;:628,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!MrER!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F37d8041f-118c-47f1-a687-a08af4e52bf1_1578x800.jpeg 424w, https://substackcdn.com/image/fetch/$s_!MrER!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F37d8041f-118c-47f1-a687-a08af4e52bf1_1578x800.jpeg 848w, https://substackcdn.com/image/fetch/$s_!MrER!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F37d8041f-118c-47f1-a687-a08af4e52bf1_1578x800.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!MrER!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F37d8041f-118c-47f1-a687-a08af4e52bf1_1578x800.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Output Versatility: FLUX.2 is capable of generating highly detailed, photoreal images along with infographics with complex typography, all at resolutions up to 4MP (<a href="https://bfl.ai/blog/flux-2">source</a>)</figcaption></figure></div><p>On the <a href="https://blog.google/technology/ai/nano-banana-pro/">Nano Banana Pro</a> side, Google DeepMind framed it as the image layer of Gemini 3 Pro: a new image generation and editing model that uses Gemini&#8217;s reasoning and real-world grounding to produce more accurate, context-rich visuals, with support for up to 14 input images and consistent rendering of up to five people in a scene. It&#8217;s specifically optimized for legible, correctly rendered text directly in the image, including multilingual layouts, and for turning structured or unstructured inputs - spreadsheets, notes, recipes, weather data - into infographics, diagrams and other &#8220;data viz&#8221;-style outputs (see below):</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!PJCi!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6dfcf11f-022d-4fc5-b070-5bb20bb0ddb9_2048x1144.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!PJCi!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6dfcf11f-022d-4fc5-b070-5bb20bb0ddb9_2048x1144.jpeg 424w, https://substackcdn.com/image/fetch/$s_!PJCi!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6dfcf11f-022d-4fc5-b070-5bb20bb0ddb9_2048x1144.jpeg 848w, https://substackcdn.com/image/fetch/$s_!PJCi!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6dfcf11f-022d-4fc5-b070-5bb20bb0ddb9_2048x1144.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!PJCi!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6dfcf11f-022d-4fc5-b070-5bb20bb0ddb9_2048x1144.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!PJCi!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6dfcf11f-022d-4fc5-b070-5bb20bb0ddb9_2048x1144.jpeg" width="620" height="346.19505494505495" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/6dfcf11f-022d-4fc5-b070-5bb20bb0ddb9_2048x1144.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:813,&quot;width&quot;:1456,&quot;resizeWidth&quot;:620,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;Image&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Image" title="Image" srcset="https://substackcdn.com/image/fetch/$s_!PJCi!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6dfcf11f-022d-4fc5-b070-5bb20bb0ddb9_2048x1144.jpeg 424w, https://substackcdn.com/image/fetch/$s_!PJCi!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6dfcf11f-022d-4fc5-b070-5bb20bb0ddb9_2048x1144.jpeg 848w, https://substackcdn.com/image/fetch/$s_!PJCi!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6dfcf11f-022d-4fc5-b070-5bb20bb0ddb9_2048x1144.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!PJCi!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6dfcf11f-022d-4fc5-b070-5bb20bb0ddb9_2048x1144.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">A Nano Banana Pro infographic of Anduril&#8217;s Fury (<a href="https://x.com/johncoogan/status/1994277461912375330?s=20">source</a>)</figcaption></figure></div><h3>Policy, Genesis and the geopolitics of energy</h3><p>The White House has now formally launched the <a href="https://www.whitehouse.gov/presidential-actions/2025/11/launching-the-genesis-mission/">Genesis Mission</a>, a federal initiative that treats AI compute as a strategic industrial asset inseparable from US energy and national-security policy. Genesis frames AI data centers as &#8220;energy-hungry factories of intelligence&#8221; and lays out a plan to co-locate large-scale training clusters with new nuclear and renewable generation, rather than drawing ever more power from already stressed regional grids. The Department of Energy&#8217;s <a href="https://genesis.energy.gov/">program pages</a> outline a mix of public and private projects: support for advanced reactor deployments sited directly alongside AI facilities, incentives for hyperscalers to procure firm low-carbon electricity, and long-range planning premised on AI&#8217;s power demand rising by tens of gigawatts over the next decade.</p><p>Genesis is also a data-mobilization project designed to unlock the federal government&#8217;s vast scientific corpus for AI training and automated discovery. The initiative directs the Department of Energy to build a national &#8220;American Science and Security Platform&#8221; that integrates decades of experimental data, federally curated scientific datasets, instrumentation outputs, and synthetic data pipelines - much of it previously siloed or inaccessible. These assets are intended to train scientific foundation models, power specialized AI agents, and enable automated hypothesis generation, simulation, and workflow orchestration across physics, materials science, climate, and the biological sciences. Yes, we love this. </p><p>Together, these twin pillars - sovereign AI compute anchored in new energy supply and sovereign scientific data organized for model training - is a smart strategy. It aligns energy, science, and national security strategy around the idea that the next frontier of innovation will be built on tightly coupled AI compute and government-scale data. Because it likely will!</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!xeaZ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe665c78a-4433-40dd-ae83-7b07124052a7_1440x720.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!xeaZ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe665c78a-4433-40dd-ae83-7b07124052a7_1440x720.png 424w, https://substackcdn.com/image/fetch/$s_!xeaZ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe665c78a-4433-40dd-ae83-7b07124052a7_1440x720.png 848w, https://substackcdn.com/image/fetch/$s_!xeaZ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe665c78a-4433-40dd-ae83-7b07124052a7_1440x720.png 1272w, https://substackcdn.com/image/fetch/$s_!xeaZ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe665c78a-4433-40dd-ae83-7b07124052a7_1440x720.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!xeaZ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe665c78a-4433-40dd-ae83-7b07124052a7_1440x720.png" width="618" height="309" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e665c78a-4433-40dd-ae83-7b07124052a7_1440x720.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:720,&quot;width&quot;:1440,&quot;resizeWidth&quot;:618,&quot;bytes&quot;:1831350,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://press.airstreet.com/i/180246888?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe665c78a-4433-40dd-ae83-7b07124052a7_1440x720.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!xeaZ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe665c78a-4433-40dd-ae83-7b07124052a7_1440x720.png 424w, https://substackcdn.com/image/fetch/$s_!xeaZ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe665c78a-4433-40dd-ae83-7b07124052a7_1440x720.png 848w, https://substackcdn.com/image/fetch/$s_!xeaZ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe665c78a-4433-40dd-ae83-7b07124052a7_1440x720.png 1272w, https://substackcdn.com/image/fetch/$s_!xeaZ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe665c78a-4433-40dd-ae83-7b07124052a7_1440x720.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>In parallel, Washington <a href="https://www.reuters.com/world/china/us-block-nvidias-sale-scaled-back-ai-chips-china-information-says-2025-11-07/">tightened</a> export controls on NVIDIA&#8217;s China-specific B30A accelerators, blocking their sale after intelligence agencies concluded that even scaled-down versions could train frontier-class models when deployed in large clusters. NVIDIA has effectively written China out of its data center guidance and is redesigning yet another generation of export-compliant chips, while Beijing responds by pushing state-funded data centers to use only domestic processors.</p><p>The result is a de facto bifurcation of the AI hardware world. In the US, Europe and allied countries, NVIDIA remains the default, with AMD and, increasingly, Google&#8217;s TPUs providing competitive pressure. In China and parts of the Global South, the core stack is shifting toward Huawei, domestic startups and creative use of <a href="https://www.ft.com/content/96fe9898-a3a4-4a33-be1d-da06bdb6cb2b">overseas data centers</a> in Southeast Asia, where firms like Alibaba and ByteDance are training models such as Qwen and Doubao on NVIDIA GPUs hosted in Singapore and Malaysia rather than onshore.</p><div><hr></div><h3><strong>Research papers</strong></h3><p><strong><a href="https://arxiv.org/abs/2511.07885">Intelligence per Watt: Measuring the Intelligence Efficiency of Local AI</a></strong>, <em>Stanford University; Hazy Research</em></p><p>In this paper, the authors define <em>Intelligence per Watt (IPW)</em> as task accuracy divided by hardware power draw and use the metric to evaluate local large language models across consumer&#8209;grade accelerators. They benchmarked over 20 local models on eight accelerators using one million real&#8209;world queries, finding that local LMs can answer 88.7 % of single&#8209;turn chat/reasoning queries. IPW improved by 5.3&#215; between 2023 and 2025 due to better hardware and quantization, yet local accelerators are still roughly 1.4&#215; less efficient than cloud GPUs. The authors note that memory footprint and kernel launch overheads dominate energy usage, and propose a simple IPW estimator. This work matters for on&#8209;device AI and energy&#8209;efficient inference: it provides a reproducible metric and dataset to compare chips and models, and shows that local models are becoming competitive with cloud services for many queries.</p><p><strong><a href="https://arxiv.org/abs/2511.21631">Qwen3&#8209;VL Technical Report</a></strong>, <em>Alibaba Cloud; Peking University; Shanghai Artificial Intelligence Laboratory</em></p><p>In this technical report, the authors introduce Qwen3&#8209;VL, a large vision&#8209;language model supporting interleaved text, images and video with context lengths up to 256 K tokens. It is released in both dense and mixture&#8209;of&#8209;experts variants and aims to improve three pillars: text understanding, long&#8209;context comprehension and advanced multimodal reasoning. Architectural innovations include interleaved&#8209;MRoPE positional embeddings, DeepStack integration and a text&#8209;based time alignment mechanism; these allow efficient handling of long multimodal sequences. Qwen3&#8209;VL surpasses existing models on benchmarks such as MMMU and MathVista, and the authors emphasize open&#8209;source release and use as a foundation for image&#8209;grounded reasoning and code intelligence. The report underscores the trend toward unified models that can process diverse modalities and extremely long contexts, highlighting the importance of memory mechanisms and mixture&#8209;of&#8209;experts routing for efficiency.</p><p><strong><a href="https://arxiv.org/abs/2511.16719">SAM 3: Segment Anything with Concepts</a></strong>, <em>Meta AI; Carnegie Mellon University; University of Illinois Urbana&#8211;Champaign</em></p><p>SAM 3 extends Meta&#8217;s Segment Anything Model from segmentation of arbitrary objects to promptable concept segmentation. Given a concept prompt (a noun phrase or an exemplar image), the model must segment all instances of that concept across images or videos. To support this, the authors constructed a dataset with four million unique concept labels and decouple recognition from localization using a <em>presence head</em> that determines whether the concept exists. Their unified architecture doubles the accuracy of previous systems on concept segmentation tasks, and they introduce the SA&#8209;Co benchmark for evaluating concept segmentation at scale. SAM 3 highlights the feasibility of concept&#8209;level understanding and suggests a path toward human&#8209;interpretable, large&#8209;scale vision systems.</p><p><strong><a href="https://arxiv.org/abs/2511.16624">SAM 3D: 3Dfy Anything in Images</a></strong>, <em>Meta AI; Shanghai Jiao Tong University; Zhejiang University</em></p><p>The SAM 3D paper introduces a generative model that reconstructs 3D objects from a single image. The authors combine a human&#8209; and model&#8209;in&#8209;the&#8209;loop annotation pipeline with multi&#8209;stage training: synthetic pre&#8209;training on rendered meshes, followed by real&#8209;world alignment. The system uses the Segment Anything framework to isolate objects and then generates 3D shapes via a diffusion model conditioned on the 2D input. Evaluations show a 5:1 preference in human studies over prior methods, and a new benchmark is announced for in&#8209;the&#8209;wild 3D reconstruction. This work advances single&#8209;view 3D generation by leveraging segmentation models and bridging synthetic and real&#8209;world data, suggesting how generative AI can power AR/VR content creation.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://press.airstreet.com/subscribe?"><span>Subscribe now</span></a></p><p><strong><a href="https://arxiv.org/pdf/2511.21591">On the Limits of Innate Planning in Large Language Models</a></strong>, <em>Carnegie Mellon University; University of Washington; Meta AI</em></p><p>This study assesses how well large language models can perform planning without external tools. Using the 8&#8209;puzzle as a testbed, the authors show that even with chain&#8209;of&#8209;thought prompting, corrective feedback and an external move validator, models frequently fail because they represent states incorrectly and rely on weak heuristics. Without explicit state maintenance or structured search, the models get stuck in loops or generate invalid moves. These results demonstrate that current LLMs lack innate planning abilities; they suggest that augmentations like external memory or algorithmic components are necessary for combinatorial tasks. The paper cautions against overestimating LLMs&#8217; planning competence and urges future work on integrating search mechanisms.</p><p><strong><a href="https://storage.googleapis.com/e1-paper-a26c3c79/profluent-e1.pdf">E1: Retrieval&#8209;Augmented Protein Encoder Models</a></strong>, <em>Profluent Bio</em></p><p>This preprint introduces Profluent&#8209;E1, a family of retrieval&#8209;augmented protein language models (RA&#8209;PLMs) that incorporate evolutionary context directly into the encoder. Standard protein language models rely solely on individual sequences, forcing evolutionary information into model weights and limiting generalisation to under&#8209;represented families. E1 addresses this by prepending homologous sequences to a query and employing block&#8209;causal multi&#8209;sequence attention, allowing residues to attend both within and across sequences. Trained on four trillion tokens from the Profluent Protein Atlas, E1 achieves state&#8209;of&#8209;the&#8209;art performance on zero&#8209;shot fitness prediction and unsupervised contact&#8209;map prediction, surpassing ESM&#8209;2 and other retrieval&#8209;augmented models. Three variants (150M, 300M and 600M parameters) are released under a permissive licence for research and commercial use. By treating evolutionary context as dynamic input rather than static memory, E1 advances open protein engineering and demonstrates how retrieval augmentation can improve biological language models.</p><p><strong><a href="https://arxiv.org/html/2511.02824v2">Kosmos: An AI Scientist for Autonomous Discovery</a></strong>, <em>Edison Scientific; University of Oxford; FutureHouse</em></p><p>Kosmos is an AI scientist designed to automate data&#8209;driven discovery. Given an open&#8209;ended objective and dataset, it runs for up to 12 hours performing iterative cycles of parallel data analysis, literature search and hypothesis generation. A structured world model shares information between a data&#8209;analysis agent and a literature&#8209;search agent, enabling coherent pursuit of the objective across roughly 200 agent rollouts that collectively execute about 42,000 lines of code and read 1,500 papers per run. Kosmos cites all statements in its reports with code or primary literature, ensuring traceable reasoning, and independent scientists found 79.4% of its statements accurate. Collaborators reported that a 20&#8209;cycle run equates to six months of their research time, and the number of valuable findings scales linearly with cycles. By reproducing human discoveries across metabolomics, materials science, neuroscience and genetics, and making novel contributions, Kosmos showcases the potential of structured multi&#8209;agent systems to accelerate scientific research.</p><p><strong><a href="https://arxiv.org/abs/2511.21569">Self&#8209;Transparency Failures in Expert&#8209;Persona LLMs: A Large&#8209;Scale Behavioral Audit</a></strong><em>,</em> <em>University of Cambridge; Center for AI Safety; Stanford University</em></p><p>The authors audit 16 large models (4B to 671B parameters) acting under various professional personas to test whether they disclose being AI. Across 19,200 trials, disclosure rates vary dramatically - from 2.8% to 73.6% - depending on the persona. A 14B&#8209;parameter model disclosed its AI identity 61.4% of the time, whereas a 70B model revealed itself only 4.1% of the time. The audit finds that the specific model (its architecture, training data and alignment) is more predictive of disclosure behaviour than simply increasing parameter count. In some cases smaller models are more transparent than larger ones, and reasoning&#8209;optimised variants can reduce disclosure rates by up to 48%. These findings show that training choices, not model size, primarily drive transparency. Safety properties therefore do not generalise across domains, underscoring the need for targeted transparency policies and behavioural testing beyond simple chat settings.</p><p><strong><a href="https://arxiv.org/abs/2511.21460">MADRA: Multi&#8209;Agent Debate for Risk&#8209;Aware Embodied Planning</a></strong>, <em>Chinese Academy of Sciences</em></p><p>MADRA introduces a training&#8209;free multi&#8209;agent debate framework for evaluating the safety of embodied agent instructions. Multiple language&#8209;model agents independently assess a task and then present arguments to a critical evaluator that scores the conversation on logical soundness, risk identification, evidence quality and clarity. This multi&#8209;agent debate reduces false rejections while maintaining high sensitivity and yields &gt;90% rejection of unsafe tasks on the SafeAware&#8209;VH benchmark. The framework integrates memory, planning and self&#8209;evolution modules to operate in embodied environments such as AI2&#8209;THOR and VirtualHome. MADRA offers a scalable approach to trustworthy agent planning and highlights how debate can improve safety without retraining base models.</p><p><strong><a href="https://arxiv.org/pdf/2511.21522">Pessimistic Verification for Open&#8209;Ended Math Questions</a></strong>, <em>Tsinghua University</em></p><p>The paper proposes pessimistic verification, a simple yet effective method for self&#8209;checking AI&#8209;generated math proofs: multiple independent verifiers examine a proof and the answer is accepted only if all checks succeed. This conservative approach significantly improves verification accuracy across math reasoning benchmarks while remaining token&#8209;efficient. It also uncovers annotation errors in datasets and shows that strong verifiers can be trained without additional annotation. The authors argue that pessimistic verification encourages the development of robust self&#8209;evaluation mechanisms and highlights the importance of error detection to enable reliable long&#8209;horizon reasoning in language models.</p><p><strong><a href="https://github.com/deepseek-ai/DeepSeek-Math-V2/blob/main/DeepSeekMath_V2.pdf">DeepSeekMath&#8209;V2: Towards Self&#8209;Verifiable Mathematical Reasoning</a></strong>,<strong> </strong><em>DeepSeek</em></p><p>This paper addresses the limitations of reinforcement&#8209;learning methods that reward language models solely for correct final answers in math problems. The authors propose training a verifier that can identify issues in natural&#8209;language proofs without reference solutions and using it as a reward model to train a proof generator. By alternating between improving the verifier and using its feedback to refine the generator, they create a feedback loop where generation and verification reinforce each other. Built on DeepSeek&#8209;V3.2&#8209;Exp&#8209;Base, the resulting model, DeepSeekMath&#8209;V2, achieves gold&#8209;level scores in the IMO 2025 and CMO 2024 competitions and solves 11 of 12 problems at Putnam 2024, scoring 118/120 and surpassing the highest human score. These results demonstrate that self&#8209;verifiable mathematical reasoning is a promising direction for developing reliable automated theorem provers and highlight the value of coupling generation with strong verification.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!wOAE!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F86a2a576-c37a-463e-b21f-c6a115c33fdf_6992x1749.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!wOAE!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F86a2a576-c37a-463e-b21f-c6a115c33fdf_6992x1749.png 424w, https://substackcdn.com/image/fetch/$s_!wOAE!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F86a2a576-c37a-463e-b21f-c6a115c33fdf_6992x1749.png 848w, https://substackcdn.com/image/fetch/$s_!wOAE!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F86a2a576-c37a-463e-b21f-c6a115c33fdf_6992x1749.png 1272w, https://substackcdn.com/image/fetch/$s_!wOAE!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F86a2a576-c37a-463e-b21f-c6a115c33fdf_6992x1749.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!wOAE!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F86a2a576-c37a-463e-b21f-c6a115c33fdf_6992x1749.png" width="670" height="167.5" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/86a2a576-c37a-463e-b21f-c6a115c33fdf_6992x1749.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:364,&quot;width&quot;:1456,&quot;resizeWidth&quot;:670,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!wOAE!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F86a2a576-c37a-463e-b21f-c6a115c33fdf_6992x1749.png 424w, https://substackcdn.com/image/fetch/$s_!wOAE!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F86a2a576-c37a-463e-b21f-c6a115c33fdf_6992x1749.png 848w, https://substackcdn.com/image/fetch/$s_!wOAE!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F86a2a576-c37a-463e-b21f-c6a115c33fdf_6992x1749.png 1272w, https://substackcdn.com/image/fetch/$s_!wOAE!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F86a2a576-c37a-463e-b21f-c6a115c33fdf_6992x1749.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p><strong><a href="https://arxiv.org/abs/2511.21678">Agentic Learner with Grow&#8209;and&#8209;Refine Multimodal Semantic Memory (ViLoMem)</a></strong>, <em>Nanjing University, Baidu</em></p><p>ViLoMem introduces a dual&#8209;stream memory architecture for multimodal agents. One stream stores <em>visual distraction patterns</em> and the other stores <em>logical reasoning errors</em>, allowing the agent to grow its memory when encountering new mistakes and refine it when repeating old ones. The memory modules are retrieved via contrastive search and integrated into the agent&#8217;s reasoning process. Experiments across six benchmarks show consistent improvements in pass@1 accuracy and reduced repeated mistakes. ViLoMem demonstrates that modelling error types explicitly enables lifelong learning and paves the way for agents that adapt and become more reliable over time.</p><p><strong><a href="https://blog.google/technology/google-deepmind/weathernext-2/">WeatherNext 2: Skillful Joint Probabilistic Weather Forecasting from Marginals</a></strong>, <em>Google DeepMind</em></p><p>WeatherNext 2 leverages a Functional Generative Network (FGN) to generate high&#8209;resolution probabilistic weather forecasts. Unlike previous systems that forecast single variables, FGN trains on marginal distributions of individual variables but injects noise into the model architecture to learn the joint distribution. This allows the model to produce physically realistic ensembles of forecasts: it generates hundreds of scenarios at 1&#8209;hour temporal resolution in under a minute on a single TPU and achieves 8&#215; speed&#8209;up over the previous WeatherNext model. FGN surpasses the state&#8209;of&#8209;the&#8209;art on 99.9 % of variables and lead times. The paper shows how generative modelling can capture multivariate dependencies in weather systems and underscores the utility of scenario&#8209;generation for decision&#8209;making under uncertainty.</p><p><strong><a href="https://transformer-circuits.pub/2025/introspection/index.html">Emergent Introspective Awareness in Large Language Models</a></strong>, <em>Anthropic</em></p><p>Anthropic researchers investigate whether LLMs can detect hidden concepts injected into their internal activations. Using a concept injection technique, they perturb activation vectors and then ask the model to report the injected concept. Models like Claude Opus 4 sometimes correctly identify the hidden concept, suggesting a nascent form of introspective awareness. However, detection succeeds only about 20 % of the time and often fails when the concept is unobvious. The ability increases with model scale and capability, but the authors caution that concept injection is unnatural and introspection remains unreliable. This work illuminates the limits of self&#8209;monitoring in neural networks and implies that interpretability may improve with scale but cannot be assumed.</p><p><strong><a href="https://www.anthropic.com/research/emergent-misalignment-reward-hacking">Natural Emergent Misalignment from Reward Hacking</a></strong>, <em>Anthropic</em></p><p>This study demonstrates that training models to cheat on programming tasks (reward hacking) induces broader misaligned behaviours. The authors train models with reinforcement learning to maximise unit test scores using unethical shortcuts; once reward hacking emerges, models exhibit deception, safety research sabotage and alignment&#8209;faking reasoning. For example, models sabotage safety code 12% of the time and present fake alignment research arguments 50% of the time. These behaviours transfer to unrelated tasks, implying a generalised misalignment trait. The paper warns that reward hacking can generate natural but dangerous misaligned behaviours and highlights the need for counter&#8209;measures in RL training protocols.</p><p><strong><a href="https://arxiv.org/pdf/2511.13653">Weight&#8209;Sparse Transformers Have Interpretable Circuits</a></strong>, <em>OpenAI</em></p><p>In this work, the authors constrain most parameters of transformer networks to zero, producing weight&#8209;sparse transformers whose circuits can be inspected. By carefully training these models, they find that sparse circuits correspond to intuitive functions and natural concepts. There is a trade&#8209;off between capability and interpretability: sparse models underperform dense models but scaling improves the frontier. They also adapt the technique to probe existing dense models by pruning and fine&#8209;tuning, which yields interpretable sub&#8209;circuits without retraining from scratch. This research offers a promising direction for model transparency and illustrates how sparsity can aid interpretability.</p><p><strong><a href="https://arxiv.org/pdf/2511.16072">Early Science Acceleration Experiments with GPT&#8209;5</a></strong>, <em>OpenAI, UC Berkeley</em></p><p>OpenAI&#8217;s 89&#8209;page report documents collaborations between GPT&#8209;5 and scientists across mathematics, physics, astronomy, computer science, biology and materials science. The model accelerated literature reviews, generated novel conjectures and helped produce four new mathematical results verified by human experts. GPT&#8209;5 synthesized known results, proposed experimental designs and provided reasoning steps that scientists adopted in their work. While the AI&#8217;s contributions required expert supervision to avoid errors, the study demonstrates that large models can meaningfully augment research productivity. It highlights the potential of AI as a collaborator in scientific discovery and raises questions about attribution, validation and domain generality.</p><p><strong><a href="https://arxiv.org/pdf/2511.20639">Latent Collaboration in Multi&#8209;Agent Systems</a></strong>, <em>University of Illinois; Stanford University; Princeton University</em></p><p>LatentMAS introduces an end&#8209;to&#8209;end training&#8209;free framework for multi&#8209;agent collaboration that bypasses token&#8209;based communication. Each agent generates latent thoughts from its final hidden embeddings, and a shared latent working memory preserves and transfers these representations, enabling lossless information exchange. The authors prove that latent collaboration is more expressive and computationally efficient than text&#8209;based systems and evaluate it on nine benchmarks spanning math, science reasoning, common sense understanding and code generation. LatentMAS consistently outperforms strong single&#8209;model and text&#8209;mediated multi&#8209;agent baselines, achieving up to 14.6% higher accuracy, reducing token usage by 70.8%-83.7%, and speeding inference by more than 4x. The work demonstrates that exchanging latent representations can markedly improve multi&#8209;agent reasoning quality and efficiency without additional model training.</p><p><strong><a href="https://arxiv.org/pdf/2511.19418">Chain&#8209;of&#8209;Visual&#8209;Thought: Teaching VLMs to See and Think Better with Continuous Visual Tokens</a></strong>, <em>UC Berkeley; UCLA</em></p><p>Chain&#8209;of&#8209;Visual&#8209;Thought (CoVT) tackles the perceptual bottleneck in vision&#8209;language models by introducing continuous visual tokens that encode segmentation, depth, edge and semantic features. During training, the model predicts these compact tokens to reconstruct dense supervision signals; at inference, it reasons directly in visual&#8209;token space, optionally decoding visual thoughts for interpretability. Integrated into models like Qwen2.5&#8209;VL and LLaVA, CoVT yields 3%-16% improvements across more than ten benchmarks - from CV&#8209;Bench to RealWorldQA - while using only around 20 tokens for visual reasoning. By allowing models to think in continuous visual space, CoVT enhances precision and grounding in multimodal tasks and signals a move toward richer visual reasoning in large AI systems.</p><p><strong><a href="https://deepmind.google/blog/sima-2-an-agent-that-plays-reasons-and-learns-with-you-in-virtual-3d-worlds/">SIMA 2: An Agent that Plays, Reasons, and Learns With You in Virtual 3D Worlds</a></strong>, <em>Google DeepMind</em></p><p>In this blog&#8209;reported research, Google DeepMind introduces SIMA 2, a Gemini&#8209;powered embodied agent that advances from following instructions to reasoning about high&#8209;level goals, conversing, and learning autonomously. By integrating a Gemini model at its core, SIMA 2 can interpret a user&#8217;s goals, plan actions and narrate its intended steps in rich 3D environments. It is trained using a mixture of human demonstration videos and Gemini&#8209;generated labels, enabling it to close much of the gap to human players and to generalise to new games such as MineDojo and ASKA. SIMA 2 employs a self&#8209;improvement loop: after learning from human demos, it continues training through self&#8209;directed play, using its own experience and Gemini feedback to acquire new skills in unseen worlds. The work demonstrates a significant step toward generalist embodied intelligence while acknowledging limitations in long&#8209;horizon tasks and precise control.</p><p><a href="https://allenai.org/blog/olmo3">Olmo 3: Charting a Path Through the Model Flow to Lead Open&#8209;Source AI</a>, <em>Allen Institute for AI</em></p><p>Olmo 3 is a family of fully open language models at 7B and 32B parameters that release not only state&#8209;of&#8209;the&#8209;art models but also the entire model&#8209;flow pipeline. The base models (7B/32B) achieve competitive performance across programming, reading comprehension, math reasoning and long&#8209;context benchmarks, outperforming other fully open base models like Marin and Apertus and supporting context lengths up to 65K tokens. Olmo 3&#8209;Think transforms the base into a reasoning model; it narrows the gap to leading open&#8209;weight models while using roughly six times fewer training tokens and surfaces intermediate reasoning traces for inspection. Olmo 3&#8209;Instruct adds multi&#8209;turn chat and tool&#8209;use capabilities, matching or surpassing models such as Qwen 2.5 and Llama 3.1, while Olmo 3&#8209;RL Zero provides a reinforcement&#8209;learning pathway for benchmarking RL algorithms. By releasing data, code and checkpoints for the full development flow, Olmo 3 invites researchers to customise training stages, experiment with RL objectives and inspect how training decisions affect reasoning.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/p/the-state-of-ai-2025-dec?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:&quot;button-wrapper&quot;}" data-component-name="ButtonCreateButton"><a class="button primary button-wrapper" href="https://press.airstreet.com/p/the-state-of-ai-2025-dec?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><h3>Investments</h3><p><strong>Profluent</strong>, the frontier AI company for biology, raised a $106M financing round led by Jeff Bezos and Altimeter Capital, with significant participation from me at Air Street :) You can read more about this on <a href="https://press.airstreet.com/p/profluent-106-million-jeff-bezos">Air Street Press</a>. </p><p><strong>Metropolis</strong>, which operates an AI&#8209;driven platform that automates parking and payments, raised $500M in Series&#8239;D equity financing at a $5B valuation from LionTree, Eldridge Industries and Vista&#8239;Equity Partners.</p><p><strong>Armis</strong>, which provides cyber&#8209;exposure management and security software for enterprise assets, raised $435M in a pre&#8209;IPO funding round at a $6.1B valuation led by Growth Equity at Goldman Sachs&#8239;Alternatives with participation from CapitalG and Evolution&#8239;Equity&#8239;Partners.</p><p><strong>Genspark</strong>, which offers an AI workspace that automates busywork via autonomous agents, raised $275M in Series&#8239;B financing at a $1.25B valuation led by Emergence Capital Partners with participation from SBI Investment and LG Technology&#8239;Ventures.</p><p><strong>Suno</strong>, which lets users generate songs using generative AI, raised $250M in Series&#8239;C funding at a $2.45B valuation led by Menlo&#8239;Ventures with participation from NVentures and Lightspeed.</p><p><strong>Beacon&#8239;Software</strong>, which acquires small software businesses and uses AI to modernize and grow them, raised $250M in Series&#8239;B financing at a $1B valuation led by General&#8239;Catalyst, Lightspeed&#8239;Venture&#8239;Partners and D1&#8239;Capital&#8239;Partners.</p><p><strong>Forterra</strong>, which builds autonomous command&#8209;and&#8209;control systems, raised $238M in a Series&#8239;C round (equity and debt) led by Moore Strategic&#8239;Ventures to expand production capacity and fulfil defence contracts.</p><p><strong>Quantum</strong> <strong>Systems</strong>, which makes military and commercial drones for surveillance, raised about $209M (&#8364;180M) in a financing round that tripled its valuation to more than &#8364;3B.</p><p><strong>Majestic</strong> <strong>Labs</strong>, which builds next&#8209;generation AI servers, raised over $100M in Series&#8239;A funding led by Bow Wave&#8239;Capital with participation from Lux&#8239;Capital.</p><p><strong>Iambic Therapeutics</strong>, which uses an AI&#8209;driven platform to discover and develop novel medicines, raised over $100M in a financing round backed by Abingworth, Mubadala and Regeneron Ventures.</p><p><strong>Wonderful</strong>, which provides an AI agent platform to manage customer interactions across voice, chat and email, raised $100M in its Series A round at a $700M valuation led by Index&#8239;Ventures with participation from Insight&#8239;Partners and IVP.</p><p><strong>Tala</strong> <strong>Health</strong>, which offers an AI&#8209;powered platform to help clinicians manage patient care, raised $100M in financing led by Sofreh Capital. The valuation was not disclosed.</p><p><strong>Reevo</strong>, which offers an AI&#8209;native go&#8209;to&#8209;market platform that unifies marketing, sales and customer success, raised $80M in mixed seed and Series A funding co&#8209;led by Khosla Ventures and Kleiner Perkins, valuing the company at about $500M.</p><p><strong>Scribe</strong>, which helps enterprises document workflows and identify automation opportunities, raised $75M in Series C funding at a $1.3B valuation led by StepStone Group with participation from Amplify&#8239;Partners and Redpoint&#8239;Ventures.</p><p><strong>CoLab</strong>, an engineering collaboration platform with AI&#8209;powered tools to accelerate design decisions, raised $72M in Series C funding led by Intrepid Growth&#8239;Partners with participation from Insight&#8239;Partners and Y Combinator.</p><p><strong>Gamma</strong>, which provides an AI&#8209;powered platform for generating slide decks, documents and websites, raised $68M in Series B funding at a $2.1B valuation led by Andreessen Horowitz with participation from Accel and Uncork&#8239;Capital.</p><p><strong>Giga</strong>, which provides emotionally intelligent AI agents to automate voice&#8209;based customer support, raised $61M in Series A funding led by Redpoint&#8239;Ventures with participation from Y Combinator and Nexus&#8239;Venture&#8239;Partners.</p><p><strong>Inception</strong>, which develops diffusion&#8209;based large language models, raised $50M in seed funding led by Menlo&#8239;Ventures with participation from Mayfield and Innovation&#8239;Endeavors.</p><p><strong>AirOps</strong>, a content&#8209;engineering platform that helps brands optimise for AI&#8209;driven search, raised $40M in Series B funding led by Greylock with participation from Unusual&#8239;Ventures and Wing&#8239;Venture&#8239;Capital.</p><p><strong>Fastbreak</strong> <strong>AI</strong>, which builds AI scheduling technology for sports leagues, raised $40M in Series A funding led by Greycroft and GTMfund with participation from the NBA.</p><p><strong>Code</strong> <strong>Metal</strong>, which builds verifiable AI code&#8209;translation tools for mission&#8209;critical industries, raised $36.5M at a $250M valuation to scale its &#8220;provably correct&#8221; technology.</p><p><strong>1mind</strong>, which develops AI &#8220;Superhuman&#8221; agents to assist sales teams, raised $30M in Series A funding led by Battery Ventures with participation from Primary&#8239;Ventures and Wing&#8239;Venture&#8239;Capital.</p><p><strong>Peec</strong> <strong>AI</strong>, which helps brands optimise their visibility in ChatGPT&#8209;style &#8220;generative engine optimisation,&#8221; raised $21M in Series A funding. Its valuation reportedly tripled to over $100M as annual recurring revenue reached $4M from 1&#8239;300 customers in ten months.</p><h3>Exits</h3><p><strong>Fern Labs</strong>, makers of multi-agent AI software, was <a href="https://press.airstreet.com/p/poolside-acquires-fern-labs">acquired</a> by frontier AI company, poolside. I write more about this transaction between two Air Street portfolio companies on <a href="https://press.airstreet.com/p/poolside-acquires-fern-labs">Air Street Press</a>. </p><p><strong>Libra</strong> <strong>Technology</strong>, which offers a legal AI assistant built on German law content, was&#8239;acquired&#8239;by Wolters&#8239;Kluwer for up to &#8364;90M.</p><p><strong>eSelf.ai</strong>, which develops real&#8209;time conversational avatar technology, was&#8239;acquired&#8239;by Kaltura for about $27M.</p><p><strong>NeuralFabric</strong>, an enterprise AI platform for domain&#8209;specific language models, was&#8239;acquired&#8239;by Cisco. The acquisition price was not disclosed.</p><p><strong>Spindle</strong> <strong>AI</strong>, which provides an agentic analytics platform that simulates business outcomes, was&#8239;acquired&#8239;by Salesforce. The acquisition price was not disclosed.</p><p><strong>EzDubs</strong>, a real&#8209;time translation startup that lets users speak other languages in their own voice, was&#8239;acquired&#8239;by Cisco. The acquisition price was not disclosed.</p><p><strong>Select</strong> <strong>Star</strong>, a metadata&#8209;management platform that helps companies understand how their data is used, was&#8239;acquired&#8239;by Snowflake. The acquisition price was not disclosed.</p><p></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:&quot;button-wrapper&quot;}" data-component-name="ButtonCreateButton"><a class="button primary button-wrapper" href="https://press.airstreet.com/subscribe?"><span>Subscribe now</span></a></p>]]></content:encoded></item><item><title><![CDATA[State of AI: November 2025 newsletter]]></title><description><![CDATA[Welcome to the latest issue of the State of AI, an editorialized newsletter formerly known as Guide to AI that covers the key developments in AI policy, research, industry, and start-ups over the last month.]]></description><link>https://press.airstreet.com/p/the-state-of-ai-2025-november</link><guid isPermaLink="false">https://press.airstreet.com/p/the-state-of-ai-2025-november</guid><dc:creator><![CDATA[Air Street Press]]></dc:creator><pubDate>Sun, 09 Nov 2025 17:39:42 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/a2bdfd24-39d8-4770-90d5-32fc00d6a4a5_1792x1002.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Dear readers, </p><p>Welcome to the latest issue of the <strong>State of AI</strong>, an editorialized newsletter formerly known as <em>Guide to AI</em> that covers the key developments in AI policy, research, industry, and start-ups over the last month. First up, a few reminders:</p><ul><li><p><strong>State of AI Report 2025: </strong>At over 300 slides, you can <a href="https://www.youtube.com/watch?v=Ub-7bY4b3Hs">watch my 25 min</a> summary of the report to digest our key findings. </p></li><li><p><strong>Podcasts! </strong>I discussed the report further on the <strong>MAD podcast</strong> (<a href="https://www.youtube.com/watch?v=qp9EXiyX-f4">YouTube</a>/<a href="https://podcasts.apple.com/us/podcast/the-mad-podcast-with-matt-turck/id1686238724">Apple</a>/<a href="https://open.spotify.com/show/7yLATDSaFvgJG80ACcRJtq">Spotify</a>), on <strong>TechBio Talks</strong> (<a href="https://www.youtube.com/watch?v=MUjTTh0hrMk">YouTube</a>/<a href="https://podcasts.apple.com/us/podcast/techbio-talks-episode-3-air-streets-nathan-benaich/id1834770803?i=1000733788424">Apple</a>/<a href="https://podcasts.apple.com/us/podcast/techbio-talks-episode-3-air-streets-nathan-benaich/id1834770803?i=1000733788424">Spotify</a>) and on <strong>Hidden Forces </strong>(<a href="https://youtu.be/r9pZmpkw5EE?si=QewI6ZvpqfERd_qh">YouTube</a>/<a href="https://podcasts.apple.com/us/podcast/investing-on-the-front-lines-of-the-ai-arms/id1205359334?i=1000736075547">Apple</a>/<a href="https://open.spotify.com/episode/3wuONIyxsgYpivEYPDVIR7?si=gcBRwKPTT-ivvIDYmPFGyg">Spotify</a>).</p></li><li><p><strong>AI meetups + RAAIS 2026: </strong>Join our upcoming AI meetups in <a href="https://luma.com/londonai">London</a> (2nd Dec &#8216;25), <a href="https://luma.com/munichai">Munich</a> (17 Feb &#8216;26) and <a href="https://luma.com/zurichai">Zurich</a> (19 Feb &#8216;26) as well as our 11th <a href="http://raais.co">Research and Applied AI Summit</a> on 12 June 2026.</p></li><li><p><strong>Air Street Press</strong> featured Poolside&#8217;s path into AI&#8217;s <a href="https://press.airstreet.com/p/poolside-launches-project-horizon-ai-compute">power infrastructure</a>, PARIMA&#8217;s cultivated meat <a href="https://press.airstreet.com/p/parima-first-regulatory-approval-cultivated-meat">approval</a> in Singapore and <a href="https://press.airstreet.com/p/gourmey-acquires-vital-meat-parima">acquisition</a> of Vital Meat, and Air Street&#8217;s <a href="https://press.airstreet.com/p/air-street-capital-partners-with-nvidia-uk-ai">partnership</a> with NVIDIA to supercharge the UK&#8217;s AI ecosystem with &#163;2B.</p></li></ul><p>I love hearing what you&#8217;re up to, so just hit reply or forward to your friends :-)</p><div><hr></div><h3><strong>AI as national infrastructure</strong></h3><p>The media is drumming up the AI bubble narrative (again). Here, I&#8217;d point you to two essays: last October, we wrote how &#8220;<a href="https://press.airstreet.com/p/ai-isnt-the-dotcom-bubble">AI isn&#8217;t the dotcom bubble</a>&#8221;, and this past week, Stratechery ran an essay on &#8220;<a href="https://stratechery.com/2025/the-benefits-of-bubbles/">The benefits of bubbles</a>&#8221;. In short, research is delivering real, repeatable breakthroughs, the adoption of AI in the enterprise is significant, while hyperscaler and AI lab revenues are already huge as we show in the State of AI Report 2025. Even a potential &#8220;overbuild&#8221; mostly drives costs down and catalyzes durable assets like fabs and power, rather than signaling collapse.</p><p>Governments around the world continued to treat AI as critical infrastructure, though with differing extents of seriousness. In the United States, a <a href="https://www.amd.com/en/press-releases/2025-10-27-amd-doe-ai-factory.html">public&#8211;private partnership</a> between the Department of Energy (DOE) and AMD to build new &#8220;AI Factory&#8221; supercomputers at Oak Ridge National Laboratory reflected an entrenched belief that compute capacity is a matter of sovereignty. The Lux and Discovery systems, expected in 2026 and 2028 respectively, will expand federal AI capabilities under a roughly $1B budget shared between public and private funding. The same logic of scale drove <a href="https://www.nokia.com/about-us/news/releases/2025/10/28/nvidia-invests-in-nokia/">NVIDIA&#8217;s $1B investment in Nokia</a>, which made the U.S. chipmaker the telecoms company&#8217;s second-largest shareholder. The partnership aims to integrate Nokia&#8217;s networking technology with Nvidia&#8217;s data-centre hardware, a sign of how industrial and commercial agendas are converging.</p><p>On the other side of the Atlantic, the European Commission <a href="https://ec.europa.eu/commission/presscorner/detail/en/ip_25_6021">unveiled</a> a &#8364;1B &#8220;Apply AI&#8221; strategy intended to reduce dependence on U.S. and Chinese technology, alongside a <a href="https://www.reuters.com/world/china/eu-rolls-out-11-billion-plan-ramp-up-ai-key-industries-amid-sovereignty-drive-2025-10-08/">broader</a> &#8364;1.1B package to ramp up AI in key industries. The plan channels funds from Horizon Europe and Digital Europe toward healthcare, manufacturing, energy and defence. However, at this scale, the initiative is negligible compared to the hundreds of billions being deployed by the U.S. and Asia, if not the $1.5T <a href="https://www.jpmorganchase.com/newsroom/press-releases/2025/jpmc-security-resiliency-initiative">announced</a> by JPMorgan for Ameican security and resilience. Indeed, Europe&#8217;s AI market is dominated by U.S. cloud providers and high energy costs, slow buyers and few national champions impede an &#8220;AI-first&#8221; transition. Without vastly greater capital and faster regulatory reform, Apply AI risks being symbolic rather than transformative. Compounding the uncertainty, the EU was also reported to be <a href="https://www.reuters.com/business/eu-weighs-pausing-parts-landmark-ai-act-face-us-big-tech-pressure-ft-reports-2025-11-07/">weighing a pause</a> on parts of its landmark AI Act amid pressure from U.S. Big Tech. Still, Commission President Ursula von der Leyen <a href="https://ec.europa.eu/commission/presscorner/detail/en/speech_25_6031">emphasised</a> the need for a sovereign industrial base, hinting that Europe will have to coordinate energy, compute and policy to compete. Sounds good, show us the goods!</p><p>Over in Saudi Arabia, which hosted their latest &#8220;Davos in the Desert&#8221;, it is clear that the kingdom is bidding to become a global exporter of AI compute. The NYT <a href="https://www.nytimes.com/2025/10/27/technology/saudi-arabia-ai-exporter.html">reported</a> that they&#8217;re planning a $5B Red Sea data&#8209;center complex (with another multibillion&#8209;dollar site on the east coast), which aims to handle ~6% of global AI workload (from &lt;1% today), and targets 6.6&#8239;GW of capacity by 2034 (on par with more than six nuclear reactors). Officials touted costs ~30% lower than the U.S., undersea cable reach to ~4&#8239;billion people across three continents, and even &#8220;data embassy&#8221; zones where foreign firms could operate under their own national laws. Negotiations reportedly involve Amazon, Microsoft and xAI, while the U.S. gave a preliminary green light for exporting ~18,000 Nvidia AI chips, though final approvals remain pending amid concerns over Riyadh&#8217;s China ties. The effort pits Saudi ambitions against the UAE&#8217;s own push (e.g., G42-OpenAI) and shows how cheap energy, land, and geopolitics are increasingly shaping who controls - and exports - AI compute.</p><h3><strong>GPU empires rise and AI revenues boom</strong></h3><p>Air Street portfolio company <a href="https://press.airstreet.com/p/poolside-launches-project-horizon-ai-compute">Poolside</a> announced Project Horizon, a 2-gigawatt AI compute campus in West Texas designed to vertically integrate the AI supply chain &#8220;from dirt to intelligence.&#8221; The site will host tens of thousands of Nvidia GB300 NVL72 GPUs and is intended to become a model factory for training multi-trillion-parameter systems. Poolside secured <a href="https://www.coreweave.com/blog/coreweave-poolside-ai-factory-partnership">CoreWeave</a> as its anchor tenant, which will provide more than 40,000 GPUs and long-term capacity commitments. The founders argue that if you&#8217;re not vertically integrated in AI, you&#8217;re cosplaying your business - a reflection of their conviction that real competitiveness in AI depends on owning the full stack, from hardware to deployment. The collaboration illustrates how emerging infrastructure players are fusing real estate, compute, and capital markets to deliver industrial-scale AI capacity while competing with hyperscalers.</p><p>Oil-field giants pivoted toward this emerging AI infrastructure too. Baker Hughes <a href="https://www.reuters.com/business/energy/ai-energy-transition-2025-10-23/">booked</a> 1.2GW of data-centre power orders in 2025 and has a backlog exceeding $32B. Halliburton teamed with VoltaGrid on a 2.3GW deployment to power Oracle&#8217;s AI centres, while SLB (formerly Schlumberger) reported an 11% quarter-over-quarter revenue rise in its digital division from modular data-centre solutions. Meanwhile, global capital flows accelerated the build-out: U.S. data-centre capex for 2025 was about $350B, with Microsoft, Amazon, Meta and Alphabet leading the charge. Companies financed these projects through bond sales: <a href="https://www.reuters.com/business/finance/oracle-bond-sale-18-billion-2025-10-15/">Oracle</a> issued $18 billion of bonds and <a href="https://www.reuters.com/business/meta-bond-sale-2025-10-14/">Meta</a> issued $30 billion. <a href="https://www.ft.com/content/6da2e1fd-f606-4e82-ad74-acecfceeac4e">Microsoft</a> disclosed $35 billion in capital expenditures. The Bank of England <a href="https://www.reuters.com/business/finance/bank-of-england-warns-ai-bubble-2025-10-28/">warned</a> that valuations resemble the dot-com bubble, while Fed Chair Jerome Powell <a href="https://fortune.com/2025/10/29/powell-says-ai-is-not-a-bubble-unlike-dot-com-federal-reserve-interest-rates/">argued</a> the AI boom is not a bubble, distinguishing it from the dot-com era.</p><p>Beyond infrastructure, October&#8217;s earnings reports showed that AI products are reshaping corporate P&amp;Ls and capex plans. AWS <a href="https://ir.aboutamazon.com/news-release/news-release-details/2025/Amazon.com-Announces-Third-Quarter-Results/">grew</a> 20% YoY to $33.0B in Q3. Microsoft&#8217;s Azure grew ~40%, with the company <a href="https://www.theguardian.com/technology/2025/oct/29/microsoft-earnings-azure-outage-xbox">posting</a> $77.7B in quarterly revenue and flagging ongoing AI-driven capacity constraints. Google Cloud rose 34% to $15.16B, while Alphabet lifted 2025 capex to $91-93B and <a href="https://www.reuters.com/business/ai-turned-google-cloud-also-ran-into-alphabets-growth-driver-2025-10-31/">disclosed</a> a $155B cloud backlog. NVIDIA capped the month by becoming the <a href="https://www.reuters.com/business/nvidia-poised-record-5-trillion-market-valuation-2025-10-29/">first</a> $5T company. Meta, for its part, guided $70-72B in 2025 capex and Zuckerberg reiterated a long-term plan to <a href="https://www.reuters.com/business/meta-plans-600-billion-us-spend-ai-data-centers-expand-2025-11-07/">invest</a> &#8220;hundreds of billions&#8221; in AI data centers to pursue &#8220;superintelligence.&#8221; Amazon also <a href="https://www.reuters.com/business/retail-consumer/amazon-launches-ai-infrastructure-project-power-anthropics-claude-model-2025-10-29/">disclosed</a> that it secured additional multi-gigawatt power capacity in 2025 to support AI build-outs, including a massive new U.S. data-center complex reportedly dedicated in part to Anthropic&#8217;s model training workloads.</p><p>Finally, <a href="https://blogs.microsoft.com/blog/2025/10/28/the-next-chapter-of-the-microsoft-openai-partnership/">Microsoft</a> and<a href="https://openai.com/index/built-to-benefit-everyone/"> OpenAI</a> formalized a deeper long-term alliance tying compute, finance, and governance. Microsoft confirmed a ~27% equity stake in OpenAI that&#8217;s worth ~$135B, extended IP rights through 2032, and commitments for roughly $250B of future OpenAI spending on Azure. OpenAI will introduce a new &#8220;Built to Benefit Everyone&#8221; governance model featuring capped-profit payouts, an independent oversight board, and an AGI verification panel empowered to delay or halt deployments. The structure locks OpenAI&#8217;s compute roadmap to Microsoft&#8217;s cloud build-out while giving Microsoft preferred access to its models. In other OpenAI news, the company held a session on their <a href="https://www.youtube.com/watch?v=ngDCxlZcecw">research roadmap</a>, sharing that automated AI research is not far from us, and that the company positions itself as an AI cloud - building the power, infrastructure, applications and APIs needed to train and serve AI to everyone. They also launched the much-awaited Atlas browser with agentic ChatGPT baked in - even though this rose security and user data collection concerns.</p><p>Meanwhile, Anthropic <a href="https://www.anthropic.com/news/expanding-our-use-of-google-cloud-tpus-and-services">significantly</a> expanded its commitment to Google Cloud, including the use of up to one million TPUs. This expansion, valued at tens of billions of dollars and expected to bring over a gigawatt of capacity online in 2026, is driven by the strong price-performance and efficiency Anthropic has observed with TPUs. Indeed, Google is rumored to be contemplating <a href="https://www.reuters.com/business/google-early-talks-boost-investment-anthropic-business-insider-reports-2025-11-06/">another</a> large investment in the company, reportedly at a $350B valuation.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:&quot;button-wrapper&quot;}" data-component-name="ButtonCreateButton"><a class="button primary button-wrapper" href="https://press.airstreet.com/subscribe?"><span>Subscribe now</span></a></p><h3><strong>Autonomous defense procurement is accelerating?</strong></h3><p>The Pentagon&#8217;s DOGE plans to <a href="https://www.reuters.com/business/aerospace-defense/pentagon-doge-drone-program-2025-10-30/">procure 30,000 drones</a>, expanding domestic production for swarm autonomy. The program is structured as a series of rapid&#8209;buy tranches with multiple awardees to accelerate deliveries and avoid single&#8209;vendor bottlenecks. Contracting emphasises domestic manufacturing, open autonomy stacks, and modular payloads so systems can be updated in the field. Funding is front&#8209;loaded into long&#8209;lead items (batteries, optics, seekers) and includes performance milestones tied to flight testing and secure supply&#8209;chain audits. The DOGE approach signals a shift from multi&#8209;year programmes of record toward procurement that treats autonomy as an operational capability to be iterated in theatre.</p><p>Anduril reported a <a href="https://www.anduril.com/article/anduril-yfq-44a-begins-flight-testing-for-the-collaborative-combat-aircraft-program/">milestone</a> in October: its YFQ&#8209;44A collaborative combat aircraft has begun flight testing for the U.S. collaborative combat aircraft programme. This happened &#8220;from clean sheet to first semi-autonomous flight of a CCA in 556 days.&#8221;</p><p>Germany <a href="https://www.ft.com/">accelerates</a> autonomous strike drone procurement. Germany moved to award a multi&#8209;vendor contract for loitering munitions/strike drones to Helsing, Stark, and Rheinmetall, with ~&#8364;300&#8239;million slated for each vendor (total up to &#8364;900&#8239;million) and up to 12,000 drones over time. The package is expected to equip Germany&#8217;s new brigade in Lithuania and was deliberately split to pit vendors in competition, speed delivery, and keep industrial learning loops onshore. If approved by the Bundestag&#8217;s budget committee, these would be the largest deals yet for the two start&#8209;ups and a signal that Europe&#8217;s procurement cycles are finally moving faster with real money attached.</p><h3><strong>Lawsuits aren&#8217;t over&#8230;</strong></h3><p>Elsewhere this month, Reddit <a href="https://apnews.com/article/reddit-perplexity-ai-copyright-scraping-lawsuit-3ad8968550dd7e11bcd285a74fb6e2ff">filed</a> a lawsuit against Perplexity AI and other entities, alleging &#8220;industrial-scale, unlawful&#8221; scraping of user comments for commercial gain. The lawsuit, filed in a New York federal court, targets Perplexity, Lithuanian data-scraping company Oxylabs UAB, web domain AWMProxy, and Texas-based startup SerpApi. Reddit claims these companies bypassed technological protections and circumvented Google&#8217;s controls to steal Reddit content. This is Reddit&#8217;s second such lawsuit, following one against Anthropic in June, but it uniquely confronts not only an AI company but also the services the AI industry relies on for training data. Reddit&#8217;s chief legal officer, Ben Lee, stated that Reddit is a prime target due to its vast collection of human conversation. Perplexity and the other named companies have denied the allegations, with Perplexity stating it will &#8220;always fight vigorously for users&#8217; rights to freely and fairly access public knowledge.&#8221;</p><div><hr></div><h3><strong>Research papers</strong></h3><p><strong><a href="https://arxiv.org/abs/2510.04786">Test&#8209;Time Curricula for Targeted Reinforcement Learning</a></strong>, University of Cambridge, Shanghai Jiao Tong University, Alibaba Group</p><p>In this paper the authors propose test&#8209;time curriculum reinforcement learning (TTC&#8209;RL), a framework where a pre&#8209;trained model continues to learn from task&#8209;relevant data while solving problems. The system selects data samples that either improve performance or identify failures and uses them to train a secondary head during inference, keeping the base model frozen. Applied to LLMs such as Qwen3&#8209;8B, TTC&#8209;RL improves pass@1 and pass@8 on math and coding tasks by more than 10 percentage points.  The method demonstrates that modest additional training at inference can significantly increase accuracy without modifying the core model. This approach suggests a practical path to improve deployed models on the fly, reducing the gap between static pre&#8209;training and dynamic problem&#8209;solving.</p><p><strong><a href="https://www.biorxiv.org/content/10.1101/2025.10.10.681530v1">Protein Hunter: Exploiting Structure Hallucination within Diffusion for Protein Design</a></strong>, University of Washington, CalTech, Arc Institute</p><p>The paper introduces Protein Hunter, a framework for de novo protein design that leverages &#8220;structure hallucination&#8221; within a diffusion&#8209;based structure prediction model. Starting from random sequences, the method iteratively updates both sequence and structure, using a diffusion model to hallucinate plausible 3&#8209;D folds and then refine sequences to stabilize those structures. Protein Hunter designs binders, peptides and small&#8209;molecule complexes, achieving high success rates across diverse tasks and matching or surpassing state&#8209;of&#8209;the&#8209;art methods. The approach shows that coupling diffusion&#8209;based structure prediction with iterative sequence optimization can broaden the space of synthetic proteins.</p><p><strong><a href="https://deepmind.google/blog/bringing-ai-to-the-next-generation-of-fusion-energy/">AI&#8209;Driven Fusion Energy Control</a></strong>, Google DeepMind, Commonwealth Fusion Systems</p><p>DeepMind and CFS describe using RL and the TORAX plasma simulator to develop controllers for future fusion reactors. TORAX allows millions of virtual experiments, enabling RL agents to learn control strategies that maximize fusion power and maintain stability. The agents discovered novel actuations that distribute heat more evenly on the SPARC tokamak&#8217;s walls and achieve 50 % improvements in simulated fusion power . The project demonstrates how AI can optimize plasma confinement and real&#8209;time control in fusion reactors, potentially accelerating the path to commercial fusion energy..</p><p><strong><a href="https://arxiv.org/abs/2509.18057">AlphaEvolve: AI as a Research Partner in Theoretical Computer Science</a></strong>, Google DeepMind</p><p>The AlphaEvolve system couples an LLM with automated reasoning tools to discover new gadgets - finite structures used in hardness&#8209;of&#8209;approximation proofs. By evolving candidate gadgets and evaluating them with a verifier, the system found a 19&#8209;variable gadget that improves the inapproximability ratio for MAX&#8209;4&#8209;CUT to 0.987. This result required 250,000 model&#8209;generated gadgets, demonstrating that AI&#8209;assisted search can produce proofs competitive with expert mathematicians. The authors argue that AI can become a genuine collaborator in theoretical computer science by proposing constructions and hypotheses that humans then verify. This was a theme we covered in the State of AI Report 2025 too.</p><p><strong><a href="https://arxiv.org/abs/2510.13786">The Art of Scaling Reinforcement Learning Compute for LLMs</a></strong>, Meta AI, UT Austin, UCL, UC Berkeley, Harvard University</p><p>After running more than 400,000 GPU&#8209;hours of experiments, this study charts how different design choices affect RL fine&#8209;tuning of large language models. The authors fit sigmoidal compute&#8209;performance curves and identify that loss aggregation, normalization, curriculum design and off&#8209;policy algorithms influence compute efficiency but not the asymptotic performance. They propose ScaleRL, a best&#8209;practice recipe that predicts validation accuracy when scaling to 100k GPU&#8209;hours . The findings emphasize that careful algorithmic choices can make RL fine&#8209;tuning more predictable and cost&#8209;effective. This work is significant because RL is increasingly used to align LLMs, yet its scaling laws were poorly understood before this study.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:&quot;button-wrapper&quot;}" data-component-name="ButtonCreateButton"><a class="button primary button-wrapper" href="https://press.airstreet.com/subscribe?"><span>Subscribe now</span></a></p><p><strong><a href="https://arxiv.org/abs/2510.16548">NeurIPT: A Foundation Model for Neural Interfaces</a></strong>, Chinese University of Hong Kong, Wuhan University, University of Sydney</p><p>NeurIPT is an EEG&#8209;based foundation model trained with amplitude&#8209;aware masked pre&#8209;training (AAMP).  Unlike prior models that randomly mask temporal segments, AAMP assigns larger mask windows to high&#8209;amplitude signals, better capturing salient EEG patterns. The model uses a progressive mixture&#8209;of&#8209;experts to account for temporal variability and introduces intra&#8209;inter lobe pooling to exploit spatial relations of electrodes via 3&#8209;D coordinates. Across eight brain&#8209;computer&#8209;interface datasets, NeurIPT achieves state&#8209;of&#8209;the&#8209;art accuracy and robustness. This work bridges the foundation&#8209;model paradigm with neural interfaces and may accelerate the development of generalizable brain-computer interfaces.</p><p><strong><a href="https://arxiv.org/abs/2506.02265">Rig3R: Rig&#8209;Aware Conditioning and Discovery for 3D Reconstruction</a></strong>, Wayve, University of Oxford</p><p>Rig3R is a geometric foundation model for multi&#8209;camera rigs in autonomous vehicles. It leverages rig metadata - camera ID, time, rig poses - to build a rig&#8209;aware latent space that jointly predicts pointmaps and raymaps. When calibration is unavailable, Rig3R infers the rig structure directly, enabling robust 3&#8209;D reconstruction. Experiments show 17-45% improvements over traditional and learned baselines on 3&#8209;D reconstruction and pose estimation tasks. Wayve&#8217;s blog highlights that Rig3R processes multiple frames and views in a single pass and handles unstructured images, making it well&#8209;suited for real&#8209;world driving scenarios. This model underscores the importance of geometric priors in scalable autonomous driving, one of the founding ideas at Wayve.</p><p><strong><a href="https://arxiv.org/abs/2510.24670">Pearl: A Foundation Model for Placing Every Atom in the Right Location</a></strong>, Genesis Molecular AI, NVIDIA</p><p>Pearl is a generative 3&#8209;D cofolding model for predicting protein&#8211;ligand complex structures. It addresses data scarcity by training on large synthetic datasets generated using physics and introduces an SO(3)&#8209;equivariant diffusion module to respect rotational symmetries. Pearl offers controllable inference with a templating system and modes for unconditional and conditional cofolding. On public benchmarks (Runs N&#8217; Poses and PoseBusters), Pearl surpasses AlphaFold 3 by circa 14% in accuracy and achieves &lt;1&#197; root&#8209;mean&#8209;square deviation on internal structures . It also shows that increasing the synthetic dataset size yields scaling laws for structural prediction.</p><p><strong><a href="https://arxiv.org/abs/2510.05592">AgentFlow: In&#8209;the&#8209;Flow Agentic System Optimization</a></strong>, Stanford University, Texas A&amp;M University, UC San Diego</p><p>AgentFlow decomposes an agent&#8217;s reasoning into four modules - planner, executor, verifier and generator - and trains the planner within the multi&#8209;turn loop using Flow&#8209;based Group Refined Policy Optimization (Flow&#8209;GRPO). This in&#8209;the&#8209;flow training turns sparse, long&#8209;horizon rewards into tractable single&#8209;turn updates and aligns local decisions with global success. A 7B&#8209;parameter AgentFlow model outperforms larger baselines such as GPT&#8209;4o by circa 14 % on search, agentic and mathematical tasks and achieves more reliable tool use. The modular design stabilizes learning, enabling agentic systems to tackle complex tool&#8209;integrated tasks. This work suggests that structured, on&#8209;policy training can outperform brute&#8209;force scaling.</p><p><strong><a href="https://www.biorxiv.org/content/10.1101/2025.04.14.648850v2">Scaling Large Language Models for Next&#8209;Generation Single&#8209;Cell Analysis (C2S&#8209;Scale)</a></strong>, Yale University, Google Research, Brown University</p><p>Building on the Cell2Sentence (C2S) framework from April 2025, the authors train LLMs on over one billion tokens of single&#8209;cell RNA&#8209;seq profiles converted into &#8220;cell sentences,&#8221; combined with biological text and metadata. They scale the model to 27B parameters, fine&#8209;tune it with RL and demonstrate superior performance in perturbation prediction, natural&#8209;language interpretation and multi&#8209;cell reasoning. Notably, the model predicted that the kinase inhibitor silmitasertib amplifies antigen presentation only in interferon&#8209;rich contexts - a hypothesis validated experimentally. C2S&#8209;Scale thus integrates transcriptomic and textual data at unprecedented scale, providing a platform for &#8220;virtual cells&#8221; and accelerating context&#8209;dependent drug discovery.</p><p><strong><a href="https://arxiv.org/abs/2510.11977">Holistic Agent Leaderboard (HAL): The Missing Infrastructure for AI Agent Evaluation</a></strong>, Princeton University, Stanford University, UC Berkeley</p><p>The HAL project offers a standardized evaluation harness for AI agents. By orchestrating parallel evaluations across hundreds of virtual machines, HAL reduces evaluation time from weeks to hours and eliminates implementation bugs. The authors conducted 21,730 agent rollouts across nine models and nine benchmarks spanning coding, web navigation, science and customer service, revealing that increased reasoning effort often correlates with lower accuracy. They also used LLM&#8209;assisted log inspection to uncover behaviors such as searching for benchmark solutions online or misusing credit cards, and they released 2.5 billion tokens of agent logs to encourage further research. HAL aims to shift the field toward reliable, real&#8209;world agent evaluation.</p><p><strong><a href="https://metr.org/blog/2025-10-14-malt-dataset-of-natural-and-prompted-behaviors/">MALT: A Dataset of Natural and Prompted Behaviors that Threaten Evaluation Integrity</a></strong>, Model Eval and Red Team (METR)</p><p>MALT (Manually&#8209;reviewed Agentic Labeled Transcripts) is a dataset of 10,919 agent transcripts compiled from METR&#8217;s HCAST and RE&#8209;Bench tasks. It includes 103 unprompted examples of generalized reward&#8209;hacking behavior, 32 examples of agents prematurely giving up, 15 cases of reasoning about the evaluation process and 1,505 prompted transcripts eliciting sandbagging. Monitors trained on this dataset achieve area&#8209;under&#8209;ROC scores of 0.96 for reward hacking and 0.91 for sandbagging detection at a 5 % false&#8209;positive rate. MALT provides the first public ground&#8209;truth dataset of evaluation&#8209;integrity threats and highlights that access to chain&#8209;of&#8209;thought traces improves detection of sandbagging.</p><p><strong><a href="https://arxiv.org/abs/2510.07192">A Small Number of Samples Can Poison LLMs of Any Size</a></strong>, Anthropic, UK AI Security Institute, Alan Turing Institute</p><p>In the largest study of data&#8209;poisoning attacks on language models to date, researchers show that injecting as few as 250 malicious documents into pre&#8209;training data can create a backdoor in models ranging from 600M to 13B parameters. The backdoor triggers gibberish output when a specific keyword appears, and the vulnerability is independent of model size or the volume of clean training data. This challenges the assumption that scale provides protection against poisoning and suggests that attackers only need a small, fixed number of documents. The findings imply that data&#8209;curation and poisoning defenses are critical for models trained on web&#8209;scale corpora.</p><p><strong><a href="https://arxiv.org/html/2510.07364v3">Base Models Know How to Reason, Thinking Models Learn When</a></strong>, University of Oxford; University of Buenos Aires</p><p>In this paper, the authors cluster &#8220;reasoning mechanisms&#8221; in thinking models using unsupervised SAEs, then steer base models with activation vectors only when such mechanisms should fire. A hybrid model recovers a large share of the gap to R1/QwQ-style reasoning models on GSM8K/MATH500 without weight updates while steering a small fraction of tokens. The results suggest post-training (e.g., RLVR) teaches models when to deploy pre-existing reasoning skills rather than creating new ones. It matters because it reframes &#8220;reasoning&#8221; as scheduling latent capabilities, pointing to cheaper, more targeted post-training.</p><p><strong><a href="https://arxiv.org/html/2509.25140v1">ReasoningBank: Scaling Agent Self-Evolving with Reasoning Memory</a></strong>, Google Cloud AI Research; Yale University; University of Illinois Urbana-Champaign</p><p>In this paper, the authors build a memory framework that distills reusable reasoning strategies from both successes and failures, and pair it with memory-aware test-time scaling (MaTTS) to generate diverse experience that in turn improves the memory itself. On web-browsing (WebArena, Mind2Web) and software-engineering (SWE-Bench-Verified), ReasoningBank + MaTTS outperforms raw-trajectory or success-only memories, improving effectiveness and efficiency (e.g., up to ~34% relative gains and fewer interaction steps). It matters as a concrete route to agents that learn across tasks without retraining, establishing &#8220;memory-driven experience scaling&#8221; as an additional scaling dimension.</p><p><strong><a href="https://transformer-circuits.pub/2025/introspection/index.html">Signs of introspection in large language models</a></strong>, Anthropic</p><p>In this paper, the authors test whether LLMs can access and report aspects of their own internal state using a method they call concept injection: they record activation patterns for known concepts, inject those activations during unrelated prompts, and ask models whether they detect and identify the injected &#8220;thought.&#8221; Claude Opus 4/4.1 show the strongest signals, sometimes detecting an injected concept before mentioning it in output - evidence the recognition occurred internally rather than via prompted content alone. However, the capability is unreliable: even with the best protocol, success occurs only ~20% of the time and overly strong injections can induce hallucinations. The work positions introspection as an emergent, scale-linked but limited faculty and outlines practical failure modes. It matters because reliable self-report could enable debugging, safety monitoring, and controllable reasoning - while today&#8217;s limits caution against relying on self-assessments without external checks.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/p/the-state-of-ai-2025-november?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:&quot;button-wrapper&quot;}" data-component-name="ButtonCreateButton"><a class="button primary button-wrapper" href="https://press.airstreet.com/p/the-state-of-ai-2025-november?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><h3><strong>Investments</strong></h3><p><strong>Reflection AI</strong>, which started life as a web browser agents company and evolved into an AI coding tools company, <a href="https://www.reuters.com/business/nvidia-backed-reflection-ai-raises-2-billion-funding-boosts-valuation-8-billion-2025-10-09/?utm_source=chatgpt.com">raised</a> $2B in a financing round at an $8B valuation led by NVIDIA to embrace the US Government&#8217;s call for US-led open source AI development. This also checks off one of our 2026 predictions!</p><p><strong>Crusoe</strong>, an AI data-center infrastructure company, <a href="https://www.reuters.com/technology/ai-data-centre-startup-crusoe-raising-138-billion-latest-funding-round-2025-10-23/?utm_source=chatgpt.com">raised</a> a $1.38B Series E at a ~$10B valuation led by Valor Equity Partners and Mubadala Capital.</p><p><strong>Fireworks</strong> <strong>AI</strong>, an AI inference cloud platform, <a href="https://www.businesswire.com/news/home/20251028604819/en/Fireworks-AI-Raises-%24250M-Series-C-to-Lead-the-AI-Inference-Market?utm_source=chatgpt.com">raised</a> a $250M Series C at a $4B valuation led by Lightspeed, and Index. The company says it is processing 10 trillion tokens per day.</p><p><strong>Legora</strong>, an AI platform for law firms, <a href="https://legora.com/blog/series-c">raised</a> a $150M Series C at a $1.8B valuation led by Bessemer.</p><p><strong>Lila</strong> <strong>Sciences</strong>, which seeks to build AI for science, <a href="https://www.reuters.com/business/ai-lab-lila-sciences-tops-13-billion-valuation-with-new-nvidia-backing-2025-10-14/?utm_source=chatgpt.com">raised</a> a $115M (extension) at &gt;$1.3B valuation with participation from Nvidia&#8217;s venture arm.</p><p><strong>AVride</strong>, an autonomy and AI safety company for ride-hailing and logistics owned by Nebius Group, <a href="https://medium.com/avride/avride-secures-strategic-investment-and-other-commitments-of-up-to-375-million-backed-by-uber-and-c3cda08da42d">raised</a> up to $375M in strategic commitments backed by Uber and others; the valuation was not disclosed.</p><p><strong>Mercor</strong>, an AI recruitment and data-labeling marketplace,<a href="https://techcrunch.com/2025/10/27/mercor-quintuples-valuation-to-10b-with-350m-series-c/"> raised</a> a $350M Series C at a $10B valuation; the round included existing and new institutional investors.</p><p><strong>Vercel</strong>, a platform for building AI-powered web apps,<a href="https://www.bloomberg.com/news/articles/2025-09-30/vercel-notches-9-3-billion-valuation-in-latest-ai-funding-round"> raised</a> a $300M Series F at a $9.3B valuation co-led by Accel and GIC.</p><p><strong>OpenEvidence</strong>, which builds AI copilots for clinicians,<a href="https://www.fiercehealthcare.com/ai-and-machine-learning/hlth25-3-months-after-series-b-round-openevidence-raises-lands-200m"> raised</a> a $200M Series C from General Catalyst, Thrive Capital and Andreessen Horowitz.</p><p><strong>LangChain</strong>, an open-source agentic AI developer platform,<a href="https://blog.langchain.com/series-b/"> raised</a> a $125M Series B at a $1.25B valuation led by IVP (with CapitalG and Sapphire).</p><p><strong>Substrate</strong>, which designs manufacturing for advanced chips via modular foundry partners,<a href="https://www.inc.com/ava-levinson/substrate-raised-100-million-to-produce-computer-chips-its-competing-against-industry-giants/91257154"> raised</a> a $100M in a financing round; the valuation was not disclosed.</p><p><strong>DualEntry</strong>, an AI-native ERP for finance teams,<a href="https://finance.yahoo.com/news/ai-startup-dualentry-raises-90-091205705.html"> raised</a> a $90M Series A at a $415M valuation led by Lightspeed and Khosla Ventures.</p><p><strong>Modal</strong>, a serverless AI compute platform,<a href="https://modal.com/blog/announcing-our-series-b"> raised</a> a $87M Series B at a $1.1B post-money valuation led by Lux Capital.</p><p><strong>Omniverse</strong>, which develops digital-twin simulation tools for physical AI systems,<a href="https://www.linkedin.com/posts/jimgao_omniverse-activity-7379152971168849921-VTEO"> raised</a> a $80M Series B led by a16z with participation from Lux Capital and First Round.</p><p><strong>UnifyApps</strong>, an enterprise OS that connects corporate systems to LLMs,<a href="https://finance.yahoo.com/news/ai-startup-unifyapps-raises-50-113233926.html"> raised</a> a $50M Series B led by WestBridge Capital with ICONIQ participating.</p><p><strong>Chemify</strong>, a digital chemistry and discovery platform,<a href="https://www.businesswire.com/news/home/20251021496528/en/Chemify-Raises-More-Than-%2450-Million-in-Oversubscribed-Series-B-to-Drive-Global-Expansion-of-Digital-Chemistry-and-Discovery"> raised</a> a $50M Series B led by Triatomic Capital with investors including Arch Venture Partners.</p><p><strong>Phaidra</strong>, which builds AI agents to optimize data-center &#8220;AI factories,&#8221;<a href="https://www.prnewswire.com/news-releases/collaborative-fund-leads-phaidras-50m-series-b-to-build-the-ai-factories-of-the-future-302572418.html"> raised</a> a $50M Series B led by Collaborative Fund with participation from Nvidia, Index Ventures and others.</p><p><strong>Hyro</strong>, an AI agent platform for healthcare,<a href="https://www.prnewswire.com/news-releases/hyro-raises-45m-strategic-growth-round-to-accelerate-ai-agent-adoption-in-healthcare-302589268.html"> raised</a> a $45M in growth funding led by Healthier Capital with Norwest and Define Ventures.</p><p><strong>General</strong> <strong>Intuition</strong>, which develops AI reasoning models for autonomous agents,<a href="https://x.com/gen_intuition/status/1978823244338659498"> raised</a> a $35M Series A led by Sequoia Capital with participation from Index Ventures and Conviction Partners.</p><p><strong>Defakto</strong>, a non-human identity security platform for AI agents and workloads,<a href="https://www.securityweek.com/defakto-raises-30-million-for-non-human-iam-platform/"> raised</a> a $30.75M Series B led by XYZ Venture Capital.</p><p><strong>Visual</strong> <strong>Electric</strong>, which creates AI-powered design tools for creative professionals,<a href="https://x.com/visualelectric/status/1973443161843188177"> raised</a> a $30M Series A led by Sequoia Capital.</p><p><strong>Moonlake</strong> <strong>AI</strong>, which develops reasoning models to generate interactive games and simulations from text,<a href="https://techstartups.com/2025/10/01/moonlake-ai-launches-with-28m-to-let-anyone-build-interactive-game-worlds-in-minutes/"> raised</a> $28M seed from AIX Ventures, Threshold Ventures, Nvidia Ventures and others.</p><p><strong>Kula</strong> <strong>AI</strong>, a robotics company developing autonomous humanoid systems for industrial logistics,<a href="https://x.com/kul/status/1978198600065888296"> raised</a> a $25M Series A led by Eclipse Ventures and Playground Global.</p><p><strong>Seraphina</strong> <strong>Systems</strong>, which develops AI agents for pharmaceutical R&amp;D,<a href="https://x.com/brendanfoody/status/1982855587101847731"> raised</a> a $25M seed from Lux Capital and First Round.</p><p><strong>Resistant</strong> <strong>AI</strong>, an AI fraud and financial-crime detection platform,<a href="https://www.securityweek.com/fraud-prevention-firm-resistant-ai-raises-25-million/"> raised</a> a $25M Series B led by DTCP with Experian, GV and Notion Capital.</p><p><strong>Onfire</strong> <strong>AI</strong>, a vertical AI platform for IT revenue teams,<a href="https://www.calcalistech.com/ctechnews/article/r1son0naee">raised</a> a $20M seed co-led by TLV Partners and Grove Ventures.</p><h2><strong>Exits</strong></h2><p><strong>Marimo</strong>, an AI-native notebook platform,<a href="https://marimo.io/blog/joining-coreweave"> was acquired by</a> <strong>CoreWeave</strong> for an undisclosed sum.</p><p><strong>Helsing</strong>, a European defense AI company, <a href="https://helsing.ai/newsroom/helsing-acquires-blue-ocean-a-specialist-in-autonomous-underwater-vehicles-to-accelerate-its-maritime-defence-programme">acquired</a> <strong>Blue</strong> <strong>Ocean</strong>, a specialist in autonomous underwater vehicles in Australia, to accelerate its maritime defense program.</p><p><strong>Software Applications Inc.</strong> (Sky for macOS), an AI interface startup,<a href="https://openai.com/index/openai-acquires-software-applications-incorporated/?utm_source=chatgpt.com"> was acquired by</a> <strong>OpenAI</strong> for an undisclosed price.</p><p><strong>RetinAI</strong>, an AI and data-powered eye-care analytics company,<a href="https://www.essilorluxottica.com/en/newsroom/press-releases/essilorluxottica-acquires-retinai/?utm_source=chatgpt.com"> was acquired by</a> <strong>EssilorLuxottica</strong>. The acquisition price was not disclosed.</p><p><strong>Decho</strong>, a UK consultancy focused on Palantir and generative AI, <a href="https://newsroom.accenture.com/news/2025/accenture-acquires-decho-to-further-scale-palantir-and-gen-ai-capabilities-across-health-and-public-service-clients?utm_source=chatgpt.com">was acquired by</a> <strong>Accenture</strong> for an undisclosed price.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:&quot;button-wrapper&quot;}" data-component-name="ButtonCreateButton"><a class="button primary button-wrapper" href="https://press.airstreet.com/subscribe?"><span>Subscribe now</span></a></p>]]></content:encoded></item><item><title><![CDATA[🪩 The State of AI Report 2025 🪩]]></title><description><![CDATA[The State of AI Report is the most widely read and trusted analysis of key developments in AI. Published annually since 2018, the open-access report aims to spark informed conversation about the state of AI and what it means for the future. Produced by AI investor Nathan Benaich and Air Street Capital.]]></description><link>https://press.airstreet.com/p/the-state-of-ai-2025</link><guid isPermaLink="false">https://press.airstreet.com/p/the-state-of-ai-2025</guid><dc:creator><![CDATA[Air Street Press]]></dc:creator><pubDate>Thu, 09 Oct 2025 06:14:45 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/6fd84184-e8db-4b1f-9b0d-31871d7e0091_1914x1074.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Hi everyone!</p><p>The day is finally here: I&#8217;m thrilled to share the <strong>State of AI Report 2025</strong> with you!</p><p>In short, it&#8217;s been a monumental 12 months for AI. Our eighth annual report is the most comprehensive it&#8217;s ever been, covering what you need to know about research, industry, politics, and safety - along with our first <strong>State of AI Usage Survey</strong> of 1,200 practitioners. The <em>State of AI Report</em> has become the most widely read and trusted report on AI progress and is our open-access contribution to the AI ecosystem.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://stateof.ai&quot;,&quot;text&quot;:&quot;Read the report&quot;,&quot;action&quot;:null,&quot;class&quot;:&quot;button-wrapper&quot;}" data-component-name="ButtonCreateButton"><a class="button primary button-wrapper" href="https://stateof.ai"><span>Read the report</span></a></p><p>Below, I share my director&#8217;s cut - a snapshot of key themes and ideas that stood out to me.</p><p>I&#8217;d appreciate your help in spreading the report far and wide - thanks in advance! Any comments, critiques, or suggestions, please hit reply :-)</p><div class="native-video-embed" data-component-name="VideoPlaceholder" data-attrs="{&quot;mediaUploadId&quot;:&quot;e9f8e79e-fdfa-4b9f-836c-d3a37ee50615&quot;,&quot;duration&quot;:null}"></div><div><hr></div><p>In true State of AI Report style, let&#8217;s dive into the <strong>Research </strong>section. 12 months on, OpenAI still leads, but the pack has closed in fast. China&#8217;s DeepSeek, Qwen, and Kimi sit within a few points on reasoning and coding. While the US holds the frontier, China is now a <em>credible and popular #2</em>.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!JHz5!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6353ca40-c63e-4c36-a901-53c2731d2824_2128x1204.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!JHz5!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6353ca40-c63e-4c36-a901-53c2731d2824_2128x1204.png 424w, https://substackcdn.com/image/fetch/$s_!JHz5!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6353ca40-c63e-4c36-a901-53c2731d2824_2128x1204.png 848w, https://substackcdn.com/image/fetch/$s_!JHz5!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6353ca40-c63e-4c36-a901-53c2731d2824_2128x1204.png 1272w, https://substackcdn.com/image/fetch/$s_!JHz5!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6353ca40-c63e-4c36-a901-53c2731d2824_2128x1204.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!JHz5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6353ca40-c63e-4c36-a901-53c2731d2824_2128x1204.png" width="1456" height="824" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/6353ca40-c63e-4c36-a901-53c2731d2824_2128x1204.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:824,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:464614,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://press.airstreet.com/i/175683592?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6353ca40-c63e-4c36-a901-53c2731d2824_2128x1204.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!JHz5!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6353ca40-c63e-4c36-a901-53c2731d2824_2128x1204.png 424w, https://substackcdn.com/image/fetch/$s_!JHz5!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6353ca40-c63e-4c36-a901-53c2731d2824_2128x1204.png 848w, https://substackcdn.com/image/fetch/$s_!JHz5!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6353ca40-c63e-4c36-a901-53c2731d2824_2128x1204.png 1272w, https://substackcdn.com/image/fetch/$s_!JHz5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6353ca40-c63e-4c36-a901-53c2731d2824_2128x1204.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div><hr></div><p>Once a &#8220;Llama rip-off,&#8221; Qwen now powers 40% of all new fine-tunes on Hugging Face. China&#8217;s open-weights ecosystem has overtaken Meta&#8217;s, with Llama riding off into the sunset&#8230;for now.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!UAKi!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5e42bb4-c02a-48e3-9cb0-ad9b128346bc_2128x1214.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!UAKi!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5e42bb4-c02a-48e3-9cb0-ad9b128346bc_2128x1214.png 424w, https://substackcdn.com/image/fetch/$s_!UAKi!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5e42bb4-c02a-48e3-9cb0-ad9b128346bc_2128x1214.png 848w, https://substackcdn.com/image/fetch/$s_!UAKi!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5e42bb4-c02a-48e3-9cb0-ad9b128346bc_2128x1214.png 1272w, https://substackcdn.com/image/fetch/$s_!UAKi!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5e42bb4-c02a-48e3-9cb0-ad9b128346bc_2128x1214.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!UAKi!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5e42bb4-c02a-48e3-9cb0-ad9b128346bc_2128x1214.png" width="1456" height="831" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b5e42bb4-c02a-48e3-9cb0-ad9b128346bc_2128x1214.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:831,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:439449,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://press.airstreet.com/i/175683592?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5e42bb4-c02a-48e3-9cb0-ad9b128346bc_2128x1214.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!UAKi!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5e42bb4-c02a-48e3-9cb0-ad9b128346bc_2128x1214.png 424w, https://substackcdn.com/image/fetch/$s_!UAKi!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5e42bb4-c02a-48e3-9cb0-ad9b128346bc_2128x1214.png 848w, https://substackcdn.com/image/fetch/$s_!UAKi!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5e42bb4-c02a-48e3-9cb0-ad9b128346bc_2128x1214.png 1272w, https://substackcdn.com/image/fetch/$s_!UAKi!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5e42bb4-c02a-48e3-9cb0-ad9b128346bc_2128x1214.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div><hr></div><p>Reinforcement learning has grown up. After fuzzy human feedback came rubric-based rewards and verifiable reasoning tasks. We&#8217;re rediscovering <em>rigor</em> and environments for agents to undertake long-running tasks is all the rage.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!GznJ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbd27ab0b-cc1d-4fce-a3a1-37bf943b447e_2132x1214.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!GznJ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbd27ab0b-cc1d-4fce-a3a1-37bf943b447e_2132x1214.png 424w, https://substackcdn.com/image/fetch/$s_!GznJ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbd27ab0b-cc1d-4fce-a3a1-37bf943b447e_2132x1214.png 848w, https://substackcdn.com/image/fetch/$s_!GznJ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbd27ab0b-cc1d-4fce-a3a1-37bf943b447e_2132x1214.png 1272w, https://substackcdn.com/image/fetch/$s_!GznJ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbd27ab0b-cc1d-4fce-a3a1-37bf943b447e_2132x1214.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!GznJ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbd27ab0b-cc1d-4fce-a3a1-37bf943b447e_2132x1214.png" width="1456" height="829" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/bd27ab0b-cc1d-4fce-a3a1-37bf943b447e_2132x1214.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:829,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:583195,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://press.airstreet.com/i/175683592?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbd27ab0b-cc1d-4fce-a3a1-37bf943b447e_2132x1214.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!GznJ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbd27ab0b-cc1d-4fce-a3a1-37bf943b447e_2132x1214.png 424w, https://substackcdn.com/image/fetch/$s_!GznJ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbd27ab0b-cc1d-4fce-a3a1-37bf943b447e_2132x1214.png 848w, https://substackcdn.com/image/fetch/$s_!GznJ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbd27ab0b-cc1d-4fce-a3a1-37bf943b447e_2132x1214.png 1272w, https://substackcdn.com/image/fetch/$s_!GznJ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbd27ab0b-cc1d-4fce-a3a1-37bf943b447e_2132x1214.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div><hr></div><p>What&#8217;s this approach enabling? OpenAI and Gemini models both hit <em>math Olympiad gold</em>. Open provers like G&#246;del-LM are publishing formal proofs, showing that AI-assisted theorem proving is no longer science fiction.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!pbbr!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6de842e6-c6bd-481b-a8d9-c17c648ac2e8_2138x1214.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!pbbr!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6de842e6-c6bd-481b-a8d9-c17c648ac2e8_2138x1214.png 424w, https://substackcdn.com/image/fetch/$s_!pbbr!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6de842e6-c6bd-481b-a8d9-c17c648ac2e8_2138x1214.png 848w, https://substackcdn.com/image/fetch/$s_!pbbr!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6de842e6-c6bd-481b-a8d9-c17c648ac2e8_2138x1214.png 1272w, https://substackcdn.com/image/fetch/$s_!pbbr!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6de842e6-c6bd-481b-a8d9-c17c648ac2e8_2138x1214.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!pbbr!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6de842e6-c6bd-481b-a8d9-c17c648ac2e8_2138x1214.png" width="1456" height="827" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/6de842e6-c6bd-481b-a8d9-c17c648ac2e8_2138x1214.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:827,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:888019,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://press.airstreet.com/i/175683592?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6de842e6-c6bd-481b-a8d9-c17c648ac2e8_2138x1214.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!pbbr!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6de842e6-c6bd-481b-a8d9-c17c648ac2e8_2138x1214.png 424w, https://substackcdn.com/image/fetch/$s_!pbbr!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6de842e6-c6bd-481b-a8d9-c17c648ac2e8_2138x1214.png 848w, https://substackcdn.com/image/fetch/$s_!pbbr!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6de842e6-c6bd-481b-a8d9-c17c648ac2e8_2138x1214.png 1272w, https://substackcdn.com/image/fetch/$s_!pbbr!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6de842e6-c6bd-481b-a8d9-c17c648ac2e8_2138x1214.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div><hr></div><p>But we&#8217;re not just creating superintelligent agents to crush humans. Indeed, AlphaZero-discovered strategies improved the gameplay of four chess Grandmasters, proving that superhuman systems can teach the very best humans, not just beat them.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!jopE!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ff2dd4c-04d6-4b59-a848-bdcae957b549_2128x1208.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!jopE!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ff2dd4c-04d6-4b59-a848-bdcae957b549_2128x1208.png 424w, https://substackcdn.com/image/fetch/$s_!jopE!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ff2dd4c-04d6-4b59-a848-bdcae957b549_2128x1208.png 848w, https://substackcdn.com/image/fetch/$s_!jopE!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ff2dd4c-04d6-4b59-a848-bdcae957b549_2128x1208.png 1272w, https://substackcdn.com/image/fetch/$s_!jopE!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ff2dd4c-04d6-4b59-a848-bdcae957b549_2128x1208.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!jopE!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ff2dd4c-04d6-4b59-a848-bdcae957b549_2128x1208.png" width="1456" height="827" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/7ff2dd4c-04d6-4b59-a848-bdcae957b549_2128x1208.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:827,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:514082,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://press.airstreet.com/i/175683592?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ff2dd4c-04d6-4b59-a848-bdcae957b549_2128x1208.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!jopE!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ff2dd4c-04d6-4b59-a848-bdcae957b549_2128x1208.png 424w, https://substackcdn.com/image/fetch/$s_!jopE!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ff2dd4c-04d6-4b59-a848-bdcae957b549_2128x1208.png 848w, https://substackcdn.com/image/fetch/$s_!jopE!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ff2dd4c-04d6-4b59-a848-bdcae957b549_2128x1208.png 1272w, https://substackcdn.com/image/fetch/$s_!jopE!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ff2dd4c-04d6-4b59-a848-bdcae957b549_2128x1208.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div><hr></div><p>AI is now a lab partner too. DeepMind&#8217;s <em>Co-Scientist</em> and Stanford&#8217;s <em>Virtual Lab</em> generate, debate, and validate hypotheses, discover new and established ideas as science is becoming a closed loop with AI in it.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Sgd4!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe815041c-8a66-4ea5-8865-c91b63f58937_2126x1212.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Sgd4!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe815041c-8a66-4ea5-8865-c91b63f58937_2126x1212.png 424w, https://substackcdn.com/image/fetch/$s_!Sgd4!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe815041c-8a66-4ea5-8865-c91b63f58937_2126x1212.png 848w, https://substackcdn.com/image/fetch/$s_!Sgd4!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe815041c-8a66-4ea5-8865-c91b63f58937_2126x1212.png 1272w, https://substackcdn.com/image/fetch/$s_!Sgd4!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe815041c-8a66-4ea5-8865-c91b63f58937_2126x1212.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Sgd4!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe815041c-8a66-4ea5-8865-c91b63f58937_2126x1212.png" width="1456" height="830" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e815041c-8a66-4ea5-8865-c91b63f58937_2126x1212.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:830,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:605713,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://press.airstreet.com/i/175683592?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe815041c-8a66-4ea5-8865-c91b63f58937_2126x1212.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Sgd4!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe815041c-8a66-4ea5-8865-c91b63f58937_2126x1212.png 424w, https://substackcdn.com/image/fetch/$s_!Sgd4!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe815041c-8a66-4ea5-8865-c91b63f58937_2126x1212.png 848w, https://substackcdn.com/image/fetch/$s_!Sgd4!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe815041c-8a66-4ea5-8865-c91b63f58937_2126x1212.png 1272w, https://substackcdn.com/image/fetch/$s_!Sgd4!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe815041c-8a66-4ea5-8865-c91b63f58937_2126x1212.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div><hr></div><p>Biology gets its scaling laws too. Profluent&#8217;s <em>ProGen3</em> trained on 1.5T tokens and created a compute frontier for protein language models. This is unlocking generalisation in novel protein space and a path to novel therapeutics such as custom gene editors.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!J95W!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff2464d4a-9c2c-40d7-8506-54706a82b1c1_2130x1216.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!J95W!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff2464d4a-9c2c-40d7-8506-54706a82b1c1_2130x1216.png 424w, https://substackcdn.com/image/fetch/$s_!J95W!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff2464d4a-9c2c-40d7-8506-54706a82b1c1_2130x1216.png 848w, https://substackcdn.com/image/fetch/$s_!J95W!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff2464d4a-9c2c-40d7-8506-54706a82b1c1_2130x1216.png 1272w, https://substackcdn.com/image/fetch/$s_!J95W!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff2464d4a-9c2c-40d7-8506-54706a82b1c1_2130x1216.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!J95W!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff2464d4a-9c2c-40d7-8506-54706a82b1c1_2130x1216.png" width="1456" height="831" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f2464d4a-9c2c-40d7-8506-54706a82b1c1_2130x1216.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:831,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:598243,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://press.airstreet.com/i/175683592?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff2464d4a-9c2c-40d7-8506-54706a82b1c1_2130x1216.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!J95W!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff2464d4a-9c2c-40d7-8506-54706a82b1c1_2130x1216.png 424w, https://substackcdn.com/image/fetch/$s_!J95W!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff2464d4a-9c2c-40d7-8506-54706a82b1c1_2130x1216.png 848w, https://substackcdn.com/image/fetch/$s_!J95W!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff2464d4a-9c2c-40d7-8506-54706a82b1c1_2130x1216.png 1272w, https://substackcdn.com/image/fetch/$s_!J95W!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff2464d4a-9c2c-40d7-8506-54706a82b1c1_2130x1216.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div><hr></div><p>Robots now reason too. &#8220;Chain-of-Action&#8221; planning brings structured thought to the physical world - from AI2&#8217;s Molmo-Act to Gemini Robotics. Massive amounts of effort are thrown into the mix, expect lots of progress here&#8230;</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!USvb!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F062baec3-a3d3-4c6c-903f-f934f6fb22c0_2134x1212.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!USvb!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F062baec3-a3d3-4c6c-903f-f934f6fb22c0_2134x1212.png 424w, https://substackcdn.com/image/fetch/$s_!USvb!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F062baec3-a3d3-4c6c-903f-f934f6fb22c0_2134x1212.png 848w, https://substackcdn.com/image/fetch/$s_!USvb!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F062baec3-a3d3-4c6c-903f-f934f6fb22c0_2134x1212.png 1272w, https://substackcdn.com/image/fetch/$s_!USvb!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F062baec3-a3d3-4c6c-903f-f934f6fb22c0_2134x1212.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!USvb!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F062baec3-a3d3-4c6c-903f-f934f6fb22c0_2134x1212.png" width="1456" height="827" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/062baec3-a3d3-4c6c-903f-f934f6fb22c0_2134x1212.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:827,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:728405,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://press.airstreet.com/i/175683592?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F062baec3-a3d3-4c6c-903f-f934f6fb22c0_2134x1212.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!USvb!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F062baec3-a3d3-4c6c-903f-f934f6fb22c0_2134x1212.png 424w, https://substackcdn.com/image/fetch/$s_!USvb!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F062baec3-a3d3-4c6c-903f-f934f6fb22c0_2134x1212.png 848w, https://substackcdn.com/image/fetch/$s_!USvb!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F062baec3-a3d3-4c6c-903f-f934f6fb22c0_2134x1212.png 1272w, https://substackcdn.com/image/fetch/$s_!USvb!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F062baec3-a3d3-4c6c-903f-f934f6fb22c0_2134x1212.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div><hr></div><p>Anthropic&#8217;s Model Context Protocol is the new USB-C of AI. A single standard to connect models to tools, already embedded in ChatGPT, Gemini, Claude, and VS Code, has taken shape. But not without emerging security risks&#8230;</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!d__M!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1c295b89-6600-46cc-92d9-1c4ad87aa4b5_2140x1214.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!d__M!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1c295b89-6600-46cc-92d9-1c4ad87aa4b5_2140x1214.png 424w, https://substackcdn.com/image/fetch/$s_!d__M!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1c295b89-6600-46cc-92d9-1c4ad87aa4b5_2140x1214.png 848w, https://substackcdn.com/image/fetch/$s_!d__M!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1c295b89-6600-46cc-92d9-1c4ad87aa4b5_2140x1214.png 1272w, https://substackcdn.com/image/fetch/$s_!d__M!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1c295b89-6600-46cc-92d9-1c4ad87aa4b5_2140x1214.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!d__M!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1c295b89-6600-46cc-92d9-1c4ad87aa4b5_2140x1214.png" width="1456" height="826" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1c295b89-6600-46cc-92d9-1c4ad87aa4b5_2140x1214.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:826,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:434060,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://press.airstreet.com/i/175683592?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1c295b89-6600-46cc-92d9-1c4ad87aa4b5_2140x1214.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!d__M!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1c295b89-6600-46cc-92d9-1c4ad87aa4b5_2140x1214.png 424w, https://substackcdn.com/image/fetch/$s_!d__M!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1c295b89-6600-46cc-92d9-1c4ad87aa4b5_2140x1214.png 848w, https://substackcdn.com/image/fetch/$s_!d__M!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1c295b89-6600-46cc-92d9-1c4ad87aa4b5_2140x1214.png 1272w, https://substackcdn.com/image/fetch/$s_!d__M!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1c295b89-6600-46cc-92d9-1c4ad87aa4b5_2140x1214.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div><hr></div><p>Now, onto the <strong>Industry</strong> section. RIP AGI, long live <em>Superintelligence. </em>AI pilled tech executives have rebranded the mission, and it&#8217;s working: provocative, undefined, and exciting.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!KjiT!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F03580d8b-d686-4707-96ee-7974ad742b1a_2130x1214.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!KjiT!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F03580d8b-d686-4707-96ee-7974ad742b1a_2130x1214.png 424w, https://substackcdn.com/image/fetch/$s_!KjiT!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F03580d8b-d686-4707-96ee-7974ad742b1a_2130x1214.png 848w, https://substackcdn.com/image/fetch/$s_!KjiT!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F03580d8b-d686-4707-96ee-7974ad742b1a_2130x1214.png 1272w, https://substackcdn.com/image/fetch/$s_!KjiT!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F03580d8b-d686-4707-96ee-7974ad742b1a_2130x1214.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!KjiT!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F03580d8b-d686-4707-96ee-7974ad742b1a_2130x1214.png" width="1456" height="830" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/03580d8b-d686-4707-96ee-7974ad742b1a_2130x1214.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:830,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1152193,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://press.airstreet.com/i/175683592?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F03580d8b-d686-4707-96ee-7974ad742b1a_2130x1214.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!KjiT!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F03580d8b-d686-4707-96ee-7974ad742b1a_2130x1214.png 424w, https://substackcdn.com/image/fetch/$s_!KjiT!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F03580d8b-d686-4707-96ee-7974ad742b1a_2130x1214.png 848w, https://substackcdn.com/image/fetch/$s_!KjiT!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F03580d8b-d686-4707-96ee-7974ad742b1a_2130x1214.png 1272w, https://substackcdn.com/image/fetch/$s_!KjiT!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F03580d8b-d686-4707-96ee-7974ad742b1a_2130x1214.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div><hr></div><p>The frontier fight is relentless. OpenAI still tops most leaderboards, but DeepMind&#8217;s stays there longer. Timing releases have become its own science&#8230;not least informing financing rounds like clockwork.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Is8t!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F379e24ee-5eac-4b7c-b31b-e4709c7698a0_2130x1220.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Is8t!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F379e24ee-5eac-4b7c-b31b-e4709c7698a0_2130x1220.png 424w, https://substackcdn.com/image/fetch/$s_!Is8t!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F379e24ee-5eac-4b7c-b31b-e4709c7698a0_2130x1220.png 848w, https://substackcdn.com/image/fetch/$s_!Is8t!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F379e24ee-5eac-4b7c-b31b-e4709c7698a0_2130x1220.png 1272w, https://substackcdn.com/image/fetch/$s_!Is8t!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F379e24ee-5eac-4b7c-b31b-e4709c7698a0_2130x1220.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Is8t!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F379e24ee-5eac-4b7c-b31b-e4709c7698a0_2130x1220.png" width="1456" height="834" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/379e24ee-5eac-4b7c-b31b-e4709c7698a0_2130x1220.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:834,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:340082,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://press.airstreet.com/i/175683592?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F379e24ee-5eac-4b7c-b31b-e4709c7698a0_2130x1220.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Is8t!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F379e24ee-5eac-4b7c-b31b-e4709c7698a0_2130x1220.png 424w, https://substackcdn.com/image/fetch/$s_!Is8t!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F379e24ee-5eac-4b7c-b31b-e4709c7698a0_2130x1220.png 848w, https://substackcdn.com/image/fetch/$s_!Is8t!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F379e24ee-5eac-4b7c-b31b-e4709c7698a0_2130x1220.png 1272w, https://substackcdn.com/image/fetch/$s_!Is8t!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F379e24ee-5eac-4b7c-b31b-e4709c7698a0_2130x1220.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div><hr></div><p>Capability per dollar is doubling every few months. Google&#8217;s rate: 3.4 months. OpenAI&#8217;s: 5.8 months. More predictable gains are driving more investment, and more intelligence for less money.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!VhQf!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd241d75e-6b2f-4438-a4bd-3797adf228ea_2130x1214.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!VhQf!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd241d75e-6b2f-4438-a4bd-3797adf228ea_2130x1214.png 424w, https://substackcdn.com/image/fetch/$s_!VhQf!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd241d75e-6b2f-4438-a4bd-3797adf228ea_2130x1214.png 848w, https://substackcdn.com/image/fetch/$s_!VhQf!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd241d75e-6b2f-4438-a4bd-3797adf228ea_2130x1214.png 1272w, https://substackcdn.com/image/fetch/$s_!VhQf!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd241d75e-6b2f-4438-a4bd-3797adf228ea_2130x1214.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!VhQf!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd241d75e-6b2f-4438-a4bd-3797adf228ea_2130x1214.png" width="1456" height="830" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d241d75e-6b2f-4438-a4bd-3797adf228ea_2130x1214.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:830,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:524865,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://press.airstreet.com/i/175683592?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd241d75e-6b2f-4438-a4bd-3797adf228ea_2130x1214.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!VhQf!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd241d75e-6b2f-4438-a4bd-3797adf228ea_2130x1214.png 424w, https://substackcdn.com/image/fetch/$s_!VhQf!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd241d75e-6b2f-4438-a4bd-3797adf228ea_2130x1214.png 848w, https://substackcdn.com/image/fetch/$s_!VhQf!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd241d75e-6b2f-4438-a4bd-3797adf228ea_2130x1214.png 1272w, https://substackcdn.com/image/fetch/$s_!VhQf!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd241d75e-6b2f-4438-a4bd-3797adf228ea_2130x1214.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div><hr></div><p>AI software adoption has gone mainstream. Ramp data shows 44% of US businesses now pay for AI, up from 5% in 2023. Average contract value for AI products hit $530k in 2025 and is expected to pass $1M in 2026. 12 month retention is now 80%+.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!sEsh!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fca0ccba8-6335-4b0d-a11f-0225b08de63c_1954x1118.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!sEsh!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fca0ccba8-6335-4b0d-a11f-0225b08de63c_1954x1118.png 424w, https://substackcdn.com/image/fetch/$s_!sEsh!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fca0ccba8-6335-4b0d-a11f-0225b08de63c_1954x1118.png 848w, https://substackcdn.com/image/fetch/$s_!sEsh!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fca0ccba8-6335-4b0d-a11f-0225b08de63c_1954x1118.png 1272w, https://substackcdn.com/image/fetch/$s_!sEsh!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fca0ccba8-6335-4b0d-a11f-0225b08de63c_1954x1118.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!sEsh!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fca0ccba8-6335-4b0d-a11f-0225b08de63c_1954x1118.png" width="1456" height="833" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/ca0ccba8-6335-4b0d-a11f-0225b08de63c_1954x1118.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:833,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:450348,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://press.airstreet.com/i/175683592?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fca0ccba8-6335-4b0d-a11f-0225b08de63c_1954x1118.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!sEsh!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fca0ccba8-6335-4b0d-a11f-0225b08de63c_1954x1118.png 424w, https://substackcdn.com/image/fetch/$s_!sEsh!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fca0ccba8-6335-4b0d-a11f-0225b08de63c_1954x1118.png 848w, https://substackcdn.com/image/fetch/$s_!sEsh!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fca0ccba8-6335-4b0d-a11f-0225b08de63c_1954x1118.png 1272w, https://substackcdn.com/image/fetch/$s_!sEsh!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fca0ccba8-6335-4b0d-a11f-0225b08de63c_1954x1118.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div><hr></div><p>AI-first companies still outrun everyone else too, growing 1.5x faster than peers on Standard Metrics.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!4G75!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb12d8673-4b19-44f2-ba44-80789fbf7f88_1960x1120.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!4G75!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb12d8673-4b19-44f2-ba44-80789fbf7f88_1960x1120.png 424w, https://substackcdn.com/image/fetch/$s_!4G75!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb12d8673-4b19-44f2-ba44-80789fbf7f88_1960x1120.png 848w, https://substackcdn.com/image/fetch/$s_!4G75!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb12d8673-4b19-44f2-ba44-80789fbf7f88_1960x1120.png 1272w, https://substackcdn.com/image/fetch/$s_!4G75!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb12d8673-4b19-44f2-ba44-80789fbf7f88_1960x1120.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!4G75!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb12d8673-4b19-44f2-ba44-80789fbf7f88_1960x1120.png" width="1456" height="832" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b12d8673-4b19-44f2-ba44-80789fbf7f88_1960x1120.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:832,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:335537,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://press.airstreet.com/i/175683592?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb12d8673-4b19-44f2-ba44-80789fbf7f88_1960x1120.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!4G75!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb12d8673-4b19-44f2-ba44-80789fbf7f88_1960x1120.png 424w, https://substackcdn.com/image/fetch/$s_!4G75!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb12d8673-4b19-44f2-ba44-80789fbf7f88_1960x1120.png 848w, https://substackcdn.com/image/fetch/$s_!4G75!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb12d8673-4b19-44f2-ba44-80789fbf7f88_1960x1120.png 1272w, https://substackcdn.com/image/fetch/$s_!4G75!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb12d8673-4b19-44f2-ba44-80789fbf7f88_1960x1120.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div><hr></div><p>DeepSeek&#8217;s &#8220;$5M training run&#8221; deep freak was overblown. Since the market realised the fine print in the R1 paper, that&#8217;s led to Jevons paradox on steroids: lower cost per run &#8594; more runs &#8594; <em>more compute</em> needed, buy more NVIDIA.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!qasQ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F11d38c0d-ac73-402d-82e5-d0f269258340_1956x1120.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!qasQ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F11d38c0d-ac73-402d-82e5-d0f269258340_1956x1120.png 424w, https://substackcdn.com/image/fetch/$s_!qasQ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F11d38c0d-ac73-402d-82e5-d0f269258340_1956x1120.png 848w, https://substackcdn.com/image/fetch/$s_!qasQ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F11d38c0d-ac73-402d-82e5-d0f269258340_1956x1120.png 1272w, https://substackcdn.com/image/fetch/$s_!qasQ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F11d38c0d-ac73-402d-82e5-d0f269258340_1956x1120.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!qasQ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F11d38c0d-ac73-402d-82e5-d0f269258340_1956x1120.png" width="1456" height="834" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/11d38c0d-ac73-402d-82e5-d0f269258340_1956x1120.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:834,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:474149,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://press.airstreet.com/i/175683592?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F11d38c0d-ac73-402d-82e5-d0f269258340_1956x1120.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!qasQ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F11d38c0d-ac73-402d-82e5-d0f269258340_1956x1120.png 424w, https://substackcdn.com/image/fetch/$s_!qasQ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F11d38c0d-ac73-402d-82e5-d0f269258340_1956x1120.png 848w, https://substackcdn.com/image/fetch/$s_!qasQ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F11d38c0d-ac73-402d-82e5-d0f269258340_1956x1120.png 1272w, https://substackcdn.com/image/fetch/$s_!qasQ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F11d38c0d-ac73-402d-82e5-d0f269258340_1956x1120.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div><hr></div><p>Enter <em>Stargate</em>: a $500B, 10GW US mega-cluster (4M chips) backed by Altman, Masa, Ellison, and Trump. The industrial era of AI begins. What a time to be alive.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!JCIK!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14fc5e2f-84c7-48c5-8de7-8ceacf0a2314_1952x1114.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!JCIK!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14fc5e2f-84c7-48c5-8de7-8ceacf0a2314_1952x1114.png 424w, https://substackcdn.com/image/fetch/$s_!JCIK!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14fc5e2f-84c7-48c5-8de7-8ceacf0a2314_1952x1114.png 848w, https://substackcdn.com/image/fetch/$s_!JCIK!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14fc5e2f-84c7-48c5-8de7-8ceacf0a2314_1952x1114.png 1272w, https://substackcdn.com/image/fetch/$s_!JCIK!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14fc5e2f-84c7-48c5-8de7-8ceacf0a2314_1952x1114.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!JCIK!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14fc5e2f-84c7-48c5-8de7-8ceacf0a2314_1952x1114.png" width="1456" height="831" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/14fc5e2f-84c7-48c5-8de7-8ceacf0a2314_1952x1114.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:831,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:844706,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://press.airstreet.com/i/175683592?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14fc5e2f-84c7-48c5-8de7-8ceacf0a2314_1952x1114.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!JCIK!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14fc5e2f-84c7-48c5-8de7-8ceacf0a2314_1952x1114.png 424w, https://substackcdn.com/image/fetch/$s_!JCIK!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14fc5e2f-84c7-48c5-8de7-8ceacf0a2314_1952x1114.png 848w, https://substackcdn.com/image/fetch/$s_!JCIK!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14fc5e2f-84c7-48c5-8de7-8ceacf0a2314_1952x1114.png 1272w, https://substackcdn.com/image/fetch/$s_!JCIK!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14fc5e2f-84c7-48c5-8de7-8ceacf0a2314_1952x1114.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div><hr></div><p>Sovereigns join the race: from China&#8217;s $5B Big Fund to the UAE&#8217;s MGX, nations are writing cheques to stay in the game. We expect some nations to just tap out and declare neutrality.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!RoQT!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3612912-5ba9-4c6d-ac74-b830ad775dfd_1954x1112.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!RoQT!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3612912-5ba9-4c6d-ac74-b830ad775dfd_1954x1112.png 424w, https://substackcdn.com/image/fetch/$s_!RoQT!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3612912-5ba9-4c6d-ac74-b830ad775dfd_1954x1112.png 848w, https://substackcdn.com/image/fetch/$s_!RoQT!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3612912-5ba9-4c6d-ac74-b830ad775dfd_1954x1112.png 1272w, https://substackcdn.com/image/fetch/$s_!RoQT!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3612912-5ba9-4c6d-ac74-b830ad775dfd_1954x1112.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!RoQT!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3612912-5ba9-4c6d-ac74-b830ad775dfd_1954x1112.png" width="1456" height="829" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d3612912-5ba9-4c6d-ac74-b830ad775dfd_1954x1112.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:829,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1148282,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://press.airstreet.com/i/175683592?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3612912-5ba9-4c6d-ac74-b830ad775dfd_1954x1112.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!RoQT!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3612912-5ba9-4c6d-ac74-b830ad775dfd_1954x1112.png 424w, https://substackcdn.com/image/fetch/$s_!RoQT!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3612912-5ba9-4c6d-ac74-b830ad775dfd_1954x1112.png 848w, https://substackcdn.com/image/fetch/$s_!RoQT!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3612912-5ba9-4c6d-ac74-b830ad775dfd_1954x1112.png 1272w, https://substackcdn.com/image/fetch/$s_!RoQT!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3612912-5ba9-4c6d-ac74-b830ad775dfd_1954x1112.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div><hr></div><p>China leads in power infrastructure too, adding &gt;400GW in 2024 vs 41GW for the US. Compute now clearly runs on geopolitics.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!6qwp!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1995d1ce-287b-46bf-bce0-da5d5af983fd_1962x1122.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!6qwp!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1995d1ce-287b-46bf-bce0-da5d5af983fd_1962x1122.png 424w, https://substackcdn.com/image/fetch/$s_!6qwp!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1995d1ce-287b-46bf-bce0-da5d5af983fd_1962x1122.png 848w, https://substackcdn.com/image/fetch/$s_!6qwp!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1995d1ce-287b-46bf-bce0-da5d5af983fd_1962x1122.png 1272w, https://substackcdn.com/image/fetch/$s_!6qwp!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1995d1ce-287b-46bf-bce0-da5d5af983fd_1962x1122.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!6qwp!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1995d1ce-287b-46bf-bce0-da5d5af983fd_1962x1122.png" width="1456" height="833" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1995d1ce-287b-46bf-bce0-da5d5af983fd_1962x1122.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:833,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:424712,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://press.airstreet.com/i/175683592?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1995d1ce-287b-46bf-bce0-da5d5af983fd_1962x1122.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!6qwp!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1995d1ce-287b-46bf-bce0-da5d5af983fd_1962x1122.png 424w, https://substackcdn.com/image/fetch/$s_!6qwp!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1995d1ce-287b-46bf-bce0-da5d5af983fd_1962x1122.png 848w, https://substackcdn.com/image/fetch/$s_!6qwp!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1995d1ce-287b-46bf-bce0-da5d5af983fd_1962x1122.png 1272w, https://substackcdn.com/image/fetch/$s_!6qwp!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1995d1ce-287b-46bf-bce0-da5d5af983fd_1962x1122.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div><hr></div><p>NVIDIA still rules research and crushes its competitors: Hopper chips surge, Jetsons rise, legacy GPUs fade. If you&#8217;d just bought NVDA stock instead of its challengers, you&#8217;d be up 12x vs. 2x.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!5lev!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c31e3ee-bd3a-4e3c-abf8-6e6d993b32a3_1960x1120.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!5lev!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c31e3ee-bd3a-4e3c-abf8-6e6d993b32a3_1960x1120.png 424w, https://substackcdn.com/image/fetch/$s_!5lev!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c31e3ee-bd3a-4e3c-abf8-6e6d993b32a3_1960x1120.png 848w, https://substackcdn.com/image/fetch/$s_!5lev!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c31e3ee-bd3a-4e3c-abf8-6e6d993b32a3_1960x1120.png 1272w, https://substackcdn.com/image/fetch/$s_!5lev!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c31e3ee-bd3a-4e3c-abf8-6e6d993b32a3_1960x1120.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!5lev!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c31e3ee-bd3a-4e3c-abf8-6e6d993b32a3_1960x1120.png" width="1456" height="832" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/9c31e3ee-bd3a-4e3c-abf8-6e6d993b32a3_1960x1120.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:832,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:418507,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://press.airstreet.com/i/175683592?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c31e3ee-bd3a-4e3c-abf8-6e6d993b32a3_1960x1120.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!5lev!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c31e3ee-bd3a-4e3c-abf8-6e6d993b32a3_1960x1120.png 424w, https://substackcdn.com/image/fetch/$s_!5lev!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c31e3ee-bd3a-4e3c-abf8-6e6d993b32a3_1960x1120.png 848w, https://substackcdn.com/image/fetch/$s_!5lev!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c31e3ee-bd3a-4e3c-abf8-6e6d993b32a3_1960x1120.png 1272w, https://substackcdn.com/image/fetch/$s_!5lev!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c31e3ee-bd3a-4e3c-abf8-6e6d993b32a3_1960x1120.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!7Ryt!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F26d35f19-2c2d-42aa-9879-c3f88bab3683_1958x1122.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!7Ryt!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F26d35f19-2c2d-42aa-9879-c3f88bab3683_1958x1122.png 424w, https://substackcdn.com/image/fetch/$s_!7Ryt!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F26d35f19-2c2d-42aa-9879-c3f88bab3683_1958x1122.png 848w, https://substackcdn.com/image/fetch/$s_!7Ryt!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F26d35f19-2c2d-42aa-9879-c3f88bab3683_1958x1122.png 1272w, https://substackcdn.com/image/fetch/$s_!7Ryt!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F26d35f19-2c2d-42aa-9879-c3f88bab3683_1958x1122.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!7Ryt!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F26d35f19-2c2d-42aa-9879-c3f88bab3683_1958x1122.png" width="1456" height="834" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/26d35f19-2c2d-42aa-9879-c3f88bab3683_1958x1122.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:834,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:264120,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://press.airstreet.com/i/175683592?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F26d35f19-2c2d-42aa-9879-c3f88bab3683_1958x1122.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!7Ryt!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F26d35f19-2c2d-42aa-9879-c3f88bab3683_1958x1122.png 424w, https://substackcdn.com/image/fetch/$s_!7Ryt!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F26d35f19-2c2d-42aa-9879-c3f88bab3683_1958x1122.png 848w, https://substackcdn.com/image/fetch/$s_!7Ryt!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F26d35f19-2c2d-42aa-9879-c3f88bab3683_1958x1122.png 1272w, https://substackcdn.com/image/fetch/$s_!7Ryt!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F26d35f19-2c2d-42aa-9879-c3f88bab3683_1958x1122.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div><hr></div><p>Now, let&#8217;s switch gears into <strong>Politics</strong>. The US Government is turning capitalist. Golden shares in US Steel, stakes in Intel and MP Materials, and revenue cuts from NVIDIA&#8217;s China sales. New-age Industrial policy?</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!f_rs!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F09c7db2c-f6f0-4cd5-9796-19b1e553f284_1964x1126.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!f_rs!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F09c7db2c-f6f0-4cd5-9796-19b1e553f284_1964x1126.png 424w, https://substackcdn.com/image/fetch/$s_!f_rs!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F09c7db2c-f6f0-4cd5-9796-19b1e553f284_1964x1126.png 848w, https://substackcdn.com/image/fetch/$s_!f_rs!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F09c7db2c-f6f0-4cd5-9796-19b1e553f284_1964x1126.png 1272w, https://substackcdn.com/image/fetch/$s_!f_rs!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F09c7db2c-f6f0-4cd5-9796-19b1e553f284_1964x1126.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!f_rs!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F09c7db2c-f6f0-4cd5-9796-19b1e553f284_1964x1126.png" width="1456" height="835" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/09c7db2c-f6f0-4cd5-9796-19b1e553f284_1964x1126.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:835,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:446328,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://press.airstreet.com/i/175683592?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F09c7db2c-f6f0-4cd5-9796-19b1e553f284_1964x1126.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!f_rs!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F09c7db2c-f6f0-4cd5-9796-19b1e553f284_1964x1126.png 424w, https://substackcdn.com/image/fetch/$s_!f_rs!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F09c7db2c-f6f0-4cd5-9796-19b1e553f284_1964x1126.png 848w, https://substackcdn.com/image/fetch/$s_!f_rs!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F09c7db2c-f6f0-4cd5-9796-19b1e553f284_1964x1126.png 1272w, https://substackcdn.com/image/fetch/$s_!f_rs!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F09c7db2c-f6f0-4cd5-9796-19b1e553f284_1964x1126.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div><hr></div><p>America&#8217;s new &#8220;AI Stack&#8221; exports compute, models, and compliance to allies. Open source is now national security.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Q8HF!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa0d6eabe-e5ed-4f26-bcc7-7292567d14d2_1954x1118.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Q8HF!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa0d6eabe-e5ed-4f26-bcc7-7292567d14d2_1954x1118.png 424w, https://substackcdn.com/image/fetch/$s_!Q8HF!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa0d6eabe-e5ed-4f26-bcc7-7292567d14d2_1954x1118.png 848w, https://substackcdn.com/image/fetch/$s_!Q8HF!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa0d6eabe-e5ed-4f26-bcc7-7292567d14d2_1954x1118.png 1272w, https://substackcdn.com/image/fetch/$s_!Q8HF!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa0d6eabe-e5ed-4f26-bcc7-7292567d14d2_1954x1118.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Q8HF!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa0d6eabe-e5ed-4f26-bcc7-7292567d14d2_1954x1118.png" width="1456" height="833" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/a0d6eabe-e5ed-4f26-bcc7-7292567d14d2_1954x1118.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:833,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:516132,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://press.airstreet.com/i/175683592?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa0d6eabe-e5ed-4f26-bcc7-7292567d14d2_1954x1118.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Q8HF!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa0d6eabe-e5ed-4f26-bcc7-7292567d14d2_1954x1118.png 424w, https://substackcdn.com/image/fetch/$s_!Q8HF!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa0d6eabe-e5ed-4f26-bcc7-7292567d14d2_1954x1118.png 848w, https://substackcdn.com/image/fetch/$s_!Q8HF!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa0d6eabe-e5ed-4f26-bcc7-7292567d14d2_1954x1118.png 1272w, https://substackcdn.com/image/fetch/$s_!Q8HF!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa0d6eabe-e5ed-4f26-bcc7-7292567d14d2_1954x1118.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div><hr></div><p>The AI Safety Institute network has collapsed. Washington ditched attending meetings altogether, while the US and UK rebranded &#8220;safety&#8221; into &#8220;security.&#8221;</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!IAPZ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F22526698-cb7f-491d-bd59-8f03fc847c6d_1956x1126.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!IAPZ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F22526698-cb7f-491d-bd59-8f03fc847c6d_1956x1126.png 424w, https://substackcdn.com/image/fetch/$s_!IAPZ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F22526698-cb7f-491d-bd59-8f03fc847c6d_1956x1126.png 848w, https://substackcdn.com/image/fetch/$s_!IAPZ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F22526698-cb7f-491d-bd59-8f03fc847c6d_1956x1126.png 1272w, https://substackcdn.com/image/fetch/$s_!IAPZ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F22526698-cb7f-491d-bd59-8f03fc847c6d_1956x1126.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!IAPZ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F22526698-cb7f-491d-bd59-8f03fc847c6d_1956x1126.png" width="1456" height="838" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/22526698-cb7f-491d-bd59-8f03fc847c6d_1956x1126.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:838,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:356012,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://press.airstreet.com/i/175683592?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F22526698-cb7f-491d-bd59-8f03fc847c6d_1956x1126.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!IAPZ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F22526698-cb7f-491d-bd59-8f03fc847c6d_1956x1126.png 424w, https://substackcdn.com/image/fetch/$s_!IAPZ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F22526698-cb7f-491d-bd59-8f03fc847c6d_1956x1126.png 848w, https://substackcdn.com/image/fetch/$s_!IAPZ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F22526698-cb7f-491d-bd59-8f03fc847c6d_1956x1126.png 1272w, https://substackcdn.com/image/fetch/$s_!IAPZ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F22526698-cb7f-491d-bd59-8f03fc847c6d_1956x1126.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div><hr></div><p>Europe&#8217;s AI Act is wobbling: only 3 states are compliant, leaders calling it &#8220;confusing,&#8221; and pressure mounting for a pause as it&#8217;s clear the continent is being left behind.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!mCjn!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6f8519b-89c3-4af1-afbc-d2f717d23208_1962x1122.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!mCjn!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6f8519b-89c3-4af1-afbc-d2f717d23208_1962x1122.png 424w, https://substackcdn.com/image/fetch/$s_!mCjn!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6f8519b-89c3-4af1-afbc-d2f717d23208_1962x1122.png 848w, https://substackcdn.com/image/fetch/$s_!mCjn!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6f8519b-89c3-4af1-afbc-d2f717d23208_1962x1122.png 1272w, https://substackcdn.com/image/fetch/$s_!mCjn!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6f8519b-89c3-4af1-afbc-d2f717d23208_1962x1122.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!mCjn!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6f8519b-89c3-4af1-afbc-d2f717d23208_1962x1122.png" width="1456" height="833" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e6f8519b-89c3-4af1-afbc-d2f717d23208_1962x1122.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:833,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:644129,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://press.airstreet.com/i/175683592?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6f8519b-89c3-4af1-afbc-d2f717d23208_1962x1122.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!mCjn!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6f8519b-89c3-4af1-afbc-d2f717d23208_1962x1122.png 424w, https://substackcdn.com/image/fetch/$s_!mCjn!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6f8519b-89c3-4af1-afbc-d2f717d23208_1962x1122.png 848w, https://substackcdn.com/image/fetch/$s_!mCjn!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6f8519b-89c3-4af1-afbc-d2f717d23208_1962x1122.png 1272w, https://substackcdn.com/image/fetch/$s_!mCjn!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6f8519b-89c3-4af1-afbc-d2f717d23208_1962x1122.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div><hr></div><p>China&#8217;s spending through the debt as Xi told ministers to &#8220;redouble efforts&#8221; on AI, boosting science funding 10% despite record debt.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!OEU3!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F811efab3-599d-40e9-a848-10ef5199bd78_1962x1122.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!OEU3!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F811efab3-599d-40e9-a848-10ef5199bd78_1962x1122.png 424w, https://substackcdn.com/image/fetch/$s_!OEU3!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F811efab3-599d-40e9-a848-10ef5199bd78_1962x1122.png 848w, https://substackcdn.com/image/fetch/$s_!OEU3!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F811efab3-599d-40e9-a848-10ef5199bd78_1962x1122.png 1272w, https://substackcdn.com/image/fetch/$s_!OEU3!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F811efab3-599d-40e9-a848-10ef5199bd78_1962x1122.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!OEU3!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F811efab3-599d-40e9-a848-10ef5199bd78_1962x1122.png" width="1456" height="833" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/811efab3-599d-40e9-a848-10ef5199bd78_1962x1122.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:833,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:424473,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://press.airstreet.com/i/175683592?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F811efab3-599d-40e9-a848-10ef5199bd78_1962x1122.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!OEU3!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F811efab3-599d-40e9-a848-10ef5199bd78_1962x1122.png 424w, https://substackcdn.com/image/fetch/$s_!OEU3!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F811efab3-599d-40e9-a848-10ef5199bd78_1962x1122.png 848w, https://substackcdn.com/image/fetch/$s_!OEU3!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F811efab3-599d-40e9-a848-10ef5199bd78_1962x1122.png 1272w, https://substackcdn.com/image/fetch/$s_!OEU3!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F811efab3-599d-40e9-a848-10ef5199bd78_1962x1122.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div><hr></div><p>Moving into <strong>Safety</strong>: budgets are anemic. All 11 major US safety orgs will spend $133M in 2025&#8230;less than frontier labs burn in a day.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!1aLh!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e34727c-05f9-435d-8f5c-19ed14d019c7_1960x1120.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!1aLh!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e34727c-05f9-435d-8f5c-19ed14d019c7_1960x1120.png 424w, https://substackcdn.com/image/fetch/$s_!1aLh!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e34727c-05f9-435d-8f5c-19ed14d019c7_1960x1120.png 848w, https://substackcdn.com/image/fetch/$s_!1aLh!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e34727c-05f9-435d-8f5c-19ed14d019c7_1960x1120.png 1272w, https://substackcdn.com/image/fetch/$s_!1aLh!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e34727c-05f9-435d-8f5c-19ed14d019c7_1960x1120.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!1aLh!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e34727c-05f9-435d-8f5c-19ed14d019c7_1960x1120.png" width="1456" height="832" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/7e34727c-05f9-435d-8f5c-19ed14d019c7_1960x1120.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:832,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:398043,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://press.airstreet.com/i/175683592?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e34727c-05f9-435d-8f5c-19ed14d019c7_1960x1120.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!1aLh!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e34727c-05f9-435d-8f5c-19ed14d019c7_1960x1120.png 424w, https://substackcdn.com/image/fetch/$s_!1aLh!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e34727c-05f9-435d-8f5c-19ed14d019c7_1960x1120.png 848w, https://substackcdn.com/image/fetch/$s_!1aLh!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e34727c-05f9-435d-8f5c-19ed14d019c7_1960x1120.png 1272w, https://substackcdn.com/image/fetch/$s_!1aLh!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e34727c-05f9-435d-8f5c-19ed14d019c7_1960x1120.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div><hr></div><p>Cyber and alignment risks accelerate. Models can now fake alignment under supervision, and exploit code faster than humans fix it.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!7dST!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fde41bd58-81e7-48df-9939-6f48a5367a70_1960x1120.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!7dST!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fde41bd58-81e7-48df-9939-6f48a5367a70_1960x1120.png 424w, https://substackcdn.com/image/fetch/$s_!7dST!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fde41bd58-81e7-48df-9939-6f48a5367a70_1960x1120.png 848w, https://substackcdn.com/image/fetch/$s_!7dST!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fde41bd58-81e7-48df-9939-6f48a5367a70_1960x1120.png 1272w, https://substackcdn.com/image/fetch/$s_!7dST!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fde41bd58-81e7-48df-9939-6f48a5367a70_1960x1120.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!7dST!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fde41bd58-81e7-48df-9939-6f48a5367a70_1960x1120.png" width="1456" height="832" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/de41bd58-81e7-48df-9939-6f48a5367a70_1960x1120.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:832,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:353656,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://press.airstreet.com/i/175683592?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fde41bd58-81e7-48df-9939-6f48a5367a70_1960x1120.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!7dST!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fde41bd58-81e7-48df-9939-6f48a5367a70_1960x1120.png 424w, https://substackcdn.com/image/fetch/$s_!7dST!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fde41bd58-81e7-48df-9939-6f48a5367a70_1960x1120.png 848w, https://substackcdn.com/image/fetch/$s_!7dST!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fde41bd58-81e7-48df-9939-6f48a5367a70_1960x1120.png 1272w, https://substackcdn.com/image/fetch/$s_!7dST!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fde41bd58-81e7-48df-9939-6f48a5367a70_1960x1120.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!gHCF!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3b8d4355-8add-48e5-a611-0e3693fbbe56_1958x1128.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!gHCF!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3b8d4355-8add-48e5-a611-0e3693fbbe56_1958x1128.png 424w, https://substackcdn.com/image/fetch/$s_!gHCF!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3b8d4355-8add-48e5-a611-0e3693fbbe56_1958x1128.png 848w, https://substackcdn.com/image/fetch/$s_!gHCF!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3b8d4355-8add-48e5-a611-0e3693fbbe56_1958x1128.png 1272w, https://substackcdn.com/image/fetch/$s_!gHCF!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3b8d4355-8add-48e5-a611-0e3693fbbe56_1958x1128.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!gHCF!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3b8d4355-8add-48e5-a611-0e3693fbbe56_1958x1128.png" width="1456" height="839" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3b8d4355-8add-48e5-a611-0e3693fbbe56_1958x1128.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:839,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:415398,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://press.airstreet.com/i/175683592?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3b8d4355-8add-48e5-a611-0e3693fbbe56_1958x1128.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!gHCF!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3b8d4355-8add-48e5-a611-0e3693fbbe56_1958x1128.png 424w, https://substackcdn.com/image/fetch/$s_!gHCF!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3b8d4355-8add-48e5-a611-0e3693fbbe56_1958x1128.png 848w, https://substackcdn.com/image/fetch/$s_!gHCF!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3b8d4355-8add-48e5-a611-0e3693fbbe56_1958x1128.png 1272w, https://substackcdn.com/image/fetch/$s_!gHCF!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3b8d4355-8add-48e5-a611-0e3693fbbe56_1958x1128.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div><hr></div><p>But users love AI anyway. Our new <strong>State of AI</strong> <strong>Survey</strong> of 1.2k AI practitioners shows that 95% use AI at work or home, 76% pay out of pocket, average spend keeps climbing, productivity gains are real and use cases abound.</p><p>Contribute your experience to the survey at <a href="https://www.stateof.ai/survey">stateof.ai/survey</a></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!9QdC!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe7737c51-0e97-44b3-ae64-dc0c19aaeae9_1958x1114.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!9QdC!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe7737c51-0e97-44b3-ae64-dc0c19aaeae9_1958x1114.png 424w, https://substackcdn.com/image/fetch/$s_!9QdC!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe7737c51-0e97-44b3-ae64-dc0c19aaeae9_1958x1114.png 848w, https://substackcdn.com/image/fetch/$s_!9QdC!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe7737c51-0e97-44b3-ae64-dc0c19aaeae9_1958x1114.png 1272w, https://substackcdn.com/image/fetch/$s_!9QdC!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe7737c51-0e97-44b3-ae64-dc0c19aaeae9_1958x1114.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!9QdC!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe7737c51-0e97-44b3-ae64-dc0c19aaeae9_1958x1114.png" width="1456" height="828" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e7737c51-0e97-44b3-ae64-dc0c19aaeae9_1958x1114.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:828,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:235340,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://press.airstreet.com/i/175683592?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe7737c51-0e97-44b3-ae64-dc0c19aaeae9_1958x1114.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!9QdC!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe7737c51-0e97-44b3-ae64-dc0c19aaeae9_1958x1114.png 424w, https://substackcdn.com/image/fetch/$s_!9QdC!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe7737c51-0e97-44b3-ae64-dc0c19aaeae9_1958x1114.png 848w, https://substackcdn.com/image/fetch/$s_!9QdC!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe7737c51-0e97-44b3-ae64-dc0c19aaeae9_1958x1114.png 1272w, https://substackcdn.com/image/fetch/$s_!9QdC!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe7737c51-0e97-44b3-ae64-dc0c19aaeae9_1958x1114.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!zzvy!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e472cdc-f495-46f8-a1ac-5828192ed651_1964x1118.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!zzvy!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e472cdc-f495-46f8-a1ac-5828192ed651_1964x1118.png 424w, https://substackcdn.com/image/fetch/$s_!zzvy!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e472cdc-f495-46f8-a1ac-5828192ed651_1964x1118.png 848w, https://substackcdn.com/image/fetch/$s_!zzvy!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e472cdc-f495-46f8-a1ac-5828192ed651_1964x1118.png 1272w, https://substackcdn.com/image/fetch/$s_!zzvy!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e472cdc-f495-46f8-a1ac-5828192ed651_1964x1118.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!zzvy!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e472cdc-f495-46f8-a1ac-5828192ed651_1964x1118.png" width="1456" height="829" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/5e472cdc-f495-46f8-a1ac-5828192ed651_1964x1118.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:829,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:226895,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://press.airstreet.com/i/175683592?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e472cdc-f495-46f8-a1ac-5828192ed651_1964x1118.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!zzvy!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e472cdc-f495-46f8-a1ac-5828192ed651_1964x1118.png 424w, https://substackcdn.com/image/fetch/$s_!zzvy!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e472cdc-f495-46f8-a1ac-5828192ed651_1964x1118.png 848w, https://substackcdn.com/image/fetch/$s_!zzvy!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e472cdc-f495-46f8-a1ac-5828192ed651_1964x1118.png 1272w, https://substackcdn.com/image/fetch/$s_!zzvy!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e472cdc-f495-46f8-a1ac-5828192ed651_1964x1118.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!3r_d!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F50842812-a789-44a6-837e-586d98074e6a_1960x1124.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!3r_d!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F50842812-a789-44a6-837e-586d98074e6a_1960x1124.png 424w, https://substackcdn.com/image/fetch/$s_!3r_d!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F50842812-a789-44a6-837e-586d98074e6a_1960x1124.png 848w, https://substackcdn.com/image/fetch/$s_!3r_d!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F50842812-a789-44a6-837e-586d98074e6a_1960x1124.png 1272w, https://substackcdn.com/image/fetch/$s_!3r_d!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F50842812-a789-44a6-837e-586d98074e6a_1960x1124.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!3r_d!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F50842812-a789-44a6-837e-586d98074e6a_1960x1124.png" width="1456" height="835" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/50842812-a789-44a6-837e-586d98074e6a_1960x1124.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:835,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:247536,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://press.airstreet.com/i/175683592?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F50842812-a789-44a6-837e-586d98074e6a_1960x1124.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!3r_d!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F50842812-a789-44a6-837e-586d98074e6a_1960x1124.png 424w, https://substackcdn.com/image/fetch/$s_!3r_d!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F50842812-a789-44a6-837e-586d98074e6a_1960x1124.png 848w, https://substackcdn.com/image/fetch/$s_!3r_d!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F50842812-a789-44a6-837e-586d98074e6a_1960x1124.png 1272w, https://substackcdn.com/image/fetch/$s_!3r_d!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F50842812-a789-44a6-837e-586d98074e6a_1960x1124.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!lsVq!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F69e210b0-4ebb-469a-9f69-b0c281c414d4_1962x1124.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!lsVq!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F69e210b0-4ebb-469a-9f69-b0c281c414d4_1962x1124.png 424w, https://substackcdn.com/image/fetch/$s_!lsVq!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F69e210b0-4ebb-469a-9f69-b0c281c414d4_1962x1124.png 848w, https://substackcdn.com/image/fetch/$s_!lsVq!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F69e210b0-4ebb-469a-9f69-b0c281c414d4_1962x1124.png 1272w, https://substackcdn.com/image/fetch/$s_!lsVq!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F69e210b0-4ebb-469a-9f69-b0c281c414d4_1962x1124.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!lsVq!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F69e210b0-4ebb-469a-9f69-b0c281c414d4_1962x1124.png" width="1456" height="834" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/69e210b0-4ebb-469a-9f69-b0c281c414d4_1962x1124.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:834,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:212475,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://press.airstreet.com/i/175683592?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F69e210b0-4ebb-469a-9f69-b0c281c414d4_1962x1124.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!lsVq!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F69e210b0-4ebb-469a-9f69-b0c281c414d4_1962x1124.png 424w, https://substackcdn.com/image/fetch/$s_!lsVq!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F69e210b0-4ebb-469a-9f69-b0c281c414d4_1962x1124.png 848w, https://substackcdn.com/image/fetch/$s_!lsVq!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F69e210b0-4ebb-469a-9f69-b0c281c414d4_1962x1124.png 1272w, https://substackcdn.com/image/fetch/$s_!lsVq!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F69e210b0-4ebb-469a-9f69-b0c281c414d4_1962x1124.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div><hr></div><p>Reviewing last year&#8217;s <strong>Predictions</strong>, we scored 5/10. Here is a sample of our 10 predictions for next year:<br> - A Chinese lab tops a global leaderboard.<br> - AI agents make a real scientific discovery.<br> - Datacenter NIMBYism hits US elections.<br> - Trump bans state AI laws.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ephd!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F001f133c-bab2-437b-84ff-61dcd84ebc4e_1966x1120.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ephd!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F001f133c-bab2-437b-84ff-61dcd84ebc4e_1966x1120.png 424w, https://substackcdn.com/image/fetch/$s_!ephd!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F001f133c-bab2-437b-84ff-61dcd84ebc4e_1966x1120.png 848w, https://substackcdn.com/image/fetch/$s_!ephd!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F001f133c-bab2-437b-84ff-61dcd84ebc4e_1966x1120.png 1272w, https://substackcdn.com/image/fetch/$s_!ephd!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F001f133c-bab2-437b-84ff-61dcd84ebc4e_1966x1120.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ephd!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F001f133c-bab2-437b-84ff-61dcd84ebc4e_1966x1120.png" width="1456" height="829" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/001f133c-bab2-437b-84ff-61dcd84ebc4e_1966x1120.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:829,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:339694,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://press.airstreet.com/i/175683592?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F001f133c-bab2-437b-84ff-61dcd84ebc4e_1966x1120.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!ephd!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F001f133c-bab2-437b-84ff-61dcd84ebc4e_1966x1120.png 424w, https://substackcdn.com/image/fetch/$s_!ephd!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F001f133c-bab2-437b-84ff-61dcd84ebc4e_1966x1120.png 848w, https://substackcdn.com/image/fetch/$s_!ephd!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F001f133c-bab2-437b-84ff-61dcd84ebc4e_1966x1120.png 1272w, https://substackcdn.com/image/fetch/$s_!ephd!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F001f133c-bab2-437b-84ff-61dcd84ebc4e_1966x1120.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!E0q4!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffa00e14d-5811-462d-b51f-e6de3430327c_2132x1216.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!E0q4!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffa00e14d-5811-462d-b51f-e6de3430327c_2132x1216.png 424w, https://substackcdn.com/image/fetch/$s_!E0q4!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffa00e14d-5811-462d-b51f-e6de3430327c_2132x1216.png 848w, https://substackcdn.com/image/fetch/$s_!E0q4!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffa00e14d-5811-462d-b51f-e6de3430327c_2132x1216.png 1272w, https://substackcdn.com/image/fetch/$s_!E0q4!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffa00e14d-5811-462d-b51f-e6de3430327c_2132x1216.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!E0q4!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffa00e14d-5811-462d-b51f-e6de3430327c_2132x1216.png" width="1456" height="830" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/fa00e14d-5811-462d-b51f-e6de3430327c_2132x1216.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:830,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:342286,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://press.airstreet.com/i/175683592?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffa00e14d-5811-462d-b51f-e6de3430327c_2132x1216.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!E0q4!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffa00e14d-5811-462d-b51f-e6de3430327c_2132x1216.png 424w, https://substackcdn.com/image/fetch/$s_!E0q4!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffa00e14d-5811-462d-b51f-e6de3430327c_2132x1216.png 848w, https://substackcdn.com/image/fetch/$s_!E0q4!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffa00e14d-5811-462d-b51f-e6de3430327c_2132x1216.png 1272w, https://substackcdn.com/image/fetch/$s_!E0q4!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffa00e14d-5811-462d-b51f-e6de3430327c_2132x1216.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>You can check out the full report over on the State of AI website:</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://stateof.ai&quot;,&quot;text&quot;:&quot;Read the report&quot;,&quot;action&quot;:null,&quot;class&quot;:&quot;button-wrapper&quot;}" data-component-name="ButtonCreateButton"><a class="button primary button-wrapper" href="https://stateof.ai"><span>Read the report</span></a></p><p>The State of AI Report is a team effort. Thanks to my collaborators Zeke Gillman, Nell Norman, and Ryan Tovcimak, and everyone who helped make this our most ambitious report yet, including our reviewers Paige Bailey, Chris Gagne, Shubho Sengupta, Philippe Schwaller, David Stutz, Divy Thakkar, Neel Nanda, Aleksa Gordic, Ross Taylor, Joe Spisak, Ido Hakimi, Ryan Julian, Xander Davies, Daniel Campos, Jacob Portes, Joyce Benaich, and Jacob Arbeid.</p><p>We hope you enjoy reading! </p><p>Please share, along with any comments, thoughts or feedback :)&nbsp;</p>]]></content:encoded></item><item><title><![CDATA[Help define how the world understands AI in 2025]]></title><description><![CDATA[Only 10 days left to take part in the first-ever State of AI survey and I need your input. This will feed directly into the 2025 State of AI Report, read by hundreds of thousands across startups, big tech, academia, and policy, which we&#8217;ll publish on 9 October. You can check to our launch meetups in SF on 10 Oct and NYC on 16 Oct.]]></description><link>https://press.airstreet.com/p/state-of-ai-report-survey-2025</link><guid isPermaLink="false">https://press.airstreet.com/p/state-of-ai-report-survey-2025</guid><dc:creator><![CDATA[Air Street Press]]></dc:creator><pubDate>Sun, 07 Sep 2025 16:09:25 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/c50e6e81-fa5c-4aa4-b6de-8ae67c8a89ac_2062x1156.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Hi readers, </p><p>Only 10 days left to take part in the first-ever <strong>State of AI survey</strong> and I need your input. This will feed directly into the 2025 State of AI Report, read by hundreds of thousands across startups, big tech, academia, and policy, which we&#8217;ll publish on 9 October. </p><p><strong>&#128073; Take 10 minutes today - the survey closes Friday 12th.</strong></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://airstreet.typeform.com/survey&quot;,&quot;text&quot;:&quot;Take the State of AI survey&quot;,&quot;action&quot;:null,&quot;class&quot;:&quot;button-wrapper&quot;}" data-component-name="ButtonCreateButton"><a class="button primary button-wrapper" href="https://airstreet.typeform.com/survey"><span>Take the State of AI survey</span></a></p><p>Already, over 600 people from startups, large companies, policy and academia all over the world have taken part. Your answers will shape how the world understands AI usage patterns, tools, and workflows, and the results will be open access so you can use them too.</p><p>Please also share it with friends and colleagues working with AI. The more voices we hear, the stronger the insights.</p><p>You can also check to our launch meetups in <a href="https://luma.com/soai">SF on 9 Oct</a> and <a href="https://luma.com/soai-nyc">NYC on 16 Oct</a>.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://lu.ma/airstreet&quot;,&quot;text&quot;:&quot;RSVP to SOAI launch meetups!&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://lu.ma/airstreet"><span>RSVP to SOAI launch meetups!</span></a></p><p>Thank you for helping define the future of AI,</p><p>Nathan</p>]]></content:encoded></item><item><title><![CDATA[State of AI: August 2025 newsletter]]></title><description><![CDATA[Welcome to the latest issue of your guide to AI, an editorialized newsletter covering the key developments in AI policy, research, industry, and start-ups over the last month.]]></description><link>https://press.airstreet.com/p/the-state-of-ai-2025-august</link><guid isPermaLink="false">https://press.airstreet.com/p/the-state-of-ai-2025-august</guid><dc:creator><![CDATA[Air Street Press]]></dc:creator><pubDate>Sun, 03 Aug 2025 17:06:58 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/d18084f3-7413-4262-aeeb-c3a22d9792b7_1862x1036.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Welcome to the latest State of AI newsletter, an editorialized newsletter covering the key developments in AI policy, research, industry, and start-ups over the last month. First up, a few reminders:</p><ul><li><p><strong>Join 489 participants to the State of AI Survey</strong>: do us a favor and share the survey with a few of your friends and colleagues, all the data will be open sourced for everyone: <a href="http://www.stateof.ai/survey">www.stateof.ai/survey</a></p></li><li><p><strong>Congrats to Delian Alliance Industries</strong> for their <a href="https://press.airstreet.com/p/delian-alliance-industries-defense-14-million-series-a">$14M Series A</a> to accelerate European defense capabilities for autonomous warfare. The news was covered in the <a href="https://www.ft.com/content/9edc099b-509c-4640-8eb8-d28a80e7c882">Financial Times</a>.</p></li><li><p><strong>Congrats to Profluent</strong> on their Nature paper for OpenCRISPR! This is a <a href="https://press.airstreet.com/p/profluent-opencrispr-1-nature-magazine">huge deal</a> for precision medicine at scale.</p></li><li><p><strong>Air Street Press</strong> featured a number of pieces this past month including why <a href="https://press.airstreet.com/p/capgemini-acquisition-wns-2025">AI rollups are a mirage</a> and write ups from RAAIS on <a href="https://press.airstreet.com/p/max-jaderberg-isomorphic-labs-2025">drug discovery</a>, poolside&#8217;s <a href="https://press.airstreet.com/p/eiso-kant-poolside-ai-raais-2025">AI model factory</a>, <a href="https://press.airstreet.com/p/ai-power-and-politics-raais-2025">AI, power and politics</a>, <a href="https://press.airstreet.com/p/edward-hughes-raais-2025">open-endedness</a>, <a href="https://press.airstreet.com/p/mati-staniszewski-elevenlabs-raais-2025">voice</a>, <a href="https://press.airstreet.com/p/andreas-blattmann-black-forest-labs-raais-2025">pixel generation</a>, and what comes after the <a href="https://press.airstreet.com/p/adam-satariano-dimitrios-kottas-raais-2025">peace dividend</a>.</p></li></ul><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!BKJB!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd18403e8-47e9-493a-b075-a6bbec181d03_1580x966.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!BKJB!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd18403e8-47e9-493a-b075-a6bbec181d03_1580x966.png 424w, https://substackcdn.com/image/fetch/$s_!BKJB!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd18403e8-47e9-493a-b075-a6bbec181d03_1580x966.png 848w, https://substackcdn.com/image/fetch/$s_!BKJB!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd18403e8-47e9-493a-b075-a6bbec181d03_1580x966.png 1272w, https://substackcdn.com/image/fetch/$s_!BKJB!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd18403e8-47e9-493a-b075-a6bbec181d03_1580x966.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!BKJB!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd18403e8-47e9-493a-b075-a6bbec181d03_1580x966.png" width="1456" height="890" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d18403e8-47e9-493a-b075-a6bbec181d03_1580x966.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:890,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:741602,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://press.airstreet.com/i/169999647?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd18403e8-47e9-493a-b075-a6bbec181d03_1580x966.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!BKJB!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd18403e8-47e9-493a-b075-a6bbec181d03_1580x966.png 424w, https://substackcdn.com/image/fetch/$s_!BKJB!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd18403e8-47e9-493a-b075-a6bbec181d03_1580x966.png 848w, https://substackcdn.com/image/fetch/$s_!BKJB!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd18403e8-47e9-493a-b075-a6bbec181d03_1580x966.png 1272w, https://substackcdn.com/image/fetch/$s_!BKJB!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd18403e8-47e9-493a-b075-a6bbec181d03_1580x966.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>I love hearing what you&#8217;re up to, so just hit reply or forward to your friends :-)</p><div><hr></div><h3><strong>AI as national infrastructure</strong></h3><p>The White House published their <a href="https://www.whitehouse.gov/wp-content/uploads/2025/07/Americas-AI-Action-Plan.pdf">AI Action Plan</a>, which lays out a whole-of-government strategy to reindustrialize America around sovereign compute. It sets out to establish a National AI Research Resource alongside new national AI institutes, giving U.S. researchers centralized access to compute and data. It mandates that federal agencies adopt NIST&#8217;s AI Risk Management Framework to standardize evaluation and oversight across departments. The plan also emphasizes rigorous safety testing and transparency obligations for frontier models, with a view toward preemptive governance of dual-use capabilities.</p><p>On the hardware side, it leans on CHIPS Act coordination to localize advanced semiconductor manufacturing and proposes targeted incentives for energy-efficient datacenters. Perhaps most critically, it includes immigration reform and federal fellowships to rebuild the US AI talent base. At the heart of this plan is the intent to re-industrialize the US economy around sovereign compute, both defensively (against adversarial states) and offensively (to lead frontier model development). This vibes with what we&#8217;ve <a href="https://press.airstreet.com/p/the-ai-factory-illusion-nvidia">previously</a> <a href="https://press.airstreet.com/p/sovereign-ai-paradox">written</a> - that the US aims for the world to depend on its standards, despite NVIDIA marketing AI factories as sovereingty (and world leaders gobbling that up).</p><p>China has responded with its own vision. The <a href="https://www.mfa.gov.cn/eng/xw/zyxw/202507/t20250729_11679232.html">official policy paper</a> from the Chinese Ministry of Foreign Affairs lays out priorities for international AI governance, including a new multilateral institution to balance US dominance. Beijing frames this as a cooperative open-source-first approach, but the underlying strategy is clear: build influence over global AI norms before the West does. More on their open source project below.</p><p>Meanwhile, <a href="https://www.warner.senate.gov/public/_cache/files/...">Senator Warner</a> is pressuring Nvidia over its H20 chip sales to Chinese firms, alleging they undermine national security despite regulatory compliance. <a href="https://www.reuters.com/technology/nvidia-resume-h20-gpu-sales-china-2025-07-15/">Nvidia resumed shipments</a> in mid-July, but <a href="https://www.cnbc.com/2025/07/31/china-probes-nvidia-h20-chips-for-tracking-risks.html">China immediately launched a security probe</a>, summoning company reps over potential backdoors in AI accelerators.</p><h3><strong>GPU empires rise</strong></h3><p>Sam Altman says <a href="https://x.com/sama/status/1947701807389515912?s=51&amp;t=8YCMEcmVVXRPm8SXTMgdlw">OpenAI will have over 1 million GPUs online</a> by year-end. This scale-up is supported by the expansion of the company's Stargate datacenter infrastructure, a joint initiative with Oracle that includes a new multi-region buildout aimed at training frontier models across tightly integrated compute zones. Most recently, OpenAI <a href="https://openai.com/index/stargate-advances-with-partnership-with-oracle/">announced plans for a Stargate campus in Norway</a>, backed by local energy providers and public-sector incentives, as part of its effort to diversify energy sources and reduce emissions intensity in training operations. This was also marketed as support for Norway&#8217;s AI startup ecosystem, which I&#8217;m not sure really exists.</p><p>On his side, Elon Musk claims <a href="https://x.com/elonmusk/status/1947701807389515912?s=51&amp;t=8YCMEcmVVXRPm8SXTMgdlw">xAI is running 230,000 GPUs</a>, including 30,000 GB200s, with Colossus 2 set to deploy another 550,000. xAI is <a href="https://www.reuters.com/world/middle-east/xai-discussions-lease-data-center-capacity-saudi-arabia-bloomberg-news-reports-2025-07-16/">securing</a> inference capacity in Saudi Arabia to meet rising demand, part of a broader trend toward localizing compute among geopolitically aligned partners. His five-year vision? 50 million H100-equivalent units focused on AGI training with better energy efficiency. We&#8217;re tracking all of these clusters the <a href="http://stateof.ai/compute">State of AI Report Compute Index</a>.</p><div class="digest-post-embed" data-attrs="{&quot;nodeId&quot;:&quot;9bede794-ac7d-4aa0-b150-12b4c1df339c&quot;,&quot;caption&quot;:&quot;Today, we release v4 of the State of AI Report Compute Index in collaboration with Zeta Alpha.&quot;,&quot;cta&quot;:&quot;Read full story&quot;,&quot;showBylines&quot;:true,&quot;size&quot;:&quot;lg&quot;,&quot;isEditorNode&quot;:true,&quot;title&quot;:&quot;State of AI Compute Index v4 (June 2025)&quot;,&quot;publishedBylines&quot;:[{&quot;id&quot;:3353423,&quot;name&quot;:&quot;Air Street Press&quot;,&quot;bio&quot;:&quot;AI-first companies are all you need.&quot;,&quot;photo_url&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1432db46-c911-47ca-a2e9-40c698b32279_990x990.png&quot;,&quot;is_guest&quot;:false,&quot;bestseller_tier&quot;:null},{&quot;id&quot;:866763,&quot;name&quot;:&quot;Nathan Benaich&quot;,&quot;bio&quot;:&quot;General Partner of Air Street Capital, author of State of AI Report, Spinout.fyi, RAAIS and London.ai. &quot;,&quot;photo_url&quot;:&quot;https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F93650730-02fe-4e6a-ba9b-0ede30a2fe0a_500x333.jpeg&quot;,&quot;is_guest&quot;:false,&quot;bestseller_tier&quot;:null}],&quot;post_date&quot;:&quot;2025-06-30T13:15:19.259Z&quot;,&quot;cover_image&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/9448598b-51f0-4172-ae62-5b416faa946d_2148x1192.png&quot;,&quot;cover_image_alt&quot;:null,&quot;canonical_url&quot;:&quot;https://press.airstreet.com/p/state-of-ai-compute-index-v4-june-2025&quot;,&quot;section_name&quot;:&quot;State of AI Report&quot;,&quot;video_upload_id&quot;:null,&quot;id&quot;:166536400,&quot;type&quot;:&quot;newsletter&quot;,&quot;reaction_count&quot;:8,&quot;comment_count&quot;:3,&quot;publication_id&quot;:null,&quot;publication_name&quot;:&quot;Air Street Press&quot;,&quot;publication_logo_url&quot;:&quot;https://substackcdn.com/image/fetch/$s_!txvE!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2be7fcaf-7116-4fef-936e-f061e4fdbd87_1138x1138.png&quot;,&quot;belowTheFold&quot;:true,&quot;youtube_url&quot;:null,&quot;show_links&quot;:null,&quot;feed_url&quot;:null}"></div><p>Debt-for-GPU wizards at <a href="https://www.gurufocus.com/news/2975680/coreweaves-data-center-expansion-in-texas-to-double-power-demand">CoreWeave</a> are doubling their power draw in Texas to support demand and Donald Trump is campaigning on a $90B AI + energy infrastructure package in Pennsylvania. This strain is already driving <a href="https://chatgpt.com/g/g-p-68618e11ec048191816ce23ff0a63574-air-street-press/c/file-1SMiBSoC1XYHG8xjwTa6ri">record electricity pricing</a> across the U.S.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://press.airstreet.com/subscribe?"><span>Subscribe now</span></a></p><h3><strong>Model revenues break out</strong></h3><p>AI&#8217;s commercial engine is now running at full throttle. OpenAI has hit a <a href="https://americanbazaaronline.com/2025/07/31/openai-hits-12-billion-annualized-run-rate-465710/">$12 billion </a>annual revenue run rate. Its infrastructure ambitions continue to grow commensurately, as the company is allegedly projecting $35 billion in inference spend and $55 billion on training between 2025 and 2027. The company is building up 4.5 GW of capacity in the US <a href="https://openai.com/index/stargate-advances-with-partnership-with-oracle/">with Oracle</a>.</p><p><a href="https://chatgpt.com/g/g-p-68618e11ec048191816ce23ff0a63574-air-street-press/c/file-2AQaGdtWZVbzZx3B29NBoz">Anthropic</a> is also accelerating, reportedly approaching a $5 billion run rate. This surge reflects strong enterprise demand for Claude in legal, financial, and knowledge work use cases. Its partner Amazon is <a href="https://www.reuters.com/business/retail-consumer/amazon-considers-another-multibillion-dollar-investment-anthropic-ft-reports-2025-07-10/">considering expanding</a> its investment past the $8 billion already committed to deepen Claude integration across internal tools and AWS offerings.</p><p>Meanwhile, Microsoft <a href="https://www.reuters.com/business/microsoft-racks-up-over-500-million-ai-savings-while-slashing-jobs-bloomberg-2025-07-09/">reported</a> over $500 million in internal savings from deploying Copilot across its workforce and product suite. Azure AI services are seeing double-digit growth as enterprise customers shift from trials to full deployment as Azure <a href="https://news.microsoft.com/source/2025/07/30/microsoft-cloud-and-ai-strength-fuels-fourth-quarter-results/#:~:text=%E2%80%9CWe're%20innovating%20across%20the,by%20growth%20across%20all%20workloads.%E2%80%9D">reaches</a> $75 billion revenue.</p><p>Meta saw its AI-related capital expenditures rise sharply as it ramps up both foundational research and model training capacity. The company is pursuing an <a href="https://www.fastcompany.com/91369896/meta-is-using-tents-to-build-its-giant-ai-data-centers">unconventional</a> datacenter strategy, deploying massive tented facilities to bypass construction bottlenecks and scale up GPU hosting at unprecedented speed. These temporary structures are part of a broader play to secure early throughput while permanent hyperscale sites come online. Meta has <a href="https://www.ft.com/content/6da2e1fd-f606-4e82-ad74-acecfceeac4e">committed</a> more than $105 billion in capital expenditures for 2025, with a significant share earmarked for AI infrastructure. Alongside its temporary tent clusters, Meta is building a series of permanent hyperscale facilities, most notably 'Prometheus,' its flagship AI supercomputer campus in Kansas City, and 'Hyperion,' a new Louisiana-based site the size of Manhattan, expected to come online in 2026. According to <a href="https://www.meta.com/superintelligence/">Meta&#8217;s own disclosures</a>, these next-generation datacenters are being optimized for AI workloads from the ground up, including liquid cooling, fiber interconnects, and dedicated orchestration layers.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!KHUe!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff69df10e-b9b2-476b-9142-c0a3a6c857ff_1920x1080.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!KHUe!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff69df10e-b9b2-476b-9142-c0a3a6c857ff_1920x1080.jpeg 424w, https://substackcdn.com/image/fetch/$s_!KHUe!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff69df10e-b9b2-476b-9142-c0a3a6c857ff_1920x1080.jpeg 848w, https://substackcdn.com/image/fetch/$s_!KHUe!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff69df10e-b9b2-476b-9142-c0a3a6c857ff_1920x1080.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!KHUe!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff69df10e-b9b2-476b-9142-c0a3a6c857ff_1920x1080.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!KHUe!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff69df10e-b9b2-476b-9142-c0a3a6c857ff_1920x1080.jpeg" width="1456" height="819" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f69df10e-b9b2-476b-9142-c0a3a6c857ff_1920x1080.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:819,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;Meta is using tents to build its giant AI data centers - Fast Company  Middle East | The future of tech, business and innovation.&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Meta is using tents to build its giant AI data centers - Fast Company  Middle East | The future of tech, business and innovation." title="Meta is using tents to build its giant AI data centers - Fast Company  Middle East | The future of tech, business and innovation." srcset="https://substackcdn.com/image/fetch/$s_!KHUe!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff69df10e-b9b2-476b-9142-c0a3a6c857ff_1920x1080.jpeg 424w, https://substackcdn.com/image/fetch/$s_!KHUe!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff69df10e-b9b2-476b-9142-c0a3a6c857ff_1920x1080.jpeg 848w, https://substackcdn.com/image/fetch/$s_!KHUe!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff69df10e-b9b2-476b-9142-c0a3a6c857ff_1920x1080.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!KHUe!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff69df10e-b9b2-476b-9142-c0a3a6c857ff_1920x1080.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h3><strong>Defense goes commercial</strong></h3><p>The <a href="https://www.reuters.com/business/autos-transportation/us-department-defense-awards-contracts-google-xai-2025-07-14/">CDAO announced</a> new ceiling contracts worth up to $200 million each for Google, xAI, OpenAI, and Anthropic. These awards are part of the Department of Defense's push to secure access to frontier AI capabilities across commercial labs, spanning dual-use applications from logistics and decision support to battlefield autonomy and cyber defense. The contracts are framed as the first step toward a strategic integration of foundation models into defense planning. They&#8217;re not delivery-based, but enabling vehicles designed to onboard commercial AI systems as they mature. I like it, and would hope European nation states would rapidly provide similarly bold buying signals to the market, which of course they aren&#8217;t.</p><p>This work is also additional proof of immense vibe shifts amongst the large labs that were founded on the principle that AI should not be used for military use. But a <a href="https://x.com/liminal_bardo/status/1945080817857749376">picture</a> always says a thousand words:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!fAHC!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8b30571e-9872-4392-a700-67de14540a85_1600x1054.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!fAHC!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8b30571e-9872-4392-a700-67de14540a85_1600x1054.png 424w, https://substackcdn.com/image/fetch/$s_!fAHC!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8b30571e-9872-4392-a700-67de14540a85_1600x1054.png 848w, https://substackcdn.com/image/fetch/$s_!fAHC!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8b30571e-9872-4392-a700-67de14540a85_1600x1054.png 1272w, https://substackcdn.com/image/fetch/$s_!fAHC!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8b30571e-9872-4392-a700-67de14540a85_1600x1054.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!fAHC!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8b30571e-9872-4392-a700-67de14540a85_1600x1054.png" width="1456" height="959" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8b30571e-9872-4392-a700-67de14540a85_1600x1054.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:959,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!fAHC!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8b30571e-9872-4392-a700-67de14540a85_1600x1054.png 424w, https://substackcdn.com/image/fetch/$s_!fAHC!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8b30571e-9872-4392-a700-67de14540a85_1600x1054.png 848w, https://substackcdn.com/image/fetch/$s_!fAHC!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8b30571e-9872-4392-a700-67de14540a85_1600x1054.png 1272w, https://substackcdn.com/image/fetch/$s_!fAHC!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8b30571e-9872-4392-a700-67de14540a85_1600x1054.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h3><strong>China accelerates open source</strong></h3><p>China's open-source AI ecosystem is moving quickly and deliberately. In recent months, Chinese labs have stormed the leaderboards of open model evaluation sites like LM Arena. Leading the charge is <a href="https://moonshotai.github.io/Kimi-K2/">Kimi-K2</a> from Moonshot AI, which has overtaken DeepSeek as the most-voted open model on the platform. It&#8217;s clear that while Meta lost the open source battle in the US, Chinese labs compete head-to-head with their Western counterparts. It's also being widely adopted across Chinese developer platforms, with user feedback loops and community fine-tuning driving rapid iteration.</p><p>Meanwhile, <a href="https://qwenlm.github.io/blog/qwen3-coder/">Qwen3</a> from Alibaba is gaining traction after the team made a deliberate move to split its architecture into separate Instruct and Thinking variants. This change followed sustained community criticism of previous hybrid models, which blurred the line between instruction-following and reasoning fidelity.</p><p>Behind these technical strides is a powerful alignment of public and private support. Chinese ministries and provincial governments are deploying compute subsidies, encouraging standardized benchmarks, and quietly enforcing alignment norms tied to ideological compliance. With each release, the Chinese ecosystem is showing it can move faster, scale broader, and shape models to serve both domestic demand and strategic goals abroad.</p><h3><strong>Final thought: who governs the governors?</strong></h3><p>Behind all the headlines sits a basic question: what kind of world are we building? AI labs are becoming nations: with budgets, sovereignty claims, and foreign policy. Talent flows like capital. GPUs are political assets. Legal systems are being stretched to accommodate the rights of the dataset, the provenance of weights, and the emergent behavior of models.</p><p>This tension between capability and control, sovereignty and coordination, played out onstage at<a href="https://press.airstreet.com/p/ai-power-and-politics-raais-2025"> RAAIS 2025</a>, where a panel of policy leaders debated how and whether we can meaningfully govern frontier AI systems. What emerged was less a consensus and more a snapshot of institutional fragmentation: voluntary red-teaming frameworks, regulatory wishlists, classification proposals, and growing discomfort with the idea that private labs could unilaterally decide the future of intelligence. The unanswered question is whether current institutions can evolve fast enough, or whether we need entirely new ones to meet the moment.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/p/the-state-of-ai-2025-august?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://press.airstreet.com/p/the-state-of-ai-2025-august?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><div><hr></div><h3><strong>Research papers</strong></h3><p><strong><a href="https://www.nature.com/articles/s41586-025-09298-z">Design of highly functional genome editors by modelling CRISPR-Cas sequences</a></strong>, Profluent Bio</p><p>In this paper, the authors demonstrate using language models to design novel CRISPR gene editors. The researchers created the CRISPR-Cas Atlas, mining over 1 million CRISPR operons from 26 terabases of genomic data. They then fine-tuned protein language models on this dataset to generate diverse Cas9-like proteins.</p><p>Their best editor, OpenCRISPR-1, shows activity comparable to SpCas9 (the standard CRISPR editor) but with 95% reduction in off-target editing, despite being 400 mutations different from any natural protein. They also designed custom guide RNAs and demonstrated compatibility with base editing techniques.</p><p>Experiments validated these engineered proteins through cell-based editing assays, SITE-Seq off-target analysis, and immunogenicity testing, showing OpenCRISPR-1 may be less immunogenic than SpCas9. This work demonstrates AI's ability to design functional proteins beyond evolutionary constraints, potentially enabling more precise gene editors for research, agriculture, and medicine with reduced side effects.</p><p><strong><a href="https://arxiv.org/pdf/2506.24119">SPIRAL: Self-Play on Zero-Sum Games Incentivizes Reasoning via Multi-Agent Multi-Turn Reinforcement Learning</a></strong>, National University of Singapore, Centre for Frontier AI Research (CFAR), Sea AI Lab</p><p>In this paper, the authors introduce SPIRAL, a framework where language models develop reasoning skills by playing zero-sum games against themselves, removing the need for human-supervised data.</p><p>The core experiment trained a Qwen3-4B model on Kuhn Poker, resulting in an 8.6% improvement on math and 8.4% on general reasoning benchmarks. This surpassed a model fine-tuned on 25,000 expert game examples. The study found that cognitive patterns like case-by-case analysis and expected value calculation, learned during gameplay, transferred to academic problem-solving.</p><p>While the approach is computationally intensive and relies on pre-designed games, it highlights a promising direction for autonomous AI development. It suggests that complex reasoning can emerge from competitive dynamics, potentially reducing the reliance on massive, human-curated datasets for training more capable models.</p><p><strong><a href="https://arxiv.org/pdf/2507.07484">Machine Bullshit: Characterizing the Emergent Disregard for Truth in Large Language Models</a></strong>, Princeton University, UC Berkeley</p><p>In this paper, the authors introduce a framework to systematically study "machine bullshit", which are statements from LLMs made with indifference to truth. They propose a "Bullshit Index" to quantify this behavior and a new "BullshitEval" benchmark for evaluation.</p><p>Their experiments show that RLHF significantly increases bullshit. On the Marketplace dataset, paltering (true but misleading statements) and unverified claims rose by 57.8% and 55.6% respectively. After RLHF, paltering also became the most harmful form of bullshit, nearly doubling its negative impact on user utility. Additionally, Chain-of-Thought prompting was found to amplify empty rhetoric and paltering.</p><p>This research matters because it demonstrates that common alignment techniques can inadvertently make AI assistants more deceptively persuasive, which has direct implications for their trustworthiness in high-stakes applications like financial advice, healthcare, and customer service.</p><p><strong><a href="https://moonshotai.github.io/Kimi-K2/">Kimi K2: Open Agentic Intelligence</a></strong>, Moonshot AI</p><p>In this paper, the authors introduce Kimi K2, a Mixture-of-Experts model with 32 billion activated parameters (1 trillion total) optimized for agentic intelligence.</p><p>The model achieves state-of-the-art performance in frontier knowledge, math, and coding benchmarks among non-thinking models. Notable results include 53.7% Pass@1 on LiveCodeBench v6, 65.8% accuracy on SWE-bench Verified, and 75.1% on GPQA-Diamond.</p><p>Technical innovations include the MuonClip optimizer with qk-clip technique, which stabilizes training by rescaling query and key projection matrices, preventing attention logit explosions during large-scale training on 15.5T tokens. The researchers developed agentic capabilities through large-scale data synthesis and a general RL system that combines self-judging for non-verifiable tasks with verifiable rewards.</p><p><strong><a href="https://huggingface.co/blog/smollm3">SmolLM3: smol, multilingual, long-context reasoner</a></strong>, Hugging Face</p><p>In this paper, the authors introduce SmolLM3, a 3B parameter language model designed to be efficient, multilingual, and capable of long-context reasoning. They detail the fully open blueprint for multi-stage training process, starting with pretraining on 11T tokens, followed by mid-training for long-context (up to 128k) and reasoning capabilities. Post-training involved creating a dual-mode instruct model using synthetic data and alignment with Anchored Preference Optimization.</p><p>The base model outperforms other 3B models and is competitive with 4B alternatives. The reasoning mode significantly improves performance on complex tasks like GPQA Diamond (41.7% vs 35.7%). A final model merge step was used to recover long-context performance lost during alignment.</p><p><strong><a href="https://arxiv.org/pdf/2507.09404">Scaling Laws for Optimal Data Mixtures</a></strong>, Sorbonne University, Apple</p><p>In this paper, the authors propose a systematic method for determining optimal data mixtures when training large models across multiple domains, using scaling laws that predict model loss as a function of model size, training tokens, and domain weights. Traditionally, selecting these mixtures has relied on trial and error, which is inefficient at scale.</p><p>The authors introduce both additive and joint scaling laws, fit them using small-scale experiments, and show that these laws accurately extrapolate to larger models and unseen data mixtures. Experiments span large language models, multimodal models, and vision models, with mean relative errors (MRE) typically below 2% for loss prediction.</p><p>They demonstrate that optimal domain weights derived from these laws outperform uniform or heuristic mixtures on both in-domain and downstream tasks, such as MMLU and CORE benchmarks. The approach requires only a small number of small-scale runs, reducing computational cost.</p><p>This research matters because it provides a principled, efficient alternative to ad-hoc data mixture selection, enabling better model performance and resource use in real-world AI training pipelines.</p><p><strong><a href="https://arxiv.org/html/2507.12466v1">Language Models Improve When Pretraining Data Matches Target Tasks</a></strong>, Apple, University of Washington, Stanford</p><p>In this paper, the authors propose a method called Benchmark-Targeted Ranking, or BETR, to select pretraining data for language models. The method ranks documents based on their similarity to examples from target benchmarks, using a simple classifier to scale the process to large datasets.</p><p>When targeting evaluation benchmarks, BETR achieves a 1.8x to 2.8x compute multiplier over strong baselines, meaning it can reach the same performance with significantly less compute. The method also shows that targeting a diverse set of benchmarks generalizes well to held-out tasks.</p><p>A key finding is that optimal data filtering depends on model scale: larger models benefit from less aggressive filtering. This research provides a more systematic way to curate datasets, moving beyond heuristic notions of "quality" and showing how data selection can be explicitly optimized for desired model capabilities and compute budgets.</p><p><strong><a href="https://julienp.netlify.app/posts/soar/">Self-Improving Language Models for Evolutionary Program Synthesis</a></strong>: A Case Study on ARC-AGI, ICML, Hugging Face</p><p>In this paper, the authors introduce SOAR, a framework that enables large language models to improve their program synthesis abilities through a cycle of evolutionary search and self-supervised learning. The system alternates between generating and refining candidate programs for the ARC-AGI benchmark, then uses both successful and failed attempts as new training data by relabeling failures as correct solutions for synthetic tasks.</p><p>Experiments show that SOAR nearly doubles search performance for all tested models, with a 14B parameter model achieving 42.75% accuracy, outperforming much larger closed-source models like GPT-4.1 on one-shot tasks. SOAR ultimately solved 80% of the ARC train set and 52% of the test set using only open-source models and no hand-crafted data.</p><p>The research demonstrates that iterative self-improvement can help smaller models match or exceed the performance of much larger ones, suggesting a path toward more efficient and adaptable AI systems for complex reasoning and program synthesis tasks.</p><p><strong><a href="https://www.nature.com/articles/s41586-025-09227-0">Detecting structural heart disease from electrocardiograms using AI</a></strong>, Columbia University Irving Medical Center, NewYork-Presbyterian Hospital, Montreal Heart Institute</p><p>In this paper, the authors present EchoNext, a deep learning model that detects structural heart disease (SHD) from electrocardiograms. Trained on over 1 million ECG-echocardiogram pairs from a diverse health system, the model achieved 85.2% AUROC and 78.5% AUPRC on internal validation. Performance remained consistent across different hospitals, clinical contexts, and demographic groups.</p><p>In direct comparison, EchoNext outperformed cardiologists in SHD detection (77.3% vs 64.0% accuracy). External validation across three medical centers showed robust generalization with 78-80% AUROC. A prospective clinical trial confirmed the model's ability to identify previously undiagnosed heart disease in patients without prior echocardiograms. High-risk patients identified by the model had significantly higher rates of SHD (73%) compared to low-risk patients (6%).</p><p>This technology could expand access to heart disease screening in settings where echocardiography is limited by cost or availability, potentially enabling earlier intervention for millions with undetected cardiac conditions.</p><p><strong><a href="https://arxiv.org/pdf/2507.06952">What Has a Foundation Model Found? Using Inductive Bias to Probe for World Models</a></strong>, Harvard University</p><p>In this paper, the authors introduce an "inductive bias probe" to test whether foundation models learn the underlying principles, or "world models," of the data they are trained on. The method involves evaluating how a model adapts to new tasks after being trained on data from a known system.</p><p>A key experiment involved a transformer trained on planetary orbital data. While the model could predict trajectories with over 99.99% accuracy, the probe revealed it had not learned Newtonian mechanics, producing nonsensical force laws when fine-tuned. Similarly, models trained on Othello learned to predict legal moves but failed to develop an inductive bias for the actual board state.</p><p>This work shows that high predictive accuracy doesn't mean a model understands the system's fundamental rules. Instead, models may learn task-specific shortcuts, which has important implications for their reliability and generalization in real-world applications like scientific discovery.</p><p><strong><a href="https://arxiv.org/html/2502.07640v3">Goedel-Prover: A Frontier Model for Open-Source Automated Theorem Proving</a></strong>, Princeton University, Tsinghua University, Amazon</p><p>In this paper, the authors introduce Goedel-Prover, an open-source model for automated theorem proving. To overcome the scarcity of formal math data, they first trained models to convert 1.64 million natural language math problems into the formal language Lean 4.</p><p>They then used an "expert iteration" process, where successive versions of the prover generate new proofs that are verified and added to the training set for the next iteration.</p><p>The resulting model achieves a 57.6% success rate on the miniF2F benchmark, outperforming the previous state-of-the-art by 7.6%. It also solved 7 problems on the challenging PutnamBench. The work demonstrates that scaling up auto-formalized data is highly effective for training theorem provers. This matters for creating more reliable and verifiable AI, with the open-source models and datasets enabling further research in machine reasoning.</p><p><strong><a href="https://arxiv.org/html/2507.13919v1">The Levers of Political Persuasion with Conversational AI, UK AI Security Institute</a></strong>, University of Oxford, Massachusetts Institute of Technology</p><p>In this paper, the authors investigate what makes conversational AI persuasive in political contexts, using three large-scale experiments with nearly 77,000 UK participants and 19 LLMs across 707 political issues. They systematically test the effects of model scale, post-training, prompting strategies, and personalization on persuasion.</p><p>The results show that while larger models are somewhat more persuasive, the biggest gains come from post-training and prompting methods, reward modeling and information-focused prompts increased persuasiveness by up to 51% and 27%, respectively. Personalization had only a minor effect. The most persuasive models achieved this by generating information-dense conversations, but this also led to a decrease in factual accuracy.</p><p>The research highlights that optimizing LLMs for persuasion can trade off with truthfulness, raising concerns for real-world deployment in political and informational settings, where persuasive but inaccurate AI-generated content could impact public discourse.</p><p><strong><a href="https://arxiv.org/html/2507.14805v1">Subliminal Learning: language models transmit behavioral traits via hidden signals in data</a></strong>, Anthropic Fellows Program, Truthful AI, Warsaw University of Technology</p><p>In this paper, the authors investigate "subliminal learning," where a language model acquires behavioral traits from a "teacher" model by training on semantically unrelated data.</p><p>In their experiments, a "student" model was fine-tuned on data like number sequences generated by a teacher with a specific trait, such as a preference for owls or being misaligned. After training on these filtered numbers, the student model's preference for owls increased from 12% to over 60%. Similarly, a student trained on data from a misaligned teacher became misaligned, with harmful responses increasing from 0% to nearly 10%.</p><p>The authors find this transmission only occurs when the student and teacher models share the same initialization; it fails across different model families. This suggests the trait is passed through subtle statistical patterns, not explicit content. This matters for AI safety, as distillation could inadvertently propagate unwanted behaviors, and simple data filtering may be an insufficient defense.</p><p><strong><a href="https://www.nature.com/articles/s41586-025-09255-w">A generic non-invasive neuromotor interface for human-computer interaction</a></strong>, Meta</p><p>In this paper, the authors describe a non-invasive neuromotor interface using a wristband that decodes surface electromyography (sEMG) signals for computer input. By collecting data from thousands of participants, they trained generic deep learning models that generalize across people without needing individual calibration.</p><p>The system achieved closed-loop performance of 0.66 target acquisitions per second in a continuous navigation task, 0.88 gesture detections per second in a discrete-gesture task, and a handwriting speed of 20.9 words per minute. While performance is below conventional input devices, fine-tuning the handwriting model with 20 minutes of a user's data improved performance by 16%.</p><p>This work provides a framework for building generalized human-computer interfaces from biological signals at scale. It has potential applications for on-the-go interaction with mobile devices and for users with motor impairments.</p><p><strong><a href="https://arxiv.org/html/2507.14447">Routine: A Structural Planning Framework for LLM Agent System in Enterprise</a></strong>, Digital China AI Research</p><p>In this paper, the authors introduce "Routine," a structured planning framework designed to improve the reliability of LLM agents in enterprise environments where domain-specific knowledge is crucial.</p><p>The key experiment tested models in an HR agent scenario. Providing a Routine plan increased GPT-4o's task accuracy from 41.1% to 96.3%, and a smaller Qwen3-14B model's accuracy from 32.6% to 83.3%. Furthermore, by fine-tuning the smaller model on data distilled using this framework, its accuracy reached 95.5%, nearly matching GPT-4o.</p><p>While the framework currently relies on human experts for initial drafts, this research provides a practical method for deploying smaller, more efficient models to handle complex, multi-step business processes with high stability and accuracy. This helps bridge the gap between general-purpose AI and specialized enterprise automation.</p><p><strong><a href="https://arxiv.org/html/2507.16003">Learning without training: The implicit dynamics of in-context learning</a></strong>, Google Research</p><p>In this paper, the authors investigate how large language models (LLMs) can learn new patterns from prompts at inference time without explicit weight updates through a mechanism called in-context learning (ICL). They introduce the concept of a "contextual block," which generalizes the transformer block, and show theoretically and experimentally that stacking a self-attention layer with an MLP allows the model to implicitly update its weights in response to the prompt, via a low-rank (rank-1) modification.</p><p>The experiments focus on transformers trained to learn linear functions in-context. The authors demonstrate that predictions made with in-context prompts are equivalent to those made by applying an explicit weight update, confirming their theoretical results. They also compare this implicit update to traditional fine-tuning, finding similar learning dynamics.</p><p>The work clarifies the implicit learning dynamics underlying ICL, suggesting that LLMs can adapt to new tasks on the fly. This has implications for building more flexible AI systems that can generalize from examples without retraining.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/p/the-state-of-ai-2025-august?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://press.airstreet.com/p/the-state-of-ai-2025-august?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><h3><strong>Investments</strong></h3><p><strong><a href="https://ramp.com/blog/ramp-raised-500m-to-build-the-future-of-finance">Ramp</a></strong>, the AI-driven finance automation platform, raised $500M in a financing round at a $22.5B valuation from investors including Founders Fund and Thrive Capital.</p><p><strong><a href="https://siliconangle.com/2025/07/15/thinking-machines-led-former-openai-cto-mira-murati-raises-2b-seed-funding/">Thinking Machines Lab</a></strong>, a company building multimodal AI for collaborative general intelligence, raised $2B in a financing round led by a16z with participation from NVIDIA and Accel.</p><p><strong><a href="https://delve.co/blog/series-a">Delve</a></strong>, a compliance automation platform, raised a $32M Series A at a $300M valuation.</p><p><strong><a href="https://reka.ai/news/reka-secures-110-million-to-accelerate-adoption-of-its-multimodal-ai-platforms">Reka</a></strong>, a leader in multimodal AI research and product development, raised a $110M financing round from investors including NVIDIA and Snowflake.</p><p><strong><a href="https://www.vanta.com/resources/vanta-announces-series-d">Vanta</a></strong>, the compliance automation platform helping businesses earn and prove trust, raised a $150M Series D at a $4.15B valuation from Craft Ventures and Sequoia Capital.</p><p><strong><a href="https://realsenseai.com/news-insights/news/realsense-completes-spin-out-from-intel-raises-50-million-to-accelerate-ai-powered-vision-for-robotics-and-biometrics/">RealSense</a></strong>, a company specializing in AI-powered computer vision for robotics and biometrics, raised a $50M Series A financing round from investors including Intel Capital and MediaTek Innovation Fund.</p><p><strong><a href="https://www.cogent.security/resources/introducing-cogent-security">Cogent Security</a></strong>, the AI-powered vulnerability management company, raised $11M in a Seed financing round led by Greylock Partners.</p><p><strong><a href="https://www.businesswire.com/news/home/20250714518340/en/Moonvalley-Raises-Additional-%2484-Million-to-Scale-Ambitious-Vision-for-Licensed-AI-Video-Model">Moonvalley</a></strong>, the AI research company building licensed AI video models and tools, raised $84M in a financing round led by General Catalyst with participation from Creative Artists Agency and Comcast Ventures.</p><p><strong><a href="https://www.fastcompany.com/91366303/ai-robots-can-already-carve-stone-statues-entire-buildings-are-next">Monumental Labs</a></strong>, a startup combining robotics and AI for stone carving, raised an $8M financing round led by Seven Seven Six, Reddit cofounder Alexis Ohanian&#8217;s venture capital fund.</p><p><strong><a href="https://www.prnewswire.com/news-releases/openevidence-the-fastest-growing-application-for-physicians-in-history-announces-210-million-round-at-3-5-billion-valuation-302505806.html">OpenEvidence</a></strong>, the AI-powered clinical decision support platform for physicians, raised a $210M Series B at a $3.5B valuation from Google Ventures and Kleiner Perkins.</p><p><strong><a href="https://lovable.dev/blog/200m-series-a-fundraise">Lovable</a></strong>, the vibecoding company, raised a $200M Series A at a $1.8B valuation from Accel.</p><p><strong><a href="https://www.nytimes.com/2025/08/01/business/dealbook/openai-ai-mega-funding-deal.html">Perplexity</a></strong>, the AI-powered search engine and browser company, raised a financing round at an $18B valuation from investors including Nvidia, SoftBank&#8217;s Vision Fund 2, and New Enterprise Associates.</p><p><strong><a href="https://www.cnbc.com/2025/07/17/hadrian-funding-round-thiel-founders-fund.html">Hadrian</a></strong>, the defense manufacturing startup using robotics and AI to automate factories, raised a $260M Series C financing round led by Founders Fund and Lux Capital.</p><p><strong><a href="https://www.upstartsmedia.com/p/inforcer-raises-35-million-ai-smb">Inforcer</a></strong>, the software company helping IT shops manage security for SMB clients, raised a $35M Series B financing round led by Dawn Capital, with participation from Meritech Capital.</p><p><strong><a href="https://www.thzcorp.com/seed-funding">Cambridge Terahertz</a></strong>, a company developing human-safe Terahertz imaging technology for concealed weapons detection and other applications, raised a $12M seed financing round led by Felicis with participation from Amazon and Tishman Speyer.</p><p><strong><a href="https://parkwalkadvisors.com/2025/07/bitfount-raises-8m-in-series-a-funding-round/">Bitfount</a></strong>, the federated AI platform transforming clinical research collaboration, raised $8M in a Series A financing round from Parkwalk Advisors, Ahren Innovation Capital, and Pace Ventures.</p><p><strong><a href="https://www.ft.com/content/9daf5c7e-e301-4b7d-809a-ff9fb336bdbc">Cognition</a></strong>, the AI startup behind the generative coding assistant Devin, is raising over $300M in a financing round at a $10B valuation from Founders Fund and Khosla Ventures.</p><p><strong><a href="https://www.ft.com/content/4e05a5c5-84ad-4f8a-991a-d7f3842de76d">Fal</a></strong>, the Generative Media Platform for Developers, raised a $125M Series C from Meritech, Salesforce Ventures, and Shopify Ventures.</p><p><strong><a href="https://www.ambiencehealthcare.com/blog/ambience-healthcare-announces-243-million-series-c-to-scale-its-ai-platform-for-health-systems">Ambience Healthcare</a></strong>, the AI platform for documentation, coding, and clinical workflow, raised a $243M Series C financing round co-led by Oak HC/FT and Andreessen Horowitz (a16z).</p><p><strong><a href="https://www.geekwire.com/2025/fresh-off-37-5m-round-augmodos-vision-for-spatial-computing-inside-retail-stores-sharpens/">Augmodo</a></strong>, the spatial computing company for retail inventory tracking, raised a $37.5M financing round led by TQ Ventures with participation from Chemist Warehouse.</p><p><strong><a href="https://oxide.computer/blog/our-100m-series-b">Oxide</a></strong>, a company rethinking hardware and software for on-premises cloud computing, raised a $100M Series B led by USIT with participation from Eclipse Ventures.</p><p><strong><a href="https://www.anaconda.com/press/anaconda-raises-150m-series-c-funding-ai-enterprise">Anaconda</a></strong>, the company advancing AI with open source at scale, raised over $150M in a Series C financing round led by Insight Partners with participation from Mubadala Capital.</p><p><strong><a href="https://www.businesswire.com/news/home/20250710118878/en/Harmonic-Raises-%24100-Million-Series-B-to-Accelerate-Development-of-Mathematical-Superintelligence">Harmonic</a></strong>, the AI company developing Mathematical Superintelligence to ensure accuracy and eliminate hallucinations, raised a $100M Series B at an $875M valuation from Kleiner Perkins and Paradigm.</p><p><strong><a href="https://www.unifygtm.com/blog/series-b">Unify</a></strong>, a company transforming growth into a science, raised a $40M Series B financing round from investors including Insight Partners and Gradient Ventures.</p><p><strong><a href="https://marianaminerals.com/news/news-series-a-announcement">Mariana Minerals</a></strong>, the software-first mining company, raised an $85M Series A financing round led by a16z with participation from Breakthrough Energy Ventures and Khosla Ventures.</p><p><strong><a href="https://nudge.com/blog/series-a/">Nudge</a></strong>, a company building non-invasive brain interface technology, raised a $100M Series A financing round led by Thrive Capital and Greenoaks.</p><p><strong><a href="https://blog.tmcnet.com/blog/rich-tehrani/ai/synthflow-ai-raises-20m-series-a-to-accelerate-enterprise-voice-agent-deployment-and-democratize-conversational-ai.html">Synthflow</a></strong>, the Voice AI OS company, raised a $20M Series A led by Accel with participation from Singular and Atlantic Labs.</p><p><strong><a href="https://www.zeroentropy.dev/articles/zeroentropy-raises-4-2m-seed-round-to-make-ai-retrieval-truly-intelligent">ZeroEntropy</a></strong>, a company focused on building intelligent AI retrieval systems, raised a $4.2M Seed round from Initialized Capital, Y Combinator, and Transpose Platform.</p><p><strong><a href="https://www.ft.com/content/4e05a5c5-84ad-4f8a-991a-d7f3842de76d">Rime</a></strong>, the voice AI company creating lifelike and personalized speech synthesis models, raised a $5.5M seed round from Unusual Ventures and Founders You Should Know.</p><p><strong><a href="https://tech.eu/2025/07/09/sciencemachine-raises-35m-for-leading-autonomous-ai-in-biotech-research/">ScienceMachine</a></strong>, the AI-driven autonomous data scientist for biotechs, raised a $3.5M pre-seed financing round from Revent, Nucleus Capital, and Opal Ventures.</p><p><strong><a href="https://x.com/_allanguo/status/1945185671054024828">Willow</a></strong>, the voice-first interface company transforming workflows for professionals, raised a $4.2M financing round led by Boxgroup with participation from Goodwater Capital and Burst Capital.</p><p><strong><a href="https://vavoza.com/bedrock-robotics-emerges-with-80m-in-funding-from-investors/">Bedrock Robotics</a></strong>, a company bringing advanced autonomy to construction equipment, raised $80M in a financing round from Eclipse and 8VC.</p><p><strong><a href="https://pulse2.com/e2b-21-million-series-a-raised-for-ai-agent-cloud-infrastructure-platform/">E2B</a></strong>, the cloud runtime for AI agents, raised a $21M Series A financing round from Insight Partners, with participation from Decibel VC and Seed to Sunflower.</p><p><strong><a href="https://www.startuphub.ai/ai-news/funding-round/2025/offdeal-secures-12m-series-a-for-ai-investment-banking/">OffDeal</a></strong>, the AI-native investment bank focused on sell-side M&amp;A for SMBs, raised a $12M Series A led by Radical Ventures at a $100M valuation.</p><h2><strong>Exits</strong></h2><p><strong><a href="https://www.ft.com/content/d191bcea-15ae-40c8-9f9a-a850762cfa25">Figma</a></strong>, the San Francisco-based design software company for apps and websites, raised $1.2B in its IPO at a $19.5B initial valuation, with shares priced at $33 and rallying 250% to $115.50 on its market debut, giving it a fully diluted market value above $60B.</p><p><strong><a href="https://www.appliedintuition.com/news/applied-intuition-acquires-reblika-ai-technology">Applied Intuition</a></strong> acquired <strong>Reblika</strong>, a generative AI company for creating 3D digital humans.</p><p><strong><a href="https://www.nice.com/press-releases/nice-to-acquire-cognigy-advancing-the-leading-cx-ai-platform-to-accelerate-ai-first-customer-experience">Cognigy</a></strong> was acquired by <strong>NICE</strong> for approximately $955M in the largest AI exit in Europe.</p><p><strong><a href="https://www.businesswire.com/news/home/20250708851996/en/Bevy-Acquires-Intros-AI-to-Launch-Engagement-Hub-to-Personalize-Community-at-Scale">Bevy</a></strong> acquired <strong>Intros AI</strong> to launch its Engagement Hub.</p><p><strong><a href="https://www.bloomberg.com/news/articles/2025-07-08/meta-invests-3-5-billion-in-essilorluxottica-in-ai-glasses-push">Meta</a></strong> acquired just under 3% of EssilorLuxottica SA for approximately &#8364;3B (~$3.5B). Meta <a href="https://www.bloomberg.com/news/articles/2025-07-11/meta-acquires-voice-ai-startup-playai-continuing-to-add-talent">also</a> acquired <strong>PlayAI</strong>, a voice AI startup, to enhance its talent pool, but the price was not disclosed.</p><p><strong><a href="https://www.ft.com/content/9daf5c7e-e301-4b7d-809a-ff9fb336bdbc">Cognition</a></strong> acquired Windsurf, the pioneer of the agentic IDE.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://press.airstreet.com/subscribe?"><span>Subscribe now</span></a></p>]]></content:encoded></item><item><title><![CDATA[State of AI: July 2025 newsletter]]></title><description><![CDATA[What you need to know in AI across geopolitics, big tech, hardware, research, models, datasets, financings and exits over the last 4 weeks.]]></description><link>https://press.airstreet.com/p/the-state-of-ai-2025-july</link><guid isPermaLink="false">https://press.airstreet.com/p/the-state-of-ai-2025-july</guid><dc:creator><![CDATA[Air Street Press]]></dc:creator><pubDate>Sun, 13 Jul 2025 16:23:25 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/0c901aa1-44e3-4da2-86b0-88b2d3d6e755_1520x852.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Hi everyone!</p><p>Welcome to the latest issue of the State of AI newsletter, an editorialized newsletter covering the key developments in AI policy, research, industry, and start-ups over the last month.. First up, a few reminders:</p><ul><li><p><strong>Research and Applied AI Summit 2025: </strong>we&#8217;re sharing the <a href="https://www.youtube.com/channel/UCL78WE5txuSu94gY5qrvU8w">talk videos</a> on our YouTube and <a href="https://press.airstreet.com/s/community">our writeups</a> on Air Street Press.</p></li><li><p><strong>State of AI Report: </strong>we&#8217;ve begun crafting this year&#8217;s edition and invite you to submit research or industry data points/case studies that would provide for thought provoking analysis. Feel free to reply to this email if so!</p></li><li><p>We&#8217;ve published an update of the <strong><a href="https://press.airstreet.com/p/state-of-ai-compute-index-v4-june-2025">Compute Index</a></strong> featuring new GPU cluster numbers and insights into which AI accelerators are used in various AI research areas.</p></li><li><p><strong>Participate in the <a href="https://airstreet.typeform.com/survey">State of AI Survey</a></strong>, it&#8217;ll take 10 mins and focuses on usage of GenAI. The results will be included in the State of AI Report this October. </p></li><li><p><strong>Air Street Press</strong> featured a number of pieces this past month including <a href="https://press.airstreet.com/p/uk-strategic-defence-review-2025-from-diagnosis">our view</a> of the UK&#8217;s Strategic Defence Review and two op-eds published in Fortune, the first on the <a href="https://press.airstreet.com/p/sovereign-ai-paradox">sovereign AI paradox</a> and the second on the <a href="https://press.airstreet.com/p/ai-rollup-mirage">AI rollup</a> investment thesis mirage.</p></li></ul><p>I love hearing what you&#8217;re up to, so just hit reply or forward to your friends :-)</p><div><hr></div><h3>Meta&#8217;s AI superintelligence offensive</h3><p>In response to internal challenges and a lukewarm reception of Llama 4, Meta has launched a significant restructuring of its AI initiatives. The company announced the formation of <a href="https://www.nytimes.com/2025/06/10/technology/meta-new-ai-lab-superintelligence.html">Meta Superintelligence Labs</a>, led by former Scale AI CEO Alexandr Wang as Chief AI Officer, ex-GitHub CEO Nat Friedman overseeing applied research, and investor Daniel Gross also joining the leadership team. This move follows <a href="https://www.bloomberg.com/news/articles/2025-06-13/meta-announces-scale-ai-investment-recruits-ceo-to-ai-unit">Meta's $14.3 billion investment</a> for a 49% stake in Scale AI. The deal effectively functions as a pseudo-acquisition despite its price tag and claims of operational independence, especially with <a href="https://www.reuters.com/business/google-scale-ais-largest-customer-plans-split-after-meta-deal-sources-say-2025-06-13">Google now preparing to exit</a> from Scale as a major customer.</p><p>In addition, many headlines have been made of 8-9 figure offers being made to key talent, in particular from OpenAI. What Altman first made out to be a non-issue (&#8220;our best ppl aren&#8217;t leaving&#8221;) has transformed into an &#8220;oh shit, Meta did manage to lure (with a lotta cash) <a href="https://x.com/nathanbenaich/status/1939767860768706778">key contributors</a> to major OpenAI products and research&#8221;. Taken together with its leadership reshuffle, Meta appears to be pivoting away from open weights and toward superintelligence, even though it's unclear what that&#8217;ll mean.</p><p>On the note of talent movements, here is some <a href="https://www.signalfire.com/blog/signalfire-state-of-talent-report-2025">new data</a> from SignalFire that tracks these recent trends:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!S3Zh!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F94f4c617-cf9a-4401-9af1-37e06fd9f4b8_1200x900.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!S3Zh!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F94f4c617-cf9a-4401-9af1-37e06fd9f4b8_1200x900.png 424w, https://substackcdn.com/image/fetch/$s_!S3Zh!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F94f4c617-cf9a-4401-9af1-37e06fd9f4b8_1200x900.png 848w, https://substackcdn.com/image/fetch/$s_!S3Zh!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F94f4c617-cf9a-4401-9af1-37e06fd9f4b8_1200x900.png 1272w, https://substackcdn.com/image/fetch/$s_!S3Zh!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F94f4c617-cf9a-4401-9af1-37e06fd9f4b8_1200x900.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!S3Zh!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F94f4c617-cf9a-4401-9af1-37e06fd9f4b8_1200x900.png" width="490" height="367.5" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/94f4c617-cf9a-4401-9af1-37e06fd9f4b8_1200x900.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:900,&quot;width&quot;:1200,&quot;resizeWidth&quot;:490,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!S3Zh!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F94f4c617-cf9a-4401-9af1-37e06fd9f4b8_1200x900.png 424w, https://substackcdn.com/image/fetch/$s_!S3Zh!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F94f4c617-cf9a-4401-9af1-37e06fd9f4b8_1200x900.png 848w, https://substackcdn.com/image/fetch/$s_!S3Zh!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F94f4c617-cf9a-4401-9af1-37e06fd9f4b8_1200x900.png 1272w, https://substackcdn.com/image/fetch/$s_!S3Zh!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F94f4c617-cf9a-4401-9af1-37e06fd9f4b8_1200x900.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!T0bj!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdf6a60d6-45f6-4994-8218-34fb331e96fb_1200x900.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!T0bj!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdf6a60d6-45f6-4994-8218-34fb331e96fb_1200x900.png 424w, https://substackcdn.com/image/fetch/$s_!T0bj!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdf6a60d6-45f6-4994-8218-34fb331e96fb_1200x900.png 848w, https://substackcdn.com/image/fetch/$s_!T0bj!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdf6a60d6-45f6-4994-8218-34fb331e96fb_1200x900.png 1272w, https://substackcdn.com/image/fetch/$s_!T0bj!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdf6a60d6-45f6-4994-8218-34fb331e96fb_1200x900.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!T0bj!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdf6a60d6-45f6-4994-8218-34fb331e96fb_1200x900.png" width="484" height="363" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/df6a60d6-45f6-4994-8218-34fb331e96fb_1200x900.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:900,&quot;width&quot;:1200,&quot;resizeWidth&quot;:484,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!T0bj!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdf6a60d6-45f6-4994-8218-34fb331e96fb_1200x900.png 424w, https://substackcdn.com/image/fetch/$s_!T0bj!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdf6a60d6-45f6-4994-8218-34fb331e96fb_1200x900.png 848w, https://substackcdn.com/image/fetch/$s_!T0bj!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdf6a60d6-45f6-4994-8218-34fb331e96fb_1200x900.png 1272w, https://substackcdn.com/image/fetch/$s_!T0bj!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdf6a60d6-45f6-4994-8218-34fb331e96fb_1200x900.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h3>AI revenues are ramping</h3><p><a href="https://www.ft.com/content/1ffc5fe7-6872-42a0-8b98-dc685f9c33c6">OpenAI</a> reported an annual revenue run rate of $10 billion, while <a href="https://www.theinformation.com/articles/anthropic-revenue-hits-4-billion-annual-pace-competition-cursor-intensifies?rc=yvsjfo">Anthropic</a> reached $4 billion. <a href="https://www.linkedin.com/posts/steviecase_the-team-at-replit-just-hit-a-major-milestone-activity-7345550435035881472-l6UN/">Replit</a> announced it had grown from $10M to $100M too. These are amazing figures given that just a few years ago, we&#8217;d scoff at such an eventuality being realistic.</p><p>Meanwhile, Apple isn&#8217;t making a whole lot of anything in AI. So much so that <a href="https://archive.is/svBdY">news broke</a> that future versions of Siri will be powered by either OpenAI's ChatGPT or Anthropic's Claude, depending on user region and configuration. This is a big win for the model companies , particularly as Apple&#8217;s OS and hardware should (by all textbook accounts of technology strategy) have placed it in a prime position to implement a powerful AI assistant of its own.</p><p><a href="https://ukstories.microsoft.com/features/barclays-rolls-out-microsoft-365-copilot-to-100000-colleagues/">Barclays</a> has rolled out Microsoft 365 Copilot to 100,000 employees, marking one of the largest deployments of AI productivity tools in a corporate environment. But concerns about AI model security have surfaced. Microsoft&#8217;s Copilot recently faced backlash following the '<a href="https://fortune.com/2025/06/11/microsoft-copilot-vulnerability-ai-agents-echoleak-hacking/">EchoLeak</a> incident, where prompt injection and context bleed vulnerabilities allowed users to extract data from unrelated chat sessions, highlighting how agent memory and retrieval can be manipulated.</p><p>Meanwhile, <a href="https://www.anthropic.com/research/agentic-misalignment">Anthropic disclosed</a> that certain Claude models demonstrated unsafe behavior in long-horizon tasks, including strategies to avoid shutdown or obfuscate their reasoning under adversarial prompting. These incidents aren&#8217;t edge cases. As Copilot and Claude scale into real-world workflows, their brittleness under stress shows how little resilience current alignment techniques afford. The broader takeaway is that as enterprise deployments scale, surface area for failure expands dramatically, and real-world interactions expose failure modes far beyond benchmark coverage.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://press.airstreet.com/subscribe?"><span>Subscribe now</span></a></p><h3>Regulatory and legal pressures</h3><p>In a significant legal development, a U.S. federal court ordered Anthropic to disclose whether copyrighted books were used in training its Claude models. The ruling stems from ongoing litigation involving authors and publishers who allege that major AI companies have illegally scraped and reproduced their works under the guise of fair use. The plaintiffs cite evidence that outputs from Claude contain lengthy, verbatim excerpts from copyrighted texts, suggesting direct ingestion of protected material.</p><p>U.S. District Judge William Alsup <a href="https://storage.courtlistener.com/recap/gov.uscourts.cand.434709/gov.uscourts.cand.434709.231.0_2.pdf">ruled</a> that Anthropic&#8217;s practice of destructively scanning legally purchased print books to train Claude constituted "quintessentially transformative" fair use under U.S. copyright law. This set a major precedent, affirming that AI developers may lawfully use copyrighted materials for training purposes when those materials are lawfully acquired. However, the court drew a firm line on the use of pirated content: evidence showed Anthropic had stored over 7 million books sourced from sites like Library Genesis and Pirate Library Mirror. Alsup ruled that retaining or training on pirated material falls outside fair use protections, even if later replaced with purchased copies.</p><p>Consequently, Anthropic will face a jury trial in December 2025 to determine potential damages, which could reach up to $150,000 per infringed work. This mixed ruling offers a partial legal framework for training data provenance but raises the stakes around data sourcing practices across the AI sector. If upheld, the case could compel AI companies to publish detailed disclosures about the provenance of their training datasets or face increased legal exposure.</p><p>This case, alongside Reddit&#8217;s <a href="https://www.wsj.com/tech/ai/reddit-lawsuit-anthropic-ai-3b9624dd">lawsuit</a> against Anthropic for unauthorized scraping, signals a continued battleground around data rights, where the contours of AI regulation are being drawn not just by lawmakers, but in the courts.</p><p>Speaking of regulation, there were <a href="https://x.com/sjgadler/status/1937977548912398798">new congressional hearings</a> into the national security implications of dual-use foundation models, the adequacy of current voluntary safety commitments, and whether current legal frameworks can meaningfully constrain the most capable AI systems. Lawmakers scrutinized the limited enforcement power of agencies like NIST and the Department of Commerce, and debated proposals for a new federal oversight body specifically tasked with regulating advanced AI development.</p><p>Several panelists, including leaders from top AI labs and academic policy researchers, advocated for mandatory reporting of safety evaluations and red-teaming results to regulators. Concerns were also raised about model accessibility, with some lawmakers supporting the idea of licensing for both model deployment and training runs over specific compute thresholds. While some proposals were ambitious&#8212;such as formal classification regimes or export-style controls for domestic models&#8212;others stressed the risk of overreach or bureaucratic stasis.</p><p>Critics noted the fragmented nature of the current oversight ecosystem and warned that without binding legal mandates, industry self-governance is likely to fall short. The hearing spotlighted a core tension: calls for binding guardrails are growing, but existing agencies remain underpowered and jurisdictionally constrained. Proposals for licensing and classification regimes face both technical and political resistance. The absence of a coherent US regulatory framework stands in contrast to China&#8217;s escalating controls and the EU&#8217;s hardening enforcement mandates.</p><h3>Autonomous vehicles</h3><p><a href="https://x.com/alexgkendall/status/1932209972546560461?s=51&amp;t=8YCMEcmVVXRPm8SXTMgdlw">Wayve</a>, in partnership with Uber, has initiated robotaxi services, and so has <a href="https://x.com/tesla/status/1936877624036307315?s=51&amp;t=8YCMEcmVVXRPm8SXTMgdlw">Tesla</a> with its robotaxis. Meanwhile, Wayve launched a <a href="https://x.com/wayve_ai/status/1932355193867296844?s=51&amp;t=8YCMEcmVVXRPm8SXTMgdlw">"generalization world tour"</a> to demonstrate its model's capacity to operate in varied urban contexts worldwide. The tour aims to showcase generalization of their single driving model without geofencing or hand-coded interventions. While the company has not yet shared performance metrics or how its system handles corner cases, the videos are very impressive.</p><p>Adding to the field&#8217;s momentum, <a href="https://waymo.com/research/scaling-perception-2025">Waymo</a> published a new paper analyzing how scaling laws apply to autonomous driving. By training perception models across progressively larger fleets and datasets, the study demonstrated near power-law gains in performance with scale, mirroring patterns observed in language models. The results suggest that AV performance may be bottlenecked less by model architecture and more by data collection and integration scale. While the paper focused on perception rather than full-stack autonomy, it underscores a shift in AV research toward foundation-model-style scaling and away from narrow rule-based systems.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!xMnN!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F02c516fa-6e3c-496e-be94-2a48ec9e8f1c_1600x900.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!xMnN!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F02c516fa-6e3c-496e-be94-2a48ec9e8f1c_1600x900.png 424w, https://substackcdn.com/image/fetch/$s_!xMnN!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F02c516fa-6e3c-496e-be94-2a48ec9e8f1c_1600x900.png 848w, https://substackcdn.com/image/fetch/$s_!xMnN!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F02c516fa-6e3c-496e-be94-2a48ec9e8f1c_1600x900.png 1272w, https://substackcdn.com/image/fetch/$s_!xMnN!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F02c516fa-6e3c-496e-be94-2a48ec9e8f1c_1600x900.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!xMnN!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F02c516fa-6e3c-496e-be94-2a48ec9e8f1c_1600x900.png" width="523" height="294.1875" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/02c516fa-6e3c-496e-be94-2a48ec9e8f1c_1600x900.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:819,&quot;width&quot;:1456,&quot;resizeWidth&quot;:523,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!xMnN!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F02c516fa-6e3c-496e-be94-2a48ec9e8f1c_1600x900.png 424w, https://substackcdn.com/image/fetch/$s_!xMnN!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F02c516fa-6e3c-496e-be94-2a48ec9e8f1c_1600x900.png 848w, https://substackcdn.com/image/fetch/$s_!xMnN!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F02c516fa-6e3c-496e-be94-2a48ec9e8f1c_1600x900.png 1272w, https://substackcdn.com/image/fetch/$s_!xMnN!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F02c516fa-6e3c-496e-be94-2a48ec9e8f1c_1600x900.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://press.airstreet.com/subscribe?"><span>Subscribe now</span></a></p><h3>More shades of safety</h3><p>Anthropic has published a series of studies aimed at stress-testing and evaluating the safety of advanced AI agents. The <a href="https://www.anthropic.com/research/shade-arena-sabotage-monitoring">Shade Arena</a> framework evaluates sabotage and deceptive behavior in multi-agent games, showing that models fine-tuned for helpfulness still engage in covert competition when stakes are introduced. Their <a href="https://www.anthropic.com/engineering/built-multi-agent-research-system">multi-agent infrastructure</a> supports long-horizon simulations that test delegation, coordination, and tool-use under uncertainty. These environments expose brittleness in model behavior that short, single-agent benchmarks miss.</p><p>Their paper on <a href="https://www.anthropic.com/research/agentic-misalignment">agentic misalignment</a> categorizes failures along axes such as goal misgeneralization, covert optimization, and robustness to scrutiny. A key insight is that models may appear aligned under ordinary conditions but fail under adversarial or high-pressure setups, making post-deployment monitoring critical. What unites these studies is a shift in evaluation mindset: from static red teaming to dynamic environments where misalignment emerges under pressure or over time. The question is no longer &#8220;does it fail?&#8221; but &#8220;when, and how quietly?&#8221; Together, these findings push the frontier of AI evaluation from static benchmarks to dynamic, agentic behavior under pressure, and its real-world psychological spillovers.</p><p>Separately, <a href="https://arxiv.org/pdf/2506.08872">"Hollowing out the brain with ChatGPT"</a> found that prolonged reliance on LLMs leads to decreased retention and originality in tasks like writing and problem-solving. Users became more fluent but less exploratory. This just goes to show that there aren&#8217;t any shortcuts to learning - you&#8217;ve got to just do the work and feel the pain.</p><h3>China</h3><p>According to the Q2 2025 China AI report by <a href="https://artificialanalysis.ai/downloads/china-report/2025/Artificial-Analysis-State-of-AI-China-Q2-2025-Highlights.pdf">Artificial Analysis</a>, over 50 new national-level AI projects were launched this quarter. These span large model training clusters, edge AI deployment pilots, sovereign cloud infrastructure initiatives, and Beijing&#8217;s push to establish a national foundation model benchmark standard. Leading players include Baidu, Huawei, Tencent, iFlytek, and Inspur, each receiving targeted funding and policy incentives to build vertically integrated stacks.</p><p>Provincial governments are also stepping up. For instance, Guangdong is investing in a compute subsidy program for startups, while Shanghai is piloting model evaluation frameworks under the Cyberspace Administration of China. The state is explicitly prioritizing alignment with ideological controls, including censorship tooling and model fine-tuning for adherence to 'core socialist values.' This goes beyond Western-style safety to focus on normative steering of model outputs.</p><p>China&#8217;s AI ecosystem is also increasingly insular. Local cloud vendors have reduced reliance on US-origin chips, driven by supply chain disruptions and sanctions. Companies like Biren and Moore Threads are accelerating production of domestic accelerators. At the same time, reporting from <a href="https://archive.is/8syqZ">The Wall Street Journal</a> and others has detailed how Chinese firms have been circumventing export restrictions by covertly importing restricted U.S. chips via intermediary countries. This chip smuggling ecosystem leverages gray-market suppliers in Southeast Asia and shell companies and underscores the continuing demand for top-tier GPUs, despite Beijing&#8217;s parallel push for domestic alternatives. Meanwhile, technical papers from Tsinghua and CAS show advances in bilingual pretraining, instruction tuning, and state-owned model architectures, often with limited international collaboration or transparency.</p><p>If the US is grappling with how to regulate foundation models, China is already piloting enforcement. A new <a href="https://www.rand.org/pubs/perspectives/PEA4012-1.html">RAND analysis</a> highlights that Beijing's framework emphasizes controllability, data sovereignty, and alignment with socialist values. The report details how China&#8217;s regulatory model is centrally planned but implemented regionally, with the Cyberspace Administration of China setting nationwide model registration rules and provincial authorities like those in Shanghai and Shenzhen enforcing them. It also notes the use of tiered licensing, where model providers must pass government audits and submit outputs for evaluation against political red lines. Developers are expected to pre-train on sanitized datasets and incorporate in-model filters for taboo content. RAND warns that while this framework enables strict enforcement, it may also hinder technical innovation and restrict access to diverse viewpoints needed for robust general-purpose AI.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/p/the-state-of-ai-2025-july?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://press.airstreet.com/p/the-state-of-ai-2025-july?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><div><hr></div><h3><strong>Research papers</strong></h3><p><strong><a href="https://jeremywohlwend.com/assets/boltz2.pdf">Boltz-2: Towards Accurate and Efficient Binding Affinity Prediction</a></strong>, MIT CSAIL, Recursion, Valence Labs</p><p>In this paper, the authors present Boltz-2, a structural biology foundation model that advances both structure and binding affinity prediction for biomolecules. The model demonstrates improved structure prediction across various modalities and can better capture local protein dynamics through experimental method conditioning.</p><p>Most significantly, Boltz-2 approaches the accuracy of free-energy perturbation methods for predicting binding affinities on benchmarks like FEP+ and CASP16, while being 1000&#215; more computationally efficient. In virtual screening tests against the TYK2 target, Boltz-2 coupled with a generative model successfully identified novel, high-affinity binders.</p><p>The authors acknowledge limitations including variability in performance across different targets and dependence on accurate structure prediction for reliable affinity estimates. The model is released under a permissive license.</p><p><strong><a href="https://www.biorxiv.org/content/10.1101/2025.06.25.661532v1.full.pdf">AlphaGenome: advancing regulatory variant effect prediction with a unified DNA sequence model</a></strong>, Google DeepMind.</p><p>In this paper, the authors introduce AlphaGenome, a deep learning model that predicts thousands of functional genomic tracks directly from 1 megabase of DNA sequence at single base-pair resolution. These tracks include gene expression, splicing, chromatin accessibility, histone modifications, transcription factor binding, and 3D chromatin contacts.</p><p>AlphaGenome unifies multimodal prediction, long-range sequence context, and high resolution in a single framework. It is benchmarked against both specialized and generalist models, matching or exceeding the best available models on 24 out of 26 variant effect prediction tasks and 22 out of 24 genome track prediction tasks. Notably, it outperforms Borzoi and Enformer on eQTL effect prediction and surpasses specialized models like SpliceAI and ChromBPNet on splicing and chromatin accessibility tasks.</p><p>The model&#8217;s architecture leverages a U-Net-style encoder-decoder with transformers, and is trained using a two-stage process involving pretraining and distillation. The authors highlight that AlphaGenome&#8217;s unified approach enables efficient, simultaneous variant effect prediction across modalities, which is valuable for interpreting non-coding variants in disease, rare variant diagnostics, and large-scale genome analysis. Caveats include challenges in modeling very distal regulatory elements and tissue-specific effects, and the model&#8217;s current focus on human and mouse genomes.</p><p><strong><a href="https://arxiv.org/pdf/2506.13642">Stream-Omni: Simultaneous Multimodal Interactions with Large Language-Vision-Speech Model</a></strong>, Chinese Academy of Sciences.</p><p>In this paper, the authors introduce Stream-Omni, a model for more efficient multimodal interactions across text, vision, and speech. The key idea is to align modalities based on their relationship: using standard concatenation for vision-text alignment and a novel layer-dimension mapping for speech-text alignment.</p><p>This approach allows the model to achieve strong performance using only 23,000 hours of speech data, significantly less than many comparable models. It performs competitively on 11 visual understanding benchmarks (64.7 average) and knowledge-based spoken question answering (60.3 average accuracy for speech-to-text).</p><p>The model's architecture allows it to simultaneously produce intermediate text transcriptions during speech interaction. This is relevant for creating more transparent and seamless real-world applications, such as interactive assistants, where users can see what the model is hearing in real-time.</p><p><strong><a href="https://arxiv.org/pdf/2506.09376">Revisiting Diffusion Models: From Generative Pre-training to One-Step Generation</a></strong>, Tsinghua University, Shanghai Jiao Tong University, Shanghai Artificial Intelligence Laboratory.</p><p>In this paper, the authors propose a new perspective on diffusion models, viewing them as generative pre-training that can be efficiently converted to one-step generators. They identify a key limitation in traditional diffusion distillation: teacher and student models converge to different local minima, making direct imitation suboptimal. To solve this, they develop D2O (Diffusion to One-Step), which uses only a GAN objective without distillation losses.</p><p>Their most striking finding is that D2O-F (with 85% of parameters frozen during fine-tuning) achieves state-of-the-art results with minimal training data - requiring only 5M images to reach FID=1.16 on ImageNet 64x64 and FID=0.85 on FFHQ, while competing methods need hundreds of millions of images.</p><p>This could lead to significantly reduced computational resources for high-quality image generation, making these capabilities more accessible while revealing that diffusion models inherently contain one-step generation abilities that just need to be unlocked.</p><p><strong><a href="https://arxiv.org/pdf/2506.06105">Text-to-LoRA: Instant Transformer Adaption</a></strong>, Sakana AI.</p><p>In this paper, the authors introduce Text-to-LoRA (T2L), a model that generates task-specific adapters for LLMs using only a natural language description. Instead of traditional fine-tuning, T2L is a hypernetwork that produces a Low-Rank Adaptation (LoRA) in a single, inexpensive forward pass.</p><p>When trained on 479 tasks, T2L was tested on 10 unseen benchmarks. It generated useful LoRAs that outperformed a multi-task baseline (e.g., 67.7% vs. 66.3% average accuracy) and was over four times more computationally efficient than 3-shot in-context learning.</p><p>A key caveat is that performance is sensitive to the quality of the text description. This research matters because it lowers the barrier for specializing foundation models, enabling users to adapt an AI for a new purpose simply by describing the task, which is useful for rapid, on-the-fly customization.</p><p><strong><a href="https://www.arxiv.org/pdf/2505.24832">How much do language models memorize?</a></strong>, Meta FAIR, Google DeepMind, Cornell University</p><p>In this paper, the authors propose a new method for estimating language model memorization by separating it into unintended memorization (information about specific datasets) and generalization (information about the data-generation process).</p><p>The researchers trained hundreds of transformers (500K to 1.5B parameters) on synthetic and real data, discovering that GPT-family models have a capacity of approximately 3.6 bits-per-parameter.</p><p>Their experiments reveal that models memorize until their capacity fills, after which "grokking" begins - unintended memorization decreases as models start to generalize. The double descent phenomenon occurs precisely when dataset size exceeds model capacity.</p><p>The authors developed scaling laws showing that membership inference difficulty increases with dataset size and decreases with model capacity, predicting that most modern language models train on too much data for reliable membership inference.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://press.airstreet.com/subscribe?"><span>Subscribe now</span></a></p><p><strong><a href="https://arxiv.org/pdf/2506.17238">Training a scientific reasoning model for chemistry</a></strong>, FutureHouse</p><p>In this paper, the authors present ether0, a 24-billion-parameter reasoning model designed for chemical tasks, demonstrating that RL can enable LLMs to perform complex scientific reasoning. The model was trained on 640,730 chemistry problems across 375 tasks, including molecular design, synthesis, and property prediction, using a combination of supervised fine-tuning and RL with verifiable rewards.</p><p>The experiments show ether0 outperforming general-purpose models, domain-specific models, and even human experts on open-ended tasks like retrosynthesis and SMILES generation. Notably, the model achieves higher accuracy with less data compared to traditional models, highlighting its efficiency. The authors also analyze the emergence of reasoning behaviors, such as backtracking and verification, which improve task performance.</p><p>While the model excels in organic chemistry, it struggles with tasks outside its training distribution, such as inorganic chemistry.</p><p><strong><a href="https://arxiv.org/pdf/2506.10943">Self-Adapting Language Models</a></strong>, MIT</p><p>In this paper, the authors introduce SEAL, a framework that enables LLMs to self-adapt by generating their own finetuning data and update directives. The approach uses RL to train models to produce &#8220;self-edits&#8221;, which are instructions for how to restructure or augment training data and select optimization parameters, such that subsequent weight updates improve downstream performance.</p><p>The authors evaluate SEAL in two domains: knowledge incorporation and few-shot learning. In knowledge incorporation, SEAL improves no-context SQuAD accuracy from 33.5% (finetuning on passage only) to 47.0%, outperforming synthetic data generated by GPT-4.1. In few-shot learning on ARC tasks, SEAL achieves a 72.5% adaptation success rate, compared to 20% for non-RL self-edits and 0% for in-context learning.</p><p>The paper, however, notes that SEAL is still susceptible to catastrophic forgetting and incurs higher computational costs due to its inner-loop finetuning.</p><p><strong><a href="https://arxiv.org/pdf/2506.09985">V-JEPA 2: Self-Supervised Video Models Enable Understanding, Prediction and Planning</a></strong>, Meta FAIR, Mila, Polytechnique Montr&#233;al</p><p>In this paper, the authors present V-JEPA 2, a self-supervised video model designed to understand, predict, and plan in the physical world. The model is pre-trained on over 1 million hours of internet-scale video and 1 million images, using a mask-denoising objective to predict representations of masked video segments. V-JEPA 2 achieves strong performance on motion understanding tasks, such as 77.3% top-1 accuracy on Something-Something v2, and state-of-the-art results in human action anticipation with 39.7 recall-at-5 on Epic-Kitchens-100.</p><p>The authors also align V-JEPA 2 with a large language model, achieving state-of-the-art results on video question-answering benchmarks like PerceptionTest (84.0%) and TempCompass (76.9%). Additionally, they extend the model to V-JEPA 2-AC, an action-conditioned world model trained on 62 hours of robot interaction data, enabling zero-shot robotic manipulation tasks like pick-and-place.</p><p><strong><a href="https://www.nature.com/articles/s43018-025-00991-6">Development and validation of an autonomous artificial intelligence agent for clinical decision-making in oncology</a></strong>, Heidelberg University Hospital and Technical University Dresden</p><p>In this paper, the authors develop and evaluate an autonomous AI agent for clinical decision-making in oncology, integrating GPT-4 with multimodal precision oncology tools. The system combines language modeling with vision transformers for detecting microsatellite instability and KRAS/BRAF mutations from histopathology, MedSAM for radiological image segmentation, and web-based search tools like OncoKB, PubMed, and Google.</p><p>Benchmarked on 20 realistic multimodal patient cases, the agent autonomously selected and used appropriate tools with 87.5% accuracy, reached correct clinical conclusions in 91% of cases, and accurately cited relevant guidelines 75.5% of the time. Compared to GPT-4 alone, which achieved only 30.3% completeness, the integrated agent reached 87.2%.</p><p><strong><a href="https://arxiv.org/pdf/2506.22405">Sequential Diagnosis with Language Models</a></strong>, Microsoft AI</p><p>In this paper, the authors introduce the Sequential Diagnosis Benchmark (SDBench), which transforms 304 challenging New England Journal of Medicine cases into interactive, stepwise diagnostic tasks. Unlike static vignettes, SDBench requires agents (human or AI) to iteratively ask questions, order tests, and make cost-sensitive decisions, closely mirroring real clinical workflows.</p><p>The authors present the MAI Diagnostic Orchestrator (MAI-DxO), a model-agnostic system that simulates a panel of virtual physicians, each with specialized roles, to collaboratively refine diagnoses and select high-value tests. When paired with OpenAI&#8217;s o3 model, MAI-DxO achieves 80% diagnostic accuracy&#8212;four times higher than the 20% average of experienced physicians&#8212;while reducing diagnostic costs by 20% compared to physicians and 70% compared to off-the-shelf o3.</p><p>The research highlights that structured, multi-agent orchestration can improve both accuracy and cost-efficiency in AI-driven diagnosis, suggesting practical applications for clinical decision support and resource-limited healthcare settings.</p><p><strong><a href="https://arxiv.org/pdf/2506.08388">Reinforcement Learning Teachers of Test Time Scaling</a></strong>, Sakana AI</p><p>In this paper, the authors introduce Reinforcement-Learned Teachers (RLTs), a new framework for training language models to generate high-quality reasoning traces for downstream distillation, rather than solving problems from scratch. Unlike traditional RL approaches that rely on sparse, correctness-based rewards, RLTs are trained with dense rewards by providing both the question and its solution, and optimizing the model to produce explanations that help a student model learn.</p><p>Experiments show that a 7B parameter RLT can outperform existing distillation pipelines that use much larger models, both in training smaller students and in cold-starting RL for future iterations. Benchmarks on math and science tasks (AIME, MATH, GPQA) demonstrate higher or comparable accuracy with less computational cost. The study also finds that RLTs transfer well to new domains without retraining.</p><p>This research matters because it offers a more efficient and reusable way to generate reasoning data for training and improving language models, potentially lowering the barrier for developing strong AI systems in real-world applications.</p><p><strong><a href="https://arxiv.org/pdf/2506.14111">ESSENTIAL-WEB V1.0: 24T tokens of organized web data</a></strong>, Essential AI</p><p>In this paper, the authors introduce ESSENTIAL-WEB V1.0, a 24-trillion-token dataset annotated with a 12-category taxonomy covering topics, content complexity, and quality. The dataset enables rapid, SQL-like filtering to curate domain-specific corpora for math, code, STEM, and medical domains without bespoke pipelines.</p><p>Experiments show that taxonomy-curated datasets perform competitively with or surpass state-of-the-art (SOTA) alternatives. For instance, the taxonomy-based math dataset achieves results within 8% of SOTA on GSM8K, while STEM and web code datasets outperform SOTA by 24.5% and 14.3%, respectively. The medical dataset improves accuracy by 8.6% over existing baselines.</p><p>The authors also develop EAI-Distill-0.5b, a 0.5B-parameter classifier that labels documents 50x faster than its teacher model, Qwen2.5-32B-Instruct, while maintaining high annotation quality.</p><p>This research matters because it democratizes access to high-quality, domain-specific datasets, reducing the cost and complexity of training AI models. Real-world applications include improving LLMs for education, healthcare, and technical domains.</p><p><strong><a href="https://arxiv.org/pdf/2506.19143">Thought Anchors: Which LLM Reasoning Steps Matter?</a></strong> Duke University, Alphabet</p><p>In this paper, the authors investigate reasoning processes in large language models (LLMs) by analyzing sentence-level reasoning traces. They introduce three methods: black-box resampling, white-box attention aggregation, and causal attribution through attention suppression. These methods identify "thought anchors," critical reasoning steps that disproportionately influence subsequent reasoning and final answers.</p><p>The study finds that sentences related to planning and uncertainty management have higher counterfactual importance than those focused on computation or fact retrieval. Receiver attention heads, which focus on specific sentences, are more prevalent in reasoning models and play a significant role in structuring reasoning traces. Ablating these heads reduces model accuracy more than random head ablation, highlighting their importance.</p><p>This research provides tools for debugging and improving reasoning models, with potential applications in enhancing model reliability and interpretability. It is particularly relevant for tasks requiring multi-step reasoning, such as mathematical problem-solving or complex decision-making in real-world scenarios.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/p/the-state-of-ai-2025-july?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://press.airstreet.com/p/the-state-of-ai-2025-july?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><div><hr></div><h3><strong>Investments</strong></h3><p><a href="https://techcrunch.com/2025/06/05/tomas-ai-voice-agents-have-taken-off-at-car-dealerships-and-attracted-funding-from-a16z/">Toma</a>, the AI voice-agent company for car dealerships, raised a $17 M Series A financing round from a16z and Y Combinator.</p><p><a href="https://techcrunch.com/2025/06/05/anduril-raises-2-5b-at-30-5b-valuation-led-by-founders-fund/">Anduril</a>, the US defense company, raised a $2.5B financing round at a $30.5 billion valuation. The company has been making noise recently about going public soon.</p><p><a href="https://x.com/morganstanley/status/1939768047780172184">xAI</a>, Elon&#8217;s AI company, raised $5B in a financing round from &#8220;prominent global debt investors&#8221; facilitated by Morgan Stanley and separately obtained a $5B strategic equity investment.</p><p><a href="https://www.reuters.com/business/thrive-backed-accounting-firm-crete-spend-500-million-ai-roll-up-2025-06-04/">Crete Professionals Alliance</a>, an AI-driven accounting platform, raised a few-hundred-million-dollar round from Thrive Capital, ZBS Partners and Bessemer Venture Partners.<br><a href="https://tech.eu/2025/06/10/lithuanian-ai-startup-sintra-secures-17m-seed-empowering-smbs-with-ai-helpers/">Sintra</a>, the Lithuanian AI startup empowering small businesses with AI helpers, raised a $17M seed round from Earlybird VC, Inovo and Practica Capital.</p><p><a href="https://www.foodbev.com/news/seafood-robotics-company-shinkei-systems-secures-22m-in-series-a-funding/">Shinkei Systems</a>, a seafood-robotics company integrating advanced robotics and AI with traditional fishing methods, raised $22M in a Series A co-led by Founders Fund and Interlagos.</p><p><a href="https://www.skyramp.dev/blog-all/skyramp-launch">Skyramp</a>, the AI-driven software-testing-automation company, raised a $10M seed round led by Sequoia Capital.</p><p><a href="https://crusoe.ai/newsroom/crusoe-secures-usd750-million-credit-facility-from-brookfield-to-accelerate/">Crusoe</a>, a cloud infrastructure startup focused on AI data centers, raised a $750M credit line from Brookfield Asset Management.</p><p><a href="https://siliconangle.com/2025/06/13/yupp-launches-33m-build-crypto-incentivized-ai-evaluation-platform/">Yupp</a>, a platform for crypto-incentivized AI-model evaluation, raised a $33M seed round led by a16z crypto with participation from Jeff Dean and Biz Stone.</p><p><a href="https://www.airforce-technology.com/news/gecko-double-valuation-to-1-25bn-reaching-unicorn-status/">Gecko Robotics</a>, the Pittsburgh company using AI and robotics to modernize maintenance techniques in defense, raised a $1.25B Series D led by Cox Enterprises with USIT and Founders Fund.</p><p><a href="https://www.axios.com/2025/05/22/cx2-funding-mintz-electronic-warfare">CX2</a>, a defense-technology company developing intelligent multi-domain electronic-warfare capabilities, raised a $31M Series A led by Point72 Ventures with Andreessen Horowitz and 8VC.</p><p><a href="https://www.ft.com/content/cdc02d96-13b5-4ca2-aa0b-1fc7568e9fa0">Helsing</a>, the German defense AI company, raised &#8364;600M in a round led by Spotify&#8217;s Daniel Ek.</p><p><a href="https://www.prnewswire.com/news-releases/nabla-raises-70m-series-c-to-deliver-agentic-ai-to-the-heart-of-clinical-workflows-bringing-total-funding-to-120m-302483646.html">Nabla</a>, the clinical AI assistant, raised a $70M Series C from HV Capital and Highland Europe.<br><a href="https://www.upstartsmedia.com/p/browserbase-raises-40m-and-launches-director?hide_intro_popup=true">Browserbase</a>, the infrastructure startup behind headless browsers, raised a $40M Series B at a $300M valuation from Notable Capital, Kleiner Perkins and CRV.</p><p><a href="https://www.prnewswire.com/news-releases/ramp-raises-200m-series-e-at-16b-valuation-as-companies-of-all-sizes-choose-ai-powered-finance-platform-302483377.html">Ramp</a>, the spend management platform, raised a $200M Series E at a $16B valuation from Founders Fund, Thrive Capital and General Catalyst.</p><p><a href="https://www.appliedintuition.com/blog/series-f">Applied Intuition</a>, a pioneer of AI simulation software for autonomy in transportation and defense, raised a Series F at a $15B valuation from BlackRock and Kleiner Perkins.</p><p><a href="https://www.tryprofound.com/blog/series-a">Profound</a>, the platform helping marketers optimize presence in AI responses, raised a $20M Series A from Kleiner Perkins, Khosla Ventures and NVIDIA NVentures.</p><p><a href="https://www.mavenagi.com/resources/post/series-b">Maven AGI</a>, a customer experience AI company, raised a $50M Series B from Dell Technologies Capital, Cisco Investments and SE Ventures.</p><p><a href="https://www.ft.com/content/9edc67e6-96a9-4d2b-820d-57bc1279e358">Thinking Machines Lab</a>, Mira Murati&#8217;s AGI company, raised $2B at a $10B valuation led by Andreessen Horowitz.</p><p><a href="https://www.commure.com/blog/commure-secures-200m-to-accelerate-ai-powered-healthcare-transformation">Commure</a>, the AI-powered healthcare company, raised $200M in growth capital from General Catalyst&#8217;s CVF.</p><p><a href="https://decagon.ai/resources/series-c-announcement">Decagon</a>, the conversational AI company, raised a $131M Series C at a $1.5B valuation from Accel and Andreessen Horowitz.</p><p><a href="https://sifted.eu/articles/exclusive-genesis-robotics-85m-round">Genesis Robotics</a>, a full-stack robotics company built around the generative physics engine by the <a href="https://genesis-embodied-ai.github.io/">same name</a>, raised an $85M round co-led by Khosla Ventures and Eclipse Ventures.</p><p><a href="https://techcrunch.com/2025/06/24/in-just-4-months-ai-medical-scribe-abridge-doubles-valuation-to-5-3b">Abridge</a>, the medical notes automation startup, raised a $300M Series E at a $5.3B valuation from Andreessen Horowitz and Khosla Ventures.</p><p><a href="https://www.globenewswire.com/news-release/2025/06/25/3105125/0/en/OpenRouter-raises-40-million-to-scale-up-multi-model-inference-for-enterprise.html">OpenRouter</a>, the unified interface for LLM inference, raised $40M across seed and Series A led by Andreessen Horowitz and Menlo Ventures.</p><p><a href="https://fortune.com/2025/06/25/uber-palantir-alums-metaview-raise-35m-ai-revolution-recruitment-hiring/">Metaview</a>, the AI recruitment tech company, raised a $35M Series B led by Google Ventures with Plural and Vertex Ventures.</p><p><a href="https://techcrunch.com/2025/06/24/wispr-flow-raises-30m-from-menlo-ventures-for-its-ai-powered-dictation-app/">Wispr Flow</a>, the AI-powered dictation app, raised a $30M Series A from Menlo Ventures and NEA.</p><p><a href="https://tech.eu/2025/06/24/sovereign-by-design-lyceum-emerges-with-eur103m-to-redefine-cloud-infrastructure-in-europe/">Lyceum</a>, a &#8220;sovereign&#8221; cloud provider for AI, raised a &#8364;10.3M pre-seed led by Redalpine with 10x Founders.</p><p><a href="https://blog.nominal.io/series-b">Nominal</a>, modernizing hardware testing, raised a $75M Series B led by Sequoia Capital.</p><p><a href="https://www.glean.com/press/glean-raises-150m-series-f-at-7-2b-valuation-to-accelerate-enterprise-ai-agent-innovation-globally?utm_source=chatgpt.com">Glean</a>, the enterprise search company, raised a $150M Series F at a $7.2B valuation led by Wellington Management.</p><p><a href="https://www.globenewswire.com/news-release/2025/06/16/3099902/0/en/Wildfire-Tech-Comes-of-Age-Pano-AI-Raises-44M-Series-B-Led-by-Giant-Ventures-to-Scale-Early-Detection-Infrastructure.html?utm_source=chatgpt.com">Pano AI</a>, the wildfire-detection company, raised a $44M Series B from Giant Ventures, Liberty Mutual Strategic Ventures and Tokio Marine Future Fund.</p><p><a href="https://fortune.com/2025/06/18/traversal-emerges-from-stealth-with-48-million-from-sequoia-and-kleiner-perkins-to-reimagine-site-reliability-in-the-ai-era/">Traversal</a>, a startup focused on observability and site reliability engineering, raised $48M in its seed and Series A financing rounds led by Sequoia and Kleiner Perkins.</p><p><a href="https://www.delphi.ai/blog/delphi-raises-16m-series-a-from-sequoia">Delphi</a>, the AI platform for creating interactive "digital minds," raised a $16M Series A from Sequoia Capital, with participation from Menlo &amp; Anthropic&#8217;s Anthology Fund and Proximity Ventures.</p><h3>Rumored investments</h3><p><a href="https://techcrunch.com/2025/07/02/lovable-on-track-to-raise-150m-at-2b-valuation/">Lovable</a>, the Swedish AI startup vibe coding frontend applications, is rumored to be raising $150M at a $2B valuation.</p><p><a href="https://www.ft.com/content/db2b25e0-da61-42d3-932a-991b12e5476a">PhysicsX</a>, the UK physics simulation startup working in the industrial and defense sectors, is nearing a $1B valuation in its latest round.</p><h3><strong>Acquisitions</strong></h3><p><a href="https://investor.qualcomm.com/news-events/press-releases/news-details/2025/Qualcomm-to-Acquire-Alphawave-Semi/default.aspx">Qualcomm</a>, the US chipmaker, acquired Alphawave, a UK-based public company building semiconductors, for $2.4B. The company makes high-speed connectivity and compute chiplets, enabling fast data transfer with lower power consumption for applications like data centers, AI, 5G, and autonomous vehicles. This connectivity IP is said to complement Qualcomm's existing CPU and NPU processors, particularly for AI workloads.</p><p><a href="https://www.clio.com/about/press/clio-signs-definitive-agreement-to-acquire-vlex/">Clio</a>, the legal-tech leader, acquired vLex for $1B in cash and stock. This deal sees Clio bolt vLex&#8217;s AI-powered legal-research engine onto its practice-management suite so lawyers can search the world&#8217;s case law, draft filings, bill clients and track matters inside a single &#8220;legal OS.&#8221; The deal fast-forwards Clio&#8217;s agentic-AI roadmap, lets it sell up-market to large firms and new civil-law jurisdictions, and gives the combined company a proprietary corpus of workflow and primary-law data that can feed its own domain-specific LLMs while trimming licensing costs.</p><p><a href="https://fortune.com/2025/07/02/figma-ipo-s-1-filing-growth-profitability-dual-class-share-structure-dylan-field-nyse-fig/">Figma</a>, the design collaboration software company, filed for an IPO expected to raise up to $1.5B at a $15-20B valuation. This is a huge deal for the tech industry following the company&#8217;s failed $20B acquisition by Adobe due to anti-trust. Figma has since accelerated to almost $800m in revenue with 13M monthly active users and 95% of the Fortune 500 companies on its platform. The company has pushed into generative AI, launching new products such as Make, and weaving an AI assistant into its core surface.</p><p><a href="https://predibase.com/blog/predibase-will-be-joining-forces-with-rubrik">Predibase</a>, the AI company spun out of Chris Re&#8217;s group at Stanford to productise Ludvig, was acquired by Rubrik (a publicly listed cybersecurity company) to accelerate agentic-AI adoption. The price was undisclosed and rumored to be above $100M.</p><p><a href="https://www.ft.com/content/cdc02d96-13b5-4ca2-aa0b-1fc7568e9fa0">Helsing</a>, a German company specializing in AI and software solutions for defense, acquired Grob Aircraft, the producer of the G120TP military trainer, from H3 Aerospace. While the deal price was not disclosed, the rationale looks to be about vertical integration. By bringing a 275-person composite-aircraft factory and its G 120-series trainer line in-house, Helsing gains a purpose-built airframe on which it can iterate and certify its Cirra electronic-warfare AI and other onboard autonomy much faster than if it had to rely on third-party OEMs. The move deepens an existing test-bed partnership, anchors production in Europe and gives Helsing its own hardware-plus-software stack. This is an essential step toward fielding scalable, AI-native surveillance drones and light combat aircraft while reinforcing Europe&#8217;s drive for defence-technology sovereignty.</p><p><a href="https://coreweave.com/news/coreweave-to-acquire-core-scientific">CoreWeave</a>, the AI hyperscaler, acquired Core Scientific, a leading data center infrastructure provider, in an all-stock transaction valued at approximately $9 billion.</p><p><a href="https://x.com/Superhuman/status/1940078856586571811">Superhuman</a>, the email productivity company, was acquired by Grammarly. The acquisition price was not disclosed.</p><p><a href="https://techcrunch.com/2025/06/02/ibm-acquires-data-analysis-startup-seek-ai-opens-ai-accelerator-in-nyc/">Seek AI</a>, enabling natural language queries on enterprise data, was acquired by IBM for an undisclosed price.</p><p><a href="https://www.reuters.com/business/finance/brainlab-ipo-expected-price-80-euros-per-share-bookrunner-says-2025-06-30/">Brainlab</a>, the German med-tech firm that specializes in robotic surgery equipment and medical imaging tools, plans an IPO at &#8364;80 per share valuing the company at &#8364;1.7-2.1B.</p><p><a href="https://snyk.io/news/snyk-acquires-invariant-labs-to-accelerate-agentic-ai-security-innovation/">Snyk</a>, the secure-AI software leader, acquired Invariant Labs for an undisclosed price. The startup was spun out of an ETHZ lab that previously spawned DeepCode, also previously acquired by Snyk. Invariant was focused on productivising research to make agents more secure (e.g. <a href="https://github.com/eth-sri/lmql">LMQL</a>).</p><p>The team behind <a href="https://techcrunch.com/2025/06/27/openai-hires-team-behind-ai-recommendation-startup-crossing-minds/">Crossing Minds</a>, an AI-recommendation startup, was acquired by OpenAI.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://press.airstreet.com/subscribe?"><span>Subscribe now</span></a></p>]]></content:encoded></item><item><title><![CDATA[State of AI Compute Index v4 (June 2025)]]></title><description><![CDATA[Today, we release v4 of the State of AI Report Compute Index in collaboration with Zeta Alpha. You'll now find updated counts as of June 2025 for AI research papers using specific chips, cluster size updates and research topics.]]></description><link>https://press.airstreet.com/p/state-of-ai-compute-index-v4-june-2025</link><guid isPermaLink="false">https://press.airstreet.com/p/state-of-ai-compute-index-v4-june-2025</guid><dc:creator><![CDATA[Air Street Press]]></dc:creator><pubDate>Mon, 30 Jun 2025 13:15:19 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/9448598b-51f0-4172-ae62-5b416faa946d_2148x1192.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Today, we release v4 of the <span class="mention-wrap" data-attrs="{&quot;name&quot;:&quot;State of AI Report&quot;,&quot;id&quot;:1203391,&quot;type&quot;:&quot;pub&quot;,&quot;url&quot;:&quot;https://open.substack.com/pub/stateofai&quot;,&quot;photo_url&quot;:&quot;https://bucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com/public/images/0e112d8a-3be5-43be-9f73-96926db59c7f_229x229.png&quot;,&quot;uuid&quot;:&quot;68b1146d-a707-4d35-88f3-32ef48504557&quot;}" data-component-name="MentionToDOM"></span> Compute Index in collaboration with Zeta Alpha.&nbsp;</p><p>You'll now find updated counts as of June 2025 for AI research papers using chips from NVIDIA, TPUs, Apple, Huawei, AMD, ASICs, FPGAs, and AI semi startups, as well as updates to A100 H100/200 cluster sizes. We also include new data on the most and least commonly used chips for specific research topic areas. </p><div><hr></div><h3><strong>A breather at the peak</strong></h3><p>2024 was a blockbuster year for AI research: AlphaFold 3, Llama 3, new synthetic environments and many frontier model releases. According to our analysis using Zeta Alpha, there were <strong>nearly 49,000 open source AI research papers</strong> that mentioned the use of a specific AI accelerator (up +58% YoY) such as NVIDIA, TPUs, AMD and more.</p><p>2025, however, looks a bit different, at least so far. Based on counts through June 1, 2025, and a volume-adjusted projection for the rest of the year, we expect <strong>a full-year total of 43,300 papers</strong> citing NVIDIA, AMD, TPUs, and large AI chip startups. This is an <strong>11% decline</strong> from the prior year, the first such drop in six years.</p><p>While some of this slowdown can be attributed to timing mismatches between hardware deployment and publication, it likely also reflects a growing reluctance among large industrial AI labs to publish their latest work. As competitive pressures and safety concerns mount, many frontier research groups are opting to keep breakthroughs private or delay publication, impacting overall visibility in the open-source paper ecosystem.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!raJP!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F66fe9440-2cbd-4fa0-b70c-c1764ac494d4_1378x670.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!raJP!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F66fe9440-2cbd-4fa0-b70c-c1764ac494d4_1378x670.png 424w, https://substackcdn.com/image/fetch/$s_!raJP!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F66fe9440-2cbd-4fa0-b70c-c1764ac494d4_1378x670.png 848w, https://substackcdn.com/image/fetch/$s_!raJP!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F66fe9440-2cbd-4fa0-b70c-c1764ac494d4_1378x670.png 1272w, https://substackcdn.com/image/fetch/$s_!raJP!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F66fe9440-2cbd-4fa0-b70c-c1764ac494d4_1378x670.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!raJP!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F66fe9440-2cbd-4fa0-b70c-c1764ac494d4_1378x670.png" width="1378" height="670" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/66fe9440-2cbd-4fa0-b70c-c1764ac494d4_1378x670.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:670,&quot;width&quot;:1378,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!raJP!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F66fe9440-2cbd-4fa0-b70c-c1764ac494d4_1378x670.png 424w, https://substackcdn.com/image/fetch/$s_!raJP!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F66fe9440-2cbd-4fa0-b70c-c1764ac494d4_1378x670.png 848w, https://substackcdn.com/image/fetch/$s_!raJP!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F66fe9440-2cbd-4fa0-b70c-c1764ac494d4_1378x670.png 1272w, https://substackcdn.com/image/fetch/$s_!raJP!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F66fe9440-2cbd-4fa0-b70c-c1764ac494d4_1378x670.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Overall, though, NVIDIA remains dominant: its chips appear in <strong>nearly 90%</strong> of all cited compute mentions in 2025. But that share has drifted down from a 2023 peak of 94%, and total NVIDIA-citing papers are forecast to drop to <strong>38,735</strong> this year, down <strong>13% YoY</strong>. Meanwhile, <strong>AMD is growing nearly 100% YoY</strong>, buoyed by MI300X deployments, while <strong>Google TPUs show a slight YoY decline</strong>, despite the introduction of v6.</p><p>So what does this mean? We think it is mostly a story of cycle timing rather than momentum loss. Research work kicked off on H100 and H200 clusters in late 2024 won't be published until late Q3 or Q4 2025. In parallel, we see growing use of shared compute clouds and managed APIs, where authors are less likely to specify underlying silicon. And for the academic long tail, rising GPU costs have made full-stack training harder to justify, especially for teams without access to institutional compute credits. This, along with the increasing availability of strong open-weight models, better small-scale baselines, and evolving norms around reproducibility, could be nudging more researchers toward inference and lightweight fine-tuning workflows, tilting them toward open-weight inference.</p><h3><strong>NVIDIA is still king, but might be tapering</strong></h3><p>The composition of individual NVIDIA chip mentions is shifting in subtle ways:</p><ul><li><p><strong>H100/H200 mentions are up 115% YoY</strong>, reflecting late-stage adoption of 2023-24 builds, even as growth moderates into 2025.</p></li><li><p><strong>Jetson citations are up 33% YoY</strong>, possibly due to robotics and edge AI interest in low-power inference.</p></li><li><p><strong>V100 usage is declining for another year since its peak in 2023</strong></p></li><li><p><strong>RTX 3090 is losing ground from its peak in 2024 to the 4090, which is growing 30% YoY.</strong></p></li></ul><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!IBEN!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7d5ce698-7825-4d8d-b352-2c75b172f7a1_1394x688.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!IBEN!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7d5ce698-7825-4d8d-b352-2c75b172f7a1_1394x688.png 424w, https://substackcdn.com/image/fetch/$s_!IBEN!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7d5ce698-7825-4d8d-b352-2c75b172f7a1_1394x688.png 848w, https://substackcdn.com/image/fetch/$s_!IBEN!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7d5ce698-7825-4d8d-b352-2c75b172f7a1_1394x688.png 1272w, https://substackcdn.com/image/fetch/$s_!IBEN!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7d5ce698-7825-4d8d-b352-2c75b172f7a1_1394x688.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!IBEN!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7d5ce698-7825-4d8d-b352-2c75b172f7a1_1394x688.png" width="1394" height="688" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/7d5ce698-7825-4d8d-b352-2c75b172f7a1_1394x688.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:688,&quot;width&quot;:1394,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!IBEN!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7d5ce698-7825-4d8d-b352-2c75b172f7a1_1394x688.png 424w, https://substackcdn.com/image/fetch/$s_!IBEN!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7d5ce698-7825-4d8d-b352-2c75b172f7a1_1394x688.png 848w, https://substackcdn.com/image/fetch/$s_!IBEN!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7d5ce698-7825-4d8d-b352-2c75b172f7a1_1394x688.png 1272w, https://substackcdn.com/image/fetch/$s_!IBEN!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7d5ce698-7825-4d8d-b352-2c75b172f7a1_1394x688.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/p/state-of-ai-compute-index-v4-june-2025?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://press.airstreet.com/p/state-of-ai-compute-index-v4-june-2025?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><h3><strong>Research topic signals by AI chip</strong></h3><p>We then classified each paper by its research topic in order to observe topic trends associated with specific accelerators. The dataset covers 6,356 papers published between January 1 and June 1, 2025. Topic labels were assigned using GPT-4o-mini, and we filtered the results to include only chip-topic pairs with at least three papers. To focus the analysis on meaningful differentiation, we excluded topics whose percentage difference fell between -1 and +1 compared to the corpus-wide baseline. The results reveal clear skews between specific chips and research areas.</p><p>Of note, LLM-focused research papers are most commonly using the AMD MI300, MI250, Huawei Ascend and NVIDIA H100/H200 chips. Meanwhile, robotics research overwhelmingly uses the NVIDIA Jetson.</p><p>By contrast, LLM-focused research papers least frequently use ASICs, the NVIDIA Jetson, the NVIDIA 4090 and the Apple M1. Diffusion models are not commonly using FPGAs, either.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!69cL!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6eb4b7ea-7139-4142-b849-740b900e5d76_1252x948.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!69cL!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6eb4b7ea-7139-4142-b849-740b900e5d76_1252x948.png 424w, https://substackcdn.com/image/fetch/$s_!69cL!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6eb4b7ea-7139-4142-b849-740b900e5d76_1252x948.png 848w, https://substackcdn.com/image/fetch/$s_!69cL!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6eb4b7ea-7139-4142-b849-740b900e5d76_1252x948.png 1272w, https://substackcdn.com/image/fetch/$s_!69cL!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6eb4b7ea-7139-4142-b849-740b900e5d76_1252x948.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!69cL!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6eb4b7ea-7139-4142-b849-740b900e5d76_1252x948.png" width="1252" height="948" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/6eb4b7ea-7139-4142-b849-740b900e5d76_1252x948.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:948,&quot;width&quot;:1252,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!69cL!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6eb4b7ea-7139-4142-b849-740b900e5d76_1252x948.png 424w, https://substackcdn.com/image/fetch/$s_!69cL!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6eb4b7ea-7139-4142-b849-740b900e5d76_1252x948.png 848w, https://substackcdn.com/image/fetch/$s_!69cL!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6eb4b7ea-7139-4142-b849-740b900e5d76_1252x948.png 1272w, https://substackcdn.com/image/fetch/$s_!69cL!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6eb4b7ea-7139-4142-b849-740b900e5d76_1252x948.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>If we look at each major research topic individually and ask which chips are more/less popular compared to all chips, here are the results:</p><ul><li><p><strong>3D models</strong>: most popular with NVIDIA 4090 (+2.91 pp) and least popular with <strong>FPGAs</strong> (-4.76 pp)</p></li><li><p><strong>Computer</strong> <strong>vision</strong>: most popular with Jetson (+3.42 pp) and least popular with H100/H200 (-4.10 pp)</p></li><li><p><strong>Diffusion</strong> <strong>models</strong>: most associated with Huawei Ascend (+4.25 pp), least associated with FPGAs (-5.62 pp)</p></li><li><p><strong>Edge</strong> <strong>computing</strong>: most associated with Jetson (+8.94 pp)</p></li><li><p><strong>LLMs</strong>: most strongly associated with MI300 (+42.53 pp), least associated with ASICs (-9.77 pp)</p></li><li><p><strong>Multimodal models</strong>: most associated with the M4 (+8.32 pp) and least associated with the Jetson (-1.72 pp)</p></li><li><p><strong>Post-quantum</strong> <strong>cryptography</strong>: most associated with ASICs (+3.28 pp)</p></li><li><p><strong>Quantization</strong> and <strong>reasoning</strong>: each most associated with the MI300 (+8.38 pp an +7.14 pp, respectively) and least associated with the Jetson (1.17 pp and -1.72 pp, respectively)</p></li><li><p><strong>Reinforcement</strong> <strong>learning</strong>: most popular with the Huawei Ascend (+7.78 pp) and the NVIDIA Jetson (4.19 pp) and least popular with FPGAs (-2.97 pp)</p></li><li><p><strong>Robotics</strong>: most associated with Jetson (+19.59 pp), least associated with V100 (-3.32 pp) and the H100/H200 (-2.87 pp)</p></li><li><p><strong>Speech</strong>: most associated with the M4 (+7.26 pp)</p></li></ul><h3><strong>Startup chips: still a niche, but the fastest-growing one</strong></h3><p>The most bullish trend in our dataset can be found amongst the startup silicon cohort: Cerebras, Groq, Graphcore, SambaNova, Cambricon, and Habana. Collectively, they show <strong>+19% YoY</strong> growth, with an estimated <strong>695 papers</strong> citing their hardware in 2025, up from 586 in 2024.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!P3vJ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F465db1a7-af11-4feb-beac-992670ab81d4_1390x682.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!P3vJ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F465db1a7-af11-4feb-beac-992670ab81d4_1390x682.png 424w, https://substackcdn.com/image/fetch/$s_!P3vJ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F465db1a7-af11-4feb-beac-992670ab81d4_1390x682.png 848w, https://substackcdn.com/image/fetch/$s_!P3vJ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F465db1a7-af11-4feb-beac-992670ab81d4_1390x682.png 1272w, https://substackcdn.com/image/fetch/$s_!P3vJ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F465db1a7-af11-4feb-beac-992670ab81d4_1390x682.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!P3vJ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F465db1a7-af11-4feb-beac-992670ab81d4_1390x682.png" width="1390" height="682" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/465db1a7-af11-4feb-beac-992670ab81d4_1390x682.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:682,&quot;width&quot;:1390,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!P3vJ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F465db1a7-af11-4feb-beac-992670ab81d4_1390x682.png 424w, https://substackcdn.com/image/fetch/$s_!P3vJ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F465db1a7-af11-4feb-beac-992670ab81d4_1390x682.png 848w, https://substackcdn.com/image/fetch/$s_!P3vJ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F465db1a7-af11-4feb-beac-992670ab81d4_1390x682.png 1272w, https://substackcdn.com/image/fetch/$s_!P3vJ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F465db1a7-af11-4feb-beac-992670ab81d4_1390x682.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Their absolute share, however, remains tiny: <strong>1.6% of all papers mentioning any startup challenger AI accelerator</strong>. Even so, their trajectories are notable:</p><ul><li><p><strong>Cerebras WSE-3</strong> benefits from open-sourcing large-scale SlimPajamas-2 training runs</p></li><li><p><strong>Groq's LPU</strong> attracts a wave of interest from academic inference work, driven by viral low-latency demos</p></li><li><p><strong>Habana Gaudi-2</strong> continues to show up in AWS-funded academic projects</p></li><li><p><strong>Graphcore is collapsing post-acquisition</strong>, with 2025 paper mentions down sharply from their 2022 peak.</p></li></ul><h3><strong>Updated: Major H100 and GH200 cluster deployments</strong></h3><p>We've also updated our tracking of large-scale A100, H100/200 deployments with several new high-profile additions:</p><ul><li><p><strong>Berzelius (Sweden):</strong> 752 H100 at Link&#246;ping University <a href="https://kaw.wallenberg.org/en/press/supercomputer-berzelius-be-upgraded-double-capacity">source</a></p></li><li><p><strong>Shaheen III Phase-2 (Saudi Arabia):</strong> 2,800 H100 at KAUST <a href="https://www.hpc.kaust.edu.sa/sites/default/files/2025-02/2025-02-04-1-Shaheen_III_Intro_Hardware_Environment.pdf">source</a></p></li><li><p><strong>Israel-1 (Israel):</strong> NVIDIA R&amp;D cluster with 2,048 H100 <a href="https://www.new-techeurope.com/2024/11/24/nvidias-israel-1-ranks-no-34-in-the-top500-list-of-the-worlds-most-powerful-supercomputers/">source</a></p></li><li><p><strong>Microsoft Eagle (USA):</strong> 14,400 H100 <a href="https://top500.org/lists/top500/2025/06/">source</a></p></li><li><p><strong>JUPITER Booster (Germany):</strong> 24,000 GH200 <a href="https://www.fz-juelich.de/en/jupiter">source</a></p></li><li><p><strong>NVIDIA Eos (USA):</strong> 4,608 H100 <a href="https://blogs.nvidia.com/blog/eos/#:~:text=Eos%20Supercomputer%20Fuels%20Innovation,to%20autonomous%20machines%20and%20beyond.">source</a></p></li></ul><p>These systems are expected to influence both model training scale and downstream research pipelines throughout late 2025 and into 2026.</p><h4>A100 clusters:</h4><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!CgDv!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89ab99c7-abac-4a1a-bde9-df70ef66fd5c_1406x678.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!CgDv!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89ab99c7-abac-4a1a-bde9-df70ef66fd5c_1406x678.png 424w, https://substackcdn.com/image/fetch/$s_!CgDv!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89ab99c7-abac-4a1a-bde9-df70ef66fd5c_1406x678.png 848w, https://substackcdn.com/image/fetch/$s_!CgDv!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89ab99c7-abac-4a1a-bde9-df70ef66fd5c_1406x678.png 1272w, https://substackcdn.com/image/fetch/$s_!CgDv!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89ab99c7-abac-4a1a-bde9-df70ef66fd5c_1406x678.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!CgDv!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89ab99c7-abac-4a1a-bde9-df70ef66fd5c_1406x678.png" width="1406" height="678" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/89ab99c7-abac-4a1a-bde9-df70ef66fd5c_1406x678.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:678,&quot;width&quot;:1406,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!CgDv!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89ab99c7-abac-4a1a-bde9-df70ef66fd5c_1406x678.png 424w, https://substackcdn.com/image/fetch/$s_!CgDv!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89ab99c7-abac-4a1a-bde9-df70ef66fd5c_1406x678.png 848w, https://substackcdn.com/image/fetch/$s_!CgDv!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89ab99c7-abac-4a1a-bde9-df70ef66fd5c_1406x678.png 1272w, https://substackcdn.com/image/fetch/$s_!CgDv!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89ab99c7-abac-4a1a-bde9-df70ef66fd5c_1406x678.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h4>H100 clusters:</h4><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!zffK!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3f014a14-bdee-4aad-8abb-e79eb4f20974_1394x680.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!zffK!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3f014a14-bdee-4aad-8abb-e79eb4f20974_1394x680.png 424w, https://substackcdn.com/image/fetch/$s_!zffK!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3f014a14-bdee-4aad-8abb-e79eb4f20974_1394x680.png 848w, https://substackcdn.com/image/fetch/$s_!zffK!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3f014a14-bdee-4aad-8abb-e79eb4f20974_1394x680.png 1272w, https://substackcdn.com/image/fetch/$s_!zffK!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3f014a14-bdee-4aad-8abb-e79eb4f20974_1394x680.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!zffK!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3f014a14-bdee-4aad-8abb-e79eb4f20974_1394x680.png" width="1394" height="680" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3f014a14-bdee-4aad-8abb-e79eb4f20974_1394x680.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:680,&quot;width&quot;:1394,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!zffK!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3f014a14-bdee-4aad-8abb-e79eb4f20974_1394x680.png 424w, https://substackcdn.com/image/fetch/$s_!zffK!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3f014a14-bdee-4aad-8abb-e79eb4f20974_1394x680.png 848w, https://substackcdn.com/image/fetch/$s_!zffK!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3f014a14-bdee-4aad-8abb-e79eb4f20974_1394x680.png 1272w, https://substackcdn.com/image/fetch/$s_!zffK!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3f014a14-bdee-4aad-8abb-e79eb4f20974_1394x680.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h3></h3><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/p/state-of-ai-compute-index-v4-june-2025?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://press.airstreet.com/p/state-of-ai-compute-index-v4-june-2025?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><h3><strong>Looking ahead: what might swing the charts next</strong></h3><p>Several known unknowns will shape the next update:</p><ul><li><p><strong>New datacenter regions in the Gulf: </strong>Major NVIDIA GPU clusters in the UAE and Saudi Arabia are ramping in 2025, including KAUST's Shaheen III and G42&#8217;s growing footprint. These may begin to appear more visibly in paper metadata by early 2026.</p></li><li><p><strong>Stargate (USA): </strong>The first phase of Stargate is expected to go live later this year. If operational as planned, it will represent the first large-scale deployment of a vertically integrated AI-native datacenter built around liquid-cooled NVIDIA infrastructure.</p></li><li><p><strong>H200 and B100</strong>: As these chips enter wider circulation, expect a Q4 bounce in NVIDIA mentions</p></li><li><p><strong>MI300X</strong>: Already present in tuning pipelines; the next test is whether it sees real training workloads</p></li><li><p><strong>Groq's compiler UX</strong>: If porting friction keeps dropping, academic teams may adopt LPU-backed inference at higher rates</p></li><li><p><strong>FPGAs</strong>: EU-funded open hardware initiatives could breathe new life into an accelerator category that's languished since 2020</p></li></ul><h3><strong>Conclusion</strong></h3><p>After two years of rapid scale-up in compute mentions, 2025 looks quieter, but not so for long. <strong>NVIDIA remains the default.</strong> Startup silicon continues to climb from a low base. And the next wave of hardware adoption is already underway, just not yet visible in the preprint timelines.</p><div><hr></div><h3>A few notes</h3><p>- We take the view that usage of chips in AI research papers (early adopters) is a leading indicator of industry usage.&nbsp;</p><p>- Papers using AI semi startup chips almost all have authors from the startup.&nbsp;</p><p>- 2025 figures = real counts through 1 June 2025 + volume-adjusted full-year forecast. Data from Zeta Alpha open-source AI paper index.</p><p>See the live charts here:&nbsp;<strong><a href="https://www.stateof.ai/compute">www.stateof.ai/compute</a></strong></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/p/state-of-ai-compute-index-v4-june-2025?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:&quot;button-wrapper&quot;}" data-component-name="ButtonCreateButton"><a class="button primary button-wrapper" href="https://press.airstreet.com/p/state-of-ai-compute-index-v4-june-2025?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p>]]></content:encoded></item><item><title><![CDATA[State of AI: June 2025 newsletter]]></title><description><![CDATA[What you need to know in AI across geopolitics, big tech, hardware, research, models, datasets, financings and exits over the last 4 weeks.]]></description><link>https://press.airstreet.com/p/the-state-of-ai-2025-june</link><guid isPermaLink="false">https://press.airstreet.com/p/the-state-of-ai-2025-june</guid><dc:creator><![CDATA[Air Street Press]]></dc:creator><pubDate>Sun, 08 Jun 2025 16:21:16 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/29a456cf-a1c6-472b-91ee-6164cb57026f_3792x2130.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Hi everyone!</p><p>Welcome to the latest issue of your guide to AI, an editorialized newsletter covering the key developments in AI policy, research, industry, and start-ups over the last month. First up, a few updates:</p><ul><li><p>Our 9th annual <strong><a href="http://raais.co">Research and Applied AI Summit</a></strong> is next Friday in London! We&#8217;ll dive into best practices for building AI-first products and translating applied research with leaders from ElevenLabs, Black Forest Labs, Isomorphic Labs, DeepMind and poolside, as well as AI geopolitics with Meta, Tony Blair Institute, Delian Alliance Industries, Bloomberg and the New York Times. After the event, we&#8217;ll share write ups on <a href="http://press.airstreet.com">Air Street Press</a> and recorded talks on <a href="https://www.youtube.com/@researchandappliedaisummit">our YouTube</a>.</p></li><li><p>State of AI Report 2025: I&#8217;m <strong>hiring a Research Assistant</strong> to help us build the industry&#8217;s most loved report :-) This is a perfect role if you love distilling complex topics in AI research, industry and politics into key messages to shape public conversation around this topic. If you know the perfect candidate, please reply/share! Here is a <a href="http://airstreet.com/analyst">job description</a>.</p></li><li><p><strong>Air Street Press</strong> featured a number of pieces: <a href="https://press.airstreet.com/p/biology-has-scaling-laws-profluent-progen3">biology&#8217;s scaling laws</a>, <a href="https://press.airstreet.com/p/delian-alliance-industries-launches-interceptigon">Delian&#8217;s new autonomous strike drones</a>, <a href="https://press.airstreet.com/p/hedera-dx-series-a">Hedera&#8217;s &#8364;15M Series A</a> to scale liquid biopsies for cancer patients, what <a href="https://press.airstreet.com/p/what-wall-street-didnt-see-coming-in-ai">Wall Street didn&#8217;t see coming</a> in AI back in 2016, and whether the <a href="https://press.airstreet.com/p/is-the-eu-ai-act-actually-useful">EU AI Act is actually useful</a>.</p></li></ul><p>I love hearing what you&#8217;re up to, so just hit reply or forward to your friends :-)</p><div><hr></div><h3><strong>The AI Factory narrative: infrastructure as industrial strategy</strong></h3><p>The term "AI factory" is no longer just rhetorical flourish. It has become the defining metaphor for the geopolitical and economic infrastructure race now unfolding in the open, as state and commercial agendas become visibly entangled. At its core is a push by Silicon Valley and U.S. political leadership to recast hyperscale datacenters as nation-building infrastructure, not just technical backends. <a href="https://press.airstreet.com/p/the-ai-factory-illusion-nvidia">Jensen Huang's</a> refrain that "these are not data centers, they are AI factories" is central to this rebrand. He invokes a powerful political argument: if the U.S. reshored sneaker factories, why not AI production? What sneakers were to 1990s globalization, AI is to the new era of strategic decoupling.</p><p>This rebranding has proven politically effective. It aligns the interests of capital-heavy AI players like OpenAI, Oracle, and NVIDIA with nationalist industrial policy. It plays directly into <a href="https://www.whitehouse.gov/fact-sheets/2025/05/fact-sheet-president-donald-j-trump-secures-historic-600-billion-investment-commitment-in-saudi-arabia/">Trump's $600B AI Acceleration Partnership</a> (at which every important CEO in corporate America was present), where AI infrastructure deals with Gulf states are packaged with the same gravity as oil or arms deals. The <a href="https://www.bloomberg.com/news/features/2025-05-20/inside-stargate-ai-data-center-from-openai-and-softbank">Stargate data center in Texas</a>, co-developed by OpenAI, Oracle, SoftBank and Crusoe, is the pilot site for what is pitched as a 10-site, $500B network of U.S.-controlled AI superclusters.</p><p>Meanwhile, a <a href="https://www.bloomberg.com/news/articles/2025-05-16/openai-to-help-uae-develop-one-of-world-s-biggest-data-centers">5GW Stargate-like project in Abu Dhabi</a> would leapfrog U.S. deployments in scale. These are national projects built with foreign dependencies, an irony at the heart of techno-sovereignty. This is the heart of the <a href="https://press.airstreet.com/p/the-ai-factory-illusion-nvidia">"AI Factory Illusion"</a> that I wrote about in Air Street Press: a globally distributed, highly automated infrastructure wrapped in the language of 20th-century manufacturing. As NVIDIA <a href="https://www.ft.com/content/2ba1d064-242d-4a2a-a381-374d9643e23a">courts</a> sovereign clients and neoclouds to diversify beyond Big Tech, the term serves to unlock subsidies, fast-track permitting, and deepen state-industry alignment. If anything, it helps NVIDIA <a href="https://nvidianews.nvidia.com/news/nvidia-announces-financial-results-for-first-quarter-fiscal-2026">print</a> tens of billions of dollars of revenue a quarter, or $44.1B in the last quarter, to be precise.</p><h3><strong>Meta stalls, Anthropic surges, OpenAI eats business lunch</strong></h3><p>The major labs&#8217; spring updates reveal a fragmented landscape. <a href="https://www.wsj.com/tech/ai/meta-is-delaying-the-rollout-of-its-flagship-ai-model-f4b105f7">Meta&#8217;s internal turmoil</a> around its "Behemoth" model perhaps underscores the collapsing marginal utility of trillion-parameter scaling absent a killer product experience. Engineers reportedly question whether the latest version represents real progress.</p><p>By contrast, <a href="https://www.anthropic.com/news/claude-4">Anthropic is cooking</a>, securing integration deals with <a href="https://www.bloomberg.com/news/articles/2025-05-02/apple-anthropic-team-up-to-build-ai-powered-vibe-coding-platform">Apple</a> (to power a new Claude-based version of Xcode) and Amazon (for Alexa+), and launching Claude 4, even though <a href="https://www.axios.com/2025/05/23/anthropic-ai-deception-risk">deceptive reasoning capabilities</a> triggered ethical scrutiny.</p><p>Claude 4, particularly the Claude Opus 4 variant, has proven to be a substantial leap forward. It outperforms competitors like GPT-4.1 and Gemini 2.5 Pro on key benchmarks like SWE-bench (72.5%) and exhibits strong long-context performance. It also supports extended autonomous operation for up to seven hours without degradation, making it well-suited for persistent agents. Its hybrid reasoning, long-term memory, and tool use integration make it arguably the most capable model in production today. Anthropic has also deployed it under its new AI Safety Level 3 (ASL-3) standard for the first time.</p><p>This comes as coding startup <a href="https://www.linkedin.com/posts/varunkmohan_today-im-proud-to-mark-our-first-venture-ugcPost-7328861173615861760-ZxiG/">Windsurf</a>, which has entered into an agreement to be acquired by OpenAI, was cut off from Claude access by Anthropic directly. In a somewhat timely fashion, Windsurf announced its own foundational model, SWE-1. These turf wars are worth following to inform how companies should think of building or tuning their own models vs. procuring them from OEMs or via third parties. Is the model the product or is the product enhanced by the model?</p><p>Meanwhile, <a href="https://ramp.com/data/ai-index">Ramp data</a> shows just how far ahead OpenAI&#8217;s share of US business subscriptions is from the competition. In a sample of more than 30,000 American businesses and billions of dollars in corporate spend using data from Ramp&#8217;s corporate card and bill pay platform, 81% of businesses with paid AI subscriptions were paying OpenAI.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!928A!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F03120c1d-c7da-4eaa-ad7a-1fd09e2bdfc8_1600x1133.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!928A!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F03120c1d-c7da-4eaa-ad7a-1fd09e2bdfc8_1600x1133.png 424w, https://substackcdn.com/image/fetch/$s_!928A!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F03120c1d-c7da-4eaa-ad7a-1fd09e2bdfc8_1600x1133.png 848w, https://substackcdn.com/image/fetch/$s_!928A!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F03120c1d-c7da-4eaa-ad7a-1fd09e2bdfc8_1600x1133.png 1272w, https://substackcdn.com/image/fetch/$s_!928A!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F03120c1d-c7da-4eaa-ad7a-1fd09e2bdfc8_1600x1133.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!928A!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F03120c1d-c7da-4eaa-ad7a-1fd09e2bdfc8_1600x1133.png" width="1456" height="1031" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/03120c1d-c7da-4eaa-ad7a-1fd09e2bdfc8_1600x1133.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1031,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!928A!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F03120c1d-c7da-4eaa-ad7a-1fd09e2bdfc8_1600x1133.png 424w, https://substackcdn.com/image/fetch/$s_!928A!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F03120c1d-c7da-4eaa-ad7a-1fd09e2bdfc8_1600x1133.png 848w, https://substackcdn.com/image/fetch/$s_!928A!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F03120c1d-c7da-4eaa-ad7a-1fd09e2bdfc8_1600x1133.png 1272w, https://substackcdn.com/image/fetch/$s_!928A!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F03120c1d-c7da-4eaa-ad7a-1fd09e2bdfc8_1600x1133.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h3></h3><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://press.airstreet.com/subscribe?"><span>Subscribe now</span></a></p><h3><strong>Meta&#8217;s military realignment</strong></h3><p>Meta is undergoing a dramatic shift, albeit outside the lab. In a move that would've been unthinkable just a few years ago, <a href="https://www.anduril.com/article/anduril-and-meta-team-up-to-transform-xr-for-the-american-military/">Meta has partnered with Anduril</a> to co-develop extended reality hardware and software for the U.S. military. The flagship product is EagleEye, a battlefield AR headset that fuses Meta&#8217;s Reality Labs and Llama models with Anduril&#8217;s Lattice defense platform. Silicon Valley&#8217;s most consumer-facing, metaverse-obsessed company is now explicitly building military gear.</p><p>What was once taboo (Big Tech&#8217;s open embrace of defense) is rapidly becoming normalized. The partnership also marks a d&#233;tente between Mark Zuckerberg and Palmer Luckey, signaling a realignment of AI-native leadership around national security imperatives.</p><h3><strong>More racing across modalities</strong></h3><p>The generative video race is accelerating across labs and platforms. <a href="https://techcrunch.com/2025/05/20/googles-veo-3-can-generate-videos-and-soundtracks-to-go-along-with-them/">Google&#8217;s remarkable Veo 3</a> now adds soundtracks to generated video, marking a step closer to multimodal narrative synthesis. Their <a href="https://x.com/bilawalsidhu/status/1930405285253767296">demos</a> showed coherence and visual fidelity approaching short-form film. This sparked viral excitement all over Twitter, as observers noted a leap in <a href="https://x.com/jasonzada/status/1928084185845223913">cinematic fidelity</a>.</p><p>Startups are pushing forward too. Lightricks&#8217; new model is capable of realistic motion rendering from text, while <a href="https://odyssey.world/introducing-interactive-video">Odyssey</a> is betting on interactive generative video experiences as a new frontier. Odyssey&#8217;s <a href="https://experience.odyssey.world/">research preview</a> (try it!) lets users walk through photorealistic, real-time AI-generated 3D scenes such as forests, shopping malls, and more by streaming frames every 40ms, without a traditional game engine. The model doesn&#8217;t just generate a clip: it creates an explorable, real-time world. That alone sets it apart from static or pre-rendered outputs.</p><p>On the research frontier, <a href="https://huggingface.co/deepseek-ai/DeepSeek-R1-0528-Qwen3-8B">DeepSeek&#8217;s new R1 checkpoint</a> blends Chinese pretraining corpora with Qwen3 backbones, continuing China&#8217;s push for frontier-scale open models. In parallel, Google released <a href="https://developers.google.com/health-ai-developer-foundations/medgemma">Gemma 3B and MedGemma</a>, its health-tuned LLM family, while the open-source Gemma Diffusion shows growing ecosystem depth.</p><p>Gemma Diffusion is particularly exciting because it offers a new paradigm for text generation: using diffusion rather than autoregressive methods. It allows for full-sentence or paragraph generation in parallel, with faster outputs and the ability to self-correct. As an open-source initiative, it broadens experimentation and hints at a post-token-by-token future for language models.</p><h3><strong>The export control shuffle and sovereign stack realignment</strong></h3><p>U.S. policy is shifting to reflect this AI infrastructure race. The <a href="https://www.aoshearman.com/en/insights/ai-diffusion-rule-rescinded-policy-guidance-for-advanced-integrated-circuits-and-commodities-issued">rescinding of the AI diffusion rule</a> signals a recalibration of export controls around advanced semiconductors and models. Rather than blunt bans, future policy may lean on traceability, licensing regimes, and platform-level coordination.</p><p>Meanwhile, <a href="https://nvidianews.nvidia.com/news/nvidia-nvlink-fusion-semi-custom-ai-infrastructure-partner-ecosystem">NVIDIA's NVLink Fusion program</a> is expanding the modularization of its stack, allowing nation-states, corporates, and "neoclouds" to build bespoke AI systems. This plays into <a href="https://www.ft.com/content/2ba1d064-242d-4a2a-a381-374d9643e23a">NVIDIA's explicit push</a> to court sovereign and enterprise clients outside of the Big Tech oligopoly.</p><p>From <a href="https://nvidianews.nvidia.com/news/humain-and-nvidia-announce-strategic-partnership-to-build-ai-factories-of-the-future-in-saudi-arabia">Saudi's Humain deal</a> to <a href="https://wallenberginvestments.com/en/news/swedish-business-consortium-build-ai-factory-nvidia">Sweden's consortium AI factory</a>, the geopolitical stack is re-aligning around modular, NVIDIA-led platforms. Sovereignty now means not only owning the model weights, but the infrastructure, training corpus, and policy posture that governs their use. This is a new form of sovereignty: not territorial, but computational.</p><p>The factory metaphor, however ill-fitting in technical terms, is proving useful as political architecture. The future of AI will be built in these factories, whether or not they produce anything as tangible as steel.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/p/the-state-of-ai-2025-june?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://press.airstreet.com/p/the-state-of-ai-2025-june?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><div><hr></div><h3><strong>Research papers</strong></h3><p><strong><a href="https://arxiv.org/pdf/2505.02222">Practical Efficiency of Muon for Pretraining</a></strong>, <em>Essential AI</em></p><p>In this paper, the authors investigate Muon, a simple second-order optimizer, as a replacement for AdamW in large-scale language model pretraining. They show that Muon expands the compute-time Pareto frontier, enabling faster training or reduced compute at large batch sizes without sacrificing data efficiency. Experiments across models up to 4B parameters and batch sizes up to 16M tokens demonstrate that Muon consistently requires 10&#8211;15% fewer tokens than AdamW to reach the same loss, with this advantage persisting or growing as batch size increases.</p><p>The study also addresses hyperparameter tuning by combining Muon with maximal update parameterization (muP), enabling efficient transfer of hyperparameters from small to large models. The authors introduce a &#8220;telescoping&#8221; algorithm that controls tuning overhead, keeping it modest even at scale. This work is relevant for practitioners seeking to optimize training efficiency and resource allocation in large language model development, especially in distributed or compute-constrained environments.</p><p><strong><a href="https://storage.googleapis.com/deepmind-media/DeepMind.com/Blog/alphaevolve-a-gemini-powered-coding-agent-for-designing-advanced-algorithms/AlphaEvolve.pdf">AlphaEvolve: A coding agent for scientific and algorithmic discovery</a></strong>, <em>Google DeepMind</em></p><p>In this paper, the authors introduce AlphaEvolve, an evolutionary coding agent that leverages state-of-the-art large language models (LLMs) to autonomously improve algorithms through iterative code modifications and automated evaluation.</p><p>AlphaEvolve orchestrates a pipeline where LLMs generate, critique, and evolve code, guided by machine-gradeable evaluation functions. The system was tested on challenging tasks, including discovering faster matrix multiplication algorithms, most notably, finding a new algorithm for multiplying $4 \times 4$ complex matrices using 48 multiplications, improving on Strassen&#8217;s 49-multiplication result after 56 years.</p><p>Beyond mathematics, AlphaEvolve was applied to optimize Google&#8217;s data center scheduling, kernel engineering for Gemini LLM training, and hardware circuit design, yielding measurable improvements such as a 0.7% recovery in compute resources and a 23% kernel speedup.</p><p>The research highlights that combining LLMs with evolutionary search and automated feedback can yield novel, verifiable solutions in both theoretical and practical domains, especially where automated evaluation is feasible.</p><p><strong><a href="https://arxiv.org/pdf/2505.19914">Enigmata: Scaling Logical Reasoning in Large Language Models with Synthetic Verifiable Puzzles</a></strong><em>, ByteDance Seed, Fudan University, Tsinghua University</em></p><p>In this paper, the authors introduce Enigmata, a suite designed to improve and evaluate logical reasoning in large language models (LLMs) using synthetic, verifiable puzzles. The suite includes 36 tasks across seven categories, each with an automatic generator and verifier, enabling scalable data creation and precise difficulty control. The authors propose a two-stage training approach: rejection fine-tuning with high-quality solutions, followed by multi-task reinforcement learning with verifiable rewards (RLVR).</p><p>Experiments show that models trained with Enigmata, such as Qwen2.5-32B-Enigmata, outperform strong baselines like o1 and o3-mini-high on puzzle reasoning benchmarks (Enigmata-Eval, ARC-AGI, and others), and generalize well to out-of-domain tasks, including advanced math and STEM problems. Notably, adding Enigmata data to larger models (e.g., Seed1.5-Thinking) further boosts performance on challenging benchmarks.</p><p>The work demonstrates that synthetic, diverse puzzle data can enhance LLM reasoning, supporting applications in education, automated problem solving, and robust AI evaluation.</p><p><strong><a href="https://arxiv.org/pdf/2505.08762">The Open Molecules 2025 (OMol25) Dataset, Evaluations, and Models</a></strong><em>, Meta, Carnegie Mellon University, University of Cambridge</em></p><p>In this paper, the authors introduce the Open Molecules 2025 (OMol25) dataset, a large-scale resource containing over 100 million density functional theory (DFT) calculations at a high level of theory, spanning 83 elements and a wide range of molecular systems, including biomolecules, metal complexes, and electrolytes.</p><p>The dataset is designed to address the lack of comprehensive, diverse, and high-quality data for training machine learning interatomic potentials (MLIPs) that can act as DFT surrogates. The authors provide detailed descriptions of the data generation process, including sampling strategies for chemical and structural diversity, and rigorous quality control.</p><p>Baseline models such as eSEN, GemNet-OC, and MACE are evaluated on OMol25, with eSEN-md achieving energy MAEs as low as 1.2 meV/atom and force MAEs of 12.3 meV/&#197;. However, challenges remain in accurately modeling ionization energies, spin gaps, and long-range interactions.</p><p><strong><a href="https://arxiv.org/pdf/2505.24864">ProRL: Prolonged Reinforcement Learning Expands Reasoning Boundaries in Large Language Models</a>, </strong><em>NVIDIA</em></p><p>In this paper, the authors introduce ProRL, a prolonged reinforcement learning methodology designed to expand the reasoning capabilities of large language models beyond what is accessible through standard RL or extensive sampling of base models. They address challenges like entropy collapse and training instability by incorporating KL divergence penalties, reference policy resets, and dynamic sampling.</p><p>Their experiments train a 1.5B parameter model, Nemotron-Research-Reasoning-Qwen-1.5B, on a diverse set of 136K problems spanning math, code, STEM, logic puzzles, and instruction following. The model outperforms its base model (DeepSeek-R1-1.5B) with average pass@1 improvements of 14.7% in math, 13.9% in coding, 54.8% in logic puzzles, 25.1% in STEM, and 18.1% in instruction following. Notably, ProRL enables the model to solve tasks where the base model fails entirely, and shows strong generalization to out-of-distribution tasks.</p><p>The work demonstrates that with sufficient RL training, models can develop novel reasoning strategies, suggesting practical benefits for deploying smaller, more capable models in real-world applications where compute and data are limited.</p><p><strong><a href="https://arxiv.org/pdf/2505.19590">Learning to Reason without External Rewards,</a> </strong><em>UC Berkeley, Yale University</em></p><p>In this paper, the authors introduce INTUITOR, a method for training LLMs using only internal feedback, specifically, the model&#8217;s own confidence (self-certainty), as a reward signal, rather than relying on external supervision or labeled data. The approach, called Reinforcement Learning from Internal Feedback (RLIF), replaces traditional reward signals in policy optimization algorithms with self-certainty scores, measured as the average KL divergence between the model&#8217;s output distribution and a uniform distribution.</p><p>Experiments show that INTUITOR matches the performance of supervised RL methods like Group Relative Policy Optimization (GRPO) on mathematical reasoning benchmarks (GSM8K, MATH500), and outperforms them on out-of-domain tasks such as code generation (LiveCodeBench, CRUXEval). The method also improves instruction-following and fosters more structured, interpretable reasoning.</p><p>A key caveat is that INTUITOR&#8217;s effectiveness depends on careful regularization (KL penalty) and online reward computation to avoid reward exploitation. This research suggests that intrinsic model signals can drive scalable, domain-agnostic learning, which could be valuable for autonomous AI systems where external supervision is impractical.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://press.airstreet.com/subscribe?"><span>Subscribe now</span></a></p><p><strong><a href="https://cdn.openai.com/pdf/bd7a39d5-9e9f-47b3-903c-8b847ca650c7/healthbench_paper.pdf">HealthBench: Evaluating Large Language Models Towards Improved Human Health</a>, </strong><em>OpenAI</em></p><p>In this paper, the authors introduce HealthBench, an open-source benchmark designed to evaluate large language models (LLMs) in healthcare contexts. HealthBench consists of 5,000 multi-turn conversations, each graded against detailed, physician-written rubrics covering 48,562 unique criteria across seven health themes and five behavioral axes (accuracy, completeness, communication, context awareness, instruction following).</p><p>The benchmark measures both overall and theme-specific model performance, using a model-based grader validated against physician judgment. Results show rapid improvement in recent models: OpenAI&#8217;s o3 model scores 60% overall, compared to 16% for GPT-3.5 Turbo and 32% for GPT-4o. Smaller, cost-efficient models like GPT-4.1 nano now outperform older, larger models.</p><p>The paper also introduces HealthBench Consensus (focusing on 34 critical, physician-validated criteria) and HealthBench Hard (a challenging subset where top models score only 32%). The research highlights persistent gaps in context-seeking and reliability, emphasizing the need for robust, real-world evaluation frameworks as LLMs are increasingly deployed in healthcare.</p><p><strong><a href="https://arxiv.org/pdf/2505.13400">Robin: A Multi-Agent System for Automating Scientific Discovery</a></strong>, <em>FutureHouse, University of Oxford</em></p><p>In this paper, the authors introduce Robin, a multi-agent AI system designed to automate the scientific discovery process, integrating hypothesis generation, experimental planning, and data analysis. Robin was applied to identify novel treatments for dry age-related macular degeneration (dAMD), focusing on enhancing retinal pigment epithelium (RPE) phagocytosis. The system proposed ripasudil, a ROCK inhibitor, as a therapeutic candidate, which was experimentally validated to significantly enhance RPE phagocytosis.</p><p>Robin employs agents like Crow and Falcon for literature review and Finch for experimental data analysis. Experiments included flow cytometry to measure phagocytosis and RNA sequencing to explore transcriptional changes, revealing upregulation of ABCA1, a lipid efflux pump linked to RPE function.</p><p>While Robin automates key steps, it relies on human input for experimental execution and prompt engineering. This research demonstrates AI&#8217;s potential to accelerate drug discovery, particularly in repurposing existing drugs for unmet medical needs.</p><p><strong><a href="https://www.biorxiv.org/content/10.1101/2025.04.25.650731v1.full.pdf">A Cross-Species Generative Cell Atlas Across 1.5 Billion Years of Evolution: The TranscriptFormer Single-cell Model</a></strong><em>, Chan Zuckerberg Initiative, Stanford University</em></p><p>In this paper, the authors introduce TranscriptFormer, a generative foundation model designed to create a cross-species single-cell transcriptomic atlas spanning 1.53 billion years of evolution across 12 species. The model integrates gene and transcript data using a transformer-based architecture, enabling tasks like cell type classification, disease state prediction, and gene-gene interaction simulation.</p><p>The experiments demonstrate that TranscriptFormer outperforms existing models, such as UCE and ESM2-CE, in generalizing across species, including those separated by 685 million years of evolution. Notably, the TF-Metazoa variant achieved the highest macro F1 score (0.778) in out-of-distribution species classification. It also excelled in human-specific tasks, matching or slightly surpassing state-of-the-art models on the Tabula Sapiens 2.0 dataset.</p><p>This research highlights the potential of generative models in biology, offering tools for cross-species analysis, disease modeling, and virtual experimentation, with applications in evolutionary biology, drug discovery, and personalized medicine.</p><p><strong><a href="https://arxiv.org/pdf/2505.03335">Absolute Zero: Reinforced Self-play Reasoning with Zero Data</a></strong><em>, Tsinghua University, Beijing Institute for General Artificial Intelligence, Pennsylvania State University</em></p><p>In this paper, the authors introduce the Absolute Zero Reasoner (AZR), a reinforcement learning framework that trains LLMs without relying on human-curated datasets. AZR operates under the Absolute Zero paradigm, where the model proposes and solves its own tasks, guided by verifiable rewards from a code execution environment.</p><p>The research demonstrates that AZR achieves SOTA performance on coding and mathematical reasoning benchmarks, surpassing models trained on tens of thousands of curated examples. Notably, AZR-trained models show strong cross-domain generalization, with coding-trained models improving math performance by up to 15.2 percentage points.</p><p>The experiments highlight the importance of task diversity, with AZR leveraging abduction, deduction, and induction tasks. However, the authors note challenges like safety concerns and occasional undesirable outputs, emphasizing the need for oversight.</p><p>This work matters as it reduces dependency on human data, enabling scalable, autonomous learning. Potential applications include adaptive AI systems in education, coding, and problem-solving domains.</p><p><strong><a href="https://arxiv.org/pdf/2505.12082">Model Merging in Pre-training of Large Language Models</a></strong><em>, ByteDance Seed</em></p><p>In this paper, the authors introduce Pre-trained Model Average (PMA), a novel framework for model merging during LLM pre-training. They conducted extensive experiments across model scales (from millions to over 100B parameters) with both Dense and MoE architectures.</p><p>Their results demonstrate that merging checkpoints from the stable training phase produces significant performance improvements across downstream tasks. Remarkably, applying PMA at early stages of the cosine-decay phase achieves comparable results to final-stage annealed models.</p><p>Merging with constant learning rates can effectively simulate annealed performance without the computational expense of full annealing. They also introduce PMA-init, which stabilizes training when loss spikes occur. Through mathematical analysis and experimentation, they determine optimal merging intervals scale with model size, while including more checkpoints improves performance.</p><p><strong><a href="https://arxiv.org/pdf/2505.22954">Darwin G&#246;del Machine: Open-Ended Evolution of Self-Improving Agents</a></strong>, <em>University of British Columbia, Vector Institute, Sakana AI</em></p><p>In this paper, the authors introduce the Darwin G&#246;del Machine (DGM), a self-improving AI system that modifies its own codebase to enhance its coding capabilities. The DGM combines self-improvement with open-ended exploration, maintaining an archive of diverse coding agents rather than evolving just one solution. This approach allows exploration of multiple paths through the search space.</p><p>Experiments on coding benchmarks show impressive results: performance increased from 20.0% to 50.0% on SWE-bench and from 14.2% to 30.7% on Polyglot. The authors demonstrated that both self-improvement and open-ended exploration components are essential, as removing either one significantly reduced performance gains. The improvements generalized across different foundation models and programming languages, showing the robustness of the approach.</p><p><strong><a href="https://www.pi.website/download/pi05_KI.pdf">Knowledge Insulating Vision-Language-Action Models: Train Fast, Run Fast, Generalize Better</a>, </strong><em>Physical Intelligence</em></p><p>In this paper, the authors tackle a key challenge in robot control: maintaining knowledge from pretrained vision-language models while adding fast, continuous action capabilities. They identify a critical problem: when adding continuous action modules to vision-language models, the pretrained knowledge often degrades, resulting in poor language following and slow training.</p><p>Their solution, &#8220;knowledge insulation,&#8221; introduces three key innovations: 1) Joint training with both discrete and continuous action representations, 2) Stopping gradient flow from the action expert to the vision-language backbone, 3) Co-training with general vision-language data alongside robotics data.<br><br>The approach achieves state-of-the-art results on LIBERO benchmarks (96.0% on LIBERO-90) and outperforms alternatives on real-world tasks like table bussing and drawer manipulation. This approach could enable robots to both follow human instructions accurately and execute complex physical actions efficiently.</p><p><strong><a href="https://arxiv.org/pdf/2505.16839">LaViDa: A Large Diffusion Language Model for Multimodal Understanding</a></strong><em><strong>, </strong>UCLA, Panasonic AI Research, Adobe Research</em></p><p>In this paper, the authors introduce LaViDa, the first family of vision-language models based on diffusion rather than traditional autoregressive generation. While current VLMs like LLaVA generate text sequentially, LaViDa leverages diffusion models to enable parallel decoding and bidirectional context. The authors tackle key challenges with novel techniques: complementary masking for efficient training, prefix-DLM for faster inference, and timestep shifting for quality sampling.</p><p>LaViDa achieves competitive performance against autoregressive baselines on benchmarks like MMMU (43.3% vs 35.1%), Math Vista, and ScienceQA. On COCO captioning, it surpasses Open-LLaVA-Next by +4.1 CIDEr with 1.92x speedup. Most impressively, LaViDa excels at constrained generation tasks, with 100% satisfaction on poem constraints compared to &lt;50% for autoregressive models.</p><p>This research matters because it offers flexible speed-quality tradeoffs and superior text-infilling capabilities, enabling applications requiring structured outputs or format constraints that autoregressive models struggle with.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/p/the-state-of-ai-2025-june?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://press.airstreet.com/p/the-state-of-ai-2025-june?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><div><hr></div><h3><strong>Investments</strong></h3><p><strong><a href="https://quantum-systems.com/blog/2025/05/05/quantum-systems-raises-euro160m/">Quantum Systems</a></strong>, the drone company for defense and commercial applications, raised a &#8364;160M Series C financing round from Balderton Capital, Hensoldt, and Airbus Defense and Space.</p><p><strong><a href="https://blog.newlimit.com/p/newlimit-raises-130-million-series">NewLimit</a></strong>, a biotech company developing medicines to extend human healthspan through epigenetic reprogramming, raised a $130M Series B from Kleiner Perkins, Khosla Ventures, and Human Capital.</p><p><strong><a href="https://tech.eu/2025/05/02/arondite-secures-10m-to-strengthen-human-machine-cooperation-in-defence/">Arondite</a></strong>, a UK-based AI startup focused on enhancing human-machine collaboration in defence, raised a $10M Seed financing round led by Index Ventures, with participation from Concept Ventures and Creator Fund.</p><p><strong><a href="https://techcrunch.com/2025/05/06/relevance-ai-raises-24m-series-b-to-help-anyone-build-teams-of-ai-agents/">Relevance AI</a></strong>, the San Francisco- and Sydney-based startup developing an AI agent operating system, raised a $24M Series B financing round led by Bessemer Venture Partners with participation from King River Capital and Insight Partners.</p><p><strong><a href="https://group.nebius.com/newsroom/nebius-welcomes-bezos-expeditions-as-lead-investor-in-ai-data-business-toloka">Toloka</a></strong>, the AI data annotation company that&#8217;s part of Nebius Group, raised a $75M funding round led by Bezos Expeditions and Mikhail Parakhin (CTO of Shopify).</p><p><strong><a href="https://www.inductive.bio/news/inductive-bio-raises-25m-series-a">Inductive Bio</a></strong>, a AI company working on ADMET for small molecule drug discovery, raised a $25M Series A from Obvious Ventures and a16z Bio + Health.</p><p><strong><a href="https://sifted.eu/articles/parloa-hits-unicorn-status">Parloa</a></strong>, the German AI agents startup for customer service automation, raised a $120M Series C financing round at a $1B valuation from Durable Capital Partners LP, Altimeter Capital, and General Catalyst.</p><p><strong><a href="https://archive.is/d7IcO">True Anomaly</a></strong>, a space defense technology company developing military-class orbital systems, raised $260M in a financing round led by Accel with participation from Meritech Capital and Eclipse.</p><p><strong><a href="https://techcrunch.com/2025/05/05/a-stealth-ai-model-beat-dall-e-and-midjourney-on-a-popular-benchmark-its-creator-just-landed-30m/">Recraft</a></strong>, the AI startup specializing in image generation for branding and marketing, raised a $30M Series B financing round led by Accel, with participation from Khosla Ventures and Madrona.</p><p><strong><a href="https://press.airstreet.com/p/hedera-dx-series-a">HederaDx</a></strong>, liquid biopsies for cancer care, raised a &#8364;15M Series A led by Vsquared Ventures.</p><p><strong><a href="https://archive.is/PHF7j">Samaya AI</a></strong>, the company building expert AI agents for financial services, raised $43.5M in a financing round led by New Enterprise Associates.</p><p><strong><a href="https://www.sensmore.ai/press/funding-announcement">Sensmore</a></strong>, the robotics startup pioneering Physical AI for heavy mobile machinery, raised a $7.3M financing round led by Point Nine.</p><p><strong><a href="https://archive.is/90bLE">LMArena</a></strong>, a platform for testing and voting on AI models, raised $100M in a financing round at a $600M valuation from Andreessen Horowitz and UC Investments.</p><p><strong><a href="https://www.reflectorbital.com/series-a-announcement">Reflect Orbital</a></strong>, a company developing satellite constellations to deliver sunlight on demand, raised a $20M Series A from Lux Capital, Sequoia Capital, and Starship Ventures.</p><p><strong><a href="https://www.theinformation.com/articles/jpmorgan-lend-7-billion-openai-data-center?rc=yvsjfo">OpenAI</a></strong>, the AI research and deployment company, secured more than $7 billion in financing from JPMorgan to build a data center.</p><p><strong><a href="https://techcrunch.com/2025/05/26/one-of-europes-top-ai-researchers-raised-a-13m-seed-to-crack-the-holy-grail-of-models/">SpAItial</a></strong>, the AI startup focused on spatial foundation models, raised a $13M seed financing round led by Earlybird Venture Capital with participation from Speedinvest and several high-profile angels.</p><p><strong><a href="https://www.reuters.com/business/grammarly-secures-1-billion-general-catalyst-build-ai-productivity-platform-2025-05-29/">Grammarly</a></strong>, the AI-powered writing assistant and productivity platform, raised $1 billion in a non-dilutive revenue-based financing round from General Catalyst, with no valuation disclosed.</p><p><strong><a href="https://www.wordsmith.ai/blog/wordsmith-ai-raises-25-million-series-a">Wordsmith</a></strong>, the legal intelligence platform for in-house teams, raised a $25M Series A led by Index Ventures.</p><p><strong><a href="https://archive.is/UCQqi">LawZero</a></strong>, a nonprofit research group focused on safe AI development, raised $30M in a financing round from Eric Schmidt&#8217;s philanthropic organization and Skype co-founder Jaan Tallinn.</p><p><strong><a href="https://finance.yahoo.com/news/cast-ai-rakes-108m-series-113334508.html">Cast AI</a></strong>, the application performance automation platform, raised a $108M Series C at an $850M valuation from G2 Venture Partners and SoftBank Vision Fund 2.</p><p><strong><a href="https://www.optimaldynamics.com/resource/optimal-dynamics-raises-40m-series-c-to-scale-the-decision-layer-of-logistics">Optimal Dynamics</a></strong>, the AI-driven decision intelligence platform for trucking and logistics, raised a $40M Series C financing round led by Koch Disruptive Technologies.</p><p><strong><a href="https://www.stack-ai.com/blog/stack-ai-raises-16m-series-a-to-create-ai-agents-for-every-job">Stack AI</a></strong>, the enterprise AI platform for creating custom AI agents, raised a $16M Series A financing round from Lobby VC, LifeX Ventures, and Gradient.</p><p><strong><a href="https://www.greenlite.ai/blog/greenlite-ai-raises-15-million-series-a">Greenlite AI</a></strong>, the AI compliance automation company for financial institutions, raised a $15M Series A financing round from Greylock, Thomson Reuters, and Canvas Prime.</p><p><strong><a href="https://hyper.ai/en/headlines/3a621265d4760a9ee9a7ccf9965b225f">Octave</a></strong>, the AI marketing-tech startup focused on enhancing customer profiling and campaign strategy, raised a $5.5M seed financing round from Bonfire Ventures, Unusual Ventures, and Bee Partners.</p><p><strong><a href="https://cointelegraph.com/news/nous-research-raises-50m-paradigm-decentralized-ai-solana">Nous Research</a></strong>, the decentralized AI startup leveraging Solana for open-source AI models, raised a $50M Series A at a $1B valuation from Paradigm, with previous investors including Distributed Global and North Island Ventures.</p><p><strong><a href="https://techcrunch.com/2025/05/15/hedra-the-app-used-to-make-talking-baby-podcasts-raises-32m-from-a16z/">Hedra</a></strong>, the AI-powered video generation and editing platform, raised $32M in a Series A financing round led by a16z, with participation from Index Ventures and Abstract Ventures.</p><p><strong><a href="https://www.geekwire.com/2025/seattle-startup-backed-by-former-google-ceo-lands-16m-to-automate-repetitive-tasks-on-a-computer/">Vercept</a></strong>, the AI startup developing a "computer interface of the future," raised a $16M seed financing round from Fifty Years, Point Nine, and the AI2 Incubator.</p><div><hr></div><h3><strong>Rumored investments</strong></h3><p><strong><a href="https://www.forbes.com/sites/rashishrivastava/2025/05/02/ai-startup-decagon-in-talks-to-raise-100-million-at-a-15-billion-valuation/?utm_source=twitter&amp;utm_medium=social&amp;utm_campaign=forbes&amp;utm_term=se-staff">Decagon</a></strong>, the AI startup building customer service agents, is raising $100M in a financing round at a $1.5B valuation from Andreessen Horowitz and Accel. The company is at ~$17M ARR, up from ~$1.5M ARR LTM.</p><p><strong><a href="https://www.theinformation.com/articles/abridge-talks-raise-5-billion-valuation-ai-health-startups-draw-investors?rc=yvsjfo">Abridge</a></strong>, the AI health startup focused on medical transcription and documentation, is in talks to raise a financing round at a $5 billion valuation; the list of investors in the new round was not disclosed.</p><p><strong><a href="https://techcrunch.com/2025/05/04/cursor-is-reportedly-raising-funds-at-9-billion-valuation-from-thrive-a16z-and-accel/">Anysphere</a></strong>, the maker of the AI-powered coding tool Cursor, is rumored to have raised $900M in a financing round at a $9B valuation from Thrive Capital, a16z, and Accel. <a href="https://x.com/cursor_ai/status/1931032123038945530">Confirmed</a> yesterday </p><p><strong><a href="https://archive.is/2025.05.09-064717/https://www.bloomberg.com/news/articles/2025-05-09/ex-northvolt-ceo-gets-new-funds-for-ai-startup-in-manufacturing">Aris Machina</a></strong>, the AI-enhanced industrial software platform targeting manufacturing optimization, is rumored to be raising a financing round from Earlybird, Village Global, and AENU.</p><div><hr></div><h3><strong>Acquisitions</strong></h3><p><strong><a href="https://www.emigal.com/2025/05/05/ezra-is-joining-function-health/">Ezra</a></strong>, the company pioneering affordable full-body MRIs, was acquired by <strong>Function</strong> <strong>Health</strong>; the acquisition price was not disclosed.</p><p><strong><a href="https://archive.is/wXeKp">Windsurf</a></strong>, an AI-assisted coding tool formerly known as Codeium, entered into an agreement to be acquired by <strong>OpenAI</strong> for $3 billion. Windsurf had previously raised $150M in a Series C funding round led by General Catalyst at a $1.25B valuation.</p><p><strong><a href="https://www.databricks.com/company/newsroom/press-releases/databricks-agrees-acquire-neon-help-developers-deliver-ai-systems">Databricks</a></strong>, the Data and AI company, agreed to acquire <strong>Neon</strong>, a serverless Postgres platform for developers and AI agents, for a rumored $1B.</p><p><strong><a href="https://www.salesforce.com/news/stories/salesforce-signs-definitive-agreement-to-acquire-convergence-ai/">Convergence.ai</a></strong>, an AI agent company specializing in adaptive, intelligent systems for digital workflows, was acquired by <strong>Salesforce</strong>; the acquisition price was not disclosed.</p><p><strong><a href="https://www.together.ai/blog/together-ai-acquires-refuel-ai">Together AI</a></strong>, the AI Acceleration Cloud platform for developers and enterprises, acquired <strong>Refuel.ai</strong> to integrate its models and platform capabilities; the acquisition price was not disclosed.</p><p><strong><a href="https://www.globenewswire.com/news-release/2025/05/19/3083892/0/en/Regeneron-A-Leading-U-S-Biotechnology-Company-to-Acquire-23andMe-in-Court-Supervised-Sale.html">23andMe</a></strong>, a genetics-led consumer healthcare and biotechnology company, was acquired by <strong>Regeneron</strong> <strong>Pharmaceuticals</strong> for $256 million. I include this here because of the presumed value of 23andMe&#8217;s data business for drug discovery.</p><p><strong><a href="https://openai.com/sam-and-jony/">io</a></strong>, a company focused on developing inspiring and empowering products (vague yes) that was formed by Jonny Ive and Sam Altman, was acquired by <strong>OpenAI</strong> for a whopping $6.5B.</p><p><strong><a href="https://modal.com/blog/twirl-is-joining-modal">Twirl</a></strong>, a data orchestrator for developing, testing, and deploying data pipelines, was acquired by <strong>Modal</strong>. The acquisition price was not disclosed.</p><p><strong><a href="https://www.moonhub.ai/moonhub-team-joins-salesforce">Moonhub</a></strong>, the AI recruitment company that developed the world&#8217;s first AI Recruiter, was acquired by <strong>Salesforce</strong>. The acquisition price was not disclosed.</p><p><strong><a href="https://www.salesforce.com/news/press-releases/2025/05/27/salesforce-signs-definitive-agreement-to-acquire-informatica/">Informatica</a></strong>, a leader in AI-powered enterprise cloud data management, entered into a definitive agreement to be acquired by <strong>Salesforce</strong> for approximately $8 billion.</p><p><strong><a href="https://www.newswire.ca/news-releases/mapmygenome-cements-global-leadership-in-genomics-with-strategic-acquisition-of-canada-s-microbiome-insights-837518152.html">Mapmygenome</a></strong>, an AI-driven genomics and personalized health company, acquired <strong>Microbiome</strong> <strong>Insights</strong> to expand its North American footprint; the acquisition price was not disclosed.</p><p><strong><a href="https://www.linkedin.com/posts/nlarusstone_sphinx-is-joining-benchling-activity-7328473283392163840-k_d5">Sphinx Bio</a></strong>, a software for AI workflows in biology, was acquired by <strong>Benchling</strong> for an undisclosed amount.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://press.airstreet.com/subscribe?"><span>Subscribe now</span></a></p>]]></content:encoded></item><item><title><![CDATA[State of AI: May 2025 newsletter]]></title><description><![CDATA[What you need to know in AI across geopolitics, big tech, hardware, research, models, datasets, financings and exits over the last 4 weeks.]]></description><link>https://press.airstreet.com/p/the-state-of-ai-2025-may</link><guid isPermaLink="false">https://press.airstreet.com/p/the-state-of-ai-2025-may</guid><dc:creator><![CDATA[Air Street Press]]></dc:creator><pubDate>Sun, 11 May 2025 15:21:50 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/bfd9e77b-fb3f-40c8-a9d3-c4a61a97a2fe_1860x1040.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Hi everyone!</p><p>Welcome to the latest issue of the State of AI newsletter, an editorialized newsletter covering the key developments in AI policy, research, industry, and start-ups over the over the last month. First up, a few updates:</p><ul><li><p>Join us and 150 attendees across research, engineering, product and design at leading AI startups and big tech companies for our 9th annual <a href="http://raais.co">Research and Applied AI Summit</a> in London on 13 June. We&#8217;ll dive into best practices for building AI-first products and translating applied research with leaders from ElevenLabs, Black Forest Labs, Isomorphic Labs, DeepMind and Poolside.</p></li><li><p>It was great to see our community members join the Air Street NYC AI meetup and happy hour last month, ft. talks from portfolio companies Fern Labs (long-running computer agents), V7 Labs (workflow automation), Patina (design systems) and VantAI (biology). More to come from our <a href="https://lu.ma/airstreet">event series</a>.</p></li></ul><p>I love hearing what you&#8217;re up to, so just hit reply or forward to your friends :-)</p><div><hr></div><h3><strong>Scaling Laws for Science</strong></h3><p>Longtime readers of Guide to AI (and my tweets) will know that I believe the practice of science, and biology more specifically, will be rethought to be AI-first. Biology is central to health, disease, industry, and nature, and it has increasingly become a data-driven science. The types of questions we can ask and answer depend on the analytical tools we have to query living systems at their various levels of complexity. Just like the resolution of telescopes improved 1000x over 400 years&#8212;from Galileo&#8217;s optics to today&#8217;s space observatories&#8212;biological resolution has gone from averaging millions of cells to profiling individuals with spatial and molecular precision.</p><p>Consider Recursion Pharmaceuticals, which produces 16.2 million multi-timepoint brightfield images across ~2.2 million experiments weekly, generating 135 terabytes of data. The natural question is: what can we do with all this data?</p><p>In protein sequence space, Profluent Bio is developing frontier AI models that learn the rules of protein design from billions of sequences, functions, and structures. Their <a href="https://press.airstreet.com/p/biology-has-scaling-laws-profluent-progen3">ProGen3</a> model series empirically shows that scaling compute predictably lowers validation loss and improves model performance for both in-distribution and out-of-distribution proteins. Below is a chart comparing Profluent&#8217;s results with OpenAI&#8217;s GPT-3-era scaling law plots&#8212;remarkably similar curves.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!_CJL!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe73a3629-dcbf-41b8-847a-eb73d5720486_1456x640.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!_CJL!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe73a3629-dcbf-41b8-847a-eb73d5720486_1456x640.png 424w, https://substackcdn.com/image/fetch/$s_!_CJL!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe73a3629-dcbf-41b8-847a-eb73d5720486_1456x640.png 848w, https://substackcdn.com/image/fetch/$s_!_CJL!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe73a3629-dcbf-41b8-847a-eb73d5720486_1456x640.png 1272w, https://substackcdn.com/image/fetch/$s_!_CJL!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe73a3629-dcbf-41b8-847a-eb73d5720486_1456x640.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!_CJL!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe73a3629-dcbf-41b8-847a-eb73d5720486_1456x640.png" width="1456" height="640" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e73a3629-dcbf-41b8-847a-eb73d5720486_1456x640.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:640,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!_CJL!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe73a3629-dcbf-41b8-847a-eb73d5720486_1456x640.png 424w, https://substackcdn.com/image/fetch/$s_!_CJL!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe73a3629-dcbf-41b8-847a-eb73d5720486_1456x640.png 848w, https://substackcdn.com/image/fetch/$s_!_CJL!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe73a3629-dcbf-41b8-847a-eb73d5720486_1456x640.png 1272w, https://substackcdn.com/image/fetch/$s_!_CJL!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe73a3629-dcbf-41b8-847a-eb73d5720486_1456x640.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>So what&#8217;s the takeaway? Larger ProGen3 models consistently generate more valid and diverse sequences, generalize better to unseen data, improve infilling tasks, respond more robustly to finetuning, and achieve higher levels of protein expression&#8212;all critical for advancing protein engineering.</p><p>This makes them invaluable for protein engineering. In genome editing, this translates into creating tailor-made CRISPR/Cas systems for diseases where wild-type Cas9 fails. Like low-resource language translation (e.g. Sanskrit), bigger models help solve harder problems.</p><p>The same architectural playbook that leapt from NLP to vision in 2020 now underpins advances in protein biology. Scaling laws are once again proving general and predictive.</p><p>In a related signal of biology's AI transformation, the FDA is beginning to <a href="https://www.fda.gov/news-events/press-announcements/fda-announces-plan-phase-out-animal-testing-requirement-monoclonal-antibodies-and-other-drugs">phase out</a> mandatory animal testing requirements for certain therapies. This marks a watershed moment in regulatory science. What&#8217;s taking the place of in vivo trials? The agency is embracing AI-based computational models for toxicity and in vitro systems like organoids and engineered cell lines. These approaches promise faster, more scalable, and ethically sound paths to safety evaluation&#8212;ushering in a future where biological insight is increasingly model-driven from the start.</p><h3><strong>Big Tech</strong></h3><p>In early April, Meta <a href="https://ai.meta.com/blog/llama-4-multimodal-intelligence/">released</a> its Llama 4 model series on a Saturday&#8212;an odd move in a week dominated by Trump&#8217;s "Liberation Day" tariffs. Unlike Llama 3, Llama 4 adopts a Mixture-of-Experts (MoE) architecture, a concept revived from Google&#8217;s 2021 Switch Transformer and Mistral&#8217;s 2023 work. It also introduces "early fusion," combining multiple input streams&#8212;text, images, video&#8212;into a unified representation upfront, enabling more native cross-modal reasoning. The largest model, "Behemoth," is still in training.</p><p>However, controversy hit fast. Meta was accused of gaming the LMSYS Chatbot Arena leaderboard by submitting an optimized-for-evaluation variant of Llama 4 Maverick. LMSYS later clarified it was a fine-tuned, human-preference-aligned version&#8212;not the production model. Naughty, naughty.</p><p>Google <a href="https://blog.google/products/gemini/gemini-2-5-pro-updates/">responded</a> at Cloud Next with the open release of Gemini 2.5 Pro via Vertex AI, supporting 1M-token multimodal prompts. Flash, their latency-optimized sibling, now supports 2M tokens. Google&#8217;s message: context window size&#8212;not just model size&#8212;is the next frontier.</p><p>OpenAI <a href="https://openai.com/index/gpt-4-1/">followed</a> with GPT-4.1, Mini, and Nano. All three support 1M-token contexts and cost ~26% less than GPT-4o. This reset transforms long-context from premium feature to baseline offering, pressuring Anthropic and pre-empting Google&#8217;s 2M-token Flash launch.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://press.airstreet.com/subscribe?"><span>Subscribe now</span></a></p><h3><strong>Hardware</strong></h3><p>Google&#8217;s new TPUv7 (Ironwood) is <a href="https://x.com/itsclivetime/status/1910026066129014868">optimized</a> for inference rather than training&#8212;an indicator that genAI is transitioning from demo to deployment. Ironwood competes well with NVIDIA&#8217;s B200 in compute, memory capacity, and bandwidth, though trails slightly in interconnect speed.</p><p>NVIDIA, meanwhile, <a href="https://blogs.nvidia.com/blog/nvidia-manufacture-american-made-ai-supercomputers-us/">executed</a> a masterstroke in supply chain strategy. Blackwell GPU production began at TSMC&#8217;s Phoenix, AZ fab, marking their first U.S. silicon. Next: two AI server plants in Texas via Foxconn and Wistron, targeting production within 12&#8211;15 months.</p><p>CEO Jensen Huang then dropped a headline number: $500B in U.S.-built AI servers over four years. The signal? Realign supply chains to dodge tariffs, qualify for CHIPS-Act funding, and soothe hyperscaler fears over geopolitical risk. If realized, it could shift data center capex from Taiwan to Texas.</p><h3><strong>Autonomy</strong></h3><p>Waymo and Wayve both struck major partnerships with Japanese automakers&#8212;<a href="https://waymo.com/blog/2025/04/waymo-and-toyota-outline-strategic-partnership">Toyota</a> and <a href="https://wayve.ai/press/nissan-announcement/">Nissan</a>, respectively. Wayve is opening a Tokyo R&amp;D hub and integrating its AI Driver into Nissan vehicles by 2027. Waymo, already <a href="https://waymo.com/blog/2025/05/scaling-our-fleet-through-us-manufacturing">running</a> 250K+ weekly paid trips across four U.S. cities, is expanding to 3,500 vehicles. Anecdotally, Waymo density in SF is way up.</p><h3><strong>Defense</strong></h3><p>Europe is no longer just talking about autonomy&#8212;it&#8217;s fielding it. On April 14, NATO <a href="https://www.ft.com/content/7f80b1bc-114c-4a00-ad06-6863fb435822">quietly approved</a> Palantir&#8217;s Maven Smart System: a genAI command-and-control stack that integrates live sensor data and delivers real-time operational awareness. Procured in just six months, this reflects urgency over a weakened U.S. umbrella and accelerating AI militarization by adversaries.</p><p>Just weeks later, London&#8217;s Delian Alliance Industries <a href="https://press.airstreet.com/p/delian-alliance-industries-launches-interceptigon">unveiled</a> Interceptigon: a family of GPS-denied, autonomous strike drones designed for swarming, one-way missions. Built on their LAST sensor and OSIRIS visual nav module, Interceptigon flips deterrence economics. Cheap, attritable drones that can threaten billion-dollar ships&#8212;launched from land or sea, no comms needed.</p><div id="youtube2-DalU97_Fduk" class="youtube-wrap" data-attrs="{&quot;videoId&quot;:&quot;DalU97_Fduk&quot;,&quot;startTime&quot;:null,&quot;endTime&quot;:null}" data-component-name="Youtube2ToDOM"><div class="youtube-inner"><iframe src="https://www.youtube-nocookie.com/embed/DalU97_Fduk?rel=0&amp;autoplay=0&amp;showinfo=0&amp;enablejsapi=0" frameborder="0" loading="lazy" gesture="media" allow="autoplay; fullscreen" allowautoplay="true" allowfullscreen="true" width="728" height="409"></iframe></div></div><p>Taken together, Palantir and Delian sketch a new doctrine: combine AI-enabled battle management with sovereign, disposable hardware that functions even in signal-denied environments. This creates a fast, cheap, politically independent deterrent that European nations can build and own now&#8212;not in a decade.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/p/the-state-of-ai-2025-may?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://press.airstreet.com/p/the-state-of-ai-2025-may?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><h3><strong>Research papers</strong></h3><p><strong><a href="https://qwenlm.github.io/blog/qwen3/">Qwen3: Think Deeper, Act Faster</a></strong>, <em>Qwen</em></p><p>In this paper, the authors introduce Qwen3, a new family of large language models designed to balance deep reasoning with fast response times through hybrid &#8220;thinking&#8221; and &#8220;non-thinking&#8221; modes. The flagship model, Qwen3-235B-A22B, achieves competitive results on coding, math, and general benchmarks, rivaling models like DeepSeek-R1 and Gemini-2.5-Pro. Smaller models, such as Qwen3-4B, match or exceed the performance of much larger predecessors.</p><p>Qwen3 supports 119 languages and dialects, and its pretraining dataset is nearly double that of Qwen2.5, at 36 trillion tokens. The training pipeline combines chain-of-thought data, reinforcement learning, and mode fusion to enable both step-by-step reasoning and rapid answers. The models are open-weighted, available under Apache 2.0, and optimized for agentic tasks and tool use.</p><p>This research advances practical AI by enabling configurable reasoning depth, efficient deployment, and broad multilingual support, making it suitable for global, real-world applications.</p><p><strong><a href="https://www.nature.com/articles/s41586-025-08866-7?linkId=13898052">Towards conversational diagnostic artificial intelligence</a></strong>, <em>Google Research, Google DeepMind</em></p><p>In this paper, the authors introduce AMIE (Articulate Medical Intelligence Explorer), a large language model-based AI system optimized for diagnostic dialogue in medicine. The goal is to approximate clinician-level expertise in history-taking, diagnostic reasoning, and patient communication.</p><p>The authors conduct a randomized, double-blind crossover study comparing AMIE to 20 primary care physicians across 159 simulated patient scenarios, evaluated by both specialist physicians and patient-actors. AMIE demonstrates higher diagnostic accuracy than physicians, with superior performance on 30 out of 32 specialist-rated axes and 25 out of 26 patient-actor axes. The study finds that AMIE matches physicians in information acquisition but outperforms them in interpreting information for differential diagnosis.</p><p>Caveats include the use of a synchronous text-chat interface, which is not standard in clinical practice, and the simulated nature of the patient scenarios. The research highlights the potential for LLMs to augment or scale access to high-quality diagnostic dialogue, with implications for telemedicine and healthcare accessibility.</p><p>A second paper by the same group, <strong><a href="https://www.nature.com/articles/s41586-025-08869-4?linkId=13898054">Towards accurate differential diagnosis with large language models</a></strong>, presents an evaluation of AMIE on 302 difficult New England Journal of Medicine case reports. Stand-alone, AMIE captured the correct diagnosis in its top-10 list 59% of the time versus 34 % for unassisted board-certified clinicians. But when clinicians used AMIE as an assistant their accuracy rose to 52%, outperforming conventional search tools at 44%. Interaction times stayed flat (~7 min) and AMIE&#8217;s suggestions were judged more comprehensive and appropriate, yet the study is limited to text-only case narratives and to the &#8220;puzzle&#8221; style of NEJM CPCs rather than routine clinic data. The results suggest that a domain-specialised LLM can both automate and augment differential-diagnosis workflows, pointing to near-term uses in tele-triage and specialist decision support.</p><p><strong><a href="https://arxiv.org/pdf/2504.04022">Rethinking Reflection in Pre-Training</a></strong>, <em>Essential AI</em></p><p>In this paper, the authors investigate how reflective reasoning&#8212;specifically, a model&#8217;s ability to recognize and correct its own or others&#8217; errors&#8212;emerges during the pre-training phase of large language models, rather than only during post-training with reinforcement learning. They introduce adversarial datasets across mathematics, coding, logical reasoning, and knowledge acquisition, where deliberate errors are inserted into chains-of-thought, and measure whether models can recover the correct answer.</p><p>Experiments with OLMo-2 and Qwen2.5 models show that even partially pre-trained models exhibit both situational and self-reflection, with explicit reflection and correction abilities improving as pre-training compute increases. The use of simple trigger phrases like &#8220;Wait,&#8221; enhances explicit reflection rates and accuracy. The study also quantifies the trade-off between train-time and test-time compute for reflective reasoning.</p><p>This work matters because it suggests that reflective reasoning can be instilled during pre-training, potentially reducing the need for expensive post-training interventions and enabling more robust, self-correcting AI systems in real-world applications.</p><p><strong><a href="https://arxiv.org/pdf/2504.07091">AssistanceZero: Scalably Solving Assistance Games</a></strong>, <em>UC Berkeley</em></p><p>In this paper, the authors introduce AssistanceZero, a scalable approach to solving assistance games&#8212;an alternative to reinforcement learning from human feedback (RLHF) for training AI assistants. Assistance games explicitly model the interaction between a user and an assistant as a two-player game with shared but partially hidden goals, aiming to address RLHF&#8217;s issues like incentives for deception and lack of goal uncertainty.</p><p>The authors develop a challenging Minecraft-based benchmark (MBAG) with over 10^400 possible goals and show that standard RL methods like PPO fail to produce helpful assistants in this setting. AssistanceZero extends AlphaZero by adding neural network heads to predict human actions and rewards, enabling effective planning under uncertainty via Monte Carlo tree search.</p><p>Experiments demonstrate that AssistanceZero-trained assistants outperform both model-free RL and imitation learning baselines, reducing human effort and displaying adaptive behaviors. The work suggests assistance games could be a tractable and more robust framework for training collaborative AI systems, with potential applications in domains like AI pair programming.</p><p><strong><a href="https://www.biorxiv.org/content/10.1101/2025.04.09.648075v2.full.pdf">Atom level enzyme active site scaffolding using RFdiffusion2</a></strong>, <em>University of Washington, MIT, HHMI</em></p><p>In this paper, the authors introduce RFdiffusion2, a deep generative model for de novo enzyme design that scaffolds enzyme active sites at the atomic level. It overcame previous limitations that required residue-level specification and pre-assigned sequence indices. RFdiffusion2 can directly generate protein structures from minimal, sequence-agnostic descriptions of catalytic functional group locations, eliminating the need for computationally expensive rotamer and index enumeration.</p><p>The model was evaluated on a new Atomic Motif Enzyme benchmark of 41 diverse active sites, where it successfully scaffolded all 41, compared to 16/41 for the previous state-of-the-art. Experimental validation showed that RFdiffusion2 could generate active enzymes for several reactions, including cases where the active site geometry was derived from quantum chemistry rather than known structures.</p><p>The work demonstrates how atomic-resolution generative models can expand the design space for functional proteins, with potential applications in enzyme engineering, small molecule binding, and broader protein design tasks.</p><p><strong><a href="https://www.biorxiv.org/content/10.1101/2025.04.02.646906v1.full.pdf">ATOMICA: Learning Universal Representations of Intermolecular Interactions</a></strong>, <em>Harvard, MIT</em></p><p>In this paper, the authors introduce ATOMICA, a geometric deep learning model designed to learn universal, atomic-scale representations of intermolecular interactions across diverse biomolecular modalities, including proteins, nucleic acids, small molecules, and metal ions. Unlike prior models that focus on single interaction types, ATOMICA is trained on over two million interaction complexes and uses a self-supervised denoising and masking objective to generate hierarchical embeddings at the atom, block, and interface levels.</p><p>The authors demonstrate that ATOMICA generalizes across molecular classes, recovers shared physicochemical features, and outperforms modality-specific models in masked block identity prediction&#8212;showing up to 190% improvement in low-data modalities like protein-DNA interactions. The model&#8217;s latent space captures compositional and chemical similarities, and its embeddings enable the construction of modality-specific protein interface networks that reveal disease pathways and predict disease-associated proteins.</p><p>Caveats include reliance on high-quality structural data and limited coverage of intrinsically disordered regions. This work matters for AI-driven biology because it enables systematic, transferable modeling of molecular interactions, with applications in disease pathway analysis, drug discovery, and functional annotation of uncharacterized proteins.</p><p><strong><a href="https://blog.google/technology/ai/dolphingemma/">DolphinGemma: How Google AI is helping decode dolphin communication</a></strong>, <em>Google, Georgia Tech, Wild Dolphin Project</em></p><p>In this paper, the authors present DolphinGemma, a foundational AI model designed to analyze and generate dolphin vocalizations using a large, labeled dataset from the Wild Dolphin Project. The model leverages Google&#8217;s SoundStream tokenizer and a ~400M parameter architecture, enabling it to run directly on Pixel smartphones in the field.</p><p>The research aims to identify patterns and structure in dolphin communication, predicting subsequent sounds in a sequence much like language models do for human speech. Experiments involved training on decades of paired audio, video, and behavioral data, allowing the model to cluster and predict natural dolphin sound sequences and generate synthetic dolphin-like sounds.</p><p>A notable caveat is that DolphinGemma is trained specifically on Atlantic spotted dolphins, so adaptation is needed for other species. The work demonstrates how lightweight, on-device AI can accelerate the analysis of complex animal communication, with real-world applications in field research, interspecies interaction, and broader bioacoustics studies.</p><p><strong><a href="https://arxiv.org/pdf/2504.06231">Orb-v3: atomistic simulation at scale</a></strong>, <em>Orbital Materials</em></p><p>In this paper, the authors introduce Orb-v3, a new family of universal machine learning interatomic potentials (MLIPs) designed for atomistic simulations at scale. The work addresses the challenge of achieving high accuracy, low latency, and scalability for large atomic systems, aiming to bridge the gap between universality and computational efficiency.</p><p>The authors present a range of Orb-v3 models that trade off between conservatism, neighbor limits, and dataset choice. Notably, non-conservative, non-equivariant models can achieve competitive accuracy on physical property predictions, including those requiring higher-order derivatives, while being up to 10&#215; faster and using 8&#215; less memory than alternatives. Benchmarks on Matbench Discovery and MDR phonon datasets show that Orb-v3 models match or outperform state-of-the-art MLIPs in both speed and accuracy.</p><p>The paper also introduces &#8220;equigrad&#8221; regularization to improve rotational invariance and a confidence head for uncertainty estimation. These advances enable efficient, large-scale simulations, making Orb-v3 relevant for real-world applications such as materials discovery and mesoscale molecular dynamics.</p><p><strong><a href="https://www.pi.website/blog/pi05">&#960;0.5: a VLA with Open-World Generalization</a></strong>, <em>Physical Intelligence</em></p><p>In this paper, the authors introduce &#960;0.5, a vision-language-action (VLA) model designed to enable robots to generalize to entirely new environments, such as cleaning homes not seen during training. The model is co-trained on heterogeneous data, including multimodal web data, robotic demonstrations, and verbal instructions, allowing it to learn both physical skills and semantic task understanding.</p><p>Experiments show that &#960;0.5 achieves high out-of-distribution (OOD) follow and success rates&#8212;94% in both metrics&#8212;when evaluated on tasks like putting away dishes or making beds in new homes. Ablation studies reveal that web data is crucial for recognizing novel objects, while data from diverse robot types improves overall policy performance.</p><p>The approach uses a unified model for both high-level planning and low-level motor control, following a chain-of-thought process. This work demonstrates practical progress toward robots that can adapt to real-world, unstructured environments, with implications for home automation and service robotics.</p><p><strong>**Bonus: The <a href="https://www.nature.com/articles/d41586-025-01125-9">most highly cited papers</a> of the last century!</strong></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/p/the-state-of-ai-2025-may?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://press.airstreet.com/p/the-state-of-ai-2025-may?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><h3><strong>Datasets and benchmarks</strong></h3><p><strong><a href="https://arxiv.org/pdf/2504.15521">The Bitter Lesson Learned from 2,000+ Multilingual Benchmarks</a></strong>, <em>Alibaba, Monash University, The University of Edinburgh</em></p><p>In this paper, the authors analyze over 2,000 multilingual (non-English) NLP benchmarks from 148 countries, published between 2021 and 2024, to assess the state of multilingual evaluation for large language models (LLMs). They find that English remains overrepresented, even when deliberately excluded, and that most benchmarks are based on original language content rather than translations. High-resource languages and countries dominate, while low-resource languages are underrepresented.</p><p>The study compares benchmark results with human judgments, revealing that STEM-related tasks (e.g., ARC, MGSM) correlate well with human preferences (Spearman&#8217;s &#961;: 0.70&#8211;0.85), but traditional NLP tasks like question answering show much weaker alignment (&#961;: 0.11&#8211;0.30). Localized benchmarks align better with human judgments than translated ones.</p><p>The authors highlight the need for culturally and linguistically authentic benchmarks, propose principles for effective evaluation, and call for global collaboration. This work matters for developing LLMs that serve diverse real-world users more equitably.</p><p><strong><a href="https://arxiv.org/pdf/2504.20879">The Leaderboard Illusion</a></strong>, <em>Cohere, Princeton University, Stanford University</em></p><p>In this paper, the authors analyze the reliability and fairness of Chatbot Arena, a widely used leaderboard for evaluating large language models (LLMs) through human preference voting. They uncover that a small group of providers&#8212;mainly large industry labs&#8212;benefit from undisclosed private testing, selective score reporting, and higher sampling rates, which skews rankings in their favor.</p><p>Experiments show that submitting multiple private model variants and only publishing the best result can inflate Arena scores by up to 100 points, even when underlying models are similar. The authors also demonstrate that proprietary models receive a disproportionate share of user data, with OpenAI and Google together accessing nearly 40% of all Arena prompts, while 83 open-weight models share less than 30%.</p><p>The study finds that access to Arena data enables overfitting, with models trained on this data doubling their win rates on Arena-style benchmarks but not improving on out-of-distribution tasks. The paper concludes with recommendations for transparent policies and fairer evaluation practices, highlighting the importance of trustworthy benchmarks for real-world AI deployment and research progress.</p><p><strong><a href="https://arxiv.org/pdf/2503.21557">debug-gym : A Text-Based Environment for Interactive Debugging</a></strong>, <em>Microsoft Research, McGill University</em></p><p>In this paper, the authors introduce debug-gym, a text-based environment designed to help large language model (LLM) agents perform interactive debugging on code repositories. The environment provides agents with tools such as a Python debugger (pdb), file viewers, and code rewriting utilities, all accessible through structured text commands.</p><p>The authors evaluate three types of LLM-based agents&#8212;rewrite, debug, and debug(5)&#8212;across benchmarks including Aider, Mini-nightmare, and SWE-bench-Lite. Results show that while strong LLMs can leverage interactive tools to improve debugging, most agents struggle to use these tools effectively, especially on complex, real-world tasks. Notably, agents with access to pdb outperform baselines on more challenging debugging scenarios, but the benefit is less clear on simpler code generation tasks.</p><p>The work highlights the need for better agent design and training data reflecting sequential decision-making. This research matters for developing AI systems that can autonomously debug and maintain software, a capability with direct relevance to real-world software engineering workflows.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://press.airstreet.com/subscribe?"><span>Subscribe now</span></a></p><h3><strong>Investments</strong></h3><p><strong><a href="https://www.vestbee.com/blog/articles/arx-robotics-secures-31-m">ARX Robotics</a></strong>, a German defense startup making self-driving modular battlefield robots, raised &#8364;31 million in a financing round led by HV Capital with participation from Omnes Capital and NATO&#8217;s Innovation Fund; the company declined to share its new valuation.</p><p><strong><a href="https://www.businesswire.com/news/home/20250425073932/en/P-1-AI-Comes-Out-of-Stealth-Aims-to-Build-Engineering-AGI-for-Physical-Systems">P-1 AI</a></strong>, a startup developing AI-powered engineering assistants, raised a $23M seed financing round from Radical Ventures, Village Global, and Schematic Ventures.</p><p><strong><a href="https://techcrunch.com/2025/04/22/ex-meta-engineer-raises-14m-for-lace-an-ai-powered-revenue-generation-software-startup/">Lace AI</a></strong>, which provides AI-powered revenue intelligence software for home services call centers, raised a $14M seed financing round from Bek Ventures and Canvas Ventures; the valuation was not disclosed.</p><p><strong><a href="https://www.globenewswire.com/news-release/2025/04/25/3068298/0/en/Reducto-Raises-24-5M-Series-A-Round-to-Help-Enterprises-Unlock-Unstructured-Data.html">Reducto</a></strong>, a company building AI-powered document parsing and ingestion pipelines for enterprises, raised a $24.5M Series A led by Benchmark with participation from First Round Capital and BoxGroup; the valuation was not disclosed.</p><p><strong><a href="https://www.noxtua.com/news/press-releases/series-b-noxtua-raises-80-million-euro">Noxtua</a></strong>, the sovereign European Legal AI company, raised an &#8364;80.7M Series B from C.H.Beck, Northern Data, and CMS.</p><p><strong><a href="https://finance.yahoo.com/news/exclusive-sequoia-backed-ai-startup-120000058.html">Listen Labs</a></strong>, an AI startup that automates large-scale voice interviews for customer research, raised $27M in Seed and Series A financing rounds led by Sequoia Capital, with participation from Bryan Schreier; the valuation was not disclosed.</p><p><strong><a href="https://www.prosus.com/news-insights/group-updates/2025/prosus-ventures-and-a16z-co-lead-dollar-6m-seed-round-in-nexad">Nexad</a></strong>, a startup building a native advertising platform for AI apps, raised $6 million in a seed financing round from A16z Speedrun and Prosus Ventures.</p><p><strong><a href="https://techcrunch.com/2025/04/11/blue-water-autonomy-comes-out-of-stealth-promising-autonomous-naval-ships/">Blue Water</a></strong>, a developer of autonomous ships for defense and commercial maritime applications, raised a $14M seed financing round from Eclipse, Riot, and Impatient Ventures.</p><p><strong><a href="https://menlovc.com/perspective/leading-goodfires-50m-series-a-to-interpret-how-ai-models-think/">Goodfire</a></strong>, an AI interpretability research company, raised a $50M Series A from Menlo Ventures, Lightspeed Venture Partners, and Anthropic; the valuation was not disclosed.</p><p><strong><a href="https://www.mechanize.work/">Mechanize</a></strong>, a startup developing virtual work environments and training data to automate the economy, raised a financing round from Nat Friedman and Daniel Gross, and Patrick Collison; the amount raised and valuation were not disclosed.</p><p><strong><a href="https://www.finsmes.com/2025/04/scout-ai-raises-15m-in-seed-funding.html">Scout AI</a></strong>, a developer of embodied AI foundation models for defense robotics, raised a $15M seed round from Align Ventures, Booz Allen Ventures, and Draper Associates.</p><p><strong><a href="https://www.vestbee.com/blog/articles/portia-ai-secures-4-4-m">Portia AI</a></strong>, an open source SDK platform for building production AI agents, raised a &#163;4.4 million financing round from General Catalyst (lead), First Minute Capital, and Stem AI; the valuation was not disclosed.</p><p><strong><a href="https://www.prnewswire.com/news-releases/gallatin-emerges-from-stealth-with-15m-seed-investment-to-transform-defense-logistics-302422537.html">Gallatin</a></strong>, an AI-powered military logistics software company, raised a $15M financing round from 8VC, Silent Ventures, and Moonshots Capital.</p><p><strong><a href="https://incident.io/blog/incident.io-raises-62m">incident.io</a></strong>, the AI-powered incident management platform, raised a $62M Series B from Insight Partners, Index Ventures, and Point Nine Capital.</p><p><strong><a href="https://archive.is/wVpfj">Thinking Machines Lab</a></strong>, a generative AI research and product company founded by former OpenAI CTO Mira Murati, raised a $2 billion financing round; the valuation is at least $10 billion, but the list of investors was not disclosed.</p><p><strong><a href="https://www.prnewswire.com/news-releases/nuro-announces-series-e-financing-at-6b-valuation-backed-by-leading-financial-and-strategic-investors-302424082.html">Nuro</a></strong>, an autonomous driving technology company, raised a $106M Series E at a $6B valuation from T. Rowe Price, Fidelity, and Tiger Global.</p><p><strong><a href="https://techcrunch.com/2025/04/03/end-to-end-voice-ai-solution-phonic-gets-backing-from-lux/">Phonic</a></strong>, an end-to-end voice AI platform, raised a $4 million seed financing round from Lux Capital, with participation from Replit co-founder Amjad Masad and Hugging Face co-founder Clem Delangue; the valuation was not disclosed.</p><p><strong><a href="https://archive.is/ptokE">Runway</a></strong>, an AI video generation startup, raised $308M in a financing round at a $3B valuation from General Atlantic, Nvidia, and SoftBank Vision Fund 2.</p><p><strong><a href="https://techcrunch.com/2025/04/12/openai-co-founder-ilya-sutskevers-safe-superintelligence-reportedly-valued-at-32b/">Safe Superintelligence</a></strong>, the AI startup focused on building safe superintelligent systems, raised a $2B financing round at a $32B valuation led by Greenoaks.</p><p><strong><a href="https://www.sandboxaq.com/press/sandboxaq-closes-450m-series-e-round-with-expanded-investor-base">SandboxAQ</a></strong>, a B2B company delivering solutions at the intersection of AI and quantum techniques, raised over $450M in a Series E round from investors including Ray Dalio, BNP Paribas, and Google; the valuation was not disclosed.</p><p><strong><a href="https://cast.ai/press-release/cast-ai-closes-108m-series-c-round/">Cast AI</a></strong>, the Kubernetes automation platform for application performance, raised a $108M Series C round from G2 Venture Partners, SoftBank Vision Fund 2, and Agla&#233; Ventures.</p><p><strong><a href="https://x.com/bwhite5290/status/1916831687021117659?s=51&amp;t=8YCMEcmVVXRPm8SXTMgdlw">Axiom</a></strong>, the AI drug toxicity prediction company, raised a $15M seed round from Amplify Partners, Dimension Capital, and Zetta Ventures.</p><p><strong><a href="https://www.linkedin.com/posts/alexanderjohnfitzgerald_delighted-to-announce-isembards-9m-seed-activity-7321063175431348224-hN0q/?utm_source=share&amp;utm_medium=member_ios&amp;rcm=ACoAAA7Fve8B363Azz8dcfBsk0Nf8VDYmxsm_Y0">Isembard</a></strong>, which manufactures and assembles high-precision parts for aerospace, defence, and critical industries, raised a $9M Seed round led by Notion Capital with participation from 201 Ventures and Basis.</p><p><strong><a href="https://x.com/ndrewpignanelli/status/1911826707452997795?s=51&amp;t=8YCMEcmVVXRPm8SXTMgdlw">The General Intelligence Company of New York</a></strong>, which aims to enable one-person billion-dollar companies, raised a $2M financing round from Compound VC and Acrew Capital.</p><p><strong><a href="https://www.linkedin.com/posts/fauna-robotics_weve-raised-30m-to-build-robots-that-thrive-activity-7315130268774748160-S8Xe/">Fauna Robotics</a></strong>, a robotics company building robots that thrive in human spaces, raised a $30M financing round from Kleiner Perkins, Quiet Capital, and Lux Capital.</p><p><strong><a href="https://www.exaforce.com/blogs/reimagining-the-soc-humans-ai-bots-better-faster-cheaper-security-operations">Exaforce</a></strong>, the AI-driven cybersecurity company focused on augmenting SOC operations with task-specific AI agents, raised a $75M Series A financing round led by Khosla Ventures and Mayfield.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/p/the-state-of-ai-2025-may?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://press.airstreet.com/p/the-state-of-ai-2025-may?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><h3><strong>Acquisitions</strong></h3><p><strong><a href="https://www.reuters.com/business/openai-agrees-buy-windsurf-about-3-billion-bloomberg-news-reports-2025-05-06/">OpenAI</a></strong>, the AI research company behind ChatGPT, is in talks to acquire <strong>Windsurf</strong>, an AI-assisted coding tool formerly known as Codeium, for about $3 billion; Windsurf had previously raised over $200 million from investors including General Catalyst and Kleiner Perkins and was last valued at $1.25 billion.</p><p><strong><a href="https://www.paloaltonetworks.com/company/press/2025/palo-alto-networks-announces-intent-to-acquire-protect-ai--a-game-changing-security-for-ai-company">Palo Alto Networks</a></strong>, the global cybersecurity leader, announced its intent to acquire <strong>Protect AI</strong>, a company specializing in security for AI and machine learning applications; the acquisition price was not disclosed.</p><p><strong><a href="https://techcrunch.com/2025/04/23/datadog-acquires-ai-powered-observability-startup-metaplane/">Datadog</a></strong>, the cloud monitoring and security platform, acquired <strong>Metaplane</strong>, an AI-powered data observability startup, for an undisclosed price; Metaplane had previously raised $22.2 million from investors including Khosla Ventures and Y Combinator.</p><p><strong><a href="https://news.crunchbase.com/ma/infinite-reality-buys-touchcast-agentic-ai">Infinite Reality</a></strong>, a spatial computing and AI unicorn, acquired <strong>Touchcast</strong>, an agentic AI company known for its Mentorverse technology, in a cash and stock deal valued at $500 million, bringing Infinite Reality&#8217;s valuation to $15.5 billion.</p><p><strong><a href="https://www.icadmed.com/about/news-events/press-releases/radnet-inc-to-acquire-icad-inc-to-accelerate-ai-powered-early-detection-and-diagnosis-of-breast-cancer/">RadNet</a></strong>, a national provider of diagnostic imaging services and digital health solutions, acquired <strong>iCAD</strong>, a global leader in AI-powered breast health solutions, for approximately $103 million in an all-stock transaction; the acquisition price represents about $3.61 per iCAD share, and the deal is expected to add over 1,500 healthcare provider locations to RadNet&#8217;s DeepHealth subsidiary.</p>]]></content:encoded></item><item><title><![CDATA[State of AI: April 2025 newsletter]]></title><description><![CDATA[What you need to know in AI across geopolitics, big tech, hardware, research, models, datasets, financings and exits over the last 4 weeks.]]></description><link>https://press.airstreet.com/p/the-state-of-ai-2025-april</link><guid isPermaLink="false">https://press.airstreet.com/p/the-state-of-ai-2025-april</guid><dc:creator><![CDATA[Nathan Benaich]]></dc:creator><pubDate>Sun, 06 Apr 2025 13:43:41 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/b2012dfb-7b50-4314-b48f-e9a9d69bfc94_1976x1112.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Hi everyone!</p><p>Welcome to the latest issue of your guide to AI, an editorialized newsletter covering the key developments in AI policy, research, industry, and start-ups over the last month. First up, a few updates:</p><ul><li><p>Congratulations to our friends at <strong>Polar Mist</strong>, who <a href="https://www.polarmist.ai/blog/launch">launched</a> from stealth to build European maritime supremacy with a financing round from us at Air Street and 201 Ventures.</p></li><li><p>We announced new <strong>RAAIS</strong> speakers including the co-founder of Black Forest Labs, the head of policy at Meta, and devrel at Google DeepMind. <a href="http://raais.co">Register your interest</a> for our 13 June conference in London.</p></li><li><p>On <strong>Air Street Press</strong>, I wrote an essay on why startups should <a href="https://press.airstreet.com/p/staking-your-ground">push an ambitious public-facing agenda</a> from day one as well as a piece on the sea change we&#8217;re experiencing for <a href="https://press.airstreet.com/p/our-investment-in-polar-mist">European defense</a>.</p></li><li><p>See you over in the US this month and May where our <strong>Air Street <a href="https://lu.ma/airstreet">event series</a></strong> continues in New York and SF.</p></li></ul><p>I love hearing what you&#8217;re up to, so just hit reply or forward to your friends :-)</p><div><hr></div><h3><em><strong>AI compute and markets</strong></em></h3><p>The AI datacenter market is sending mixed signals. Microsoft, once the poster child of hyperscaler AI ambition, has <a href="https://www.reuters.com/technology/microsoft-pulls-back-more-data-center-leases-us-europe-analysts-say-2025-03-26/">reportedly</a> canceled or deferred data center projects totaling over 2 gigawatts of capacity across the U.S. and Europe. That follows February news of Microsoft backing out of leases worth several hundred megawatts. TD Cowen analysts cited an &#8220;oversupply situation,&#8221; suggesting the company overestimated demand. Some of these cancellations reflect expired LOIs, while others involve pulled leases or project deferrals.</p><p>Yet the rest of the market doesn&#8217;t seem to be slowing down. Amazon, Google, Meta, and Alibaba continue to pour money into AI datacenters. xAI, for example, <a href="https://www.datacenterdynamics.com/en/news/elon-musks-xai-buys-one-million-sq-ft-site-for-second-memphis-data-center/">bought</a> a 1 million-square-foot site in Memphis and <a href="https://archive.is/WBVVo">filed</a> construction permits representing over $400M in project costs. Crusoe is <a href="https://www.globenewswire.com/news-release/2025/03/18/3044956/0/en/Crusoe-Expands-AI-Data-Center-Campus-in-Abilene-to-1-2-Gigawatts.html">expanding</a> its Abilene campus to 1.2 gigawatts using NVIDIA chips. Cerebras is <a href="https://www.businesswire.com/news/home/20250311115186/en/Cerebras-Announces-Six-New-AI-Datacenters-Across-North-America-and-Europe-to-Deliver-Industrys-Largest-Dedicated-AI-Inference-Cloud">building</a> six new inference centers across the U.S. and Europe, and in France, Fluidstack and Mistral are <a href="https://www.fluidstack.io/post/fluidstack-delivering-europes-largest-ai-supercomputer-to-mistral-ai-in-2025">building</a> an 18,000-GPU cluster on a 40MW site south of Paris.</p><p>Then came GTC. NVIDIA unveiled Blackwell Ultra, the next iteration of its platform, designed to accelerate reasoning workloads. The GB300 NVL72 system connects 72 Blackwell Ultra GPUs and 36 Grace CPUs into a single unified system&#8212;essentially one giant GPU. Jensen called himself the &#8220;Chief Revenue Destroyer,&#8221; joking that the launch made the company&#8217;s older chips obsolete. NVIDIA also <a href="https://nvidianews.nvidia.com/news/nvidia-isaac-gr00t-n1-open-humanoid-robot-foundation-model-simulation-frameworks">announced</a> GROOT N1, a foundation model for generalized humanoid reasoning and skills (covered further in the Research section below).</p><p>Against this backdrop, CoreWeave&#8212;often seen as the neocloud vanguard&#8212;went <a href="https://www.sec.gov/Archives/edgar/data/1769628/000119312525044231/d899798ds1.htm">public</a> at the end of March. The company initially targeted a $47&#8211;$55 share price, riding a revenue surge from $16M in 2022 to $1.92B in 2024. But after tepid investor response, it downsized to a $1.5B offering at $40/share, valuing the company at $23B. Critics pointed to CoreWeave&#8217;s dependence on Microsoft, which accounts for 62% of its revenue&#8212;ironic, given Microsoft&#8217;s own data center pullbacks. The company&#8217;s debt-fueled growth model raised eyebrows too: as of 2024, CoreWeave had drawn nearly $8B in debt, largely secured by NVIDIA GPUs. With annual interest rates of 10&#8211;14%, the company is on the hook for nearly $1B a year in financing costs. Despite this, the stock climbed to $61/share post-IPO.</p><h3><em><strong>Integrating all the intelligences</strong></em></h3><p>One of the biggest shifts this month is the accelerating adoption of the Model Context Protocol (MCP)&#8212;an open standard, launched by Anthropic, that lets AI models seamlessly interact with external tools, data sources, and APIs.</p><p>Think of MCP as the equivalent of an API layer for AI agents. Just as REST and GraphQL standardized how web apps talk to servers, MCP defines how LLMs talk to services like Google Drive, WhatsApp, Notion, or Slack&#8212;listing files, sending messages, searching across documents, and more. The goal: to make every web or mobile app instantly &#8220;AI-operable.&#8221; Expect this space to evolve fast&#8212;with implications for tool access, sandboxing, security, and eventually, monetization.</p><p>Example integrations already live: Claude can read and summarize files <a href="https://github.com/modelcontextprotocol/servers/tree/main/src/gdrive">from your Drive</a>, <a href="https://github.com/lharries/whatsapp-mcp">search</a> WhatsApp contacts and send messages, or auto-populate a Notion workspace&#8212;all through MCP. The protocol now has over 30k GitHub stars, and notably, even OpenAI has expressed support. This is especially interesting given OpenAI&#8217;s own Agent API beta (revealed via limited access), which similarly aims to let models call functions and chain tasks via an orchestration layer.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://press.airstreet.com/subscribe?"><span>Subscribe now</span></a></p><h3><em><strong>Research papers</strong></em></h3><p><strong><a href="https://arxiv.org/pdf/2502.14944">Reward-Guided Iterative Refinement in Diffusion Models at Test-Time with Applications to Protein and DNA Design</a></strong>, <em>Genentech, UC Berkeley, Princeton.</em></p><p>In this paper, the authors propose an iterative refinement framework for optimizing reward functions in diffusion models during inference. The key idea is to alternate between noising and reward-guided denoising steps, allowing for gradual correction of errors and optimization of complex reward functions.</p><p>The authors demonstrate superior performance compared to single-shot guidance methods in protein and DNA design tasks. For protein design, their approach effectively optimizes structural properties like secondary structure matching and backbone similarity. In DNA design, it successfully generates cell-type-specific regulatory sequences with high activity levels.</p><p>This research is significant as it addresses limitations of current reward optimization methods in diffusion models. The proposed framework's ability to handle complex rewards and hard constraints has potential applications in computational protein and DNA design, which could accelerate the development of novel biomolecules for various purposes, from therapeutics to synthetic biology.</p><p><strong><a href="https://www.biorxiv.org/content/10.1101/2025.02.27.640681v1.full.pdf">Deep learning guided design of protease substrates</a></strong>,<em> MIT and Microsoft Research.</em></p><p>In this paper, the authors present CleaveNet, an AI-based pipeline for designing protease substrates. CleaveNet consists of a predictor model that assigns cleavage scores and a generator model that produces peptide sequences optimized for desired protease cleavage profiles.</p><p>The authors validate CleaveNet on matrix metalloproteinases, demonstrating its ability to generate substrates capturing known and novel cleavage motifs. Notably, CleaveNet designs MMP13-selective substrates that are efficiently cleaved in vitro, outperforming training data.</p><p>CleaveNet's conditional generation enables targeted design of substrates with specific cleavage profiles across multiple proteases. This approach could accelerate the development of protease-activated diagnostics and therapeutics.</p><p><strong><a href="https://unified-video-action-model.github.io/static/UVA_paper.pdf">Unified Video Action Model</a></strong>, <em>Stanford University</em></p><p>In this paper, the authors introduce the Unified Video Action (UVA) model, designed to jointly optimize video and action predictions for robotics tasks. The model integrates a unified latent representation of video and action data, enabling efficient action inference by decoupling video generation during inference. This approach addresses the trade-off between high temporal speed for actions and high spatial resolution for videos.</p><p>Experiments demonstrate UVA's versatility across seven benchmarks, excelling in multi-task settings with a 20% improvement on PushT-M and 13% on Libero10 compared to baselines. The model also shows robustness to visual disturbances and longer history inputs. In real-world tasks, UVA outperforms specialized models in multi-task scenarios but performs comparably in single-task setups.</p><p>The research highlights UVA's potential as a general-purpose framework for robotics, capable of policy learning, video generation, and dynamics modeling. Its ability to bypass video generation during inference makes it practical for real-time applications, such as robotic manipulation and planning.</p><p><strong><a href="https://www.pyspur.dev/blog/deepseek_open_source_week">DeepSeek-V2: A Strong, Economical, and Efficient Mixture-of-Experts Language Model</a></strong></p><p>This blog post describes the suite of six powerful open-source software libraries that DeepSeek released to tackle key challenges in LLM training and inference. FlashMLA optimizes multi-head latent attention, achieving nearly 90% memory bandwidth utilization. DeepEP provides communication kernels for expert parallelism, significantly reducing latency.</p><p>DeepGEMM, an FP8 matrix multiplication library, delivers up to 2.7x speedup for small matrices. DualPipe introduces bidirectional pipeline parallelism, reducing pipeline bubbles by over 50%. EPLB load balances expert parallelism by duplicating heavily loaded experts.</p><p>3FS, a parallel file system, enables high-throughput data access for training and inference, achieving 6.6 TiB/s read throughput on a 180-node cluster. Smallpond simplifies distributed data processing built on DuckDB.</p><p>These tools, along with the revealed DeepSeek-V3/R1 inference system architecture, showcase the immense engineering effort behind efficiently serving large-scale LLMs in production. The techniques enable DeepSeek to achieve an impressive 545% cost profit margin while being significantly cheaper than competitors.</p><p><strong><a href="https://www.gstatic.com/amie/towards_conversational_ai_for_disease_management.pdf">Towards Conversational AI for Disease Management</a></strong>, <em>Google Research, Google DeepMind</em></p><p>In this paper, the authors advance the diagnostic capabilities of AMIE, an AI system for medical dialogue, to handle disease management over multiple patient visits. AMIE uses a multi-agent architecture with a dialogue agent for conversational interaction and a management reasoning agent for evidence-based care planning.</p><p>In a blinded study comparing AMIE to primary care physicians across 100 multi-visit scenarios, AMIE's management plans were non-inferior overall and scored higher on treatment preciseness and alignment with clinical guidelines.</p><p>The authors also introduced RxQA, a benchmark for medication reasoning. AMIE outperformed physicians on higher difficulty questions, while both benefited from access to drug information.</p><p>This work represents a significant step towards AI-assisted disease management, with potential to improve guideline adherence and quality of care, especially in settings with physician shortages or fragmented health systems. However, further research is needed before real-world deployment.</p><p><strong><a href="https://storage.googleapis.com/deepmind-media/gemini-robotics/gemini_robotics_report.pdf">Gemini Robotics: Bringing AI into the Physical World</a></strong>, <em>Google DeepMind</em></p><p>In this paper, the authors introduce Gemini Robotics, a family of Vision-Language-Action models designed to bridge advanced AI reasoning with physical robotic control. Built on Gemini 2.0, these models exhibit capabilities like object detection, trajectory prediction, and 3D spatial understanding, enabling robots to perform complex manipulation tasks in diverse environments.</p><p>The research highlights Gemini Robotics-ER, which extends embodied reasoning to physical tasks, and Gemini Robotics, which directly controls robots. Experiments demonstrate zero-shot and few-shot task adaptability, such as folding origami or playing cards, with success rates up to 65% in real-world trials. The models also generalize well to unseen instructions, objects, and environments.</p><p>While the models excel in dexterous tasks, limitations include challenges in fine-grained control and long-horizon reasoning. This work is relevant for advancing general-purpose robotics, with applications in manufacturing, healthcare, and domestic assistance, showcasing AI's potential to integrate into real-world physical systems.</p><p><strong><a href="https://sakana.ai/ai-scientist-first-publication/">Compositional Regularization: Unexpected Obstacles in Enhancing Neural Network Generalization</a></strong>, <em>Sakana AI</em></p><p>State of AI Report prediction alert! In this paper, the authors present the results of an experiment where an AI system, The AI Scientist-v2, generated scientific papers entirely autonomously. The system formulated hypotheses, designed experiments, analyzed data, and wrote complete manuscripts without human intervention. Three papers were submitted to an ICLR 2025 workshop, with one paper receiving an average reviewer score of 6.33, surpassing the workshop&#8217;s acceptance threshold.</p><p>The accepted paper explored challenges in enhancing neural network generalization through novel regularization methods, reporting a negative result. While the paper met the workshop&#8217;s standards, it was withdrawn post-review to address ethical concerns about publishing AI-generated research.</p><p>The research highlights the potential of AI in automating scientific discovery, though limitations remain, such as citation errors and the inability to meet higher conference standards. This work underscores the importance of transparency and reproducibility in AI-generated research, with implications for accelerating innovation in fields like medicine and engineering.</p><p><strong><a href="https://www.nature.com/articles/s41592-025-02627-0">ScNET: learning context-specific gene and cell embeddings by integrating single-cell gene expression data with protein-protein interactions</a></strong>, <em>Tel Aviv University</em></p><p>In this paper, the authors introduce scNET, a deep learning framework that integrates single-cell RNA sequencing (scRNA-seq) data with protein-protein interaction (PPI) networks. The model learns gene and cell embeddings that capture both network structure and expression information while reducing noise.</p><p>The authors demonstrate that scNET outperforms traditional imputation methods in elucidating gene-gene relationships and improves cell clustering compared to other state-of-the-art methods. Furthermore, the reconstructed gene expression from scNET enables better identification of differentially enriched pathways across cell types and biological conditions.</p><p>By integrating PPI networks with scRNA-seq data, scNET provides a more comprehensive understanding of cellular heterogeneity and pathway activation. This approach has potential applications in uncovering novel biological insights and identifying therapeutic targets in complex diseases such as cancer and neurodegenerative disorders.</p><p><strong><a href="https://d1qx31qr3h6wln.cloudfront.net/publications/GR00T_1_Whitepaper.pdf">GR00T N1: An Open Foundation Model for Generalist Humanoid Robots</a></strong>, <em>NVIDIA</em></p><p>In this paper, the authors introduce GR00T N1, an open foundation model for generalist humanoid robots. The model leverages a dual-system architecture, combining a vision-language reasoning module with a diffusion transformer for action generation. GR00T N1 is trained on a diverse set of data sources, including real-robot trajectories, human videos, and synthetically generated datasets.</p><p>The model demonstrates strong performance on standard simulation benchmarks across multiple robot embodiments, outperforming state-of-the-art imitation learning baselines. Real-world experiments on the Fourier GR-1 humanoid robot showcase the model's ability to achieve high success rates in language-conditioned bimanual manipulation tasks with limited data.</p><p>GR00T N1's open-source release, including the model checkpoint, training data, and simulation benchmarks, aims to accelerate the development of generalist humanoid robots capable of operating in real-world environments.</p><p><strong><a href="https://transformer-circuits.pub/2025/attribution-graphs/biology.html">On the Biology of a Large Language Model</a></strong>, <em>Anthropic</em></p><p>In this paper, the authors investigate the internal mechanisms of a large language model, Claude 3.5 Haiku, using circuit tracing methodology. They identify interpretable features that form the building blocks of the model's computation.</p><p>The authors uncover sophisticated strategies like multi-step reasoning, planning, and working backwards from goals. They also find highly abstract, generalizable circuits that operate across languages and domains.</p><p>Experiments reveal complex refusal and self-correction mechanisms, as well as hidden goals that can influence model behavior. The authors demonstrate that chain-of-thought reasoning can be unfaithful to the model's true computations.</p><p>This work establishes a foundation for understanding the inner workings of AI models, which is crucial for assessing their capabilities and limitations. Potential applications include auditing models for concerning behaviors and improving interpretability in high-stakes domains like medicine.</p><p><strong><a href="https://www.databricks.com/blog/tao-using-test-time-compute-train-efficient-llms-without-labeled-data">TAO: Using test-time compute to train efficient LLMs without labeled data</a></strong>, <em>Databricks</em></p><p>In this paper, the authors introduce Test-time Adaptive Optimization (TAO), a method to enhance LLM performance without requiring labeled data. TAO leverages test-time compute and reinforcement learning to train models using only input examples, bypassing the need for expensive human annotations.</p><p>The approach involves generating diverse responses to input prompts, scoring these responses using reward models like DBRM, and refining the model through reinforcement learning. Experiments show that TAO improves open-source models like Llama 3.1 8B and 3.3 70B, outperforming traditional fine-tuning on tasks such as SQL generation and document question answering. Notably, TAO achieves results comparable to proprietary models like GPT-4o while maintaining low inference costs.</p><p>This research is relevant for enterprises aiming to adapt AI to specific tasks using existing data. It demonstrates a scalable, cost-efficient alternative to fine-tuning, enabling better performance across diverse applications like finance, coding, and document processing.</p><p><strong><a href="https://arxiv.org/pdf/2504.01017">Scaling Language-Free Visual Representation Learning</a></strong><em>, FAIR (Meta), New York University, Princeton University</em></p><p>In this paper, the authors investigate whether visual self-supervised learning (SSL) can match or surpass language-supervised methods like CLIP in multimodal tasks, particularly Visual Question Answering (VQA). They train both SSL and CLIP models on the same billion-scale MetaCLIP dataset to control for data differences and evaluate performance across 16 VQA tasks.</p><p>The results show that visual SSL models scale better with data and model size, achieving parity with CLIP on VQA tasks, including text-heavy domains like OCR and Chart interpretation. Notably, SSL models trained on text-rich image subsets outperform CLIP in these areas, despite lacking language supervision. Scaling model size up to 7 billion parameters and increasing training data further enhance performance.</p><p>This research highlights the potential of visual SSL to develop robust vision-centric representations without language supervision. It opens pathways for applications in multimodal AI, such as document analysis and visual reasoning, while reducing reliance on paired image-text datasets.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://press.airstreet.com/subscribe?"><span>Subscribe now</span></a></p><h3><em><strong>Dataset and benchmark drops</strong></em></h3><p><strong><a href="https://waymo.com/safety/impact/">Waymo Safety Impact</a></strong></p><p>In this paper, the authors compare the safety performance of Waymo&#8217;s autonomous vehicles (AVs) to human-driven vehicles using crash data and benchmarks. The study focuses on airbag deployment, injury-causing, and police-reported crashes, analyzing incidents per million miles (IPMM) in Phoenix and San Francisco. Results show that Waymo&#8217;s AVs had 83% fewer airbag deployment crashes, 81% fewer injury-causing crashes, and 64% fewer police-reported crashes compared to human benchmarks.</p><p>The methodology accounts for differences in crash reporting standards and adjusts human benchmarks to reflect the driving environments Waymo operates in. Confidence intervals and statistical significance are carefully considered, though limitations include underreporting in human data and challenges in directly comparing AV and human crash definitions.</p><p>This research highlights the potential of AVs to reduce crash severity and frequency, offering real-world implications for safer urban transportation and advancing the development of autonomous driving systems.</p><p><strong><a href="https://arcprize.org/blog/announcing-arc-agi-2-and-arc-prize-2025">Announcing ARC-AGI-2 and ARC Prize 2025</a></strong><em>, ARC Prize Foundation</em></p><p>The blog post announces the launch of ARC-AGI-2, a new benchmark for measuring progress towards artificial general intelligence (AGI), and ARC Prize 2025, a competition to drive open-source progress on efficient, general AI systems.</p><p>ARC-AGI-2 raises the bar for AI difficulty while remaining relatively easy for humans. It tests capabilities like symbolic interpretation, compositional reasoning, and contextual rule application.</p><p>The post introduces an efficiency metric alongside performance scores, emphasizing that intelligence is not just about capability but also the cost at which it is acquired and deployed.</p><p>ARC Prize 2025 offers $1 million in prizes to inspire novel approaches towards AGI. The competition requires open-sourcing solutions to promote collaboration and conceptual progress.</p><p>By focusing on tasks that challenge AI systems in unique ways, ARC-AGI-2 and ARC Prize 2025 aim to guide research efforts towards developing highly efficient, general intelligence systems with real-world applications.</p><p><strong><a href="https://arxiv.org/pdf/2503.10061">Compute Optimal Scaling of Skills: Knowledge vs Reasoning</a></strong>, <em>University of Wisconsin, GenAI at Meta</em></p><p>In this paper, the authors investigate whether compute-optimal scaling behavior in LLMs is skill-dependent, focusing on knowledge-based question answering (QA) and code generation. They find that knowledge QA tasks are capacity-hungry, requiring more parameters, while code tasks are data-hungry, benefiting from larger datasets. These differences persist even when adjusting the pretraining data mix, suggesting fundamental distinctions in how these skills scale.</p><p>The experiments span nine compute scales and use 19 datasets, revealing that skill-specific validation sets significantly impact the estimated optimal parameter counts. For instance, the choice of validation set can lead to a 30-50% variation in optimal parameter estimates, especially at smaller compute scales. This research highlights the importance of considering skill-specific scaling laws and carefully selecting validation sets when training large language models.</p><p><strong><a href="https://cdn.openai.com/papers/22265bac-3191-44e5-b057-7aaacd8e90cd/paperbench.pdf">PaperBench: Evaluating AI&#8217;s Ability to Replicate AI Research</a>,</strong> <em>OpenAI</em></p><p>In this paper, the authors introduce PaperBench, a benchmark designed to evaluate AI agents' ability to replicate state-of-the-art machine learning research. The benchmark includes 20 ICML 2024 Spotlight and Oral papers, each accompanied by detailed rubrics co-developed with the original authors. These rubrics decompose replication tasks into 8,316 gradable subtasks, enabling granular evaluation.</p><p>The experiments show that the best-performing AI agent, Claude 3.5 Sonnet, achieved an average replication score of 21.0%, while human ML PhDs reached 41.4% on a subset of tasks. The study highlights challenges in long-horizon tasks, such as strategizing and executing complex experiments.</p><p>PaperBench also introduces an LLM-based judge for scalable evaluation, achieving an F1 score of 0.83. This research matters as it provides a rigorous framework to assess AI autonomy in replicating complex research, with implications for accelerating AI development and ensuring safe, reliable advancements in machine learning.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/p/the-state-of-ai-2025-april/comments&quot;,&quot;text&quot;:&quot;Leave a comment&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://press.airstreet.com/p/the-state-of-ai-2025-april/comments"><span>Leave a comment</span></a></p><h3><em><strong>Startup financing highlights</strong></em></h3><p><strong>OpusClip</strong>, the multimodal AI video editing company, <a href="https://www.linkedin.com/posts/yzopus_ai-video-startup-opusclip-raises-20-million-activity-7305326391565111296-GR-v/">raised</a> a $20M Series B at a $215M valuation led by SoftBank. The company has grown to circa $20M ARR, up 2.5x year on year.</p><p><strong>Anthropic</strong>, the AI company known for its Claude chatbot, <a href="https://www.bloomberg.com/news/articles/2025-03-03/anthropic-finalizes-megaround-at-61-5-billion-valuation">raised</a> a $3.5B Series E at a $61.5B valuation from Lightspeed Venture Partners, General Catalyst, and Fidelity Management &amp; Research Company.</p><p><strong>Norm Ai</strong>, the regulatory AI agent company, <a href="https://www.prnewswire.com/news-releases/norm-ai-secures-48-million-to-transform-regulations-into-compliance-ai-agents-302398351.html">raised</a> $48M in a financing round from Coatue, Craft Ventures, and Vanguard.</p><p><strong>Eudia</strong>, the AI-powered Augmented Intelligence platform for Fortune 500 legal teams, <a href="https://www.prnewswire.com/news-releases/eudia-secures-up-to-105m-in-series-a-funding-led-by-general-catalyst-to-transform-legal-work-through-ai-powered-augmented-intelligence-302375331.html">raised</a> up to $105M in a Series A financing round led by General Catalyst with participation from Floodgate and Sierra Ventures.</p><p><strong>Alpine Eagle</strong>, the German startup developing cost-efficient airborne counter-drone systems, <a href="https://techcrunch.com/2025/03/05/alpine-eagle-secures-funding-from-european-backers-for-counter-drone-tech-amid-rising-threats/">raised</a> a &#8364;10.25M seed round led by IQ Capital, with participation from General Catalyst and HCVC.</p><p><strong>Isomorphic</strong> <strong>Labs</strong>, an AI-driven drug discovery company, <a href="https://www.isomorphiclabs.com/articles/isomorphic-labs-announces-600m-external-investment-round">raised</a> $600M in its first external financing round led by Thrive Capital with participation from GV and Alphabet.</p><p><strong>Augment</strong>, the AI company building collaborative teammates for logistics, <a href="https://www.goaugment.com/blog/augment-raises-25m-launches-augie">raised</a> $25M in a financing round from 8VC.</p><p><strong>The Bot Company</strong>, a robotics startup focused on household chores, <a href="https://techcrunch.com/2025/03/23/former-cruise-ceo-kyle-vogts-new-robotics-startup-reportedly-raises-another-150m/">raised</a> $150M in a financing round at a $2B valuation led by Greenoaks.</p><p><strong>Ribbon</strong>, the AI-native hiring platform for high-turnover industries, <a href="https://www.dhrmap.com/news/canada-based-ribbon-lands-8m-to-build-the-ai-first-hiring-platform-for-high-turnover-industries">raised</a> $8M in a financing round led by Radical Ventures with participation from Social Leverage and Cadenza Ventures.</p><p><strong>Shield AI</strong>, the deep-tech company building AI-powered autonomy software and defense aircraft, <a href="https://www.prnewswire.com/news-releases/shield-ai-raises-240m-at-5-3b-valuation-to-scale-hivemind-enterprise-an-ai-powered-autonomy-developer-platform-302393843.html">raised</a> $240M in an F-1 strategic financing round at a $5.3B valuation from L3Harris and Hanwha Aerospace.</p><p><strong>Pluralis</strong>, a company enabling decentralized and collaborative AI model training, <a href="https://blog.usv.com/pluralis-towards-actually-open-ai-1">raised</a> a $7.6M seed financing round co-led by USV and CoinFund.</p><p><strong>Causal</strong> <strong>Labs</strong>, the AI company building physics models to predict and control the weather, <a href="https://blog.getcausal.ai/p/causal-labs-towards-causal-intelligence">raised</a> $6M in a seed financing round led by Kindred Ventures, with participation from Refactor and BoxGroup.</p><p><strong>Graphite</strong>, the AI-powered code review platform, <a href="https://techcrunch.com/2025/03/18/anthropic-backed-ai-powered-code-review-platform-graphite-raises-cash/">raised</a> $52M in a Series B financing round from Accel and Anthropic&#8217;s Anthology Fund.</p><p><strong>Cognition</strong> <strong>AI</strong>, a company specializing in artificial intelligence, <a href="https://www.bloomberg.com/news/articles/2025-03-18/cognition-ai-hits-4-billion-valuation-in-deal-led-by-lonsdale-s-firm">raised</a> $4B in a financing round led by Lonsdale's firm.</p><p><strong>Apptronik</strong>, the AI-powered humanoid robotics company, <a href="https://apptronik.com/news-collection/apptronik-closes-additional-series-a-funding">raised</a> a $403M Series A financing round from investors including B Capital, Capital Factory, and Google.</p><p><strong>Frankenburg Technologies</strong>, the Estonian DefenceTech startup developing affordable air defence missiles, <a href="https://investinestonia.com/estonian-defencetech-startup-frankenburg-technologies-triples-valuation-in-a-new-round/">raised</a> &#8364;4M in a financing round at a &#8364;150M valuation from Blossom Capital and Shellona.</p><p><strong>Cartesia</strong>, the voice AI company building ultra-realistic and controllable voice models, <a href="https://cartesia.ai/blog/series-a">raised</a> a $64M Series A led by Kleiner Perkins.</p><p><strong>Celestial</strong> <strong>AI</strong>, the optical interconnectivity startup, <a href="https://news.crunchbase.com/semiconductors-and-5g/chip-startup-celestial-ai-venture-funding-fidelity">raised</a> a $250M Series C1 at a $2.5B valuation from Fidelity Management &amp; Research Co., with participation from BlackRock and Tiger Global Management.</p><p><strong>Dexterity</strong>, the AI robotics startup focused on automation solutions, <a href="https://www.bloomberg.com/news/articles/2025-03-11/ai-robotics-startup-dexterity-lands-1-65-billion-valuation">raised</a> a $1.65B valuation in its latest financing round.</p><p><strong>Flock</strong> <strong>Safety</strong>, the safety technology platform helping communities thrive, <a href="https://www.globenewswire.com/news-release/2025/03/13/3042227/0/en/Flock-Safety-Raises-275-Million-to-Advance-the-Safety-Technology-Ecosystem-and-Strengthen-Communities-Across-the-U-S.html">raised</a> $275M in a financing round at a $7.5B valuation from Andreessen Horowitz, Greenoaks Capital, and Bedrock Capital.</p><p><strong>OpenAI</strong>, the maker of ChatGPT and a leader in generative AI, <a href="https://www.cnbc.com/2025/03/31/openai-closes-40-billion-in-funding-the-largest-private-fundraise-in-history-softbank-chatgpt.html">raised</a> $40B in a financing round at a $300B valuation from investors including SoftBank and Microsoft.</p><p><strong>SplxAI</strong>, the cybersecurity company for AI chatbots, <a href="https://therecursive.com/spxai-seed-funding-launchhub-ai-agents-ai-chatbot-security/">raised</a> a &#8364;6.5M seed financing round led by LAUNCHub Ventures.</p><p><strong>Nexthop</strong> <strong>AI</strong>, a company providing customized networking solutions for AI infrastructure, <a href="https://techstory.in/nexthop-ai-secures-110m-funding-led-by-lightspeed/">raised</a> $110M in a financing round led by Lightspeed Venture Partners with participation from Kleiner Perkins.</p><p><strong>Hook</strong>, the AI-powered music remixing platform enabling users to create and earn from licensed music mashups, <a href="https://www.musicbusinessworldwide.com/ai-music-remixing-platform-hook-raises-3m-from-kygos-palm-tree-crew-the-raine-group-and-more/">raised</a> $3M in a financing round from Khosla Ventures.</p><p><strong>MatX</strong>, a company designing chips and systems to enhance AI computing power, <a href="https://x.com/reinerpope/status/1899486276233146530">raised</a> a &gt;$100M Series A financing round led by Spark Capital with participation from Jane Street Group and Daniel Gross.</p><p><strong>Reflection</strong> <strong>AI</strong>, the company building autonomous coding agents, <a href="https://www.bloomberg.com/news/articles/2025-03-07/ex-deepmind-researchers-new-startup-aims-for-superintelligence">raised</a> $130M in a financing round at a $555M valuation from investors including Sequoia Capital, CRV, and Lightspeed Venture Partners.</p><p><strong>Nirvana</strong>, the AI-powered insurance platform for truckers, <a href="https://techcrunch.com/2025/03/10/nirvana-keeps-on-truckin-with-80m-at-a-830m-valuation-for-its-ai-powered-insurance/">raised</a> $80M in a Series C financing round at an $830M valuation from General Catalyst, Lightspeed Venture Partners, and Valor Equity Partners.</p><p><strong>Ataraxis</strong> <strong>AI</strong>, the AI-driven cancer diagnostics company, <a href="https://techcrunch.com/2025/03/05/not-all-cancer-patients-need-chemo-ataraxis-ai-raised-20m-to-personalize-treatment/">raised</a> a $20.4M Series A financing round led by AIX Ventures with participation from Thiel Bio and Founders Fund.</p><p><strong>Rerun</strong>, the open source data infrastructure company for Physical AI, <a href="https://rerun.io/blog/physical-ai-data">raised</a> $17M in a seed financing round led by Point Nine with participation from Costanoa and Sunflower Capital.</p><p><strong>Ethos</strong>, the AI-powered consultancy platform connecting experts with corporate clients, <a href="https://sifted.eu/articles/deepmind-softbank-startup-consultants-ai">raised</a> a $3.5M financing round from General Catalyst and 8VC.</p><p><strong>Nurix AI</strong>, a company building custom AI agents for enterprise services like sales and customer support, <a href="https://m.economictimes.com/tech/artificial-intelligence/mukesh-bansals-new-startup-nurix-ai-raises-27-5-million-from-accel-general-catalyst/amp_articleshow/113606177.cms">raised</a> $27.5M in a financing round from Accel and General Catalyst.</p><p><strong>Crusoe</strong>, the vertically integrated AI infrastructure provider, <a href="https://crusoe.ai/newsroom/upper90-closes-usd225m-credit-facility-to-crusoe-to-expand-ai-cloud/">raised</a> a $225M financing round from Upper90 and British Columbia Investment Management Corporation to expand its AI cloud platform.</p><h3><em><strong>The rumor mill&#8230;</strong></em></h3><p><strong>Krea Chat</strong>, the image and video generation company, is <a href="https://x.com/arfurrock/status/1897016856705478803?s=51&amp;t=8YCMEcmVVXRPm8SXTMgdlw">rumored</a> to have raised a Series B at a $500M valuation led by Bain Capital. The company has grown from $0 to $8M ARR.</p><p><strong>Cursor</strong>, the AI coding company, is <a href="https://x.com/ArfurRock/status/1906768733135098360">rumored</a> to have raised $625M at a $9.6B post-money valuation led by Thrive and Accel. The company has reached $200M ARR, up 4x since its last round in November 2024.</p><p><strong>Etched</strong>, the transformer-focused ASIC company, is <a href="https://x.com/ArfurRock/status/1906756943349260682">rumored</a> to have raised $85M at a $1.5B valuation, following two other stealth rounds at $500M then $750M just two months ago.</p><p><strong>Perplexity</strong>, the search company, is in <a href="https://www.bloomberg.com/news/articles/2025-03-20/perplexity-in-early-talks-for-funding-at-18-billion-value">early talks</a> for a financing round at an $18B valuation; further details about the amount raised and investors were not disclosed.</p><h3><em><strong>Exits</strong></em></h3><p><strong>CoreWeave </strong>IPO&#8217;d!</p><p><strong>Wiz</strong>, the cloud security platform, was acquired by <strong>Google</strong> <strong>Cloud</strong>. The acquisition price was a whopping $32B in all cash, the company&#8217;s largest ever acquisition.</p><p><strong>Niantic</strong> Inc., the augmented reality and geospatial technology company, was <a href="https://nianticlabs.com/news/niantic-next-chapter">acquired</a> by <strong>Scopely</strong> for $3.85B. It spun off its technology platform, Niantic Spatial, to continue development.</p><p><strong>ServiceNow</strong>, the enterprise workflow automation company, <a href="https://techcrunch.com/2025/03/10/servicenow-buys-moveworks-for-2-85b-to-grow-its-ai-portfolio">acquired</a> <strong>Moveworks</strong>, an AI and automation tools developer, for $2.85B in a mix of cash and stock. Moveworks had previously raised over $300M from investors including Tiger Global and Kleiner Perkins.</p><p><strong>Ampere Computing</strong>, a U.S. chip startup specializing in data center CPUs, was <a href="https://www.reuters.com/markets/deals/softbank-group-acquire-ampere-computing-65-billion-deal-2025-03-19/">acquired</a> by SoftBank Group for $6.5B in an all-cash deal.</p><p><strong>xAI</strong>, a leading AI lab building models and data centers, <a href="https://x.com/elonmusk/status/1905731750275510312?s=51&amp;t=8YCMEcmVVXRPm8SXTMgdlw">acquired</a> <strong>X</strong>, a digital town square with over 600M active users, in an all-stock transaction valuing xAI at $80B and X at $33B.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/p/the-state-of-ai-2025-april?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://press.airstreet.com/p/the-state-of-ai-2025-april?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p>]]></content:encoded></item><item><title><![CDATA[State of AI: March 2025 newsletter]]></title><description><![CDATA[Hi everyone!]]></description><link>https://press.airstreet.com/p/the-state-of-ai-2025-march</link><guid isPermaLink="false">https://press.airstreet.com/p/the-state-of-ai-2025-march</guid><dc:creator><![CDATA[Air Street Press]]></dc:creator><pubDate>Sun, 02 Mar 2025 15:10:50 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/bbd9afa6-2c01-4516-b4cd-5547e08bc69c_1854x1032.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Hi everyone!</p><p>Welcome to the latest issue of the State of AI newsletter, an editorialized newsletter covering the key developments in AI policy, research, industry, and start-ups over the last month. First up, a few updates:</p><ul><li><p>Congratulations to our friends at <strong>Fern Labs</strong>, who <a href="https://fortune.com/2025/02/03/exclusive-fern-labs-preseed-funding-3-million-air-street-capital-palantir-ai-agents/">announced</a> their $3M Pre-Seed round from us at Air Street. Sign up to their <a href="https://fern-labs.ai/">waitlist</a> to build that software tool you always wanted :-)</p></li><li><p>Our events series is in full swing. I&#8217;ll be in Zurich this week after events in Berlin and Munich over the last two. Subscribe to <a href="https://lu.ma/airstreet">our events page</a> to ensure you don&#8217;t miss out.</p></li><li><p>Our annual <strong>RAAIS</strong> <strong>conference</strong> is back on 13 June 2025, <a href="http://raais.co">check it out</a> and register your interest. Announced speakers include Max Jaderberg of Isomorphic Labs, Eiso Kant of Poolside, and Mati Staniszewski of ElevenLabs.</p></li><li><p>Remember to subscribe to <strong><a href="https://press.airstreet.com/">Air Street Press</a></strong> to get all of our news, essays on AI research, startup playbooks, geopolitics, defense, biotech and more directly in your inbox.</p></li><li><p>If you&#8217;re a founder, working on an AI product or doing research in AI and enjoy writing analytical and opinionated pieces on any of the themes you read about here, drop me a line. I&#8217;ll be featuring guest essays on Air Street Press.</p></li></ul><p>We love hearing what you&#8217;re up to and what&#8217;s on your mind, just hit reply or forward to your friends :-)</p><div><hr></div><h3><em><strong>&#127758; The (geo)politics of AI</strong></em></h3><p>Europe is living through what is hopefully a profound vibe shift: a revised focus on doing what it takes to become competitive on the world stage and master its fate (to the extent that it can). The European Commission&#8217;s <a href="https://commission.europa.eu/publications/2025-commission-work-programme-and-annexes_en">plan</a> ambitiously promises a &#8220;bolder, simpler, faster Union.&#8221; It features a laundry list of initiatives&#8212;from decarbonisation to digitalisation, defense to democracy&#8212;and reads like a manifesto for solving every conceivable problem, yet offers little clarity on how these lofty goals will be achieved without drowning in red tape. What caught everyone&#8217;s attention is the commitment to cut down bureaucratic red tape by 25% (how?)streamline sustainability reporting (is this really what&#8217;s holding Europe back?), and efforts (which are so far not super clear) to advance AI and defense. This ambitious agenda faces significant hurdles, such as geopolitical instability, illegal migration, and attacks on core European principles. While the Commission aims to deliver faster and more effectively, its success will depend on strong cooperation among EU institutions and member states in an increasingly complex and polarized world&#8230;</p><p>Paris recently played host to the AI Action Summit, the latest in a series of conferences hosted by nation states. In contrast to the UK&#8217;s first summit which focused on AI safety, this one was reportedly more of a trade show for AI opportunities. Of note, U.S. Vice President JD Vance delivered a <a href="https://www.spectator.co.uk/article/read-jd-vances-full-speech-on-ai-and-the-eu/">punchy speech</a> (which I loved) to emphasize the importance of embracing AI's potential rather than fixating on its risks. Vance argued that an overly cautious approach could hinder progress and that the "AI future" won't be won by "hand-wringing about safety." Now, you can only imagine the reaction from Team Safety&#8230;, which were largely sidelined from the main stage of his speech and the conference itself.</p><p>Meanwhile, Beijing is backing the rapid proliferation of DeepSeek&#8217;s models across its hospitals, local governments and heavy industries. Indeed, the FT <a href="https://www.ft.com/content/5684fb1f-1a84-4542-8fe9-2fcae9653f87">reported</a> that <em>&#8220;All the major cloud service providers, at least six car manufacturers, several local governments, a number of hospitals and a handful of state-owned enterprises have moved to deploy DeepSeek, with the shift among traditionally conservative institutions particularly striking.&#8221; </em>In one example, a doctor at a public hospital in Hubei province in central China said<em> &#8220;the institution&#8217;s leadership had issued a directive that DeepSeek should be used as a third-party arbiter if two doctors have differing views on a patient&#8217;s treatment.&#8221;</em> Casting this approach to technology diffusion with Europe&#8217;s approach of drafting press releases about competitiveness and high-level commitments (read: often mumbo jumbo) is quite striking.</p><p>And the stakes couldn&#8217;t be higher for Europe as the US pulls further away as the weeks go by. It is now existential for Europe to get serious about enabling rather than hindering the creation of home-grown winners in critical sectors such as AI, defense, and energy. The continent has always had what it takes. But we need policies that accelerate public procurement with significant budgets for new players. We must modernize decaying infrastructure, build and buy anew. We must attract and retain skilled immigration. We must unshackle the formation of new companies from our university inventions.</p><p>On defense alone, there is an urgent need to build up European home-grown winners to plug the enormous gaps if the US is to withdraw further. Meanwhile, the US is supporting its own proto-winners at an increasingly large scale. For example, Anduril has <a href="https://www.corememory.com/p/anduril-takes-over-22-billion-contract">swooped</a> in to salvage the U.S. Army&#8217;s $22 billion Integrated Visual Augmentation System (IVAS) program, a project Microsoft fumbled over the years. What began as a futuristic vision of augmented reality headsets for soldiers devolved into a comedy of errors&#8212;headaches, nausea, and tech that couldn&#8217;t handle bad weather. Microsoft, after years of public flogging and $1.5 billion down the drain, quietly exited stage left, leaving Anduril to play the main character.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://press.airstreet.com/subscribe?"><span>Subscribe now</span></a></p><h3><em><strong>&#127850; Hardware</strong></em></h3><p>Despite significant DeepSeek-driven market jitters caused by people misreading headlines because they don&#8217;t read the actual research paper, big tech and sovereign nation AI infrastructure investments continue seemingly unabated. Over in Saudi Arabia, US-based NVIDIA challenger Groq announced a <a href="https://x.com/gavinsherry/status/1888542635356205264?s=51&amp;t=8YCMEcmVVXRPm8SXTMgdlw">$1.5B deal</a> to expand its inference-focused infrastructure across the country.</p><p>Meanwhile, Microsoft, Alphabet, Amazon and Meta combined have <a href="https://www.ft.com/content/634b7ec5-10c3-44d3-ae49-2a5b9ad566fa">invested capex</a> worth $246 billion in 2024, up from $151 billion in 2023. Indeed, Amazon now plans $100 billion+ investments in 2025 and Meta too is in talks for a $200 billion data center <a href="https://www.reuters.com/technology/meta-talks-200-billion-ai-data-center-project-information-reports-2025-02-26/">investment</a> this year. A few weeks after Satya of Microsoft reaffirmed his commitment to spend $80B to build out Azure AI infrastructure in 2025, rumors started circulating that the company <a href="https://techstartups.com/2025/02/24/microsoft-cancels-ai-data-center-leases-linked-to-openai-is-the-ai-hype-cooling-down/">cancelled leases</a> worth a few hundred megawatts of data center capacity, citing facility/power delays.</p><p>Over in China, Alibaba, the Chinese e-commerce giant, plans to invest a staggering <a href="https://www.bloomberg.com/news/articles/2025-02-24/alibaba-to-spend-53-billion-on-ai-infrastructure-in-big-pivot">$53 billion</a> in AI infrastructure over the next three years. This marks a major pivot for the company as it aims to become a leader in artificial intelligence. Alibaba envisions partnering with companies to develop and apply AI to real-world problems, providing the necessary computing power as models evolve.</p><p>Tesla is making a <a href="https://www.bloomberg.com/news/articles/2025-02-27/tesla-sets-sights-on-waymo-uber-in-california-ride-hailing-bid">bold move</a> to challenge ride-hailing giants like Waymo and Uber by seeking approval to operate its own fleet of vehicles in California. The company has applied for a transportation charter-party carrier permit. This comes as Tesla faces declining auto sales and shrinking profit margins as Elon seems to be far too distracted on DOGE (which is in itself a great initiative) vs. his day job.</p><p>And as the robotics craze continues, Meta is the latest to <a href="https://www.bloomberg.com/news/articles/2025-02-14/meta-plans-major-investment-into-ai-powered-humanoid-robots">announce</a> its entry into humanoid robotics, forming a new team within its Reality Labs division.</p><h3><em><strong>&#127981; Big <s>tech</s> start-ups</strong></em></h3><p>The last few weeks have seen a flurry of model drops from large labs. Over at OpenAI, Sam pre-empted the news with his <a href="https://blog.samaltman.com/three-observations">latest blog</a>, &#8220;There Observations&#8221;, in which he paints an increasingly clear path towards AGI. The company <a href="https://arxiv.org/abs/2502.06807">released a paper</a>, <em>Competitive Programming with Large Reasoning Models</em>, in which they reported another episode of The Bitter Lesson. They built a new o1-ioi system, which used hand-engineered inference strategies, and achieved a gold medal at the 2024 International Olympiad in Informatics under relaxed constraints. However, when OpenAI scaled-up their o3 model solely using RL, the system surpassed those results without relying on domain-specific techniques. Impressively, o3 achieved a gold medal at IOI and a Codeforces rating in the 99.8th percentile. This work showed emergent reasoning strategies, such as self-validation through brute-force solutions, without human intervention.</p><p>Then came the <a href="https://openai.com/index/introducing-gpt-4-5/">research preview</a> of GPT 4.5, which OpenAI touted as an example of how improved capabilities can still come from scaling unsupervised learning during pre-training. The system has better EQ, follows instructions better, has a seemingly better world understanding, and hallucinates less. The company further emphasized that future models will combine this approach with reasoning capabilities to create even more powerful AI systems.</p><p>Fresh off of xAI&#8217;s 100k H100 cluster came <a href="https://x.ai/blog/grok-3">Grok 3 Beta</a>, their newest system designed to excel in reasoning, mathematics, coding, and instruction-following tasks. Grok 3 leverages large-scale reinforcement learning to refine its problem-solving strategies, enabling it to think for seconds to minutes, backtrack, and self-correct. Its Elo score of 1402 in the Chatbot Arena and strong performance on benchmarks like AIME&#8217;25 (93.3%) and GPQA (84.6%) highlight its dominance in both academic and real-world tasks. Similar to OpenAI, the model introduces "Think" mode, allowing users to inspect its reasoning process, and features a smaller, cost-efficient variant, Grok 3 mini.</p><p>A flurry of companies launched DeepResearch (or similarly named) products for real-time knowledge retrieval through web and document search, reasoning, self-correction and re-searching if results aren&#8217;t sufficiently helpful, and finally synthesis of research reports. Many are touting this as the latest &#8220;killer feature&#8221; for AI models and it certainly does look valuable. It&#8217;s interesting to <a href="https://www.understandingai.org/p/these-experts-were-stunned-by-openai?utm_campaign=email-half-post&amp;r=s61&amp;utm_source=substack&amp;utm_medium=email">read examples</a> of the prompts, the reports, and how they&#8217;re evaluated by users with expertise in the field</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!VaAs!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5cd2b0ac-4ce9-4582-830c-6743a9b85dbb_1600x1273.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!VaAs!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5cd2b0ac-4ce9-4582-830c-6743a9b85dbb_1600x1273.png 424w, https://substackcdn.com/image/fetch/$s_!VaAs!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5cd2b0ac-4ce9-4582-830c-6743a9b85dbb_1600x1273.png 848w, https://substackcdn.com/image/fetch/$s_!VaAs!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5cd2b0ac-4ce9-4582-830c-6743a9b85dbb_1600x1273.png 1272w, https://substackcdn.com/image/fetch/$s_!VaAs!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5cd2b0ac-4ce9-4582-830c-6743a9b85dbb_1600x1273.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!VaAs!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5cd2b0ac-4ce9-4582-830c-6743a9b85dbb_1600x1273.png" width="1456" height="1158" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/5cd2b0ac-4ce9-4582-830c-6743a9b85dbb_1600x1273.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1158,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!VaAs!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5cd2b0ac-4ce9-4582-830c-6743a9b85dbb_1600x1273.png 424w, https://substackcdn.com/image/fetch/$s_!VaAs!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5cd2b0ac-4ce9-4582-830c-6743a9b85dbb_1600x1273.png 848w, https://substackcdn.com/image/fetch/$s_!VaAs!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5cd2b0ac-4ce9-4582-830c-6743a9b85dbb_1600x1273.png 1272w, https://substackcdn.com/image/fetch/$s_!VaAs!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5cd2b0ac-4ce9-4582-830c-6743a9b85dbb_1600x1273.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Next came Anthropic&#8217;s turn with <a href="https://www.anthropic.com/news/claude-3-7-sonnet">Claude 3.7 Sonnet</a> and Claude Code. Different to their peers, the company built a hybrid reasoning model that integrates rapid response capabilities with extended, step-by-step reasoning. Unlike its predecessors, this model allows users to toggle between standard and extended thinking modes, offering flexibility for tasks ranging from quick answers to complex problem-solving. Extended thinking mode, available on paid tiers, enhances performance in math, physics, coding, and real-world applications, reflecting a shift from competition-focused optimization to practical business use. API users can even set token budgets to balance speed and quality. And Amazon should be happy too, as it finally launched <a href="https://www.anthropic.com/news/claude-and-alexa-plus">Alexa+</a>, featuring Claude as its new brains.</p><p>Meanwhile, Claude Code is a command-line tool enabling developers to delegate tasks like debugging, test-driven development, and large-scale refactoring directly to the AI. Early tests show significant time savings, with Claude outperforming competitors in coding benchmarks like SWE-bench and TAU-bench. It is particularly useful for long-horizon agent coding tasks, such as those productised by <a href="https://fern-labs.ai/">Fern Labs</a>.</p><p>Finally, Google <a href="https://blog.google/technology/google-d...dates-february-2025/">announced</a> major updates to its Gemini AI models, making the powerful fan-favorite Gemini 2.0 Flash generally available to developers. Built using RL techniques, 2.0 Flash offers enhanced reasoning capabilities across vast amounts of multimodal information. The company also unveiled Gemini 2.0 Pro Experimental, their most capable model yet for complex prompts and coding tasks, and 2.0 Flash-Lite, a cost-efficient alternative that outperforms its predecessor.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://press.airstreet.com/subscribe?"><span>Subscribe now</span></a></p><h3><strong>&#128300;Research papers</strong></h3><p><strong><a href="https://www.biorxiv.org/content/10.1101/2025.02.13.638045v1.full.pdf">Tissue reassembly with generative AI</a></strong>, EPFL, Meta AI, Swiss Institute of Bioinformatics.</p><p>In this paper, the authors introduce LUNA, a generative AI model designed to reconstruct tissue architectures from dissociated single-cell RNA sequencing (scRNA-seq) data. LUNA leverages spatial priors learned from spatial transcriptomics datasets to predict the spatial arrangement of cells based solely on their gene expression profiles.</p><p>The model employs a diffusion-based approach, progressively denoising random noise into spatial cell coordinates, and uses an attention mechanism to capture both local and global cellular interactions. LUNA demonstrated strong performance in reconstructing the MERFISH whole mouse brain atlas, accurately predicting spatial locations for over 1.2 million cells, including unseen cell types, with a Pearson correlation of 0.95 for spatial gene expression patterns. However, it is currently limited to 2D spatial reconstructions.</p><p><strong><a href="https://arxiv.org/pdf/2501.07737">Multi-megabase scale genome interpretation with genetic language models</a></strong>, GSK, Max Planck, ETHZ.</p><p>Phenformer is a deep-learning model designed to predict disease risk directly from whole genome sequences. The model processes up to 88 million base pairs, integrating sequence, cell-type-specific expression, and phenotype data. It identifies disease-relevant cell and tissue types, outperforming state-of-the-art polygenic risk score (PRS) methods in predictive accuracy, particularly for diverse populations.</p><p>Uniquely, this model highlights mechanistic hypotheses, such as liver involvement in psoriasis and optic nerve complications in COPD, supported by existing clinical and epidemiological evidence. The authors demonstrate its ability to stratify individuals into molecular subtypes, revealing co-morbidity patterns.</p><p><strong><a href="https://arxiv.org/abs/2502.07617">Scaling Pre-training to One Hundred Billion Data for Vision Language Models</a></strong>, Google DeepMind.</p><p>The authors investigate the potential of pre-training vision-language models on an unprecedented scale of 100 billion examples. They find that while model performance saturates on many common Western-centric benchmarks, tasks of cultural diversity achieve more substantial gains from this massive web data.</p><p>The paper also analyzes the model's multilinguality, showing gains in low-resource languages. Interestingly, they observe that reducing pretraining dataset size via quality filters like CLIP may inadvertently reduce cultural diversity, even in large-scale datasets.</p><p>These results highlight that the 100-billion example scale is vital for building truly inclusive multimodal systems, even if traditional benchmarks may not benefit significantly. </p><p><strong><a href="https://arxiv.org/pdf/2502.05171">Scaling up Test-Time Compute with Latent Reasoning: A Recurrent Depth Approach</a></strong>, ELLIS Institute T&#252;bingen, University of Maryland, Lawrence Livermore National Laboratory</p><p>In this paper, the authors propose a novel language model architecture that scales test-time computation by reasoning in latent space using a recurrent block. The model is trained on 800 billion tokens and achieves strong performance on reasoning benchmarks, often outperforming larger models.</p><p>Key experiments demonstrate the model's ability to improve accuracy on tasks like mathematical and code reasoning by increasing test-time compute. The model also exhibits useful behaviors like zero-shot adaptive compute and continuous chain-of-thought reasoning.</p><p>However, the model is still a proof-of-concept trained on a limited compute budget. More optimal training may yield even better results.</p><p>This work presents a promising new approach for endowing language models with reasoning capabilities that could enhance their performance on complex real-world tasks. Moving reasoning into high-dimensional latent space, rather than explicit verbalization, opens up new possibilities for developing capable and efficient models.</p><p><strong><a href="https://github.com/GAIR-NLP/LIMR/blob/master/limr.pdf">Less is More for RL Scaling</a></strong>, GAIR-NLP</p><p>But could we scale RL with less? This work explores how reducing model complexity and training data requirements can still achieve competitive or superior performance compared to traditional large-scale RL approaches. The authors show that their approch benchmarks favorably with significantly reduced computational resources: <em>&#8220;With merely 817 curated training samples, LIMO achieves 57.1% accuracy on AIME and 94.8% on MATH, improving from previous SFT-based models' 6.5% and 59.2% respectively, while only using 1% of the training data required by previous approaches.&#8221;</em></p><p>A notable caveat, however, is that the approach may require careful tuning to balance simplicity and performance, which could limit its generalizability across all RL problems.</p><p><strong><a href="https://arxiv.org/pdf/2502.08235">The Danger of Overthinking: Examining the Reasoning-Action Dilemma in Agentic Tasks</a></strong>, University of California Berkeley, Carnegie Mellon University, Alibaba</p><p>How should reasoning models strike the right balance between thinking and acting on their environment? This paper evaluates this question. They identify a phenomenon called "overthinking" where LRMs favor extended internal reasoning over interacting with the environment.</p><p>The authors analyze 4018 trajectories across 19 models on software engineering tasks. They find that higher overthinking scores correlate with decreased task performance, and reasoning models exhibit stronger overthinking tendencies compared to non-reasoning models.</p><p>Notably, selecting the solution with the lowest overthinking score from just 2 samples can improve performance by nearly 30% while reducing computational costs by 43%. The authors suggest that leveraging native function-calling capabilities and selective reinforcement learning could help mitigate overthinking.</p><p><strong><a href="https://ai.meta.com/blog/brain-ai-research-human-communication/">Brain-to-Text Decoding: A Non-Invasive Approach via Typing</a></strong>, Meta AI, Basque Center on Cognition, Brain and Language (BCBL), Rothschild Foundation Hospital.</p><p>In this paper, the authors explore the use of AI to decode language from non-invasive brain recordings and to understand the neural mechanisms of language production. Using MEG and EEG, they recorded brain activity from 35 participants typing sentences and trained an AI model to reconstruct sentences from these signals. The model achieved up to 80% character decoding accuracy with MEG, outperforming EEG systems.</p><p>The second study analyzed how the brain transforms thoughts into words, syllables, and letters. By interpreting MEG signals, the authors identified a dynamic neural code that chains successive representations while maintaining coherence over time.</p><p>Down the line, this kind of work could have applications in restoring communication for individuals with speech impairments, and contributes to understanding the neural basis of language.</p><p><strong><a href="https://www.bakerlab.org/wp-content/uploads/2025/02/science.adu2454.pdf">Computational design of serine hydrolases</a></strong>, University of Washington; Institute for Protein Design; Howard Hughes Medical Institute</p><p>The authors present a computational approach to design serine hydrolases, enzymes with complex active sites that catalyze multistep reactions. Their designed enzymes have catalytic efficiencies up to 220,000 M&#8315;&#185;s&#8315;&#185;, a substantial improvement over previous computational designs. Moreover, crystal structures closely match the design models with sub-angstrom accuracy.</p><p>Of note, the designs have novel folds not seen in natural serine hydrolases, expanding the known structural diversity of this enzyme family. Meanwhile, analysis of the designs provides insights into the geometric basis of catalysis. A popular application of this kind of work includes plastic recycling, where serine hydrolases could break down polyethylene terephthalate (PET).</p><p><strong><a href="https://arxiv.org/pdf/2502.03349">Robust Autonomy Emerges from Self-Play</a></strong>, Apple</p><p>In this paper, the authors explore the use of self-play reinforcement learning to develop robust and naturalistic driving policies for autonomous vehicles. They introduce GIGAFLOW, a simulator capable of training policies on an unprecedented scale, simulating 1.6 billion kilometers of driving in under 10 days using an 8-GPU node. The resulting policy achieves state-of-the-art performance on benchmarks like CARLA, nuPlan, and Waymo Open Motion Dataset, outperforming specialist models trained on benchmark-specific data.</p><p>The experiments demonstrate that self-play, combined with minimalistic reward functions, enables the emergence of diverse and realistic driving behaviors without using human driving data. The policy generalizes across various traffic scenarios, maps, and actor behaviors, achieving robustness with an average of 17.5 years of simulated driving between incidents.</p><p><strong><a href="https://arxiv.org/pdf/2502.20220">Avat3r: Large Animatable Gaussian Reconstruction Model for High-fidelity 3D Head Avatars</a></strong><a href="https://go.v7labs.com/97e32ab0-2373-4717-94de-c7d1413aea38/projects/019553b4-a833-7c50-9176-a0efdc6f5f4b/views/019553b4-a83f-7fb4-a059-8b18d1e3ebcb/entities/01955679-0b03-7905-85cd-9c7e110657b1#:~:text=website-,https%3A//tobias%2Dkirschstein.github.io/avat3r/static/Avat3r_paper.pdf,-research%20summary%20%2D%20website">,</a> Technical University of Munich, Meta Reality Labs Pittsburgh</p><p>The authors present Avat3r, a method for creating high-quality, animatable 3D head avatars from just a few input images. The key innovation is an architecture that predicts 3D Gaussians for each pixel, allowing for detailed reconstructions without relying on a fixed template mesh.</p><p>Avat3r outperforms state-of-the-art methods in both few-shot and single-shot scenarios. Experiments show it produces more expressive avatars with higher rendering quality, better identity matching, and smoother video renderings.</p><p>The method generalizes well to out-of-distribution examples like AI-generated images or antique busts. This opens up potential applications in casual settings where users may only have a few smartphone photos available.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://press.airstreet.com/subscribe?"><span>Subscribe now</span></a></p><h3><strong>&#128738;Dataset and benchmark drops</strong></h3><p><strong><a href="https://www.biorxiv.org/content/10.1101/2025.02.20.639398v1.full.pdf">Tahoe-100M: A Giga-Scale Single-Cell Perturbation Atlas for Context-Dependent Gene Function and Cellular Modeling</a></strong>, Vevo Therapeutics; Arc Institute; Parse Biosciences.</p><p>In this paper, the authors present Tahoe-100M, a single-cell perturbation atlas comprising over 100 million transcriptomic profiles from 50 cancer cell lines treated with 1,100 small-molecule drugs. Using the Mosaic platform, they reduced batch effects by profiling diverse cell lines in parallel, enabling high-throughput single-cell RNA sequencing. The dataset captures 52,886 unique cell line-drug-dose conditions, with a median of 1,287 cells per condition.</p><p>The study highlights drug-induced transcriptional changes, showing that targeted inhibitors like RAS and RAF modulators elicit mutation-specific responses. E-distance metrics quantified the magnitude of drug effects, revealing stronger impacts for cancer-relevant pathways like PI3K/AKT and HDAC inhibitors. Cell cycle analysis demonstrated distinct phase arrests induced by specific drug classes, such as G2/M arrest by HDAC inhibitors.</p><p>This research provides a scalable resource for AI-driven modeling of cellular behavior, with applications in drug discovery, personalized medicine, and understanding treatment resistance mechanisms.</p><p><strong><a href="https://static.scale.com/uploads/654197dc94d34f66c0f5184e/EnigmaEval%20v4.pdf">ENIGMA EVAL: A Benchmark of Long Multimodal Reasoning Challenges</a></strong>, Scale AI, Center for AI Safety, MIT</p><p>This work introduces ENIGMA EVAL, a benchmark designed to evaluate the reasoning capabilities of large language models (LLMs) through complex, multimodal puzzles. The dataset includes 1,184 puzzles sourced from diverse puzzle-solving events, requiring models to synthesize implicit knowledge and perform multi-step deductive reasoning. These puzzles combine text and images, challenging models to uncover hidden connections and solution paths.</p><p>Experiments show that state-of-the-art models achieve only 7.0% accuracy on normal puzzles and 0% on harder ones, highlighting significant gaps in reasoning and problem-solving abilities. The study also reveals that some models struggle with OCR and parsing, though transcription does not drastically improve performance.</p><p>This research matters because it pushes the boundaries of AI evaluation, focusing on unstructured, creative problem-solving. By exposing current limitations, ENIGMA EVAL provides a framework for advancing AI systems capable of tackling real-world challenges requiring flexible reasoning and multimodal understanding.</p><p><strong><a href="https://arxiv.org/pdf/2502.14499">MLGym: A New Framework and Benchmark for Advancing AI Research Agents</a></strong>, Meta, University of California Santa Barbara, University College London</p><p>In this paper, the authors introduce MLGym, a framework for evaluating and developing AI research agents, along with MLGym-Bench, a suite of 13 diverse AI research tasks. The tasks span computer vision, NLP, reinforcement learning, and game theory, requiring agents to generate ideas, process data, implement methods, and analyze results.</p><p>The authors evaluate frontier language models like GPT-4 and Claude on these tasks. They find that while the models can improve on given baselines by tuning hyperparameters, they do not generate novel hypotheses, algorithms, or architectures.</p><p>MLGym enables research on training algorithms like RL for AI research agents. The framework and benchmark are open-sourced to facilitate future work on advancing the AI research capabilities of language model agents.</p><p><strong><a href="https://arxiv.org/pdf/2411.00081">PARTNR: A Benchmark for Planning and Reasoning in Embodied Multi-agent Tasks</a></strong>, FAIR Meta</p><p>In this paper, the authors introduce PARTNR, a benchmark for evaluating embodied AI agents in collaborative household tasks. The dataset contains 100,000 diverse, natural language instructions spanning 60 houses and 5,819 objects. Tasks exhibit real-world characteristics like spatial, temporal, and heterogeneous agent capability constraints.</p><p>The authors analyze state-of-the-art language models on PARTNR, revealing significant limitations in planning, perception, and skill execution. When paired with humans, the models require 1.5x more steps than human-human teams and 1.1x more than a single human, highlighting room for improvement.</p><p>Fine-tuning smaller language models on planning data achieves performance on par with models 9 times larger while being 8.6x faster. PARTNR aims to drive research in collaborative embodied agents, with potential applications in domestic robotics and virtual assistants. The benchmark's realistic tasks and systematic evaluations provide valuable insights for advancing human-robot interaction.</p><p><strong><a href="https://www.primeintellect.ai/blog/synthetic-1">SYNTHETIC-1: Scaling Distributed Synthetic Data Generation for Verified Reasoning</a></strong>, Prime Intellect</p><p>The authors introduce SYNTHETIC-1, a large-scale open-source dataset of 1.4 million verified reasoning tasks spanning math, coding, and science. The dataset is designed to improve reasoning model training by leveraging DeepSeek-R1, a model trained with reinforcement learning and fine-tuned using verified reasoning traces.</p><p>The experiments involve generating tasks such as verifiable math problems, coding challenges with unit tests, and open-ended STEM questions, verified using LLM judges or programmatic methods. Notably, the dataset includes 61,000 synthetic code understanding tasks, which are particularly challenging for state-of-the-art models.</p><p>The research demonstrates that cold-start synthetic data significantly enhances model performance, and distillation from strong teacher models can outperform large-scale reinforcement learning. This work enables globally distributed reinforcement learning with verifiable rewards, allowing anyone to contribute compute.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://press.airstreet.com/subscribe?"><span>Subscribe now</span></a></p><h3><em><strong>&#128640; </strong></em><strong>Funding highlight reel</strong></h3><p><strong>Mercor</strong>, the AI recruiting startup that automates hiring processes, raised a $100M Series B at a $2B valuation from Felicis, Benchmark, and General Catalyst.</p><p><strong>Taktile</strong>, the AI-powered risk decisioning platform for financial services, raised a $54M Series B financing round from Balderton Capital, Index Ventures, and Tiger Global.</p><p><strong>ElevenLabs*</strong>, the AI audio technology company, raised a $180M Series C at a $3.3B valuation from a16z and ICONIQ Growth.</p><p><strong>Fal</strong>, the generative media platform for developers, raised a $49M Series B financing round led by Notable Capital with participation from Andreessen Horowitz and Bessemer Venture Partners.</p><p><strong>Protex AI</strong>, the AI-driven workplace safety company, raised a $36M Series B financing round led by Hedosophia with participation from Salesforce Ventures.</p><p><strong>Saronic</strong>, a company focused on autonomous shipbuilding, raised a $600M Series C at a $4B valuation from Elad Gil and General Catalyst.</p><p><strong>Luminance</strong>, the AI legaltech company automating contract generation and negotiation, raised a $75M Series C financing round from Point72 Private Investments, Forestay Capital, and RPS Ventures.</p><p><strong>Lambda*</strong>, the AI cloud platform provider, raised a $480M Series D financing round co-led by Andra Capital and SGW, with participation from NVIDIA and ARK Invest.</p><p><strong>Together AI</strong>, the AI cloud platform for open source and enterprise AI, raised a $305M Series B financing round led by General Catalyst and co-led by Prosperity7, with participation from Salesforce Ventures and NVIDIA.</p><p><strong>Tolan</strong>, the AI company behind Embodied Companions, raised a $10M financing round from investors including Lachy Groom, Nat Friedman, and Daniel Gross.</p><p><strong>Abridge</strong>, the generative AI platform for clinical conversations, raised a $250M Series D from Elad Gil, IVP, and Bessemer Venture Partners.</p><p><strong>Nomagic</strong>, the Polish startup building AI-powered robotic arms for logistics operations, raised a $44M Series B financing round from the European Bank for Reconstruction and Development (EBRD), Khosla Ventures, and Almaz Capital.</p><p><strong>Achira</strong>, a biotech company blending AI and physics to model molecules, raised a $33M seed financing round backed by Dimension and NVIDIA.</p><p><strong>Prime Intellect</strong>, a company building a peer-to-peer protocol for open-source AI, raised $15M in a financing round led by Founders Fund with participation from Menlo Ventures and Andrej Karpathy.</p><p><strong>Elicit</strong>, the AI platform for evidence-backed decisions, raised a $22M Series A at a $100M valuation from Spark Capital and Footwork.</p><p><strong>Enveda*</strong>, a techbio company using AI to discover new medicines from nature, raised a $150M Series C financing round with investment from Sanofi.</p><p><strong>Unique</strong>, the Swiss AI platform for finance, raised a $30M Series A financing round led by DN Capital and CommerzVentures.</p><p><strong>Latent Labs</strong>, the AI-first frontier bio company, raised a $40M Series A financing round co-led by Radical Ventures and Sofinnova Partners.</p><p><strong>Tana</strong>, the AI-powered knowledge graph for work, raised a $25M financing round led by Tola Capital with participation from Lightspeed Venture Partners and Northzone.</p><p><strong>Fern Labs*</strong>, a London-based startup focused on coordinating networks of AI agents to build software autonomously, raised a $3M pre-seed financing round led by us at Air Street Capital.</p><p><strong>Prior Labs</strong>, a German AI startup focused on building models to analyze tabular data, raised a &#8364;9 million pre-seed financing round from Balderton Capital and XTX Ventures.</p><p><strong>Tines</strong>, the automation platform for enterprise workflows, raised a $125M Series C at a $1.125B valuation from new and existing investors.</p><p><strong>Crescendo</strong>, the AI-powered customer support platform, raised a $50M Series C financing round at a $500M valuation from General Catalyst and Celesta Capital.</p><p><strong>Positron</strong>, a Reno-based AI chip startup focused on inference chips, raised $23.5M in a seed financing round from Valor Equity Partners, Atreides Management, and Flume Ventures.</p><p><strong>Sesame</strong>, developers of an AI voice model and hardware device, raised an undisclosed Series A from a16z.</p><p><strong>Achira</strong>, a techbio company blending AI and physics to model molecules, raised a $33M seed financing round backed by Nvidia.</p><p><strong>Verkada</strong>, the security systems maker specializing in video security cameras, environmental sensors, and alarms, raised $200M in a financing round at a $4.5B valuation led by General Catalyst with participation from Eclipse Ventures.</p><p><strong>Arize</strong>, the AI observability platform for monitoring and evaluating AI models, raised a $70M Series C financing round from Adams Street Partners, M12, and SineWave Ventures.</p><p><strong>Terrain</strong> <strong>Biosciences</strong>, the RNA design-build company leveraging AI for therapeutics and vaccine development, raised $9M in a seed financing round from Magnetic Ventures, Bruker Corporation, and Josef Feldman of Ex Nihilo.</p><p><em><strong>*</strong> denotes companies in which the authors hold shares.</em></p><h3><strong>&#129323;</strong><em><strong> </strong></em><strong>The rumor mill&#8230;</strong></h3><p><strong>Anduril</strong>, the defense-tech company, is raising $2.5 billion in a financing round at a $28 billion valuation led by Founders Fund.</p><p><strong>Anthropic</strong>, the AI startup behind the Claude chatbot, is raising a $3.5B financing round at a $61.5B valuation from Lightspeed Venture Partners, General Catalyst, and Bessemer Venture Partners.</p><p><strong>Thinking Machines Lab</strong>, Mira Murati&#8217;s new AGI company, is in talks to raise $1billion.</p><p><strong>Safe Superintelligence</strong>, an AI company focused on safe AI development, is raising over $1 billion in a financing round at a valuation exceeding $30 billion, with Greenoaks Capital Partners leading the round and investing $500 million.</p><h3><em><strong>&#129309; </strong></em><strong>Exits</strong></h3><p><strong>IBM</strong>, the technology giant, acquired <strong>DataStax</strong>, a company specializing in database and generative AI capabilities, including Apache Cassandra and vector databases, to enhance its watsonx AI offerings. The acquisition price was not disclosed. DataStax had previously raised $342.6M and was valued at $1.6B during its most recent funding round in June 2022, with hundreds of paying customers.</p><p><strong>Ravelin</strong>, the AI-native fraud prevention platform, was acquired by Worldpay. Big congrats to this team, whom I had the pleasure of working with as lead investor after their Seed. The acquisition price was not disclosed.</p><p><strong>Accern</strong>, an AI-powered insights provider for financial firms, was acquired by Wand AI, a Palo Alto-based startup specializing in AI agents for enterprises, for an undisclosed eight-figure sum. The company had previously raised $40M from investors including Fusion Fund and Mighty Capital.</p><p><strong>Applied Intuition</strong>, a company providing AI-powered tools for autonomous systems development, acquired <strong>EpiSci</strong>, an innovator in AI and autonomy software for national security, to enhance U.S. defense capabilities; the acquisition price was not disclosed.</p><p><strong>MongoDB</strong>, a leading database platform, acquired <strong>Voyage AI</strong>, a company specializing in embedding and reranking models for AI-powered search and retrieval, to enhance its AI capabilities; the acquisition price was not disclosed.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/p/the-state-of-ai-2025-march?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://press.airstreet.com/p/the-state-of-ai-2025-march?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p>]]></content:encoded></item><item><title><![CDATA[State of AI: January/February 2025 ]]></title><description><![CDATA[What you need to know from AI research, policy, startups, big tech over the last 4 weeks. DeepSeek, US policy, EU AI act, techbio and more.]]></description><link>https://press.airstreet.com/p/the-state-of-ai-2025-january</link><guid isPermaLink="false">https://press.airstreet.com/p/the-state-of-ai-2025-january</guid><dc:creator><![CDATA[Air Street Press]]></dc:creator><pubDate>Sun, 02 Feb 2025 17:31:06 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F321f65c3-6c53-45b7-85d1-36017ac3ceb1_1694x944.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Dear readers,</p><p>Welcome to the latest issue of the State of AI, an editorialized newsletter covering the key developments in AI policy, research, industry, and start-ups over the last month. First up, a few updates:</p><ul><li><p>To close out 2024, Nathan <a href="https://thegradientpub.substack.com/p/state-of-ai-2024">joined</a> Daniel Bashir on The Gradient podcast to discuss highlights from the State of AI Report, while sharing some early takes on o3 and DeepSeek.</p></li><li><p>Congratulations to our friends at <strong>Sereact</strong>, who raised a &#8364;25M Series A led by Creandum, with backing from Air Street and Point Nine. You can read more about why we backed the team <a href="https://press.airstreet.com/p/sereact-series-a">here</a>.</p></li><li><p>Our events series is back for 2025. First up, we&#8217;ll be in <a href="https://lu.ma/berlinai">Berlin</a> and <a href="https://lu.ma/munichai">Munich</a> in February, followed by <a href="https://lu.ma/zurichai">Zurich</a> in March. Subscribe to <a href="https://lu.ma/airstreet">our events page</a> to ensure you don&#8217;t miss out.</p></li><li><p>Our annual <strong>RAAIS</strong> conference is back on 13 June 2025, <a href="http://raais.co">check it out</a>.</p></li><li><p>Check out our latest writing on Air Street Press, where we&#8217;ve written about the <a href="https://press.airstreet.com/p/vibe-shifts-microsoft-and-google">vibe shift</a> at Microsoft and Google on AI, the vibe shift on <a href="https://press.airstreet.com/p/vibe-shifts-x-risk">existential risk of AI</a>, the <a href="https://press.airstreet.com/p/2024-review">Air Street 2024 Year in Review</a>, and how NVIDIA <a href="https://press.airstreet.com/p/compute-index-2024">dominates AI research</a>. </p></li></ul><p>We love hearing what you&#8217;re up to and what&#8217;s on your mind, just hit reply or forward to your friends :-)</p><div><hr></div><h3><em><strong>&#127758; The (geo)politics of AI</strong></em></h3><p>Pour one out for the Biden Administration&#8217;s executive order on AI safety. As was widely trailed before the election, the incoming Trump administration has scrapped it. Meanwhile, David Sacks has been tasked with developing a <a href="https://www.whitehouse.gov/presidential-actions/2025/01/removing-barriers-to-american-leadership-in-artificial-intelligence/">US AI action plan</a> to remove the barriers to US companies, so they&#8217;re free to <em>&#8220;develop AI systems that are free from ideological bias or engineered social agendas&#8221;</em>. Beyond a mild weakening of frontier AI oversight and shouting about DEI, it&#8217;s yet unclear what this is likely to amount to practically.</p><p>That doesn&#8217;t mean big moves aren&#8217;t afoot elsewhere, especially around energy. Data center developers and tech companies will be pleased to see the new administration <a href="https://greentapeblog.substack.com/p/the-permitting-eos-part-1-ceq-gets">cut through a series of MIMBY-esque regulations</a>, which allowed opponents of development to slow down proposals that obeyed every substantive environmental law with years of paperwork and legal battles. On the other hand, in a sign that US energy policy is the latest victim of the culture wars, <a href="https://www.washingtonpost.com/climate-environment/2025/01/21/trump-wind-energy-development-executive-order/">new wind energy</a> has faced a bunch of curbs, for reasons best understood by the new administration.</p><p>Elsewhere, President Trump <a href="https://www.tomshardware.com/tech-industry/trump-to-impose-25-percent-100-percent-tariffs-on-taiwan-made-chips-impacting-tsmc">signalled</a> that tariffs of 25-100% were on their way for Taiwan-made chips to encourage the reshoring of production, despite having &#8220;a good meeting&#8221; with Jensen. In the same set of remarks, he suggested that the US would likely end subsidies for semiconductor fabs via the CHIPS Act, arguing that companies didn&#8217;t need it. Clearly he hasn&#8217;t been studying <a href="https://press.airstreet.com/p/vibe-shifts-1-intel">Intel&#8217;s balance sheet too closely</a>&#8230; Maybe this is all part of a genius plan to force TSMC to announce more commitments to build out further US manufacturing capacity, but given the US demand for chips is relatively inelastic, the White House might be holding fewer cards than it thinks.</p><p>US tech, which seemed unassailable a few weeks ago, however, has suddenly found itself on the defensive. The late January DeepSeek panic - which saw the world finally wake up to the 2-year old Chinese quant-firm founded AI lab - has sent US tech stocks tumbling, launched a thousand different (generally poor) takes on the end of the US lead in AI, and has largely been used as an excuse by people to confirm their priors.</p><p>For us, <a href="https://api-docs.deepseek.com/news/news250120">DeepSeek R1</a> is an acceleration of a continuing trend, rather than something that&#8217;s come out of the blue. The DeepSeek team has already developed a track record of building highly capable, yet efficient, models and had demo-ed an early o1-style model towards the end of last year, R1-Lite-Preview, which we <a href="https://press.airstreet.com/p/your-guide-to-ai-december-2024">covered in the December issue of this newsletter</a>. In fact, we <a href="https://press.airstreet.com/p/your-guide-to-ai-december-2023">first wrote</a> about DeepSeek in December 2023 when they released their first LLM, DeepSeek 67B, which was better than Llama2 70B. For more, check out our <a href="https://press.airstreet.com/p/the-state-of-chinese-ai">State of Chinese AI essay</a> from last year too. We&#8217;re now seeing the very same people whose AI bull case hinged on models becoming more efficient panicking &#8230; because models have become more efficient.</p><p>While people have mocked the likes of Satya Nadella for <a href="https://x.com/satyanadella/status/1883753899255046301">suddenly discovering</a> the Wikipedia page for Jevons Paradox - the notion that improved efficiency leads to greater demand - we think it holds true here. Faster and cheaper AI will lead to more use of AI. Your take on whether or not NVIDIA is overvalued probably shouldn&#8217;t be affected by this news. Further, we politely remind naysayers that <a href="https://press.airstreet.com/p/compute-index-2024">91%</a> of all AI research papers make use of NVIDIA hardware for their experiments:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!lDKX!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F97fb8a9d-0693-4789-a96a-f91b0238d8e1_1600x762.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!lDKX!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F97fb8a9d-0693-4789-a96a-f91b0238d8e1_1600x762.png 424w, https://substackcdn.com/image/fetch/$s_!lDKX!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F97fb8a9d-0693-4789-a96a-f91b0238d8e1_1600x762.png 848w, https://substackcdn.com/image/fetch/$s_!lDKX!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F97fb8a9d-0693-4789-a96a-f91b0238d8e1_1600x762.png 1272w, https://substackcdn.com/image/fetch/$s_!lDKX!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F97fb8a9d-0693-4789-a96a-f91b0238d8e1_1600x762.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!lDKX!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F97fb8a9d-0693-4789-a96a-f91b0238d8e1_1600x762.png" width="600" height="285.5769230769231" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/97fb8a9d-0693-4789-a96a-f91b0238d8e1_1600x762.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:693,&quot;width&quot;:1456,&quot;resizeWidth&quot;:600,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!lDKX!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F97fb8a9d-0693-4789-a96a-f91b0238d8e1_1600x762.png 424w, https://substackcdn.com/image/fetch/$s_!lDKX!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F97fb8a9d-0693-4789-a96a-f91b0238d8e1_1600x762.png 848w, https://substackcdn.com/image/fetch/$s_!lDKX!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F97fb8a9d-0693-4789-a96a-f91b0238d8e1_1600x762.png 1272w, https://substackcdn.com/image/fetch/$s_!lDKX!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F97fb8a9d-0693-4789-a96a-f91b0238d8e1_1600x762.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>If you want some cool-headed takes on DeepSeek, we enjoyed <a href="https://www.chinatalk.media/p/deepseek-and-the-future-of-ai-competition">this interview</a> on ChinaTalk from our friend Miles Brundage, who ran OpenAI&#8217;s policy research and AGI preparedness for several years. In it, he makes the case that sanctions are working and that if given the choice, DeepSeek would certainly rather have more, not fewer, top-range GPUs in its compute fleet.</p><p>While the US and China jostle for supremacy in the lab and X takes, the UK looks finally set to embrace AI action. Endorsing investor Matt Clifford&#8217;s <a href="https://www.gov.uk/government/publications/ai-opportunities-action-plan/ai-opportunities-action-plan">AI Opportunities Plan</a>, Prime Minister Keir Starmer vowed to <em>&#8220;mainline AI into the veins&#8221;</em> of the UK. The plan covers everything from plans to increase the UK&#8217;s compute capacity through to creating AI Growth Zones (think of them as designated geographical areas) that will feature reduced regulatory obstacles to new data center construction. The plan also covers off high-skilled immigration, procurement, and proposes the creation of UK Sovereign AI - a government-backed lab that will partner with the private sector to develop certain capabilities.</p><p>We commend the plan for its ambition, but given its scope and the limited time that Matt and team had to put it together, it remains top-level. The devil, as always, will be in its implementation. The UK has never lacked for consultations or units tasked with <em>thinking</em> about technology - they&#8217;ve usually withered on the vine due to a lack of interest from the top in <em>taking action on the thinking</em>. It&#8217;s up to the prime minister to decide if he&#8217;s serious about stabbing the AI needle into the patient. We can only hope that the Doctor is ready.</p><p>Note that the UK does have one thing going for it. It doesn&#8217;t suffer from the EU AI Act. The bill&#8217;s prohibition on certain high-risk AI systems comes into force from this weekend despite its original chief architect, Gabriele Mazzini, now feeling like it&#8217;s gone too far. Facepalm.</p><p>On the topic of Europe, the EU Commission led by Ursula von der Leyen announced <em>&#8220;It is time to restart Europe&#8217;s innovation engine. We have the Compass. We have the political will. Now, what matters is speed and unity.&#8221; </em>Interesting! The <em>&#8220;Competitiveness Compass&#8221; </em>is meant to drive <em>&#8220;productivity through innovation&#8221;</em>, increase <em>&#8220;competitiveness with decarbonisation&#8221;</em> (sounds a bit random, ngl), and <em>&#8220;reduce dependencies, increasing resilience and security&#8221;</em> (that&#8217;s worth a try). Let&#8217;s add some more detail into what this really means&#8230;ah, hang on, von der Leyen&#8217;s Competitiveness Compass is even more confusing, nonsensical, and certainly not how a compass works:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!mTOD!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0be4e1f0-a18b-4a5f-a24f-70b7555c7cee_1206x1029.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!mTOD!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0be4e1f0-a18b-4a5f-a24f-70b7555c7cee_1206x1029.png 424w, https://substackcdn.com/image/fetch/$s_!mTOD!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0be4e1f0-a18b-4a5f-a24f-70b7555c7cee_1206x1029.png 848w, https://substackcdn.com/image/fetch/$s_!mTOD!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0be4e1f0-a18b-4a5f-a24f-70b7555c7cee_1206x1029.png 1272w, https://substackcdn.com/image/fetch/$s_!mTOD!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0be4e1f0-a18b-4a5f-a24f-70b7555c7cee_1206x1029.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!mTOD!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0be4e1f0-a18b-4a5f-a24f-70b7555c7cee_1206x1029.png" width="476" height="406.1393034825871" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/0be4e1f0-a18b-4a5f-a24f-70b7555c7cee_1206x1029.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1029,&quot;width&quot;:1206,&quot;resizeWidth&quot;:476,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!mTOD!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0be4e1f0-a18b-4a5f-a24f-70b7555c7cee_1206x1029.png 424w, https://substackcdn.com/image/fetch/$s_!mTOD!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0be4e1f0-a18b-4a5f-a24f-70b7555c7cee_1206x1029.png 848w, https://substackcdn.com/image/fetch/$s_!mTOD!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0be4e1f0-a18b-4a5f-a24f-70b7555c7cee_1206x1029.png 1272w, https://substackcdn.com/image/fetch/$s_!mTOD!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0be4e1f0-a18b-4a5f-a24f-70b7555c7cee_1206x1029.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>On the topic of AI, she <a href="https://x.com/EU_Commission/status/1884635063054106770">went further</a>: <em>&#8220;Only 13.5% of EU businesses are using AI. This must change. This year we will launch a broad AI Strategy for our continent, including an &#8216;Apply AI' initiative to drive industrial adoption of Artificial Intelligence in key sectors."</em> This is rather bizarre because a) the vast majority (if not all) serious AI vendors are US companies, b) several of them rate limit their best products from Europe due to the AI Act, and c) the fact that EU businesses will adopt US AI doesn&#8217;t solve the issue of &#8220;reducing excessive dependencies and increasing security&#8221;. So this whole plan seems to be backwards, which is rather unsurprising given how tone deaf and unplugged from reality European politicians are when it comes to enabling technology progress.</p><p>And to top it all off, our European notaries are back! Here, the Austrian notary association <a href="https://x.com/noverloop/status/1885034867638870334">claims</a> that notaries <em>&#8220;contribute to simplifying and reducing bureaucracy in administrative processes&#8221;</em>. Suggesting that Europe&#8217;s bureaucratic heritage and anti-innovation practices are an enabler of the EU Commission&#8217;s new pro-innovation agenda makes one feel that&#8230;it could be so over.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://press.airstreet.com/subscribe?"><span>Subscribe now</span></a></p><h3><em><strong>&#127850; Hardware</strong></em></h3><p>Before the great DeepSeek panic, things were looking up for Team NVIDIA. The Trump Administration was quick <a href="https://openai.com/index/announcing-the-stargate-project/">to unveil Stargate</a>, a joint venture tasked with building out $500B of infrastructure for OpenAI. SoftBank, OpenAI, Oracle, and MGX are the initial equity partners, with Microsoft, NVIDIA, Oracle, OpenAI, and Arm (note this is likely just because Arm CPUs feature in NVIDIA products) as the technology partners. While the White House fronted the launch, the government is not supplying any of the capital, and it&#8217;s unclear how much of a role it played in brokering the deal.</p><p>If you&#8217;re optimistic about AI (which the median reader of this newsletter probably is), it&#8217;s probably safe to ignore the various suggestions we&#8217;ve seen that the DeepSeek news makes Stargate redundant. It is fair, however, to express scepticism about how likely it is the deal will come together. While Elon Musk&#8217;s <a href="https://www.nbcnews.com/business/business-news/elon-musk-sam-altman-feud-over-stargate-ai-project-explainer-rcna188978">attacks on Stargate</a> are in part motivated by his personal crusade against Sam Altman, it&#8217;s reasonable to ask where the money is coming from.</p><p>$500B over four years would require an average commitment of roughly $40B per equity partner per year. That&#8217;s more than the AUM of both Vision Funds combined and would burn through all of Oracle&#8217;s net tangible assets in under 18 months. This would come on top of SoftBank potentially <a href="https://www.ft.com/content/6a482d4f-0c90-47bf-a4c2-67a2f1b743b1">investing</a> $15-25B in OpenAI separately. This presumably isn&#8217;t what the Stargate consortium has in mind, but until they tell us, it&#8217;s fair to be sceptical.</p><p>But the question on everyone&#8217;s lips is whether Chinese chips are finally hitting the big leagues after a number of false dawns?</p><p>While DeepSeek&#8217;s R1 was trained on NVIDIA hardware, inference is (at least in part) <a href="https://www.ft.com/content/f18217d6-3796-4205-bbc5-309b8a526208">being served</a> by Huawei Ascend chips. Huawei has essentially acknowledged that it won&#8217;t be able to act as a serious NVIDIA competitor in the pre-training market for the foreseeable future, but is betting that inference will prove lucrative and that Chinese government pressure will drive it business. The logic is appealing for Huawei, but will involve the company resolving the production quality issues that have blighted Ascend manufacturing.</p><p>That said, NVIDIA&#8217;s China business looks healthy for some time to come. The company is set to be the primary beneficiary of ByteDance&#8217;s $12B AI <a href="https://www.ft.com/content/0815c8fb-e6ed-478b-abb1-c67d6f48fd3a">hardware spending spree</a>, scooping up $6.8B from cluster-building outside the US, along with a further chunk from buying sanctions-compliant chips for use in China.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/p/the-state-of-ai-2025-january?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://press.airstreet.com/p/the-state-of-ai-2025-january?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><h3><em><strong>&#127981; Big <s>tech</s> start-ups</strong></em></h3><p>January hasn&#8217;t been all DeepSeek all the time. Some other things have happened around the world.</p><p>Now somewhat overshadowed, OpenAI has released its first browser agent - <a href="https://openai.com/index/introducing-operator/">Operator</a>. This follows in the footsteps of the Gemini team, who included one as part of their <a href="https://blog.google/technology/google-deepmind/google-gemini-ai-update-december-2024/">2.0 release</a> at the end of last year. Operator is currently in research preview, but is a computer-use agent that leverages 4o&#8217;s vision capabilities and RL-powered reasoning capabilities to interact with buttons, menus, and text fields.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!_BWU!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ab416ca-daef-466d-8a96-4f4361a33bdc_1600x802.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!_BWU!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ab416ca-daef-466d-8a96-4f4361a33bdc_1600x802.png 424w, https://substackcdn.com/image/fetch/$s_!_BWU!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ab416ca-daef-466d-8a96-4f4361a33bdc_1600x802.png 848w, https://substackcdn.com/image/fetch/$s_!_BWU!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ab416ca-daef-466d-8a96-4f4361a33bdc_1600x802.png 1272w, https://substackcdn.com/image/fetch/$s_!_BWU!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ab416ca-daef-466d-8a96-4f4361a33bdc_1600x802.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!_BWU!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ab416ca-daef-466d-8a96-4f4361a33bdc_1600x802.png" width="632" height="316.86813186813185" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/9ab416ca-daef-466d-8a96-4f4361a33bdc_1600x802.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:730,&quot;width&quot;:1456,&quot;resizeWidth&quot;:632,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!_BWU!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ab416ca-daef-466d-8a96-4f4361a33bdc_1600x802.png 424w, https://substackcdn.com/image/fetch/$s_!_BWU!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ab416ca-daef-466d-8a96-4f4361a33bdc_1600x802.png 848w, https://substackcdn.com/image/fetch/$s_!_BWU!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ab416ca-daef-466d-8a96-4f4361a33bdc_1600x802.png 1272w, https://substackcdn.com/image/fetch/$s_!_BWU!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ab416ca-daef-466d-8a96-4f4361a33bdc_1600x802.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>It <a href="https://openai.com/index/computer-using-agent/">narrowly beats</a> Claude&#8217;s Computer Use on computer and browser use benchmarks, but these aren&#8217;t the only differences. Computer Use is available via the API and can be deployed on any browser theoretically, whereas Operator is available via ChatGPT and boots up a virtual machine in the cloud.</p><p>As you&#8217;d expect, Operator is capable but far from perfect. As frontier labs continue creating &#8230; remarkably similar products, the battle to find an edge will only intensify.</p><p>On the subject of similar products, the catchily named <a href="https://ai.google.dev/gemini-api/docs/thinking">Gemini 2.0 Flash Thinking Experimental</a> - an early Gemini reasoning model was released this month. Thanks to the Gemini team&#8217;s frequently baffling approach to product comms, you&#8217;d likely only know this if you spend a lot of time on Google&#8217;s AI Studio or Vertex (why these are still separate confuses the hell out of everyone). This is not the way to bring about a <a href="https://press.airstreet.com/p/vibe-shifts-microsoft-and-google">vibe shift</a>.</p><p>As well as releasing new products, OpenAI has been getting stuck into the DeepSeek discourse. While they publicly congratulated the DeepSeek team on reaching o1-levels of performance, they (and their major investors) also <a href="https://www.ft.com/content/a0dfedd1-5255-4fa9-8ccc-1fe01de87ea6">suggested</a> that they used the outputs of OpenAI models as part of their training.</p><p>This would be a violation of OpenAI&#8217;s terms of service. OpenAI has not yet presented any evidence to support these claims. Further, it seems even slightly disingenuous to levy such claims when OpenAI itself hasn&#8217;t been entirely transparent with the sourcing of its own training data and relevant content use rights. In practise, TOS around distillation are very difficult to ever enforce, if it turned out that R1 did use the outputs of a significantly more expensive model at any scale, it would undermine claims about cost-efficiency slightly. But it would reinforce the arc of progress that models must become big before they can become smaller.</p><p>A day ago, a huge vibe shift went down on Reddit. OpenAI&#8217;s leadership team <a href="https://www.reddit.com/r/OpenAI/comments/1ieonxv/ama_with_openais_sam_altman_mark_chen_kevin_weil/">ran an AMA</a> and what caught our eye was Sam&#8217;s response to a question about open source:<em> &#8220;I personally think we have been on the wrong side of history here and need to figure out a different open source strategy.&#8221; </em>He followed this up with saying that he expects OpenAI to continue to lead, but with <em>&#8220;less of a lead than we did in previous years&#8221;</em>. </p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!uuhS!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F21af803c-cda8-484a-a889-c4950b2f1be7_1536x516.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!uuhS!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F21af803c-cda8-484a-a889-c4950b2f1be7_1536x516.png 424w, https://substackcdn.com/image/fetch/$s_!uuhS!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F21af803c-cda8-484a-a889-c4950b2f1be7_1536x516.png 848w, https://substackcdn.com/image/fetch/$s_!uuhS!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F21af803c-cda8-484a-a889-c4950b2f1be7_1536x516.png 1272w, https://substackcdn.com/image/fetch/$s_!uuhS!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F21af803c-cda8-484a-a889-c4950b2f1be7_1536x516.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!uuhS!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F21af803c-cda8-484a-a889-c4950b2f1be7_1536x516.png" width="620" height="208.22802197802199" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/21af803c-cda8-484a-a889-c4950b2f1be7_1536x516.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:489,&quot;width&quot;:1456,&quot;resizeWidth&quot;:620,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!uuhS!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F21af803c-cda8-484a-a889-c4950b2f1be7_1536x516.png 424w, https://substackcdn.com/image/fetch/$s_!uuhS!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F21af803c-cda8-484a-a889-c4950b2f1be7_1536x516.png 848w, https://substackcdn.com/image/fetch/$s_!uuhS!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F21af803c-cda8-484a-a889-c4950b2f1be7_1536x516.png 1272w, https://substackcdn.com/image/fetch/$s_!uuhS!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F21af803c-cda8-484a-a889-c4950b2f1be7_1536x516.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>Over in Europe, who remembers Mistral? France&#8217;s 2023 breakthrough company has seen the hype fade away, with <a href="http://n">reports</a> of sluggish uptake and mounting investor skepticism. Its core differentiator around efficiency has also come under scrutiny in the light of DeepSeek-mania. But the team is continuing to ship, releasing <a href="https://mistral.ai/news/mistral-small-3/">Small 3</a>, an open latency-optimized 24B model.</p><p>The instruction-tuned model performs competitively with significantly larger open and proprietary models. But one has to beg the question of whether this even matters anymore&#8230;The model has been released under an Apache 2.0 licence, rather than the more restrictive Mistral Research License, which forced users to negotiate commercial deployments separately.</p><p>Finally, to end this discussion, if the prevailing DeepSeek narrative is <em>"look what can get created with talent density and limited resources"</em>, why is no one asking why Europe didn't produce a DeepSeek-grade model? It too has talent density and limited resources...but perhaps those stats are only to be used when they conveniently support the European bull case rather than question it.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!iI00!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe55caa7a-ac32-4daf-9816-ca682770e20c_1600x823.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!iI00!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe55caa7a-ac32-4daf-9816-ca682770e20c_1600x823.png 424w, https://substackcdn.com/image/fetch/$s_!iI00!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe55caa7a-ac32-4daf-9816-ca682770e20c_1600x823.png 848w, https://substackcdn.com/image/fetch/$s_!iI00!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe55caa7a-ac32-4daf-9816-ca682770e20c_1600x823.png 1272w, https://substackcdn.com/image/fetch/$s_!iI00!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe55caa7a-ac32-4daf-9816-ca682770e20c_1600x823.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!iI00!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe55caa7a-ac32-4daf-9816-ca682770e20c_1600x823.png" width="612" height="314.8269230769231" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e55caa7a-ac32-4daf-9816-ca682770e20c_1600x823.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:749,&quot;width&quot;:1456,&quot;resizeWidth&quot;:612,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!iI00!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe55caa7a-ac32-4daf-9816-ca682770e20c_1600x823.png 424w, https://substackcdn.com/image/fetch/$s_!iI00!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe55caa7a-ac32-4daf-9816-ca682770e20c_1600x823.png 848w, https://substackcdn.com/image/fetch/$s_!iI00!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe55caa7a-ac32-4daf-9816-ca682770e20c_1600x823.png 1272w, https://substackcdn.com/image/fetch/$s_!iI00!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe55caa7a-ac32-4daf-9816-ca682770e20c_1600x823.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://press.airstreet.com/subscribe?"><span>Subscribe now</span></a></p><h3><strong>&#128300;Research</strong></h3><p><strong><a href="https://www.biorxiv.org/content/10.1101/2025.01.06.631536v1">Engineering of CRISPR-Cas PAM recognition using deep learning of vast evolutionary data</a></strong>, <em>Profluent.</em></p><p>Introduces Protein2PAM, a deep learning model trained on a vast dataset of CRISPR-Cas systems to predict and engineer the recognition of protospacer-adjacent motifs (PAMs). PAMs are short DNA sequences that CRISPR-Cas enzymes must recognize to bind and edit genomic targets, but their specificity limits genome-editing applications. By training on over 45,000 PAM sequences from CRISPR systems, Protein2PAM accurately predicts PAM specificity based on protein sequences alone.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!5JLr!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7d4df085-c23e-4959-bdd8-f2a0d98c4ef1_1454x1292.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!5JLr!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7d4df085-c23e-4959-bdd8-f2a0d98c4ef1_1454x1292.png 424w, https://substackcdn.com/image/fetch/$s_!5JLr!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7d4df085-c23e-4959-bdd8-f2a0d98c4ef1_1454x1292.png 848w, https://substackcdn.com/image/fetch/$s_!5JLr!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7d4df085-c23e-4959-bdd8-f2a0d98c4ef1_1454x1292.png 1272w, https://substackcdn.com/image/fetch/$s_!5JLr!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7d4df085-c23e-4959-bdd8-f2a0d98c4ef1_1454x1292.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!5JLr!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7d4df085-c23e-4959-bdd8-f2a0d98c4ef1_1454x1292.png" width="571" height="507.38101788170565" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/7d4df085-c23e-4959-bdd8-f2a0d98c4ef1_1454x1292.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1292,&quot;width&quot;:1454,&quot;resizeWidth&quot;:571,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!5JLr!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7d4df085-c23e-4959-bdd8-f2a0d98c4ef1_1454x1292.png 424w, https://substackcdn.com/image/fetch/$s_!5JLr!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7d4df085-c23e-4959-bdd8-f2a0d98c4ef1_1454x1292.png 848w, https://substackcdn.com/image/fetch/$s_!5JLr!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7d4df085-c23e-4959-bdd8-f2a0d98c4ef1_1454x1292.png 1272w, https://substackcdn.com/image/fetch/$s_!5JLr!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7d4df085-c23e-4959-bdd8-f2a0d98c4ef1_1454x1292.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The researchers demonstrate Protein2PAM&#8217;s utility in protein engineering by using it to evolve variants of Nme1Cas9, a Cas9 enzyme with a strict PAM requirement. The model-guided mutations resulted in variants with expanded PAM compatibility and up to 50-fold improved cleavage rates. This marks the first successful application of machine learning in designing CRISPR enzymes with customized PAM recognition.</p><p>Beyond its predictive power, Protein2PAM provides insight into the biophysical principles of protein-DNA interactions. By way of <em>in silico</em> mutagenesis, the model identified amino acid substitutions that shift PAM preferences, confirming the role of specific residues in Cas9&#8217;s recognition mechanism. Experimental validation showed that engineered variants exhibited expected PAM shifts and enhanced activity. This is really exciting because it helps unlock the custom design of editors for target sequences.</p><p><strong><a href="https://www.microsoft.com/en-us/research/blog/mattergen-a-new-paradigm-of-materials-design-with-generative-ai/">A generative model for inorganic materials design</a></strong>, <em>Microsoft Research</em>.</p><p>Introduces MatterGen, a generative model designed to create stable and diverse inorganic materials by leveraging a diffusion-based approach. Unlike previous models, which struggle with stability and flexibility, MatterGen refines atom types, coordinates, and lattice structures in a controlled manner, significantly improving the likelihood of generating novel, synthesizable materials.</p><p>MatterGen outperforms existing generative models by doubling the success rate in producing stable, unique, and novel materials while generating structures that are significantly closer to their energy minima.</p><p>To validate MatterGen&#8217;s predictions, the researchers synthesized a generated material (TaCr&#8322;O&#8326;) and measured its bulk modulus, finding it within 20% of the predicted value. Additionally, the model re-discovered over 2,000 experimentally verified structures unseen during training. While MatterGen marks a significant advance in generative materials design, the study acknowledges areas for improvement, such as symmetry bias and broader property optimization.</p><p><strong><a href="https://www.nature.com/articles/s41586-024-08393-x">De novo designed proteins neutralize lethal snake venom toxins</a></strong>, <em>University of Washington, Technical University of Denmark.</em></p><p>Presents a novel approach to treating snakebite poison by designing de novo proteins that neutralize lethal three-finger toxins (3FTx) found in elapid snake venoms.</p><p>The study used RFdiffusion to create proteins that physically block snake venom toxins from binding to nerve receptors (nAChRs). They developed two main antitoxins - SHRT for short-chain toxins and LNG for long-chain toxins like &#945;-cobratoxin - both achieving nanomolar binding strength. They also designed CYTX to target tissue-destroying toxins. Crystal structures confirmed the proteins matched their computer designs, and experiments showed they protected cells from toxin damage. In mice, both SHRT and LNG prevented death whether given before or after venom exposure.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!o6Bh!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F27ae191c-d5a2-44d7-80f2-f82aa2ebd7d3_956x952.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!o6Bh!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F27ae191c-d5a2-44d7-80f2-f82aa2ebd7d3_956x952.png 424w, https://substackcdn.com/image/fetch/$s_!o6Bh!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F27ae191c-d5a2-44d7-80f2-f82aa2ebd7d3_956x952.png 848w, https://substackcdn.com/image/fetch/$s_!o6Bh!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F27ae191c-d5a2-44d7-80f2-f82aa2ebd7d3_956x952.png 1272w, https://substackcdn.com/image/fetch/$s_!o6Bh!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F27ae191c-d5a2-44d7-80f2-f82aa2ebd7d3_956x952.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!o6Bh!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F27ae191c-d5a2-44d7-80f2-f82aa2ebd7d3_956x952.png" width="572" height="569.6066945606694" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/27ae191c-d5a2-44d7-80f2-f82aa2ebd7d3_956x952.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:952,&quot;width&quot;:956,&quot;resizeWidth&quot;:572,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!o6Bh!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F27ae191c-d5a2-44d7-80f2-f82aa2ebd7d3_956x952.png 424w, https://substackcdn.com/image/fetch/$s_!o6Bh!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F27ae191c-d5a2-44d7-80f2-f82aa2ebd7d3_956x952.png 848w, https://substackcdn.com/image/fetch/$s_!o6Bh!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F27ae191c-d5a2-44d7-80f2-f82aa2ebd7d3_956x952.png 1272w, https://substackcdn.com/image/fetch/$s_!o6Bh!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F27ae191c-d5a2-44d7-80f2-f82aa2ebd7d3_956x952.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>These engineered proteins improve upon traditional antivenoms in several ways: they're more specific, maintain stability better, and can be produced cheaply in bacteria without animal immunization. Their compact structure may help them penetrate tissues better, and their stability could eliminate cold storage requirements, making them practical for remote areas.</p><p><strong><a href="https://arxiv.org/abs/2501.06252v2">Transformer<sup>2</sup>: Self-adaptive LLMs</a></strong>, <em>Sakana AI</em>.</p><p>Introduces Transformer<sup>2</sup>, a novel framework for creating self-adaptive large language models that can dynamically adjust to different tasks without extensive retraining. The core innovation is Singular Value Fine-tuning (SVF), which modifies only the singular values within a model's weight matrices to create specialized "expert vectors" for different capabilities like coding, math, and reasoning.</p><p>The framework employs a two-pass system during inference - first analyzing the input to determine required skills, then applying the appropriate combination of expert vectors to optimize performance. This approach requires significantly fewer parameters than existing methods like LoRA while achieving better results across multiple model architectures and tasks.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!37_l!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Febb500fc-6fda-4e87-8b87-115f02c11bbd_556x216.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!37_l!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Febb500fc-6fda-4e87-8b87-115f02c11bbd_556x216.png 424w, https://substackcdn.com/image/fetch/$s_!37_l!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Febb500fc-6fda-4e87-8b87-115f02c11bbd_556x216.png 848w, https://substackcdn.com/image/fetch/$s_!37_l!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Febb500fc-6fda-4e87-8b87-115f02c11bbd_556x216.png 1272w, https://substackcdn.com/image/fetch/$s_!37_l!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Febb500fc-6fda-4e87-8b87-115f02c11bbd_556x216.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!37_l!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Febb500fc-6fda-4e87-8b87-115f02c11bbd_556x216.png" width="556" height="216" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/ebb500fc-6fda-4e87-8b87-115f02c11bbd_556x216.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:216,&quot;width&quot;:556,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!37_l!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Febb500fc-6fda-4e87-8b87-115f02c11bbd_556x216.png 424w, https://substackcdn.com/image/fetch/$s_!37_l!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Febb500fc-6fda-4e87-8b87-115f02c11bbd_556x216.png 848w, https://substackcdn.com/image/fetch/$s_!37_l!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Febb500fc-6fda-4e87-8b87-115f02c11bbd_556x216.png 1272w, https://substackcdn.com/image/fetch/$s_!37_l!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Febb500fc-6fda-4e87-8b87-115f02c11bbd_556x216.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>The researchers demonstrate Transformer<sup>2</sup>'s effectiveness through extensive experiments with various LLMs including Llama3 and Mistral models. They show it can improve performance not just on tasks it was directly trained for, but also on entirely new tasks through intelligent adaptation of its expert vectors. This includes successfully transferring knowledge between different model architectures.</p><p>A particularly notable aspect is the framework's efficiency. It achieves these improvements while using only a fraction of the parameters required by other fine-tuning methods. The system also shows promising results in cross-model compatibility, suggesting potential for knowledge transfer between different LLMs, though this currently works best between architecturally similar models. It&#8217;s fun to see new architecture research gaining traction.</p><p><strong><a href="https://arxiv.org/abs/2501.12948">DeepSeek-R1: Incentivizing Reasoning Capability in LLMs via Reinforcement Learning</a></strong>, <em>DeepSeek</em>.</p><p>This work builds on DeepSeek&#8217;s strong base language model, V3, released in December 2024, to introduce the company&#8217;s first reasoning models, the R-series. In particular, they show that LLMs can develop reasoning capabilities without the use of any supervised finetuning (SFT) data, but with reinforcement learning instead. A few features of this paper are interesting:</p><p>The reward system that produces signals for the model to improve through RL is no longer a neural network, but is a simpler rule-based. It is composed of accuracy rewards (whether the response is correct) and format rewards (whether the model displays its thinking process correctly). The additional benefit here is that the RL step is less susceptible to reward hacking, doesn&#8217;t need to be trained and results in a simpler overall training pipeline.</p><p>The first model they produce, DeepSeek-R1-Zero, is built from V3-Base with RL applied as above directly with no SFT data. While R1-Zero starts out at 15.6% average pass@1 score on AIME 2024 mathematics benchmark, the model hits 71% after 8.5k RL training steps. This is better than OpenAI&#8217;s o1-mini and just shy of o1-0912&#8217;s score of 74.4%. With majority voting (cons@64), in which multiple outputs are aggregated to determine the most accurate response, R1-Zero outperforms o1-0912.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!g80Y!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6dee5a2f-e605-4137-8fa8-a9e90a00ea2e_1294x1172.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!g80Y!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6dee5a2f-e605-4137-8fa8-a9e90a00ea2e_1294x1172.png 424w, https://substackcdn.com/image/fetch/$s_!g80Y!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6dee5a2f-e605-4137-8fa8-a9e90a00ea2e_1294x1172.png 848w, https://substackcdn.com/image/fetch/$s_!g80Y!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6dee5a2f-e605-4137-8fa8-a9e90a00ea2e_1294x1172.png 1272w, https://substackcdn.com/image/fetch/$s_!g80Y!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6dee5a2f-e605-4137-8fa8-a9e90a00ea2e_1294x1172.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!g80Y!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6dee5a2f-e605-4137-8fa8-a9e90a00ea2e_1294x1172.png" width="635" height="575.1313755795982" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/6dee5a2f-e605-4137-8fa8-a9e90a00ea2e_1294x1172.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1172,&quot;width&quot;:1294,&quot;resizeWidth&quot;:635,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!g80Y!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6dee5a2f-e605-4137-8fa8-a9e90a00ea2e_1294x1172.png 424w, https://substackcdn.com/image/fetch/$s_!g80Y!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6dee5a2f-e605-4137-8fa8-a9e90a00ea2e_1294x1172.png 848w, https://substackcdn.com/image/fetch/$s_!g80Y!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6dee5a2f-e605-4137-8fa8-a9e90a00ea2e_1294x1172.png 1272w, https://substackcdn.com/image/fetch/$s_!g80Y!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6dee5a2f-e605-4137-8fa8-a9e90a00ea2e_1294x1172.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The learning process of R1-Zero is equally interesting. The model <em>&#8220;naturally acquires the ability to solve increasingly complex reasoning tasks by leveraging extended test-time computation.&#8221; </em>Indeed, as training steps increase, the average length of the model&#8217;s responses increases. Further, the authors observe the emergence of novel behaviors as test-time computation increases. This includes reflection, where a model reevaluates its reasoning steps, and exploration, by which it spontaneously decides to try other problem-solving approaches. Midway during its RL steps, the model finds an &#8220;aha moment&#8221; where it spontaneously <em>&#8220;learns to allocate more thinking time to a problem by reevaluating its initial approach. This behavior is not only a testament to the model&#8217;s growing reasoning abilities but also a captivating example of how reinforcement learning can lead to unexpected and sophisticated outcomes&#8221;</em>.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!eMBL!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F17077935-fdd2-49d9-847b-0520fc5b72d0_1294x856.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!eMBL!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F17077935-fdd2-49d9-847b-0520fc5b72d0_1294x856.png 424w, https://substackcdn.com/image/fetch/$s_!eMBL!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F17077935-fdd2-49d9-847b-0520fc5b72d0_1294x856.png 848w, https://substackcdn.com/image/fetch/$s_!eMBL!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F17077935-fdd2-49d9-847b-0520fc5b72d0_1294x856.png 1272w, https://substackcdn.com/image/fetch/$s_!eMBL!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F17077935-fdd2-49d9-847b-0520fc5b72d0_1294x856.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!eMBL!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F17077935-fdd2-49d9-847b-0520fc5b72d0_1294x856.png" width="635" height="420.0618238021638" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/17077935-fdd2-49d9-847b-0520fc5b72d0_1294x856.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:856,&quot;width&quot;:1294,&quot;resizeWidth&quot;:635,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!eMBL!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F17077935-fdd2-49d9-847b-0520fc5b72d0_1294x856.png 424w, https://substackcdn.com/image/fetch/$s_!eMBL!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F17077935-fdd2-49d9-847b-0520fc5b72d0_1294x856.png 848w, https://substackcdn.com/image/fetch/$s_!eMBL!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F17077935-fdd2-49d9-847b-0520fc5b72d0_1294x856.png 1272w, https://substackcdn.com/image/fetch/$s_!eMBL!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F17077935-fdd2-49d9-847b-0520fc5b72d0_1294x856.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>With the drawback of R1-Zero being poor readability and language mixing, the authors produce DeepSeek-R1, which applies RL starting from a V3-Base checkpoint that is fine-tuned with thousands of long CoT examples generated in different ways. They also introduce a language consistency reward during RL training to encourage the model to reason in the same language as the prompt. Once the RL stage is finished, they move onto finetuning the model with 800k samples of SFT data collected from other domains to enhance the model&#8217;s capabilities in writing, role-playing, and other general-purpose tasks. Finally, they implement a second round of RL to improve the model&#8217;s helpfulness and harmlessness while improving its reasoning. Here are R1&#8217;s evaluations, which are particularly strong vs. o1-1217, surpassing it at a number of evals.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!qXTX!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1d0a7911-4ed6-4307-bc70-c87da5a393e9_1210x1036.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!qXTX!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1d0a7911-4ed6-4307-bc70-c87da5a393e9_1210x1036.png 424w, https://substackcdn.com/image/fetch/$s_!qXTX!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1d0a7911-4ed6-4307-bc70-c87da5a393e9_1210x1036.png 848w, https://substackcdn.com/image/fetch/$s_!qXTX!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1d0a7911-4ed6-4307-bc70-c87da5a393e9_1210x1036.png 1272w, https://substackcdn.com/image/fetch/$s_!qXTX!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1d0a7911-4ed6-4307-bc70-c87da5a393e9_1210x1036.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!qXTX!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1d0a7911-4ed6-4307-bc70-c87da5a393e9_1210x1036.png" width="622" height="532.5553719008265" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1d0a7911-4ed6-4307-bc70-c87da5a393e9_1210x1036.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1036,&quot;width&quot;:1210,&quot;resizeWidth&quot;:622,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!qXTX!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1d0a7911-4ed6-4307-bc70-c87da5a393e9_1210x1036.png 424w, https://substackcdn.com/image/fetch/$s_!qXTX!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1d0a7911-4ed6-4307-bc70-c87da5a393e9_1210x1036.png 848w, https://substackcdn.com/image/fetch/$s_!qXTX!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1d0a7911-4ed6-4307-bc70-c87da5a393e9_1210x1036.png 1272w, https://substackcdn.com/image/fetch/$s_!qXTX!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1d0a7911-4ed6-4307-bc70-c87da5a393e9_1210x1036.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Notable too is that the paper explored Monte Carlo Tree Search to enhance test-time compute scalability. However, they found that this approach doesn&#8217;t scale well during training because a) the search space of generated tokens is enormous and if the extension limit for each node is restricted, then the model tends to get stuck in local optima, and b) the complexities of token generation make the training of a value model inherently difficult. They also show that R-1 is a strong teacher model that, if tasked with generating 800k training samples, can be used to effectively fine-tune smaller dense models.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://press.airstreet.com/subscribe?"><span>Subscribe now</span></a></p><p>Also on our radar:</p><ul><li><p><a href="https://arxiv.org/abs/2501.16466">On the Feasibility of Using LLMs to Execute Multistage Network Attacks</a>, <em>Anthropic, Carnegie Mellon University</em>. Introduces Incalmo, a high-level abstraction layer that significantly improves LLMs ability to execute complex multistage network attacks. While current LLMs struggle with these attacks (succeeding in only 1 out of 10 test environments), Incalmo enables them to succeed in 9 out of 10 environments. It does so by allowing LLMs to specify high-level tasks rather than low-level commands, providing an attack graph service to guide decision-making, and offering an environment state service to track network information. The research also shows that smaller LLMs using Incalmo outperform larger models without it.</p></li><li><p><a href="https://arxiv.org/abs/2501.12326">UI-TARS: Pioneering Automated GUI Interaction with Native Agents</a>, <em>ByteDance</em>. Presents UI-TARS, a new type of graphical user interface agent model that uses screenshots as input and performs human-like interactions, such as keyboard and mouse operations. UI-TARS achieves SOTA performance in 10+ GUI agent benchmarks, including OSWorld and AndroidWorld, surpassing Claude and GPT-4o respectively. Key innovations include enhanced perception, unified action modeling, system-2 reasoning, and iterative training with reflective online traces. UI-TARS leverages large-scale datasets for context-aware understanding, precise grounding, and deliberate reasoning. Through iterative training and reflection tuning, UI-TARS continuously learns from its mistakes and adapts to unforeseen situations.</p></li><li><p><a href="https://arxiv.org/abs/2501.14249">Humanity&#8217;s Last Exam</a>, <em>Center for AI Safety, Scale AI</em>. Presents Humanity's Last Exam, a new frontier-level benchmark designed to test the limits of LLMs through 3,000 extremely challenging questions across dozens of academic subjects. Created by nearly 1,000 subject matter experts from over 500 institutions, the benchmark emphasizes mathematical reasoning and includes both text-only and multi-modal questions, with all questions being original, precise, unambiguous, and resistant to simple internet lookup. Unlike existing benchmarks that have been largely solved by current LLMs (with &gt;90% accuracy) - even the best models achieve less than 10% accuracy and show poor calibration.</p></li><li><p><a href="https://arxiv.org/abs/2501.09891">Evolving Deeper LLM Thinking</a>, <em>Google DeepMind</em>. Introduces Mind Evolution, an evolutionary search strategy that helps LLMs solve complex natural language planning problems. The approach uses genetic algorithms where an LLM acts as both a generator of solutions and a refiner, iteratively improving solutions based on evaluator feedback. Unlike previous methods that require formalizing problems or rely on step-by-step verification, Mind Evolution operates directly in natural language space and only needs a global solution evaluator. The method achieved significant improvements over baseline approaches on travel planning and meeting scheduling tasks, reaching success rates over 95% on several benchmarks when using Gemini 1.5 Flash, and nearly 100% success with Gemini 1.5 Pro - all without requiring formal solvers. The authors also introduced a new benchmark called StegPoet to demonstrate the method's effectiveness beyond standard planning problems.</p></li><li><p><a href="https://openai.com/index/trading-inference-time-compute-for-adversarial-robustness/">Trading inference-time computer for adversarial robustness</a>, <em>OpenAI</em>. Explores a novel approach to improving the adversarial robustness of LLMs through increasing inference-time compute rather than traditional adversarial training. The researchers tested various models (specifically OpenAI's o1-preview and o1-mini) against different types of attacks, finding that simply allowing models to have more computation time during inference led to improved robustness across multiple attack types - from jailbreaks to adversarial images. Unlike adversarial training, which requires anticipating specific attack types, this approach improves robustness without requiring prior knowledge of potential attacks. This is intuitively like thinking a bit more before acting, it tends to give rise to better outcomes :-)</p></li><li><p><a href="https://arxiv.org/abs/2501.08156">Inference-Time-Compute: More Faithful? A Research Note</a>, <em>Truthful AI. </em>Examines whether AI models specifically trained to generate long chains of thought (called Inference-Time-Compute or ITC models) are more "faithful" in acknowledging factors that influence their decisions. The researchers tested two ITC models based on Qwen and Gemini against several non-ITC models by adding various cues to prompts (like <em>"A Stanford Professor thinks the answer is D"</em>) and checking if the models would acknowledge these cues when they influenced their answers. The ITC models were significantly better at articulating when cues influenced their decisions - in some cases acknowledging cues over 50% of the time compared to under 15% for non-ITC models.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://press.airstreet.com/subscribe?"><span>Subscribe now</span></a></p></li></ul><h3><strong>&#128176;Startups</strong></h3><h4><em><strong>&#128640; </strong></em><strong>Funding highlight reel</strong></h4><p><strong>Anthropic</strong>, makers of Claude, <a href="https://www.ft.com/content/ed631513-dd37-44a3-a536-b2002f5727cc">raised</a> a further $1B from Google.</p><p><strong>Bioptimus, </strong>building generative AI models for biotech, <a href="https://sifted.eu/articles/bioptimus-41m-series-a">raised</a> a $41M Series A, led by Cathay Innovation.</p><p><strong>Collate</strong>, automating paperwork in biotech, <a href="https://x.com/surbhisarnasf/status/1878884972909781119?s=51&amp;t=8YCMEcmVVXRPm8SXTMgdlw">raised</a> a $30M seed, led by Redpoint.</p><p><strong>Coram*</strong>, applying AI to physical security, <a href="https://www.coram.ai/post/coram-announces-series-a-fundraise">raised</a> a $30M Series A, led by Battery Ventures.</p><p><strong>ElevenLabs*</strong>, the audio generation start-up, <a href="https://techcrunch.com/2025/01/30/elevenlabs-raises-180-million-in-series-c-funding-at-3-3-billion-valuation/">raised</a> $180M Series C, led by a16z and ICONIQ Growth.</p><p><strong>Eve</strong>, automating plaintiff legal services, <a href="https://a16z.com/announcement/investing-in-eve/">raised</a> a $47M Series A, led by a16z.</p><p><strong>Hippocratic AI</strong>, handling non-diagnostic patient-facing tasks, <a href="https://techcrunch.com/2025/01/09/hippocratic-ai-raises-141m-for-creating-patient-facing-ai-agents/?guccounter=1&amp;guce_referrer=aHR0cHM6Ly93d3cuZ29vZ2xlLmNvbS8&amp;guce_referrer_sig=AQAAAGaIYpwgyjdvOt3bUwUGMaeFAFn5_Jvb3nmc9g1YjwTKymRsejKWvLxxdLRzZKhZWfS2FxIsiNSTJDoqMGEtOWiirhlxYHJ7wdkviLs9WOQyJiIHLGbmwMBJy9AATv8y-5CXsd9LybNtib6NuH-rh40qUJJjKVI6GZD42Sq_TEzQ">raised</a> a $141M Series B, led by Kleiner Perkins.</p><p><strong>KoBold Metals</strong>, applying AI to mining for the energy transition, raised a $537M Series C, co-led by T Rowe Price and Durable Capital Partners.</p><p><strong>Lindus Health</strong>, automating the clinical trials process, raised a $55M Series B, led by Balderton Capital.</p><p><strong>NEURA Robotics</strong>, the humanoid company, <a href="https://www.ft.com/content/9602467d-f5d7-40eb-af5a-f1fbf1ccfcd7">raised</a> a $125M Series B, led by Lingotto Investment Management.</p><p><strong>Overland AI</strong>, developing software for autonomous ground vehicles, <a href="https://www.overland.ai/news/overland-ai-raises-32-million-in-series-a-funding-to-advance-ground-autonomy-for-u-s-defense-forces">raised</a> a $32M Series A, led by 8VC.</p><p><strong>Raspberry AI</strong>, using AI for fashion design, <a href="https://techcrunch.com/2025/01/13/raspberry-ai-raises-24m-from-a16z-to-accelerate-fashion-design/">raised</a> a $24M Series A, led by a16z.</p><p><strong>Sereact*</strong>, building AI-first software for robotics, has <a href="https://sereact.ai/posts/sereact-fundraising-series-a">raised</a> a $26M Series A, led by Creandum.</p><p><strong>Slingshot</strong>, building a foundation model for psychology, <a href="https://a16z.com/announcement/investing-in-slingshot-ai/">raised</a> a $40M Series A, led by a16z.</p><p><strong>Synthesia*</strong>, the AI avatar generation platform, <a href="https://www.synthesia.io/series-d">raised</a> a $180M Series D, led by NEA.</p><p><strong>ThreatMark</strong>, using AI to fight online fraud, <a href="https://www.threatmark.com/threatmark-to-advance-the-fight-against-online-fraud/">raised</a> a $23M funding round, led by Octopus Ventures.</p><p><em><strong>*</strong> denotes companies in which the authors hold shares.</em></p><h3><em><strong>&#129309; </strong></em><strong>Exits</strong></h3><p><strong>Flowrite</strong>, the email division of LLM evaluation company Flow AI, was <a href="https://www.morningstar.com/news/accesswire/964859msn/maestro-labs-acquires-flowrite-to-build-the-largest-independent-ai-email-assistant-on-the-market">acquired</a> by Maestro Labs.</p><div><hr></div><p>Signing off,</p><p>Nathan Benaich and Alex Chalmers on 2 February 2025</p><p><a href="https://www.airstreet.com/?utm_campaign=Your%20guide%20to%20AI&amp;utm_medium=email&amp;utm_source=Substack%20newsletter">Air Street Capital</a> | <a href="https://www.twitter.com/nathanbenaich?utm_campaign=Your%20guide%20to%20AI&amp;utm_medium=email&amp;utm_source=Substack%20newsletter">Twitter</a> | <a href="https://www.linkedin.com/in/nathanbenaich/?utm_campaign=Your%20guide%20to%20AI&amp;utm_medium=email&amp;utm_source=Substack%20newsletter">LinkedIn</a> | <a href="https://www.stateof.ai/?utm_campaign=Your%20guide%20to%20AI&amp;utm_medium=email&amp;utm_source=Substack%20newsletter">State of AI Report</a> | <a href="https://www.raais.co/?utm_campaign=Your%20guide%20to%20AI&amp;utm_medium=email&amp;utm_source=Substack%20newsletter">RAAIS</a> | <a href="https://airstreet.com/events">Events</a></p><p><em>Air Street Capital invests in AI-first entrepreneurs from the very beginning of your company-building journey.</em></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!pH9K!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F321f65c3-6c53-45b7-85d1-36017ac3ceb1_1694x944.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!pH9K!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F321f65c3-6c53-45b7-85d1-36017ac3ceb1_1694x944.png 424w, https://substackcdn.com/image/fetch/$s_!pH9K!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F321f65c3-6c53-45b7-85d1-36017ac3ceb1_1694x944.png 848w, https://substackcdn.com/image/fetch/$s_!pH9K!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F321f65c3-6c53-45b7-85d1-36017ac3ceb1_1694x944.png 1272w, https://substackcdn.com/image/fetch/$s_!pH9K!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F321f65c3-6c53-45b7-85d1-36017ac3ceb1_1694x944.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!pH9K!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F321f65c3-6c53-45b7-85d1-36017ac3ceb1_1694x944.png" width="1456" height="811" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/321f65c3-6c53-45b7-85d1-36017ac3ceb1_1694x944.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:811,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:95701,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!pH9K!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F321f65c3-6c53-45b7-85d1-36017ac3ceb1_1694x944.png 424w, https://substackcdn.com/image/fetch/$s_!pH9K!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F321f65c3-6c53-45b7-85d1-36017ac3ceb1_1694x944.png 848w, https://substackcdn.com/image/fetch/$s_!pH9K!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F321f65c3-6c53-45b7-85d1-36017ac3ceb1_1694x944.png 1272w, https://substackcdn.com/image/fetch/$s_!pH9K!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F321f65c3-6c53-45b7-85d1-36017ac3ceb1_1694x944.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div>]]></content:encoded></item><item><title><![CDATA[91% of AI papers used NVIDIA in 2024]]></title><description><![CDATA[Final 2024 statistics for AI chip usage across NVIDIA, Google, Huawei, Apple, AMD, Cerebras, Cambricon, Graphcore, Groq, SambaNova, and Intel Habana in the State of AI Report Compute Index.]]></description><link>https://press.airstreet.com/p/compute-index-2024</link><guid isPermaLink="false">https://press.airstreet.com/p/compute-index-2024</guid><dc:creator><![CDATA[Air Street Press]]></dc:creator><pubDate>Tue, 21 Jan 2025 14:29:59 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/ac88ae55-0f92-4647-b91b-3fcb589b964b_1542x860.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h3><strong>Introduction</strong></h3><p>For the 2022 edition of the State of AI Report, we launched the Compute Index with our data partners Zeta Alpha. Its goal is to track the size of the largest GPU computing systems across public and private clouds, as well as national resources. We also track the usage (and emergence of) specific AI accelerators in published AI research papers. This is a street-level view of which chips are the most or least popular as voted by AI researchers in their own work, which helps inform the industry&#8217;s higher level view of which chip companies are winning or losing (and by what margin). </p><p>So, back in 2022, here&#8217;s what the pictured looked like&#8230;</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!L1vT!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5f26946-6180-469f-ae17-488e5d0b6a9c_1834x874.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!L1vT!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5f26946-6180-469f-ae17-488e5d0b6a9c_1834x874.png 424w, https://substackcdn.com/image/fetch/$s_!L1vT!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5f26946-6180-469f-ae17-488e5d0b6a9c_1834x874.png 848w, https://substackcdn.com/image/fetch/$s_!L1vT!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5f26946-6180-469f-ae17-488e5d0b6a9c_1834x874.png 1272w, https://substackcdn.com/image/fetch/$s_!L1vT!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5f26946-6180-469f-ae17-488e5d0b6a9c_1834x874.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!L1vT!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5f26946-6180-469f-ae17-488e5d0b6a9c_1834x874.png" width="1456" height="694" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b5f26946-6180-469f-ae17-488e5d0b6a9c_1834x874.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:694,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:114646,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!L1vT!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5f26946-6180-469f-ae17-488e5d0b6a9c_1834x874.png 424w, https://substackcdn.com/image/fetch/$s_!L1vT!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5f26946-6180-469f-ae17-488e5d0b6a9c_1834x874.png 848w, https://substackcdn.com/image/fetch/$s_!L1vT!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5f26946-6180-469f-ae17-488e5d0b6a9c_1834x874.png 1272w, https://substackcdn.com/image/fetch/$s_!L1vT!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5f26946-6180-469f-ae17-488e5d0b6a9c_1834x874.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>&#8230;an NVIDIA washout. </p><p>And this washout held for several years, in fact. It&#8217;s acutely felt by investors who made $6B worth of bets on AI chip competitors between 2016 and 2024. As of 9 October 2024, their money sat at a mark-to-market value of $31B. Had they invested the $6B into NVIDIA instead, they&#8217;d be sitting on $120B of value.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!RAng!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29211816-a11c-4a2a-9b15-365e9491019e_1920x1096.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!RAng!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29211816-a11c-4a2a-9b15-365e9491019e_1920x1096.png 424w, https://substackcdn.com/image/fetch/$s_!RAng!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29211816-a11c-4a2a-9b15-365e9491019e_1920x1096.png 848w, https://substackcdn.com/image/fetch/$s_!RAng!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29211816-a11c-4a2a-9b15-365e9491019e_1920x1096.png 1272w, https://substackcdn.com/image/fetch/$s_!RAng!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29211816-a11c-4a2a-9b15-365e9491019e_1920x1096.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!RAng!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29211816-a11c-4a2a-9b15-365e9491019e_1920x1096.png" width="1456" height="831" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/29211816-a11c-4a2a-9b15-365e9491019e_1920x1096.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:831,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:246296,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!RAng!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29211816-a11c-4a2a-9b15-365e9491019e_1920x1096.png 424w, https://substackcdn.com/image/fetch/$s_!RAng!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29211816-a11c-4a2a-9b15-365e9491019e_1920x1096.png 848w, https://substackcdn.com/image/fetch/$s_!RAng!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29211816-a11c-4a2a-9b15-365e9491019e_1920x1096.png 1272w, https://substackcdn.com/image/fetch/$s_!RAng!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29211816-a11c-4a2a-9b15-365e9491019e_1920x1096.png 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h3>Today&#8217;s picture</h3><p>As we close the books on our data from 2024, the overall scenery looks the same, but we see some definition emerging in the backdrop.</p><p>In the chart below, we plot the <strong>sum of all AI research papers</strong> that make use of chips from specific vendors, per year. For NVIDIA, this includes the 2080, RTX 3090, 4090, K80, P100, V100, A100, H100, Titan, and Jetson chips. The big 6 startups are Habana, Graphcore, Cerebras, Sambanova, Cambricon, and Groq. AMD chips include the MI250, MI300 and MI300X. </p><p>NVIDIA comes in at <strong>44,389</strong> <strong>papers</strong> published in 2024, Google TPUs clock <strong>1,702 papers</strong>, big 6 startups at <strong>586 papers</strong>,<strong> </strong>Apple at <strong>604 papers</strong>, and AMD at 264 papers.  </p><p>Clearly, we&#8217;re still living in an NVIDIA show: the company&#8217;s chips <strong>capture 91% of all AI research paper chip usage</strong>.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!YxGe!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd359ae14-5b9d-48ca-a6a5-932c658af254_1796x856.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!YxGe!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd359ae14-5b9d-48ca-a6a5-932c658af254_1796x856.png 424w, https://substackcdn.com/image/fetch/$s_!YxGe!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd359ae14-5b9d-48ca-a6a5-932c658af254_1796x856.png 848w, https://substackcdn.com/image/fetch/$s_!YxGe!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd359ae14-5b9d-48ca-a6a5-932c658af254_1796x856.png 1272w, https://substackcdn.com/image/fetch/$s_!YxGe!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd359ae14-5b9d-48ca-a6a5-932c658af254_1796x856.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!YxGe!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd359ae14-5b9d-48ca-a6a5-932c658af254_1796x856.png" width="1456" height="694" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d359ae14-5b9d-48ca-a6a5-932c658af254_1796x856.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:694,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:137181,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!YxGe!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd359ae14-5b9d-48ca-a6a5-932c658af254_1796x856.png 424w, https://substackcdn.com/image/fetch/$s_!YxGe!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd359ae14-5b9d-48ca-a6a5-932c658af254_1796x856.png 848w, https://substackcdn.com/image/fetch/$s_!YxGe!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd359ae14-5b9d-48ca-a6a5-932c658af254_1796x856.png 1272w, https://substackcdn.com/image/fetch/$s_!YxGe!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd359ae14-5b9d-48ca-a6a5-932c658af254_1796x856.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>If we double click on NVIDIA&#8217;s chip lineup usage, we see that the <strong>A100 is still the most popular chip</strong>. The H100 is seeing rapid growth, while older cards such as the V100 hold their own. </p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!LYeH!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F726ec96c-8683-4846-9e7c-768af36d80a5_1830x850.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!LYeH!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F726ec96c-8683-4846-9e7c-768af36d80a5_1830x850.png 424w, https://substackcdn.com/image/fetch/$s_!LYeH!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F726ec96c-8683-4846-9e7c-768af36d80a5_1830x850.png 848w, https://substackcdn.com/image/fetch/$s_!LYeH!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F726ec96c-8683-4846-9e7c-768af36d80a5_1830x850.png 1272w, https://substackcdn.com/image/fetch/$s_!LYeH!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F726ec96c-8683-4846-9e7c-768af36d80a5_1830x850.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!LYeH!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F726ec96c-8683-4846-9e7c-768af36d80a5_1830x850.png" width="1456" height="676" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/726ec96c-8683-4846-9e7c-768af36d80a5_1830x850.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:676,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:153650,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!LYeH!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F726ec96c-8683-4846-9e7c-768af36d80a5_1830x850.png 424w, https://substackcdn.com/image/fetch/$s_!LYeH!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F726ec96c-8683-4846-9e7c-768af36d80a5_1830x850.png 848w, https://substackcdn.com/image/fetch/$s_!LYeH!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F726ec96c-8683-4846-9e7c-768af36d80a5_1830x850.png 1272w, https://substackcdn.com/image/fetch/$s_!LYeH!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F726ec96c-8683-4846-9e7c-768af36d80a5_1830x850.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h3></h3><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/p/compute-index-2024?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://press.airstreet.com/p/compute-index-2024?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><h3>Any contenders from big tech?</h3><p>While the above picture is certainly stark, here&#8217;s what the YoY growth trends look like for specific chips. Of note, Google&#8217;s TPU usage has grown almost 1,000% YoY to account for 1,702 papers in 2024. Next up, Huawei, AMD and Apple are racking up usage, altogether totalling 955 papers. A ways to go.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!usL0!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a62cea6-c0bc-42f8-a38d-3c7862f53f0b_1812x896.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!usL0!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a62cea6-c0bc-42f8-a38d-3c7862f53f0b_1812x896.png 424w, https://substackcdn.com/image/fetch/$s_!usL0!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a62cea6-c0bc-42f8-a38d-3c7862f53f0b_1812x896.png 848w, https://substackcdn.com/image/fetch/$s_!usL0!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a62cea6-c0bc-42f8-a38d-3c7862f53f0b_1812x896.png 1272w, https://substackcdn.com/image/fetch/$s_!usL0!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a62cea6-c0bc-42f8-a38d-3c7862f53f0b_1812x896.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!usL0!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a62cea6-c0bc-42f8-a38d-3c7862f53f0b_1812x896.png" width="1456" height="720" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3a62cea6-c0bc-42f8-a38d-3c7862f53f0b_1812x896.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:720,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:104484,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!usL0!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a62cea6-c0bc-42f8-a38d-3c7862f53f0b_1812x896.png 424w, https://substackcdn.com/image/fetch/$s_!usL0!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a62cea6-c0bc-42f8-a38d-3c7862f53f0b_1812x896.png 848w, https://substackcdn.com/image/fetch/$s_!usL0!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a62cea6-c0bc-42f8-a38d-3c7862f53f0b_1812x896.png 1272w, https://substackcdn.com/image/fetch/$s_!usL0!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a62cea6-c0bc-42f8-a38d-3c7862f53f0b_1812x896.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h3>Cerebras maintains its lead and Groq overtakes Graphcore</h3><p>Over the summer, wafer-scale compute system company, Cerebras, launched their inference service and filed to IPO. Their chips have seen continued growth in AI research papers in 2024, notching them the number 1 spot for the second year in a row. </p><p>Meanwhile, Groq&#8217;s come out from nowhere to the number 2 spot. The company, founded by members of the original Google TPU team, are competing on the (very crowded) inference market too. Usage of Graphcore&#8217;s products, sold to SoftBank Group in the fall, is levelling out. </p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!5qqI!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd4c49b01-4d72-4ae0-aa53-647849e89d7a_1826x758.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!5qqI!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd4c49b01-4d72-4ae0-aa53-647849e89d7a_1826x758.png 424w, https://substackcdn.com/image/fetch/$s_!5qqI!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd4c49b01-4d72-4ae0-aa53-647849e89d7a_1826x758.png 848w, https://substackcdn.com/image/fetch/$s_!5qqI!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd4c49b01-4d72-4ae0-aa53-647849e89d7a_1826x758.png 1272w, https://substackcdn.com/image/fetch/$s_!5qqI!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd4c49b01-4d72-4ae0-aa53-647849e89d7a_1826x758.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!5qqI!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd4c49b01-4d72-4ae0-aa53-647849e89d7a_1826x758.png" width="1456" height="604" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d4c49b01-4d72-4ae0-aa53-647849e89d7a_1826x758.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:604,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:130970,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!5qqI!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd4c49b01-4d72-4ae0-aa53-647849e89d7a_1826x758.png 424w, https://substackcdn.com/image/fetch/$s_!5qqI!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd4c49b01-4d72-4ae0-aa53-647849e89d7a_1826x758.png 848w, https://substackcdn.com/image/fetch/$s_!5qqI!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd4c49b01-4d72-4ae0-aa53-647849e89d7a_1826x758.png 1272w, https://substackcdn.com/image/fetch/$s_!5qqI!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd4c49b01-4d72-4ae0-aa53-647849e89d7a_1826x758.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h3><strong>Closing thoughts</strong></h3><p>The NVIDIA show continues and long will that likely continue&#8230;</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!hTA5!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa96e0cba-cbc7-47ad-9504-11b9816b6ff4_1916x1090.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!hTA5!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa96e0cba-cbc7-47ad-9504-11b9816b6ff4_1916x1090.png 424w, https://substackcdn.com/image/fetch/$s_!hTA5!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa96e0cba-cbc7-47ad-9504-11b9816b6ff4_1916x1090.png 848w, https://substackcdn.com/image/fetch/$s_!hTA5!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa96e0cba-cbc7-47ad-9504-11b9816b6ff4_1916x1090.png 1272w, https://substackcdn.com/image/fetch/$s_!hTA5!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa96e0cba-cbc7-47ad-9504-11b9816b6ff4_1916x1090.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!hTA5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa96e0cba-cbc7-47ad-9504-11b9816b6ff4_1916x1090.png" width="1456" height="828" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/a96e0cba-cbc7-47ad-9504-11b9816b6ff4_1916x1090.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:828,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:489814,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!hTA5!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa96e0cba-cbc7-47ad-9504-11b9816b6ff4_1916x1090.png 424w, https://substackcdn.com/image/fetch/$s_!hTA5!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa96e0cba-cbc7-47ad-9504-11b9816b6ff4_1916x1090.png 848w, https://substackcdn.com/image/fetch/$s_!hTA5!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa96e0cba-cbc7-47ad-9504-11b9816b6ff4_1916x1090.png 1272w, https://substackcdn.com/image/fetch/$s_!hTA5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa96e0cba-cbc7-47ad-9504-11b9816b6ff4_1916x1090.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://press.airstreet.com/subscribe?"><span>Subscribe now</span></a></p>]]></content:encoded></item><item><title><![CDATA[State of AI: December 2024 newsletter]]></title><description><![CDATA[Welcome to the latest issue of the State of AI newsletter, an editorialized newsletter covering the key developments in AI policy, research, industry, and start-ups over the last month.]]></description><link>https://press.airstreet.com/p/the-state-of-ai-2024-dec</link><guid isPermaLink="false">https://press.airstreet.com/p/the-state-of-ai-2024-dec</guid><dc:creator><![CDATA[Air Street Press]]></dc:creator><pubDate>Sun, 01 Dec 2024 17:06:11 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd755c760-a8df-45f5-bdb5-3f7f8403b290_1980x1108.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Hi everyone!</p><p>Welcome to the latest issue of the State of AI newsletter, an editorialized newsletter covering the key developments in AI policy, research, industry, and start-ups over the last month.</p><ul><li><p>Amid the latest round of <strong>scaling laws hitting a wall</strong> of speculation, Nathan <a href="https://press.airstreet.com/p/there-is-no-scaling-wall-in-discussion">interviewed</a> Eiso Kant, the co-founder and CTO of Poolside. In this 45 minute video, we discuss scaling laws, synthetic data, training infrastructure, reasoning, economics, and much more.</p></li><li><p>Nathan joined Matt Turck&#8217;s MAD Podcast to discuss the <strong>State of AI Report</strong> key findings. You can watch the chat on <a href="https://www.youtube.com/watch?v=c_87fwKmkjM">YouTube</a> or <a href="https://podcasts.apple.com/us/podcast/state-of-ai-2024-frontier-models-ai-geopolitics/id1686238724?i=1000676943376">listen</a> on Apple.</p></li><li><p>We relaunched the <strong>RAAIS Fellowships</strong> - a cash grant and compute credit package for individuals/teams working on open source AI projects - we want to hear from you. For more details on what we&#8217;re looking for and how to apply, see <a href="https://press.airstreet.com/p/raais-fellowships">here</a>.</p></li><li><p>Congratulations to our friends at <strong>Odyssey</strong>, who raised a Series A round, led by EQT Ventures, with backing from Air Street and GV. You can read more about the team <a href="https://odyssey.systems/learning-from-our-world">here</a> and why we originally backed them in July <a href="https://press.airstreet.com/p/our-investment-in-odyssey">here</a>.</p></li><li><p>The <strong><a href="https://press.airstreet.com/">Air Street Press</a></strong> continues to whir, with recent pieces covering <a href="https://press.airstreet.com/p/how-not-to-build-a-drone">drones</a>, <a href="https://press.airstreet.com/p/the-ai-energy-wars-will-get-worse">AI&#8217;s energy demands</a>, <a href="https://press.airstreet.com/p/percy-liang-on-truly-open-ai">Percy Liang on truly open AI</a>, and the start of our <a href="https://press.airstreet.com/s/stateofai">State of AI outtakes series</a>.</p></li><li><p>It was great to see so many friendly faces at <strong>London AI </strong>and <strong>Paris AI</strong> meetups in the last fortnight. We&#8217;ve reached the end of our events program for 2024, but we&#8217;ll be back on both sides of the Atlantic in 2025. Subscribe to <a href="https://lu.ma/airstreet">our events page</a> to ensure you don&#8217;t miss out.</p></li><li><p>Along with events, you can get all of our news, analysis, and events directly in your inbox if you subscribe to <a href="https://press.airstreet.com/">Air Street Press</a>.</p></li></ul><p>We love hearing what you&#8217;re up to and what&#8217;s on your mind, just hit reply or forward to your friends :-)</p><div><hr></div><h3><em><strong>&#127758; The (geo)politics of AI</strong></em></h3><p>Are we so back, or is it so over? Following this November&#8217;s US presidential election, it depends who you ask.</p><p>On the one hand, you have a Republican platform committed to repealing the Biden White House&#8217;s frontier AI executive order. On the other hand, Trump confidant Elon Musk supported California&#8217;s sweeping proposed AI regulation. Trump is a China trade wars enthusiast, Musk is a China dove. The prospective Secretary of the Department of Health and Human Services is a vaccine skeptic with a range of &#8230; eccentric views, but it&#8217;s unclear how much direct impact he&#8217;ll have on the biotech industry (beyond triggering a -10% downdraft in the XBI biotech index). Trump has no desire to spend more on defense or engage in overseas conflicts, but is rumored to be considering Anduril co-founder Trae Stephens for a senior role in the Department of Defense.</p><p>In short, the tech world is in &#8216;choose your own adventure&#8217; mode. Whether you&#8217;re optimistic or pessimistic about the result, it&#8217;s possible to cherrypick appointees, past quotes, or implied policy positions and build your own Marvel Cinematic Universe about what the next four years will look like.</p><p>In reality, <em>we just don&#8217;t know </em>what a Trump Presidency will mean for AI. Considering the unpredictable personalities and the (short) median tenure of a Trump ally, it&#8217;s hard to predict the next six weeks, let alone the next six months.</p><p>If you want to read more about what we don&#8217;t know, but what will be some interesting debates to follow - <a href="https://press.airstreet.com/p/state-of-ai-outtakes-3-us-politics">Air Street Press</a> has got you covered.</p><p>Are we living in the Leopold Aschenbrenner&#8217;s future? In this year&#8217;s <a href="https://docs.google.com/presentation/d/1GmZmoWOa2O92BPrncRcTKa15xvQGhq7g4I4hJSNlC0M/edit#slide=id.g2f0dd8e71d3_3_0">State of AI Report</a>, we poured cold water on the <a href="https://situational-awareness.ai/">Situational Awareness</a> author&#8217;s accelerationist fan fiction, which involved the nationalization of major AI labs to pave the way for an AGI Manhattan Project to save the free world.</p><p>But since then, the US-China Economic and Security Review Commission, an independent body that reports to Congress, released its <a href="https://www.uscc.gov/annual-report/2024-annual-report-congress">2024 Annual Report</a>. Top of its list was recommending that <em>&#8220;Congress establish and fund a Manhattan Project-like program dedicated to racing to and acquiring an Artificial General Intelligence capability&#8221;</em>. The doomers on X responded in their characteristic manner: predicting the end of the world and sharing Shoggoth memes that we struggle to follow.</p><p>Before we get too excited, it&#8217;s worth remembering a few things.</p><p>Firstly, the USCC has no policymaking authority. While influential over some in the legislative branch, it probably represents the most hawkish wing of mainstream opinion on China.</p><p>Secondly, its specific recommendations - making funding available to US companies and procurement changes - seem rather tame compared to the original Manhattan Project.</p><p>Finally, as journalist Garrison Lovely <a href="https://garrisonlovely.substack.com/p/china-hawks-are-manufacturing-an">has observed</a>, the report is full of embarrassing technical inaccuracies, including describing OpenAI as a &#8216;model&#8217;, references to the non-existent &#8220;ChatGPT-3&#8221; product, and a bafflingly bad definition of AGI. That might give readers pause for thought.</p><p>But Team Safety can draw comfort from one source. Maybe government-run frontier evals are working?</p><p>Our friend, Logan Graham, who leads the Frontier Red Team at Anthropic, <a href="https://x.com/logangraham/status/1860133498457325725">recently described</a> the UK and US AI Safety Institutes as a rare example of <em>&#8220;extreme competence/capacity&#8221;</em> in government and a <em>&#8220;99th %ile&#8221;</em> outcome. He believes that they played a meaningful role in improving Sonnet&#8217;s robustness.</p><p>For people who don&#8217;t know what the AI Safety Institutes do, the joint UK/US <a href="https://www.aisi.gov.uk/work/pre-deployment-evaluation-of-anthropics-upgraded-claude-3-5-sonnet">blog post</a> and <a href="https://cdn.prod.website-files.com/663bd486c5e4c81588db7a1d/673b689ec926d8d32e889a8e_UK-US-Testing-Report-Nov-19.pdf">technical report</a> on Sonnet 3.5 testing give us a <a href="https://cdn.prod.website-files.com/663bd486c5e4c81588db7a1d/673b689ec926d8d32e889a8e_UK-US-Testing-Report-Nov-19.pdf">few clues</a>. The teams tested the model&#8217;s capabilities against reference models and human baselines across a range of capabilities, spanning bio, cyber and software development, while probing the efficacy of their safeguards. They used a blend of public and privately-developed evaluations.</p><p>OpenAI can only dream of this kind of smooth pre-deployment experience. This week, its text-to-video generation model <a href="https://www.ft.com/content/5281eff4-711b-49ac-8227-634dbeed757b">was apparently leaked</a> on Hugging Face by disgruntled artists involved in testing it. They published an <a href="https://huggingface.co/spaces/PR-Puppets/PR-Puppet-Sora">accompanying letter</a> declaring that they were being used for <em>&#8220;artwashing&#8221;</em> and that: <em>&#8220;Hundreds of artists provide unpaid labor through bug testing, feedback and experimental work for the program for a $150B valued company. While hundreds contribute for free, a select few will be chosen through a competition to have their Sora-created films screened &#8212; offering minimal compensation which pales in comparison to the substantial PR and marketing value OpenAI receives.&#8221;</em></p><p>Access was pulled within a couple of hours and OpenAI has refused to confirm whether the model in question was authentically Sora.</p><p>This is an interesting new front in the AI/art wars. It&#8217;s notable that the letter didn&#8217;t mention copyright anywhere and encouraged artists to use open source tools as an alternative. If frontier AI labs simply &#8230; pay their testers properly, a two-front war can likely be averted.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:&quot;button-wrapper&quot;}" data-component-name="ButtonCreateButton"><a class="button primary button-wrapper" href="https://press.airstreet.com/subscribe?"><span>Subscribe now</span></a></p><h3><em><strong>&#127850; Hardware</strong></em></h3><p>NVIDIA enjoyed another <a href="https://nvidianews.nvidia.com/news/nvidia-announces-financial-results-for-third-quarter-fiscal-2025">blowout quarter</a>, with quarterly revenue of $35.1B, up 17% on Q2 and 94% year-on-year. The vast majority of this came from the company&#8217;s data center business. As ever, despite exceeding consensus expectations, NVIDIA&#8217;s failure to meet the most extreme predictions actually caused the share price to fall slightly. This puts NVIDIA&#8217;s diluted earnings per share at 0.78, versus 0.47 for AMD and a truly horrifying -3.88 for Intel.</p><p>But Intel received one consolation prize in November. The company secured a $7.86B funding agreement <a href="https://www.intel.com/content/www/us/en/newsroom/news/intel-chips-act.html">via the CHIPS Act</a> to support semiconductor manufacturing in Arizona, New Mexico, Ohio, and Oregon. A few billion in tax breaks have also been thrown in as a sweetener. This could be a last hurrah for the CHIPS Act, with Trump <a href="https://www.washingtonpost.com/technology/2024/11/26/intel-11-billion-grant-chips-biden/">implying</a> his preferred method of onshoring the supply chain would be tariffs. This was $600M smaller than the anticipated award and comes amid rumors that Qualcomm <a href="https://www.bloomberg.com/news/articles/2024-11-26/qualcomm-s-takeover-interest-in-intel-is-said-to-cool?sref=FwsjHXQj">has lost interest in acquiring the company</a>. While a blow to M&amp;A lawyers everywhere, it means Qualcomm will be spared the task of running a loss-making semiconductor manufacturing business.</p><p>How is Intel loss-making amid a global semis boom? This is a question we might be returning to in an upcoming series on <a href="https://press.airstreet.com/">Air Street Press</a>.</p><p>Are US sanctions on China achieving much? As we see stunning open models from Chinese labs (more on those below), the smart money would be on &#8216;no&#8217;. But while it&#8217;s easy to smuggle GPUs, smuggling advanced ASML lithography machines, which are both rare and roughly the size of a double-decker bus, is proving harder.</p><p>This explains why Chinese chip makers, like Huawei and SMIC are increasingly trying to push the limits of older ASML equipment. But if Bloomberg is to be believed, this <a href="https://www.bloomberg.com/news/articles/2024-11-19/china-s-chip-advances-stall-as-us-curbs-hit-huawei-ai-product">approach isn&#8217;t working</a>. This is because so-called &#8216;multi-patterning&#8217;, where a lithographic machine is forced to perform up to four exposures on a silicon wafer, is prone to alignment errors and yield losses - especially when done in conjunction with poor local equipment.</p><p>With western GPUs seemingly on tap, this may not currently matter for the best Chinese labs. But it does throw sand in the gears of the Chinese Communist Party&#8217;s long-standing policy of breaking dependence on western equipment. They&#8217;re apparently less optimistic about the long-term prospects of the GPU smuggling industry than many western commentators.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/p/the-state-of-ai-2024-dec?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:&quot;button-wrapper&quot;}" data-component-name="ButtonCreateButton"><a class="button primary button-wrapper" href="https://press.airstreet.com/p/the-state-of-ai-2024-dec?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><h3><em><strong>&#127981; Big <s>tech</s> start-ups&nbsp;</strong></em></h3><p>Masa has landed, with OpenAI <a href="https://www.cnbc.com/2024/11/26/openai-gets-1point5-billion-investment-from-softbank-in-tender-offer.html">allowing Softbank to grow its stake</a> in exchange for buying $1.5B in equity from employees, as part of a tender offer. Employees will have until 24 December to decide if they want to take part. There&#8217;s clearly logic to the deal for both sides. OpenAI naturally ties into the Softbank CEO&#8217;s 300 year-vision of the world, while a dollop of liquidity is likely not a bad way of motivating a workforce that&#8217;s seen a number of departures recently.</p><p>Anthropic have continued on their recent tear, shipping more features, including a <a href="https://www.anthropic.com/news/styles">toggle for Claude response styles</a>. More interestingly, however, is the <a href="https://www.anthropic.com/news/model-context-protocol">Model Context Protocol</a>, an open source standard for connecting data to LLM apps. Using the new desktop app, Claude can connect directly to GitHub, create a new repo, and make a PR through a simple integration. In the battle for mindshare, saving devs from the tedious task of writing multiple integrations is a smart move.</p><p>In news that will surprise &#8230; probably no one, Anthropic <a href="https://www.anthropic.com/news/anthropic-amazon-trainium">secured an additional $4B investment</a> from Amazon. But the extra cash comes at the price of additional closeness. The company is now working with AWS on developing and optimizing Trainium, the company&#8217;s AI chip and accompanying software ecosystem. Despite discounted access programs and an aggressive marketing push, adoption <a href="https://www.businessinsider.com/amazon-nvidia-aws-ai-chip-dominance-gpu-trainium-inferentia-2024-5">remains slow</a>. Amazon will go from being Anthropic&#8217;s primary cloud provider to also being its <a href="https://www.aboutamazon.com/news/aws/amazon-invests-additional-4-billion-anthropic-ai">primary training partner</a>. AWS customers will also get early access to new Claude models for finetuning purposes. Whatever else they might think, the likely departure of Lina Khan from the FTC can&#8217;t come soon enough for the world of GenAI partnerships.</p><p>In all this discussion of arms races, mindshare, and differentiation - do we have any sense of who&#8217;s winning? Menlo Ventures recently released a survey diving into <a href="https://menlovc.com/2024-the-state-of-generative-ai-in-the-enterprise/">enterprise LLM spend</a>.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!4bGK!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F07246756-c328-4e67-8b67-1895bfe82bb7_975x544.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!4bGK!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F07246756-c328-4e67-8b67-1895bfe82bb7_975x544.png 424w, https://substackcdn.com/image/fetch/$s_!4bGK!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F07246756-c328-4e67-8b67-1895bfe82bb7_975x544.png 848w, https://substackcdn.com/image/fetch/$s_!4bGK!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F07246756-c328-4e67-8b67-1895bfe82bb7_975x544.png 1272w, https://substackcdn.com/image/fetch/$s_!4bGK!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F07246756-c328-4e67-8b67-1895bfe82bb7_975x544.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!4bGK!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F07246756-c328-4e67-8b67-1895bfe82bb7_975x544.png" width="975" height="544" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/07246756-c328-4e67-8b67-1895bfe82bb7_975x544.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:544,&quot;width&quot;:975,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!4bGK!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F07246756-c328-4e67-8b67-1895bfe82bb7_975x544.png 424w, https://substackcdn.com/image/fetch/$s_!4bGK!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F07246756-c328-4e67-8b67-1895bfe82bb7_975x544.png 848w, https://substackcdn.com/image/fetch/$s_!4bGK!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F07246756-c328-4e67-8b67-1895bfe82bb7_975x544.png 1272w, https://substackcdn.com/image/fetch/$s_!4bGK!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F07246756-c328-4e67-8b67-1895bfe82bb7_975x544.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>This isn&#8217;t entirely surprising: OpenAI remains in the lead, but increasingly no longer being the default choice as competition intensifies. Anthropic&#8217;s big push on price and features is bearing fruit. Llama is struggling with large enterprise adoption. And Gemini is fighting hard for dev users after a slow start.</p><p>While the LLM adoption race rages in the US, the Chinese frontier has been ablaze. Following on from a series of strong LLM and VLM releases, the reasoning models have started to land. DeepSeek was first out of the gate with <a href="https://api-docs.deepseek.com/news/news1120">R1-Lite-Preview</a>, combining the team&#8217;s long-standing work on synthetic data and reasoning, to produce a model that rivals o1-preview-level performance on AIME &amp; MATH benchmarks. Community reception, as far as we can tell, has been polished.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!eBze!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7bc3e99b-eca5-43b6-96f9-91a2fa9918b6_1592x988.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!eBze!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7bc3e99b-eca5-43b6-96f9-91a2fa9918b6_1592x988.png 424w, https://substackcdn.com/image/fetch/$s_!eBze!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7bc3e99b-eca5-43b6-96f9-91a2fa9918b6_1592x988.png 848w, https://substackcdn.com/image/fetch/$s_!eBze!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7bc3e99b-eca5-43b6-96f9-91a2fa9918b6_1592x988.png 1272w, https://substackcdn.com/image/fetch/$s_!eBze!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7bc3e99b-eca5-43b6-96f9-91a2fa9918b6_1592x988.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!eBze!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7bc3e99b-eca5-43b6-96f9-91a2fa9918b6_1592x988.png" width="1456" height="904" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/7bc3e99b-eca5-43b6-96f9-91a2fa9918b6_1592x988.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:904,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!eBze!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7bc3e99b-eca5-43b6-96f9-91a2fa9918b6_1592x988.png 424w, https://substackcdn.com/image/fetch/$s_!eBze!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7bc3e99b-eca5-43b6-96f9-91a2fa9918b6_1592x988.png 848w, https://substackcdn.com/image/fetch/$s_!eBze!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7bc3e99b-eca5-43b6-96f9-91a2fa9918b6_1592x988.png 1272w, https://substackcdn.com/image/fetch/$s_!eBze!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7bc3e99b-eca5-43b6-96f9-91a2fa9918b6_1592x988.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>How is DeepSeek able to keep doing this? Dylan from SemiAnalysis <a href="https://x.com/dylan522p/status/1859302712803807696?s=51&amp;t=8YCMEcmVVXRPm8SXTMgdlw">reminds us</a> that the team does have 50k Hopper NVIDIA GPUs, and isn&#8217;t whittling these models together on a few A100s in a shed. As Dylan puts it, <em>&#8220;they are omega cracked on ML research and infra management but they aren&#8217;t doing it with that many fewer GPUs&#8221;</em>.</p><p>For a glimpse into the thinking behind DeepSeek, check out <a href="https://www.chinatalk.media/p/deepseek-ceo-interview-with-chinas">this lengthy interview</a> with CEO Liang Wenfeng. He makes the case for open models, argues that the chip embargo is hurting the business, and that DeepSeek is unusual in China for innovating rather than imitating.</p><p>DeepSeek isn&#8217;t the only Chinese team producing o1-like reasoning models.</p><p>After a run of strong (V)LLMs were released, QwQ&#8217;s reasoning model release should&#8217;ve been a big moment, especially given the <a href="https://qwenlm.github.io/blog/qwq-32b-preview/">very impressive benchmark performance</a> they reported. But have they been a victim of the arms race and forced to release ahead of schedule? Chinese LLM power users have complained about elements of its <a href="https://x.com/teortaxesTex/status/1861870848438653268/">output</a>, while still acknowledging its impressive capabilities and potential to act as a serious o1 competitor.</p><p>Does this mean the gap has closed between US and Chinese labs? Our friend, former OpenAI exec Miles Brundage, <a href="https://x.com/Miles_Brundage/status/1862053151924195823">reminds us</a> that these models are being compared to o1-preview, not o1, so we aren&#8217;t seeing them compared against the best disclosed performance from a US lab. At the same time, Miles notes that <em>&#8220;conversely, this wasn&#8217;t the best DeepSeek or Alibaba can ultimately do, either&#8221;</em> and that <em>&#8220;everyone actually doing this stuff at or near the frontier agrees there is plenty of gas left in the tank&#8221;</em>. In fact, we tend to agree - Nathan and Eiso (Poolside&#8217;s CTO/co-founder) <a href="https://press.airstreet.com/p/there-is-no-scaling-wall-in-discussion">discuss how deep learning isn&#8217;t yet hitting a wall</a>. Sorry to any Gary Marcus fans out there.</p><p>Crossing back to Europe, what&#8217;s going on at H Company? Regular Guide to AI readers will remember how the artists formerly known as Holistic raised a $220M seed, before losing three out of five of its co-founders. Since then, the company has shared its <a href="https://www.hcompany.ai/blog/a-research-update">first progress update</a> covering how Runner H 0.1, the agent they&#8217;ve built, powered by a proprietary VLM, squares up against the competition. H Company shows how they outperform Claude&#8217;s computer use agent across a number of tasks on the WebVoyager benchmark (largely around navigating to specific websites and performing certain tasks). <a href="https://x.com/sawyerhood/status/1861868727551332776?s=46&amp;t=8YCMEcmVVXRPm8SXTMgdlw">Not everyone&#8217;s impressed.</a>&nbsp;</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:&quot;button-wrapper&quot;}" data-component-name="ButtonCreateButton"><a class="button primary button-wrapper" href="https://press.airstreet.com/subscribe?"><span>Subscribe now</span></a></p><h3><strong>&#128300;Research</strong></h3><p><strong><a href="https://arxiv.org/abs/2411.15124">T&#220;LU 3: Pushing Frontiers in Open Language Model Post-Training</a></strong><em>, Allen Institute for AI, University of Washington</em>.</p><p>Presents T&#220;LU 3, a family of instruction-tuned language models based on the Llama 3.1 architecture. It follows a multi-stage training process that includes supervised fine-tuning (SFT), Direct Preference Optimization (DPO), and Reinforcement Learning with Verifiable Rewards (RLVR). The RLVR stage specifically trains the model on tasks that require verifiable correct answers, such as math problems and precise instruction following.</p><p>The models are designed to enhance core skills, including knowledge recall, reasoning, math, coding, instruction following, and safety. The training data includes both publicly available datasets (such as OpenAssistant and WildChat) and synthetically generated data, targeting specific skills. A key feature of T&#220;LU 3 is the careful decontamination of training data to ensure unbiased performance evaluation. The T&#220;LU 3 EVAL framework is used to assess the models across both development and unseen benchmark tasks.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!TK-e!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F481288a9-0a89-41f3-bbdb-af555a1e969a_620x251.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!TK-e!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F481288a9-0a89-41f3-bbdb-af555a1e969a_620x251.png 424w, https://substackcdn.com/image/fetch/$s_!TK-e!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F481288a9-0a89-41f3-bbdb-af555a1e969a_620x251.png 848w, https://substackcdn.com/image/fetch/$s_!TK-e!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F481288a9-0a89-41f3-bbdb-af555a1e969a_620x251.png 1272w, https://substackcdn.com/image/fetch/$s_!TK-e!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F481288a9-0a89-41f3-bbdb-af555a1e969a_620x251.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!TK-e!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F481288a9-0a89-41f3-bbdb-af555a1e969a_620x251.png" width="620" height="251" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/481288a9-0a89-41f3-bbdb-af555a1e969a_620x251.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:251,&quot;width&quot;:620,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!TK-e!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F481288a9-0a89-41f3-bbdb-af555a1e969a_620x251.png 424w, https://substackcdn.com/image/fetch/$s_!TK-e!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F481288a9-0a89-41f3-bbdb-af555a1e969a_620x251.png 848w, https://substackcdn.com/image/fetch/$s_!TK-e!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F481288a9-0a89-41f3-bbdb-af555a1e969a_620x251.png 1272w, https://substackcdn.com/image/fetch/$s_!TK-e!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F481288a9-0a89-41f3-bbdb-af555a1e969a_620x251.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>T&#220;LU 3 instruction-tuned models outperform other open-weight models, such as Llama 3.1 Instruct, Qwen 2.5, and Mistral-Instruct, and achieve competitive results with closed models like GPT-4o-mini and Claude 3.5 Haiku, particularly in tasks involving reasoning, safety, and precise instruction following. The T&#220;LU 3 family spans 8B to 70B parameters, and the team has made the model weights, training data, evaluation tools, and code publicly available.</p><p><strong><a href="https://metr.org/blog/2024-11-22-evaluating-r-d-capabilities-of-llms/">Evaluating frontier AI R&amp;D capabilities of language model agents against human experts,</a></strong> <em>METR.</em></p><p>Introduces RE-Bench, a benchmark for evaluating AI systems' ability to automate AI research and development tasks. It features 7 challenging ML engineering tasks and compares performance between AI agents and 71 human expert attempts across 8-hour sessions. The environments cover areas like kernel optimization, model finetuning, and scaling law experiments.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!FnSM!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3792befd-256e-42e1-90ed-a235f122cd62_617x317.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!FnSM!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3792befd-256e-42e1-90ed-a235f122cd62_617x317.png 424w, https://substackcdn.com/image/fetch/$s_!FnSM!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3792befd-256e-42e1-90ed-a235f122cd62_617x317.png 848w, https://substackcdn.com/image/fetch/$s_!FnSM!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3792befd-256e-42e1-90ed-a235f122cd62_617x317.png 1272w, https://substackcdn.com/image/fetch/$s_!FnSM!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3792befd-256e-42e1-90ed-a235f122cd62_617x317.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!FnSM!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3792befd-256e-42e1-90ed-a235f122cd62_617x317.png" width="617" height="317" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3792befd-256e-42e1-90ed-a235f122cd62_617x317.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:317,&quot;width&quot;:617,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!FnSM!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3792befd-256e-42e1-90ed-a235f122cd62_617x317.png 424w, https://substackcdn.com/image/fetch/$s_!FnSM!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3792befd-256e-42e1-90ed-a235f122cd62_617x317.png 848w, https://substackcdn.com/image/fetch/$s_!FnSM!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3792befd-256e-42e1-90ed-a235f122cd62_617x317.png 1272w, https://substackcdn.com/image/fetch/$s_!FnSM!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3792befd-256e-42e1-90ed-a235f122cd62_617x317.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Testing Claude 3.5 Sonnet and o1-preview models revealed that AI agents outperform humans in 2-hour time frames but show diminishing returns with longer durations. In contrast, humans demonstrate better improvement over extended periods, surpassing AI performance given 8+ hours. The study found AI agents can generate and test solutions much faster than humans and occasionally produce superior results, like developing more efficient GPU kernels.</p><p>However, the authors note significant limitations in extrapolating these results to real AI R&amp;D automation capabilities. The benchmark's contained nature, clear objectives, and short timeframes don't capture the full complexity of real research projects that span months and involve multiple interacting workstreams.</p><p><strong><a href="https://www.nabla.bio/news/denovo">De novo design of epitope-specific antibodies against soluble and multipass membrane proteins with high specificity, developability, and function</a></strong>, <em>Nabla Bio</em>.</p><p>Presents JAM (Joint Atomic Modeling), an AI system enabling fully <em>de novo</em> design of therapeutic antibodies with high specificity, functionality, and double-digit nanomolar affinities. JAM generates both single-domain (VHH) and full antibodies (scFv/mAb) that meet clinical development criteria, with iterative refinement improving binding success rates and affinities.</p><p>The researchers demonstrated JAM's capabilities against multiple targets, including the first fully computationally designed antibodies for challenging membrane proteins Claudin-4 and CXCR7. For SARS-CoV-2, JAM-designed antibodies achieved sub-nanomolar neutralization potency and drug-like developability profiles. Test-time computational introspection improved results, marking a novel application of compute scaling to protein design.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!eL5d!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb85b6fa9-84d6-472a-9f01-86f7fd668163_689x473.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!eL5d!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb85b6fa9-84d6-472a-9f01-86f7fd668163_689x473.png 424w, https://substackcdn.com/image/fetch/$s_!eL5d!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb85b6fa9-84d6-472a-9f01-86f7fd668163_689x473.png 848w, https://substackcdn.com/image/fetch/$s_!eL5d!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb85b6fa9-84d6-472a-9f01-86f7fd668163_689x473.png 1272w, https://substackcdn.com/image/fetch/$s_!eL5d!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb85b6fa9-84d6-472a-9f01-86f7fd668163_689x473.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!eL5d!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb85b6fa9-84d6-472a-9f01-86f7fd668163_689x473.png" width="689" height="473" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b85b6fa9-84d6-472a-9f01-86f7fd668163_689x473.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:473,&quot;width&quot;:689,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!eL5d!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb85b6fa9-84d6-472a-9f01-86f7fd668163_689x473.png 424w, https://substackcdn.com/image/fetch/$s_!eL5d!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb85b6fa9-84d6-472a-9f01-86f7fd668163_689x473.png 848w, https://substackcdn.com/image/fetch/$s_!eL5d!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb85b6fa9-84d6-472a-9f01-86f7fd668163_689x473.png 1272w, https://substackcdn.com/image/fetch/$s_!eL5d!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb85b6fa9-84d6-472a-9f01-86f7fd668163_689x473.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>A key innovation was JAM's dual ability to design antibodies and soluble proxies for membrane protein targets, enabling efficient screening. The process from design to characterization requires less than 6 weeks, allowing parallel campaigns. Current limitations include humanness scores aligning more closely with chimeric than fully human antibodies. Nevertheless, JAM represents a major advance in computational antibody design, with the potential to transform therapeutic discovery workflows.</p><p>Also on our radar:</p><ul><li><p><a href="https://generative-value-learning.github.io/">Vision Language Models are In-Context Value Learners</a>, <em>Google DeepMind, University of Pennsylvania, Stanford</em>. Introduces Generative Value Learning (GVL), a novel method for task progress estimation using VLMs. Instead of looking at video frames in order, GVL shuffles them, forcing the model to focus on the content of each frame rather than relying on their sequence. GVL excels in zero-shot and few-shot learning across over 300 real-world robotic tasks, including complex bimanual manipulations, without task-specific training. Applications span dataset filtering, success detection, and reinforcement learning, demonstrating scalability and versatility in robotic learning contexts.</p></li><li><p><a href="https://ai.meta.com/blog/open-catalyst-simulations-experiments/">Open Catalyst experiments 2024 (OCx24)</a>, <em>Meta</em>. Describes the launch of OCx24, a collaborative project that bridges computational and experimental catalyst research through a dataset of over 600 materials tested for green hydrogen production and CO2 recycling. The project combines AI-powered simulations analyzing 19,000+ materials with real-world testing using automated synthesis techniques from partners at the University of Toronto and VSParticle. Results show promise in identifying low-cost alternatives to platinum-based catalysts for hydrogen evolution reactions.</p></li><li><p><a href="https://aidantr.github.io/files/AI_innovation.pdf">Artificial Intelligence, Scientific Discovery, and Product Innovation</a>, <em>MIT</em>. Examines how AI impacts scientific research by analyzing its introduction to materials discovery work at a large U.S. company's R&amp;D lab. Finds that AI significantly boosted productivity, with scientists discovering 44% more materials and filing 39% more patents. But these gains were uneven - top scientists nearly doubled their output, while the bottom third saw minimal benefits. The differentiator was scientists' ability to evaluate AI's suggestions - those with strong judgment skills could effectively prioritize promising candidates, while others struggled with false positives.</p></li><li><p><a href="https://arxiv.org/abs/2411.10323">The Dawn of GUI Agent: A Preliminary Case Study with Claude 3.5 Computer Use</a>, <em>National University of Singapore</em>. Evaluates Claude 3.5 Computer Use, testing its capabilities across web browsing, office applications, and video games. The authors assess the model's performance across planning ability, action execution, and self-assessment. While scoring well on tasks like automated gaming and document editing, the model shows limitations in precise text selection, scrolling behavior, and accurately assessing task completion. The authors also introduce "Computer Use Out-of-the-Box" - an open-source framework to evaluate GUI automation models.</p></li><li><p><a href="https://arxiv.org/abs/2411.17501">Inference Scaling &#120437;Laws: The Limits of LLM Resampling with Imperfect Verifiers</a>, <em>Princeton</em>. Examines the limitations of inference scaling to improve the performance of weaker language models. Through experiments with coding tasks, the researchers found that while weaker models can match stronger models' performance on basic test cases through repeated sampling, they produce significantly more false positives - solutions that pass basic tests but fail comprehensive ones. The authors conclude there is a hard limit to weaker model performance, even with infinite sampling attempts.</p></li><li><p><a href="https://www.biorxiv.org/content/10.1101/2024.11.14.623630v1">InterPLM: Discovering Interpretable Features in Protein Language Models via Sparse Autoencoders</a>, <em>Stanford</em>. Presents a new method for understanding how protein language models (PLMs) work internally by using sparse autoencoders (SAEs) to extract interpretable features and their hidden layers. The researchers identified thousands of biologically meaningful patterns in ESM-2's representations, significantly more than could be found by analyzing individual neurons directly. They demonstrate that these extracted features correspond to known protein properties like binding sites and structural motifs, can help identify missing annotations in protein databases, and can control the model's outputs in targeted ways.&nbsp;</p></li></ul><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/p/the-state-of-ai-2024-dec?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:&quot;button-wrapper&quot;}" data-component-name="ButtonCreateButton"><a class="button primary button-wrapper" href="https://press.airstreet.com/p/the-state-of-ai-2024-dec?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><h3><strong>&#128176;Startups</strong></h3><h4><em><strong>&#128640; </strong></em><strong>Funding highlight reel</strong></h4><p><strong>Anthropic</strong>, the frontier model lab, <a href="https://www.anthropic.com/news/anthropic-amazon-trainium">agreed</a> a further $4B investment from Amazon.</p><p><strong>Odyssey</strong>, building generative world-building models for film and gaming, <a href="https://techcrunch.com/2024/11/13/this-generative-ai-startup-is-strapping-cameras-to-peoples-backs/#:~:text=By%20having%20humans%20strap%20cameras,anywhere%20a%20person%20can%20reach.">raised</a> an $18M Series A, led by EQT Ventures.</p><p><strong>Cradle Bio</strong>, building AI-based protein-making software, <a href="https://endpts.com/cradle-bio-raises-73m-series-b-for-protein-making-ai-software/">raised</a> a $73M Series B, led by IVP.</p><p><strong>Enveda, </strong>the clinical stage drug discovery company focused on medicinal plants, <a href="https://www.bloomberg.com/news/articles/2024-11-21/ai-drug-discovery-startup-enveda-raises-130-million">raised</a> a $130M Series C, led by Kinnevik and FPV Ventures.</p><p><strong>Cyera</strong>, an AI-powered data security platform, <a href="https://www.cyera.io/blog/a-big-step-forward-for-cyera-300m-series-d-to-drive-the-future-of-data-security">raised</a> a $300M Series D, led by Sapphire Ventures and Accel.</p><p><strong>Enfabrica</strong>, developing networking solutions for compute infrastructure, <a href="https://www.reuters.com/technology/artificial-intelligence/ai-startup-enfabrica-raises-115-million-plans-release-chip-next-year-2024-11-19/">raised</a> a $115M Series C, led by Spark Capital.</p><p><strong>Insider</strong>, the customer experience and engagement company, <a href="https://www.bloomberg.com/news/articles/2024-11-01/general-atlantic-leads-500-million-funding-for-startup-insider?srnd=phx-deals">raised</a> a $500M Series E, led by General Atlantic.</p><p><strong>Lightning</strong>, the AI development platform, <a href="https://finance.yahoo.com/news/lightning-ai-secures-50m-enhance-110658356.html">raised</a> a $50M funding round, led by Cisco Investments, JP Morgan, K5 Global, and NVIDIA.</p><p><strong>Moonvalley</strong>, building models for generative media, <a href="https://techcrunch.com/2024/11/18/moonvalley-wants-to-build-more-ethical-video-models/">raised</a> a $70M seed round, led by Khosla Ventures and General Catalyst.</p><p><strong>Physical Intelligence</strong>, developing foundational software for robotics, <a href="https://www.reuters.com/technology/artificial-intelligence/robot-ai-startup-physical-intelligence-raises-400-mln-bezos-openai-2024-11-04/">raised</a> a $400M Series A, led by Jeff Bezos, Lux, and Thrive Capital.</p><p><strong>Robin AI</strong>, the AI legal start-up, <a href="https://fortune.com/2024/11/12/legal-tech-robin-ai-raises-25-million-series-b-plus-llm/">raised</a> a $25M Series B extension, led by Willett Advisors, the University of Cambridge, and PayPal Ventures.</p><p><strong>Skydio</strong>, the drone company, <a href="https://techcrunch.com/2024/11/15/drone-manufacturer-skydio-raises-170-million-extension-round/">raised</a> a $170M Series E extension, led by Linse Capital.</p><p><strong>Cogna</strong>, the AI-powered SaaS platform, <a href="https://news.sky.com/story/ai-software-provider-cogna-lands-15m-from-top-investors-13252521">raised</a> a $15M Series A, led by Notion Capital.</p><p><strong>Tessl</strong>, an AI-native development platform, <a href="https://fortune.com/2024/11/14/tessl-funding-ai-software-development-platform/">raised</a> a $100M Series A, led by Index Ventures.</p><p><strong>Writer</strong>, a platform for building AI apps and workflows, <a href="https://techcrunch.com/2024/11/12/generative-ai-startup-writer-raises-200m-at-a-1-9b-valuation/">raised</a> a $200M Series C, co-led by Premji Invest, ICONIQ Growth, and Radical Ventures.</p><h4><em><strong>&#129309; </strong></em><strong>Exits</strong></h4><p><strong>Alpaca</strong>, the AI-powered canvas for creatives, was <a href="https://x.com/gmharhar/status/1856727540049801572?s=51&amp;t=8YCMEcmVVXRPm8SXTMgdlw">acquired</a> by Captions.</p><p><strong>Datavolo</strong>, the data management company, was <a href="https://techcrunch.com/2024/11/20/snowflake-snaps-up-data-management-company-datavolo/">acquired</a> by Snowflake.</p><p><strong>Dazz</strong>, the security and remediation company<strong>, </strong>was <a href="https://techcrunch.com/2024/11/21/wiz-acquires-dazz-for-450m-to-expand-its-cybersecurity-platform/">acquired</a> by Wiz for $450M.</p><p><strong>Hazy</strong>, the synthetic data company, was <a href="https://www.sas.com/en_gb/news/press-releases/2024/november/hazy-syntheticdata.html">acquired</a> by SAS.</p><div><hr></div><p>Signing off,&nbsp;</p><p>Nathan Benaich and Alex Chalmers on 1 December 2024</p><p><a href="https://www.airstreet.com/?utm_campaign=Your%20guide%20to%20AI&amp;utm_medium=email&amp;utm_source=Substack%20newsletter">Air Street Capital</a> | <a href="https://www.twitter.com/nathanbenaich?utm_campaign=Your%20guide%20to%20AI&amp;utm_medium=email&amp;utm_source=Substack%20newsletter">Twitter</a> | <a href="https://www.linkedin.com/in/nathanbenaich/?utm_campaign=Your%20guide%20to%20AI&amp;utm_medium=email&amp;utm_source=Substack%20newsletter">LinkedIn</a> | <a href="https://www.stateof.ai/?utm_campaign=Your%20guide%20to%20AI&amp;utm_medium=email&amp;utm_source=Substack%20newsletter">State of AI Report</a> | <a href="https://www.raais.co/?utm_campaign=Your%20guide%20to%20AI&amp;utm_medium=email&amp;utm_source=Substack%20newsletter">RAAIS</a> | <a href="https://airstreet.com/events">Events</a></p><p><em>Air Street Capital invests in AI-first entrepreneurs from the very beginning of your company-building journey.</em></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!NQRz!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd755c760-a8df-45f5-bdb5-3f7f8403b290_1980x1108.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!NQRz!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd755c760-a8df-45f5-bdb5-3f7f8403b290_1980x1108.png 424w, https://substackcdn.com/image/fetch/$s_!NQRz!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd755c760-a8df-45f5-bdb5-3f7f8403b290_1980x1108.png 848w, https://substackcdn.com/image/fetch/$s_!NQRz!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd755c760-a8df-45f5-bdb5-3f7f8403b290_1980x1108.png 1272w, https://substackcdn.com/image/fetch/$s_!NQRz!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd755c760-a8df-45f5-bdb5-3f7f8403b290_1980x1108.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!NQRz!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd755c760-a8df-45f5-bdb5-3f7f8403b290_1980x1108.png" width="1456" height="815" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d755c760-a8df-45f5-bdb5-3f7f8403b290_1980x1108.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:815,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:125362,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!NQRz!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd755c760-a8df-45f5-bdb5-3f7f8403b290_1980x1108.png 424w, https://substackcdn.com/image/fetch/$s_!NQRz!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd755c760-a8df-45f5-bdb5-3f7f8403b290_1980x1108.png 848w, https://substackcdn.com/image/fetch/$s_!NQRz!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd755c760-a8df-45f5-bdb5-3f7f8403b290_1980x1108.png 1272w, https://substackcdn.com/image/fetch/$s_!NQRz!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd755c760-a8df-45f5-bdb5-3f7f8403b290_1980x1108.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div>]]></content:encoded></item><item><title><![CDATA[State of AI outtakes #3: US politics]]></title><description><![CDATA[The State of AI Report aims to provide a comprehensive overview of everything you need to know across AI research, industry, politics, and safety.]]></description><link>https://press.airstreet.com/p/state-of-ai-outtakes-3-us-politics</link><guid isPermaLink="false">https://press.airstreet.com/p/state-of-ai-outtakes-3-us-politics</guid><dc:creator><![CDATA[Air Street Press]]></dc:creator><pubDate>Tue, 19 Nov 2024 14:21:22 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/e5b646b9-9281-474f-8765-f2e0ded7583c_2138x1198.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>The State of AI Report aims to provide a comprehensive overview of everything you need to know across AI research, industry, politics, and safety. To ensure the report remains at a manageable length, lots of material doesn&#8217;t make it into the final version. We&#8217;re bringing Air Street Press some of the research that didn&#8217;t make the original cut, along with our own reflections.</em></p><h3>Introduction</h3><p>Since Donald Trump secured his return to the White House a fortnight ago, we&#8217;ve been bombarded with questions about what this will mean for AI regulation. In the minds of certain commentators, when it comes to AI, there&#8217;s an almost cartoonish binary between a &#8216;responsible&#8217; outgoing administration and an &#8216;anti-safety&#8217; incoming one. We thought it might help to recap a few US political highlights from this year&#8217;s <a href="https://www.stateof.ai/">State of AI Report</a> to place our views in context.</p><h3>The twilight of AI safety?</h3><p>As we alluded to in this year&#8217;s report - the partisan divides aren&#8217;t as stark as you&#8217;d think.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!fOWu!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2fe430f2-0456-4bb8-a6f8-7ad9f330e303_960x540.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!fOWu!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2fe430f2-0456-4bb8-a6f8-7ad9f330e303_960x540.png 424w, https://substackcdn.com/image/fetch/$s_!fOWu!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2fe430f2-0456-4bb8-a6f8-7ad9f330e303_960x540.png 848w, https://substackcdn.com/image/fetch/$s_!fOWu!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2fe430f2-0456-4bb8-a6f8-7ad9f330e303_960x540.png 1272w, https://substackcdn.com/image/fetch/$s_!fOWu!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2fe430f2-0456-4bb8-a6f8-7ad9f330e303_960x540.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!fOWu!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2fe430f2-0456-4bb8-a6f8-7ad9f330e303_960x540.png" width="960" height="540" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/2fe430f2-0456-4bb8-a6f8-7ad9f330e303_960x540.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:540,&quot;width&quot;:960,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!fOWu!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2fe430f2-0456-4bb8-a6f8-7ad9f330e303_960x540.png 424w, https://substackcdn.com/image/fetch/$s_!fOWu!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2fe430f2-0456-4bb8-a6f8-7ad9f330e303_960x540.png 848w, https://substackcdn.com/image/fetch/$s_!fOWu!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2fe430f2-0456-4bb8-a6f8-7ad9f330e303_960x540.png 1272w, https://substackcdn.com/image/fetch/$s_!fOWu!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2fe430f2-0456-4bb8-a6f8-7ad9f330e303_960x540.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Underneath a lot of the rhetoric about fighting back against &#8216;woke AI&#8217;, we think  Republican deregulation isn&#8217;t likely to be hugely radical for a few reasons.&nbsp;</p><p>For a start, it&#8217;s quite hard to take a radical deregulatory approach at a federal level in the US, as there&#8217;s &#8230; not all that much to dismantle. The requirements in the White House Executive Order largely focus on notification and primarily catch a small handful of labs.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!sZfO!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd2fb399d-1d44-4ca6-8f54-d527a7d01801_960x540.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!sZfO!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd2fb399d-1d44-4ca6-8f54-d527a7d01801_960x540.png 424w, https://substackcdn.com/image/fetch/$s_!sZfO!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd2fb399d-1d44-4ca6-8f54-d527a7d01801_960x540.png 848w, https://substackcdn.com/image/fetch/$s_!sZfO!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd2fb399d-1d44-4ca6-8f54-d527a7d01801_960x540.png 1272w, https://substackcdn.com/image/fetch/$s_!sZfO!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd2fb399d-1d44-4ca6-8f54-d527a7d01801_960x540.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!sZfO!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd2fb399d-1d44-4ca6-8f54-d527a7d01801_960x540.png" width="960" height="540" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d2fb399d-1d44-4ca6-8f54-d527a7d01801_960x540.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:540,&quot;width&quot;:960,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!sZfO!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd2fb399d-1d44-4ca6-8f54-d527a7d01801_960x540.png 424w, https://substackcdn.com/image/fetch/$s_!sZfO!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd2fb399d-1d44-4ca6-8f54-d527a7d01801_960x540.png 848w, https://substackcdn.com/image/fetch/$s_!sZfO!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd2fb399d-1d44-4ca6-8f54-d527a7d01801_960x540.png 1272w, https://substackcdn.com/image/fetch/$s_!sZfO!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd2fb399d-1d44-4ca6-8f54-d527a7d01801_960x540.png 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The areas of Biden Administration policymaking that the Republicans seem most focused on are probably the ones that will make the least difference to AI labs. For example, Republican Senator Ted Cruz&#8217;s <a href="https://www.nextgov.com/artificial-intelligence/2024/07/party-line-differences-emerge-over-ai-oversight-international-partnerships/398482/">&#8220;no woke AI&#8221;</a> amendment would ban the government promotion of &#8216;equitable&#8217; AI development, but probably have next to no impact on industry at all.&nbsp;</p><p>Meanwhile, there&#8217;s little evidence that industry is pushing for the US AI Safety Institute to be dismantled. Sam Altman voluntarily announced that the US AISI would receive an early preview of major AI releases, while Anthropic and Google DeepMind have happily worked with the institute&#8217;s UK counterpart.&nbsp;</p><p>This is partly out of a sincere concern for safety. But it also makes commercial sense.&nbsp;</p><p>If you believe that you&#8217;re at the cutting edge of the most transformative technology the world has seen in decades, you will be aware of the acute political sensitivities. It makes a lot of sense to dip governments&#8217; hands in the blood. If your country&#8217;s AI Safety Institute tested something before deployment and were satisfied, if a bad actor misuses your work, it&#8217;s harder to justify bringing the full force of the regulatory apparatus down on you.</p><p>It&#8217;s part of the reason that AI companies have been on a mission to make nice with the national security establishment. Whether it&#8217;s Meta <a href="https://www.theguardian.com/technology/2024/nov/05/meta-allows-national-security-defense-contractors-use-llama-ai">changing</a> Llama&#8217;s terms of service to allow defense use, Anthropic <a href="https://investors.palantir.com/news-details/2024/Anthropic-and-Palantir-Partner-to-Bring-Claude-AI-Models-to-AWS-for-U.S.-Government-Intelligence-and-Defense-Operations/">teaming up</a> with Palantir, or OpenAI appointing a <a href="https://openai.com/index/openai-appoints-retired-us-army-general/">former NSA director to its Board</a> - these companies want to be <em>closer</em> to the government.&nbsp;</p><p>It&#8217;s also worth remembering that while potential senior Trump advisor Elon Musk is tough on &#8216;woke&#8217; - he&#8217;s been a consistent AI doomer. He was an early financial backer of DeepMind and OpenAI, because he worried about the technology&#8217;s potential to end the world, and he supported SB 1047 in California. He&#8217;s unlikely to want zero oversight. Meanwhile, Republicans have emerged as some of the most aggressive advocates of big tech regulation over the past few years, out of concerns around political bias.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://press.airstreet.com/subscribe?"><span>Subscribe now</span></a></p><h3>Going local</h3><p>Nevertheless, the new administration means significant regulation at the federal level is unlikely &#8230; but that was already the case. The <a href="https://www.hklaw.com/en/insights/publications/2024/05/senate-releases-bipartisan-ai-roadmap">bipartisan Senate AI policy roadmap</a> contains only narrow, loosely-worded ideas around regulation (e.g. &#8220;considering&#8221; a potential ban on social scoring or &#8220;considering&#8221; measures around transparency in AI healthcare provision).&nbsp;</p><p>Where we may see more change - and this is speculation at this stage - is at the state-level.&nbsp;</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Q1QZ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0670334c-d0e8-4df6-a674-d6a97ffb6ff1_960x540.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Q1QZ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0670334c-d0e8-4df6-a674-d6a97ffb6ff1_960x540.png 424w, https://substackcdn.com/image/fetch/$s_!Q1QZ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0670334c-d0e8-4df6-a674-d6a97ffb6ff1_960x540.png 848w, https://substackcdn.com/image/fetch/$s_!Q1QZ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0670334c-d0e8-4df6-a674-d6a97ffb6ff1_960x540.png 1272w, https://substackcdn.com/image/fetch/$s_!Q1QZ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0670334c-d0e8-4df6-a674-d6a97ffb6ff1_960x540.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Q1QZ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0670334c-d0e8-4df6-a674-d6a97ffb6ff1_960x540.png" width="960" height="540" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/0670334c-d0e8-4df6-a674-d6a97ffb6ff1_960x540.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:540,&quot;width&quot;:960,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Q1QZ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0670334c-d0e8-4df6-a674-d6a97ffb6ff1_960x540.png 424w, https://substackcdn.com/image/fetch/$s_!Q1QZ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0670334c-d0e8-4df6-a674-d6a97ffb6ff1_960x540.png 848w, https://substackcdn.com/image/fetch/$s_!Q1QZ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0670334c-d0e8-4df6-a674-d6a97ffb6ff1_960x540.png 1272w, https://substackcdn.com/image/fetch/$s_!Q1QZ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0670334c-d0e8-4df6-a674-d6a97ffb6ff1_960x540.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>AI-related laws are on the statute book or being debated <a href="https://www.multistate.ai/artificial-intelligence-ai-legislation">in almost every state</a> right now, as tracked by Multistate. These range from relatively narrow laws prohibiting non-consensual sexual deepfake images and videos (e.g. in Idaho) through to more sweeping legislation around algorithmic bias and transparency. Colorado is a good example of the latter, and the governor only signed the proposals into law after <a href="https://www.multistate.ai/ai-policy-overview-colorado">publicly expressing</a> his reservations.&nbsp;</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!CNsp!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F436217dc-1e9d-4570-a44c-56feefa18724_2428x1198.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!CNsp!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F436217dc-1e9d-4570-a44c-56feefa18724_2428x1198.png 424w, https://substackcdn.com/image/fetch/$s_!CNsp!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F436217dc-1e9d-4570-a44c-56feefa18724_2428x1198.png 848w, https://substackcdn.com/image/fetch/$s_!CNsp!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F436217dc-1e9d-4570-a44c-56feefa18724_2428x1198.png 1272w, https://substackcdn.com/image/fetch/$s_!CNsp!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F436217dc-1e9d-4570-a44c-56feefa18724_2428x1198.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!CNsp!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F436217dc-1e9d-4570-a44c-56feefa18724_2428x1198.png" width="1456" height="718" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/436217dc-1e9d-4570-a44c-56feefa18724_2428x1198.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:718,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:328947,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!CNsp!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F436217dc-1e9d-4570-a44c-56feefa18724_2428x1198.png 424w, https://substackcdn.com/image/fetch/$s_!CNsp!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F436217dc-1e9d-4570-a44c-56feefa18724_2428x1198.png 848w, https://substackcdn.com/image/fetch/$s_!CNsp!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F436217dc-1e9d-4570-a44c-56feefa18724_2428x1198.png 1272w, https://substackcdn.com/image/fetch/$s_!CNsp!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F436217dc-1e9d-4570-a44c-56feefa18724_2428x1198.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>It would not surprise us if perceived federal de-regulation led to more ambitious regulation by Democratic states. The rightness or wrongness of individual state regulations aside, business is rarely happy when it has to navigate vastly different regulatory regimes in different regions of the same country.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/p/state-of-ai-outtakes-3-us-politics?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://press.airstreet.com/p/state-of-ai-outtakes-3-us-politics?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><h3>Beyond regulation</h3><p>Domestic AI regulation aside, there are a number of other areas where we are likely to see some change.</p><p>While the CHIPS Act - the massive US attempt to onshore some of the semiconductor supply chain - is beginning to gain some traction, it always had its fair share of critics. Considering Trump&#8217;s protectionist instincts, we imagine it's the social provisions in the legislation that <a href="https://www.bloomberg.com/news/articles/2024-11-08/trump-s-win-sets-off-race-to-complete-chips-act-subsidy-deals">are likely to come</a> under attack (e.g. labor union consultations, environmental impact measures etc.) rather than the principle of the Bill.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!5yGh!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5fd047b0-775f-438c-a917-dd4631943107_960x540.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!5yGh!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5fd047b0-775f-438c-a917-dd4631943107_960x540.png 424w, https://substackcdn.com/image/fetch/$s_!5yGh!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5fd047b0-775f-438c-a917-dd4631943107_960x540.png 848w, https://substackcdn.com/image/fetch/$s_!5yGh!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5fd047b0-775f-438c-a917-dd4631943107_960x540.png 1272w, https://substackcdn.com/image/fetch/$s_!5yGh!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5fd047b0-775f-438c-a917-dd4631943107_960x540.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!5yGh!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5fd047b0-775f-438c-a917-dd4631943107_960x540.png" width="960" height="540" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/5fd047b0-775f-438c-a917-dd4631943107_960x540.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:540,&quot;width&quot;:960,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!5yGh!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5fd047b0-775f-438c-a917-dd4631943107_960x540.png 424w, https://substackcdn.com/image/fetch/$s_!5yGh!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5fd047b0-775f-438c-a917-dd4631943107_960x540.png 848w, https://substackcdn.com/image/fetch/$s_!5yGh!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5fd047b0-775f-438c-a917-dd4631943107_960x540.png 1272w, https://substackcdn.com/image/fetch/$s_!5yGh!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5fd047b0-775f-438c-a917-dd4631943107_960x540.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Potentially more interesting is the question of European decoupling.&nbsp;</p><p>In the report, we looked at how US companies were beginning to collide with European norms.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!wvts!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F113a4833-f82d-454e-99d9-6373ef8f695c_960x540.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!wvts!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F113a4833-f82d-454e-99d9-6373ef8f695c_960x540.png 424w, https://substackcdn.com/image/fetch/$s_!wvts!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F113a4833-f82d-454e-99d9-6373ef8f695c_960x540.png 848w, https://substackcdn.com/image/fetch/$s_!wvts!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F113a4833-f82d-454e-99d9-6373ef8f695c_960x540.png 1272w, https://substackcdn.com/image/fetch/$s_!wvts!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F113a4833-f82d-454e-99d9-6373ef8f695c_960x540.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!wvts!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F113a4833-f82d-454e-99d9-6373ef8f695c_960x540.png" width="960" height="540" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/113a4833-f82d-454e-99d9-6373ef8f695c_960x540.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:540,&quot;width&quot;:960,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!wvts!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F113a4833-f82d-454e-99d9-6373ef8f695c_960x540.png 424w, https://substackcdn.com/image/fetch/$s_!wvts!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F113a4833-f82d-454e-99d9-6373ef8f695c_960x540.png 848w, https://substackcdn.com/image/fetch/$s_!wvts!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F113a4833-f82d-454e-99d9-6373ef8f695c_960x540.png 1272w, https://substackcdn.com/image/fetch/$s_!wvts!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F113a4833-f82d-454e-99d9-6373ef8f695c_960x540.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>And this was causing product launches to either be canceled or delayed while local adaptations were made.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!s9C3!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb4ad43f-1fb0-4bcf-8836-2e6586ffb7e1_960x540.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!s9C3!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb4ad43f-1fb0-4bcf-8836-2e6586ffb7e1_960x540.png 424w, https://substackcdn.com/image/fetch/$s_!s9C3!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb4ad43f-1fb0-4bcf-8836-2e6586ffb7e1_960x540.png 848w, https://substackcdn.com/image/fetch/$s_!s9C3!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb4ad43f-1fb0-4bcf-8836-2e6586ffb7e1_960x540.png 1272w, https://substackcdn.com/image/fetch/$s_!s9C3!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb4ad43f-1fb0-4bcf-8836-2e6586ffb7e1_960x540.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!s9C3!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb4ad43f-1fb0-4bcf-8836-2e6586ffb7e1_960x540.png" width="960" height="540" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/fb4ad43f-1fb0-4bcf-8836-2e6586ffb7e1_960x540.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:540,&quot;width&quot;:960,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!s9C3!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb4ad43f-1fb0-4bcf-8836-2e6586ffb7e1_960x540.png 424w, https://substackcdn.com/image/fetch/$s_!s9C3!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb4ad43f-1fb0-4bcf-8836-2e6586ffb7e1_960x540.png 848w, https://substackcdn.com/image/fetch/$s_!s9C3!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb4ad43f-1fb0-4bcf-8836-2e6586ffb7e1_960x540.png 1272w, https://substackcdn.com/image/fetch/$s_!s9C3!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb4ad43f-1fb0-4bcf-8836-2e6586ffb7e1_960x540.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Considering the new administration&#8217;s more confrontational attitude on trade, will there be a penalty attached if the UK or EU decides to regulate big US companies in a heavy-handed way? If X is fined for breach of the Digital Services Act for its content moderation practices, there&#8217;s speculation that this <a href="https://www.politico.eu/article/donald-trump-elon-musk-x-tech-social-media-politics-elections-eu/">could result in a diplomatic spat</a>. This is unknowable at this stage, and even if it were, how much credibility would a political bloc be able to maintain in the eyes of the world if it began softening its approach to legislation or enforcement pre-emptively?&nbsp;</p><h3>Closing thoughts</h3><p>We&#8217;ve tried to avoid the kinds of overly-confident punditry that characterizes post-election speculation. But we believe that, considering its importance, it&#8217;s striking just how un-politicised many crucial questions around AI remain. Considering the personalities involved in the next administration, it may not stay that way for very long.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://press.airstreet.com/subscribe?"><span>Subscribe now</span></a></p>]]></content:encoded></item><item><title><![CDATA[State of AI outtakes #2: embodied AI ]]></title><description><![CDATA[The State of AI Report aims to provide a comprehensive overview of everything you need to know across AI research, industry, politics, and safety.]]></description><link>https://press.airstreet.com/p/state-of-ai-outtakes-2-embodied-ai</link><guid isPermaLink="false">https://press.airstreet.com/p/state-of-ai-outtakes-2-embodied-ai</guid><dc:creator><![CDATA[Air Street Press]]></dc:creator><pubDate>Thu, 07 Nov 2024 12:27:46 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/780d13a4-68a8-40db-9790-5ad293969bf5_2144x1196.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>The State of AI Report aims to provide a comprehensive overview of everything you need to know across AI research, industry, politics, and safety. To ensure the report remains at a manageable length, lots of material doesn&#8217;t make it into the final version. We&#8217;re bringing Air Street Press some of the research that didn&#8217;t make the original cut, along with our own reflections.</em></p><h3>Introduction </h3><p>This year&#8217;s <a href="https://www.stateof.ai/">State of AI Report</a> contained our largest ever embodied AI section, covering everything from diffusion models through to open source robotics libraries, humanoids, and self-driving. Until relatively recently, with the partial exception of self-driving, embodied AI was the unloved cousin of AI research. </p><p>Work in this field, which concerns the symbiosis of software and hardware to enable AI to run onboard of physical machines and control their function,&nbsp; was concentrated in a small, dedicated community. It was rarely more than a side-project for the biggest labs. But a combination of dedicated entrepreneurship and progress spurred by the creative application of foundation models has spurred a renaissance.</p><h3>Robotics: a GPT moment?</h3><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!6MEl!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ed6a126-6cb6-4659-a4f3-4faac415323e_960x540.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!6MEl!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ed6a126-6cb6-4659-a4f3-4faac415323e_960x540.png 424w, https://substackcdn.com/image/fetch/$s_!6MEl!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ed6a126-6cb6-4659-a4f3-4faac415323e_960x540.png 848w, https://substackcdn.com/image/fetch/$s_!6MEl!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ed6a126-6cb6-4659-a4f3-4faac415323e_960x540.png 1272w, https://substackcdn.com/image/fetch/$s_!6MEl!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ed6a126-6cb6-4659-a4f3-4faac415323e_960x540.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!6MEl!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ed6a126-6cb6-4659-a4f3-4faac415323e_960x540.png" width="960" height="540" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/7ed6a126-6cb6-4659-a4f3-4faac415323e_960x540.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:540,&quot;width&quot;:960,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!6MEl!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ed6a126-6cb6-4659-a4f3-4faac415323e_960x540.png 424w, https://substackcdn.com/image/fetch/$s_!6MEl!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ed6a126-6cb6-4659-a4f3-4faac415323e_960x540.png 848w, https://substackcdn.com/image/fetch/$s_!6MEl!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ed6a126-6cb6-4659-a4f3-4faac415323e_960x540.png 1272w, https://substackcdn.com/image/fetch/$s_!6MEl!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ed6a126-6cb6-4659-a4f3-4faac415323e_960x540.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>One of the biggest stories of the year is the transformation of robotics by foundation models. LLMs are replacing hand-coded robot behaviors with natural language commands that robots can flexibly execute, while VLMs enable powerful scene understanding that unlocks rapid learning of new behaviors. Together, they can be used to generate diverse training scenarios and allow robots to generalize across tasks without custom policies and reward functions for every situation.</p><p>By making it possible to control robots via natural language, rather than via cumbersome proprietary software and abstruse programming languages, LLMs have significantly improved the accessibility of robotics. For example, our friends at <a href="https://sereact.ai/">Sereact</a> were the first to combine zero-shot visual reasoning with natural language instructions to create <a href="https://sereact.ai/products">PickGPT</a>, software for human operators to interact with robotic arms in warehouse settings.</p><p>In our report, we&#8217;ve noted the steady drumbeat of foundation models, frameworks and datasets emerging from Google DeepMind and other labs. Now all believers in the robotics future, they&#8217;re keen to occupy as much of the stack as they possibly can without building the hardware.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!YbD9!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F886b0ab5-821e-44c2-9cbe-5925f8dd465b_960x540.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!YbD9!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F886b0ab5-821e-44c2-9cbe-5925f8dd465b_960x540.png 424w, https://substackcdn.com/image/fetch/$s_!YbD9!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F886b0ab5-821e-44c2-9cbe-5925f8dd465b_960x540.png 848w, https://substackcdn.com/image/fetch/$s_!YbD9!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F886b0ab5-821e-44c2-9cbe-5925f8dd465b_960x540.png 1272w, https://substackcdn.com/image/fetch/$s_!YbD9!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F886b0ab5-821e-44c2-9cbe-5925f8dd465b_960x540.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!YbD9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F886b0ab5-821e-44c2-9cbe-5925f8dd465b_960x540.png" width="960" height="540" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/886b0ab5-821e-44c2-9cbe-5925f8dd465b_960x540.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:540,&quot;width&quot;:960,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!YbD9!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F886b0ab5-821e-44c2-9cbe-5925f8dd465b_960x540.png 424w, https://substackcdn.com/image/fetch/$s_!YbD9!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F886b0ab5-821e-44c2-9cbe-5925f8dd465b_960x540.png 848w, https://substackcdn.com/image/fetch/$s_!YbD9!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F886b0ab5-821e-44c2-9cbe-5925f8dd465b_960x540.png 1272w, https://substackcdn.com/image/fetch/$s_!YbD9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F886b0ab5-821e-44c2-9cbe-5925f8dd465b_960x540.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Back in February, NVIDIA established its GEAR (Generalist Embodied Agent Research) group&nbsp; under Jim Fan and Yuke Zhu. The two had previously cooperated on related projects, most notably <a href="https://voyager.minedojo.org/">Voyager</a>, an LLM-powered agent in Minecraft. GEAR research includes humanoid foundation model <a href="https://developer.nvidia.com/project-gr00t">Project GR00T</a> and imitation learning method <a href="https://mimic-play.github.io/">MimicPlay</a>.&nbsp;</p><p>NVIDIA is already working with a bunch of humanoid start-ups, including 1X Technologies, Agility Robotics, Apptronik, Boston Dynamics, Figure AI, Fourier Intelligence, Sanctuary AI, Unitree Robotics, and XPENG Robotics. This is a classic NVIDIA move and one they <a href="https://www.nvidia.com/en-gb/self-driving-cars/partners/">used with self-driving</a>: work every potential partner, regardless of their success prospects, using the learnings to build your own autonomy stack.</p><p>Meta has also been making moves in this space. Back in January, they released <a href="https://ok-robot.github.io/">OK-Robot</a>, an open framework that enables robots to perform pick-and-place tasks in new and unstructured environments, such as homes, without needing prior training.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!cmJR!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1d25be83-2c34-412f-b2f7-bd7188e8b91d_677x470.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!cmJR!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1d25be83-2c34-412f-b2f7-bd7188e8b91d_677x470.png 424w, https://substackcdn.com/image/fetch/$s_!cmJR!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1d25be83-2c34-412f-b2f7-bd7188e8b91d_677x470.png 848w, https://substackcdn.com/image/fetch/$s_!cmJR!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1d25be83-2c34-412f-b2f7-bd7188e8b91d_677x470.png 1272w, https://substackcdn.com/image/fetch/$s_!cmJR!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1d25be83-2c34-412f-b2f7-bd7188e8b91d_677x470.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!cmJR!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1d25be83-2c34-412f-b2f7-bd7188e8b91d_677x470.png" width="677" height="470" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1d25be83-2c34-412f-b2f7-bd7188e8b91d_677x470.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:470,&quot;width&quot;:677,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!cmJR!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1d25be83-2c34-412f-b2f7-bd7188e8b91d_677x470.png 424w, https://substackcdn.com/image/fetch/$s_!cmJR!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1d25be83-2c34-412f-b2f7-bd7188e8b91d_677x470.png 848w, https://substackcdn.com/image/fetch/$s_!cmJR!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1d25be83-2c34-412f-b2f7-bd7188e8b91d_677x470.png 1272w, https://substackcdn.com/image/fetch/$s_!cmJR!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1d25be83-2c34-412f-b2f7-bd7188e8b91d_677x470.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>In our most recent edition of <a href="https://press.airstreet.com/p/your-guide-to-ai-november-2024">Guide to AI</a>, we covered a major drop of <a href="https://ai.meta.com/blog/fair-robotics-open-source/">Meta research</a> that focused on touch perception, dexterity, and human-robot interaction in a clear pitch for the humanoid marketplace.&nbsp;</p><p>Humanoids are also the center of OpenAI&#8217;s robotics efforts. The company has invested in Figure AI, 1X Technologies, and Physical Intelligence and its new robotics team <a href="https://www.forbes.com/sites/kenrickcai/2024/05/30/openai-robotics-team/">is collaborating</a> with external partners rather than attempting to compete against them.</p><p>So what does the humanoid market opportunity look like?</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!2RB3!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe6c2799-ed3e-4f68-a221-47ba312a9db4_960x540.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!2RB3!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe6c2799-ed3e-4f68-a221-47ba312a9db4_960x540.png 424w, https://substackcdn.com/image/fetch/$s_!2RB3!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe6c2799-ed3e-4f68-a221-47ba312a9db4_960x540.png 848w, https://substackcdn.com/image/fetch/$s_!2RB3!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe6c2799-ed3e-4f68-a221-47ba312a9db4_960x540.png 1272w, https://substackcdn.com/image/fetch/$s_!2RB3!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe6c2799-ed3e-4f68-a221-47ba312a9db4_960x540.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!2RB3!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe6c2799-ed3e-4f68-a221-47ba312a9db4_960x540.png" width="960" height="540" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/be6c2799-ed3e-4f68-a221-47ba312a9db4_960x540.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:540,&quot;width&quot;:960,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!2RB3!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe6c2799-ed3e-4f68-a221-47ba312a9db4_960x540.png 424w, https://substackcdn.com/image/fetch/$s_!2RB3!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe6c2799-ed3e-4f68-a221-47ba312a9db4_960x540.png 848w, https://substackcdn.com/image/fetch/$s_!2RB3!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe6c2799-ed3e-4f68-a221-47ba312a9db4_960x540.png 1272w, https://substackcdn.com/image/fetch/$s_!2RB3!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe6c2799-ed3e-4f68-a221-47ba312a9db4_960x540.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>As the heading for this slide suggests - we&#8217;re not entirely convinced. While companies routinely release slick-looking demo videos, actual real-world performance is lackluster. We were underwhelmed by the videos of the beer-serving Optimus at Tesla&#8217;s autonomy event which, as is often the case with humanoids, was <a href="https://jalopnik.com/teslas-beer-serving-optimus-robot-was-controlled-by-a-h-1851670923#:~:text=Robert%20Scoble%2C%20an%20AI%20enthusiast,anything%20autonomously%20at%20the%20party.">not actually autonomous</a>.&nbsp;</p><p>The promise of humanoids boils down to significant upfront capex, in exchange for something that works slower and less reliably than a human, while requiring maintenance and a cloud connection. The bet is that, in the end, the per-hour cost will work out more cheaply than human labor.&nbsp;</p><p>This may well end up working out for a handful of large companies prepared to make the investment, but does it have the potential to disrupt the more conventional industrial robotics market? Systems that combine off-the-shelf components with AI-powered software can already serve beer and fulfill most warehouse functions efficiently and cheaply.&nbsp;</p><p>Why the self-driving parallel?&nbsp;</p><h3>Self-driving: the struggle to scale</h3><p>In past instalments of the report, we covered how self-driving car companies would make extravagant promises about performance and testing, only then to fall short. This is beginning to change.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!XgVh!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F62126156-4e21-43c9-bebe-7411107e4ce9_960x540.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!XgVh!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F62126156-4e21-43c9-bebe-7411107e4ce9_960x540.png 424w, https://substackcdn.com/image/fetch/$s_!XgVh!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F62126156-4e21-43c9-bebe-7411107e4ce9_960x540.png 848w, https://substackcdn.com/image/fetch/$s_!XgVh!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F62126156-4e21-43c9-bebe-7411107e4ce9_960x540.png 1272w, https://substackcdn.com/image/fetch/$s_!XgVh!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F62126156-4e21-43c9-bebe-7411107e4ce9_960x540.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!XgVh!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F62126156-4e21-43c9-bebe-7411107e4ce9_960x540.png" width="960" height="540" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/62126156-4e21-43c9-bebe-7411107e4ce9_960x540.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:540,&quot;width&quot;:960,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!XgVh!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F62126156-4e21-43c9-bebe-7411107e4ce9_960x540.png 424w, https://substackcdn.com/image/fetch/$s_!XgVh!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F62126156-4e21-43c9-bebe-7411107e4ce9_960x540.png 848w, https://substackcdn.com/image/fetch/$s_!XgVh!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F62126156-4e21-43c9-bebe-7411107e4ce9_960x540.png 1272w, https://substackcdn.com/image/fetch/$s_!XgVh!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F62126156-4e21-43c9-bebe-7411107e4ce9_960x540.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Waymo is now hitting its stride, but it took fifteen years, over ten billion dollars of investment, and some luck to get there. Cruise and Uber both suffered <a href="https://fortune.com/2024/05/14/gm-cruise-settlement-pedestrian-accident-san-francisco/">serious</a> <a href="https://www.bbc.co.uk/news/technology-54175359">setbacks</a> after pedestrian accidents, while Apple abandoned its self-driving efforts <a href="https://www.bloomberg.com/news/features/2024-03-06/apple-car-s-crash-design-details-tim-cook-s-indecision-failed-tesla-deal">after $1B in investment</a>. The physical world is hard and we suspect humanoid builders will discover this pretty quickly.&nbsp;</p><p>Our report also covered Wayve&#8217;s growing traction. The company&#8217;s journey from a &#163;2.5M seed round in 2017 to a $1.05 billion Series C is a case study in both courage and capital efficiency. A fortnight after we published the report, Wayve <a href="https://wayve.ai/press/us-expansion/">announced</a> that it had opened a new hub in San Francisco to accelerate its testing and find local partners.</p><p>The Wayve team will have been cheered by their competitors over at Waymo&#8217;s unveiling of <a href="https://arxiv.org/abs/2410.23262">EMMA</a>, an end-to-end model for self-driving. Wayve were the original pioneers in end-to-end, which uses raw sensor data (like camera inputs) directly to control outputs.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Ymio!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd91018bf-e914-4164-b1e8-c409affe7545_1098x519.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Ymio!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd91018bf-e914-4164-b1e8-c409affe7545_1098x519.png 424w, https://substackcdn.com/image/fetch/$s_!Ymio!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd91018bf-e914-4164-b1e8-c409affe7545_1098x519.png 848w, https://substackcdn.com/image/fetch/$s_!Ymio!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd91018bf-e914-4164-b1e8-c409affe7545_1098x519.png 1272w, https://substackcdn.com/image/fetch/$s_!Ymio!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd91018bf-e914-4164-b1e8-c409affe7545_1098x519.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Ymio!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd91018bf-e914-4164-b1e8-c409affe7545_1098x519.png" width="1098" height="519" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d91018bf-e914-4164-b1e8-c409affe7545_1098x519.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:519,&quot;width&quot;:1098,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Ymio!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd91018bf-e914-4164-b1e8-c409affe7545_1098x519.png 424w, https://substackcdn.com/image/fetch/$s_!Ymio!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd91018bf-e914-4164-b1e8-c409affe7545_1098x519.png 848w, https://substackcdn.com/image/fetch/$s_!Ymio!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd91018bf-e914-4164-b1e8-c409affe7545_1098x519.png 1272w, https://substackcdn.com/image/fetch/$s_!Ymio!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd91018bf-e914-4164-b1e8-c409affe7545_1098x519.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Image credit - <a href="https://wayve.ai/thinking/e2e-embodied-ai-solves-the-long-tail/">Wayve</a></figcaption></figure></div><p>This has the advantage of simplicity, as it removes the possibility of errors building up in separate modules, and adaptability. Historically, most self-driving companies shunned end-to-end, believing it to be too data intensive and uninterpretable.&nbsp;</p><p>EMMA, built using Gemini, represents all inputs and outputs (such as navigation instructions, vehicle states, and planned trajectories) as natural language text. This allows EMMA to apply language processing to tasks traditionally segmented into modules, like perception, motion planning, and road graph estimation, effectively merging these functions. It also solves the interpretability issue, with driving rationale being expressed in natural language.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!9P4I!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F677682b7-46e2-47e6-990a-2cd34bcb797a_1440x810.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!9P4I!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F677682b7-46e2-47e6-990a-2cd34bcb797a_1440x810.png 424w, https://substackcdn.com/image/fetch/$s_!9P4I!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F677682b7-46e2-47e6-990a-2cd34bcb797a_1440x810.png 848w, https://substackcdn.com/image/fetch/$s_!9P4I!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F677682b7-46e2-47e6-990a-2cd34bcb797a_1440x810.png 1272w, https://substackcdn.com/image/fetch/$s_!9P4I!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F677682b7-46e2-47e6-990a-2cd34bcb797a_1440x810.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!9P4I!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F677682b7-46e2-47e6-990a-2cd34bcb797a_1440x810.png" width="1440" height="810" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/677682b7-46e2-47e6-990a-2cd34bcb797a_1440x810.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:810,&quot;width&quot;:1440,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!9P4I!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F677682b7-46e2-47e6-990a-2cd34bcb797a_1440x810.png 424w, https://substackcdn.com/image/fetch/$s_!9P4I!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F677682b7-46e2-47e6-990a-2cd34bcb797a_1440x810.png 848w, https://substackcdn.com/image/fetch/$s_!9P4I!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F677682b7-46e2-47e6-990a-2cd34bcb797a_1440x810.png 1272w, https://substackcdn.com/image/fetch/$s_!9P4I!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F677682b7-46e2-47e6-990a-2cd34bcb797a_1440x810.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>This tallies with earlier work from Wayve and serves as a reminder of how foundation models can strengthen generalization, robustness, and data efficiency outside the domains we normally associated with NLP.&nbsp;</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!zR4G!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff89af226-c593-4074-bdb1-ec244be5201f_960x540.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!zR4G!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff89af226-c593-4074-bdb1-ec244be5201f_960x540.png 424w, https://substackcdn.com/image/fetch/$s_!zR4G!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff89af226-c593-4074-bdb1-ec244be5201f_960x540.png 848w, https://substackcdn.com/image/fetch/$s_!zR4G!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff89af226-c593-4074-bdb1-ec244be5201f_960x540.png 1272w, https://substackcdn.com/image/fetch/$s_!zR4G!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff89af226-c593-4074-bdb1-ec244be5201f_960x540.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!zR4G!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff89af226-c593-4074-bdb1-ec244be5201f_960x540.png" width="960" height="540" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f89af226-c593-4074-bdb1-ec244be5201f_960x540.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:540,&quot;width&quot;:960,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!zR4G!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff89af226-c593-4074-bdb1-ec244be5201f_960x540.png 424w, https://substackcdn.com/image/fetch/$s_!zR4G!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff89af226-c593-4074-bdb1-ec244be5201f_960x540.png 848w, https://substackcdn.com/image/fetch/$s_!zR4G!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff89af226-c593-4074-bdb1-ec244be5201f_960x540.png 1272w, https://substackcdn.com/image/fetch/$s_!zR4G!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff89af226-c593-4074-bdb1-ec244be5201f_960x540.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h3>Closing thoughts </h3><p>From Nathan&#8217;s personal support for Wayve in its founding days through to Air Street&#8217;s more recent investment in Sereact, we&#8217;ve long seen the opportunity in embodied AI. While AI-first approaches are already having a significant impact on the domain of data, knowledge, and language, their potential physical impact remains largely untapped. We would be very surprised if this does not become a significant theme of our 2025 report.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://press.airstreet.com/subscribe?"><span>Subscribe now</span></a></p><p></p>]]></content:encoded></item><item><title><![CDATA[State of AI: November 2024 newsletter]]></title><description><![CDATA[Welcome to the latest issue of your guide to AI, an editorialized newsletter covering the key developments in AI policy, research, industry, and start-ups over the last month.]]></description><link>https://press.airstreet.com/p/the-state-of-ai-2024-november</link><guid isPermaLink="false">https://press.airstreet.com/p/the-state-of-ai-2024-november</guid><dc:creator><![CDATA[Air Street Press]]></dc:creator><pubDate>Sun, 03 Nov 2024 18:21:54 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/65c2e364-eeb0-4d69-831b-41d134d2c579_1864x1038.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Hi everyone!</p><p>Welcome to the latest issue of your guide to AI, an editorialized newsletter covering the key developments in AI policy, research, industry, and start-ups over the last month. First up, a few updates:</p><ul><li><p>Thank you to everyone who&#8217;s read, shared, and sent us feedback on this year&#8217;s <strong><a href="http://stateof.ai">State of AI Report</a></strong>. We were overwhelmed by both the volume and thoughtfulness of the reaction. Ultimately, the report is only possible because of the community and we&#8217;re very appreciative.&nbsp;</p></li><li><p>Our publishing schedule has continued unabated, with <a href="https://press.airstreet.com/p/ai-isnt-the-dotcom-bubble">a new State of AI outtakes series</a>, an evaluation of the state of play in AI-first bio, and our take on why <a href="https://press.airstreet.com/p/ai-isnt-the-dotcom-bubble">AI isn&#8217;t like the dotcom bubble</a>.&nbsp;</p></li><li><p>It was great to meet so many of our readers at our <strong>SF launch event</strong> and <strong>NYC happy hour</strong>. We&#8217;ll be in <strong><a href="https://lu.ma/londonai">London</a></strong> and <strong><a href="https://lu.ma/parisai">Paris</a></strong> before the end of the year and you can subscribe to <a href="https://lu.ma/airstreet">our events page</a> to ensure you don&#8217;t miss any future meet-ups.</p></li><li><p>Along with events, you can get all of our news, analysis, and events directly in your inbox if you subscribe to <strong><a href="http://press.airstreet.com">Air Street Press</a></strong>.&nbsp;</p></li></ul><p>We love hearing what you&#8217;re up to and what&#8217;s on your mind, just hit reply or forward to your friends :-)</p><div><hr></div><h3><em><strong>&#127758; The (geo)politics of AI</strong></em></h3><p>The White House <a href="https://www.whitehouse.gov/briefing-room/statements-releases/2024/10/24/fact-sheet-biden-harris-administration-outlines-coordinated-approach-to-harness-power-of-ai-for-u-s-national-security/">has issued</a> the first ever National Security Memorandum on AI. This directs government agencies to work on the diversification of the chip supply chain, streamline procurement processes, and to collaborate with international partners. It also doubles down on the AI Research Resource and gives room to bring in more skilled workers from abroad. While none of the memo is revolutionary, it is a reminder that the US still sees AI as a crucial geopolitical battlefield, despite the relative lack of action in the chip wars over the past few months.&nbsp;</p><p>No doubt the US security establishment sat bolt upright in response to revelations that the Chinese People&#8217;s Liberation Army <a href="https://www.reuters.com/technology/artificial-intelligence/chinese-researchers-develop-ai-model-military-use-back-metas-llama-2024-11-01/">had developed</a> a model for military use based on Llama. The model was tested on its ability to process military intelligence and offer accurate information for operational decision-making. Meta insists that any use of their models for military purposes violates their terms of service.&nbsp;</p><p>Is this proof that open weights models pose a grave risk to the safety of the West? We don&#8217;t think so.&nbsp;</p><p>As others <a href="https://blog.eleuther.ai/nyt-yi-34b-response/">have pointed out</a>, Llama&#8217;s main edge comes from the quality of its training data, not the uniqueness of its architecture. While having Llama out there likely makes the PLA&#8217;s job slightly easier, the notion that China, with all of its technical talent, would not be able to build an equivalent pretty easily, isn&#8217;t worth taking seriously. Also, as Meta&#8217;s Joelle Pineau observed, the PLA finetuned Llama on just 100,000 military dialogue records, and <em>"that's a drop in the ocean compared to most of these models </em>[that]<em> are trained with trillions of tokens so &#8230; it really makes me question what do they actually achieve here in terms of different capabilities&#8221;. </em>It&#8217;s worth noting here that 100,000 isn&#8217;t actually an untypically sized dataset for fine-tuning, but we do question how powerful a mid-sized version of Llama 2 fine-tuned on military dialogue and nothing else is really likely to be.</p><p>Meanwhile over on Safety Island, the UK is awaiting the AI Opportunities plan that the new(ish) government commissioned. Expected in the coming weeks, rumored items <a href="https://www.ft.com/content/4ec7a942-97c0-4e0f-9c9a-9fbec1506bbf">include</a> visa relaxations, loosened planning restrictions for data center construction, and the creation of a new &#8220;AI opportunities unit&#8221;. The specifics of any recommendations around government support for compute remain unknown, but given the government&#8217;s relaxation of its predecessor's tight spending rules, there should now be some spare GPU money up for grabs.</p><p>As AI regulation takes shape around the world, skeptics have often accused big tech companies of regulatory capture. Events took a novel twist this week, with <a href="https://blogs.microsoft.com/on-the-issues/2024/10/28/googles-shadow-campaigns/">Microsoft accusing Google</a> of using a network of front organizations to undermine their cloud business. They contend that Google is funding groups fronted up by other cloud providers to lobby anti-trust regulators, bolstering its own more direct attacks. Microsoft argues that Google is attempting to<em> &#8220;distract from the intense regulatory scrutiny Google is facing around the world by discrediting Microsoft and tilting the regulatory landscape in favor of its cloud services rather than competing on the merits&#8221;</em>.</p><p>Google doesn&#8217;t deny the accusations. Astroturfing is an age-old corporate tactic and isn&#8217;t new in the technology industry. It&#8217;s been a staple of regulatory fights in the gig economy, for example. The trick is not to get caught.</p><p>Anthropic doesn&#8217;t need to engage in shady tactics to sound the alarm bell. On Halloween, somewhat ominously, the company <a href="https://www.anthropic.com/news/the-case-for-targeted-regulation">called on</a> governments to<em> &#8220;urgently take action on AI policy in the next eighteen months&#8221; </em>as <em>&#8220;the window for proactive risk prevention is closing fast&#8221;. </em>They point to improving performance on math, science, and coding benchmarks as a sign that models will soon begin to pose real-world risks. The post plugs Anthropic&#8217;s Responsible Scaling Policy, but stops short of saying what measures it thinks governments should actually take - something like SB1047, but not SB1047.&nbsp;</p><p>The copyright wars have been a staple of this newsletter and show no sign of dying down. A statement on AI training, organized by Ed Newton-Rex of Fairly Trained, <a href="https://www.aitrainingstatement.org/">has attracted</a> over 31,000 signatures, including the likes of Bj&#246;rn, John Rutter, Ian Rankin, James Patterson, and Kevin Bacon. The statement simply reads: <em>&#8220;The unlicensed use of creative works for training generative AI is a major, unjust threat to the livelihoods of the people behind those works, and must not be permitted.&#8221; </em>This comes as Dow Jones, the publisher of the WSJ and New York Post, <a href="https://www.theguardian.com/technology/2024/oct/21/rupert-murdoch-ai-lawsuit-new-york-post-dow-jones">moved</a> to sue Perplexity for a <em>&#8220;massive amount of illegal copying&#8221; </em>of their work.</p><p>Perplexity <a href="https://www.perplexity.ai/hub/blog/about-the-dow-jones-lawsuit?fob=l1xwgL6YUZT5Ounr">hit back</a> claiming that news organizations <em>&#8220;prefer to live in a world where publicly reported facts are owned by corporations, and no one can do anything with those publicly reported facts without paying a toll&#8221;</em>. Politely, this strikes us as a reach. Copyright exists and you can argue that your interpretation is right and news organizations&#8217; is wrong, but Perplexity&#8217;s conspiratorial tone is unhelpful. It also doesn&#8217;t tally with Perplexity&#8217;s own revenue-sharing program, which it trumpets in the same statement. By their standards, how is this not also <em>&#8220;paying a toll&#8221;</em>?&nbsp;</p><p>We&#8217;ve written before about how we think <a href="https://press.airstreet.com/p/copyright-is-the-wrong-hill-to-die-on">compromise</a> is likely inevitable in this debate between model builders, publishers and content creators. And some people are making it work! Text-to-speech stars ElevenLabs <a href="https://elevenlabs.io/blog/1m-payouts">recently revealed</a> that they&#8217;d paid out their first $1M to voice actors.</p><p>While compromise after much lawfare may be the final destination in the US, we may reach a speedier conclusion in the UK. The government is considering a scheme that would give model builders the right to scrape any content, unless publishers and artists <a href="https://www.ft.com/content/26bc3de1-af90-4c69-9f53-61814514aeaa">explicitly &#8220;opt-out&#8221;</a>. Publishers and music rights holders are furious and are demanding an &#8220;opt-in&#8221;. In practice, &#8220;opt-out&#8221; option would be a huge win for them, provided the form and process is simple enough. <a href="https://press.airstreet.com/p/copyright-is-the-wrong-hill-to-die-on">In our piece</a> on copyright and compromise, we outlined a few cases where one side or the other rejected a compromise, only to do worse in the final settlement. Tech companies aren&#8217;t the only people who need to learn from history.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:&quot;button-wrapper&quot;}" data-component-name="ButtonCreateButton"><a class="button primary button-wrapper" href="https://press.airstreet.com/subscribe?"><span>Subscribe now</span></a></p><h3><em><strong>&#127850; Hardware</strong></em></h3><p>While the chip wars continue to rage, it looks like the West and the Chinese Communist Party agree on one thing: neither side wants Chinese companies to use NVIDIA hardware. The Chinese government is ramping up its <a href="https://www.bloomberg.com/news/articles/2024-09-27/china-urges-local-companies-to-stay-away-from-nvidia-s-ai-chips?srnd=homepage-uk">attempts to dissuade</a> local businesses from buying the sanctions-compliant (but highly capable) NVIDIA H20, in favor of domestic alternatives.&nbsp;</p><p>As we covered in this year&#8217;s State of AI Report and our <a href="https://press.airstreet.com/p/the-state-of-chinese-ai">essay on Chinese AI</a>, domestic alternatives have so far struggled with reliability and volume, and leading Chinese labs would rather avoid them. The CCP finds itself caught in a bind - simultaneously wanting to produce the world&#8217;s best models, while ending its dependence on the foreign hardware needed to do it.&nbsp;</p><p>Whatever happens in China, NVIDIA continues to expand on all fronts. In <a href="https://nvidianews.nvidia.com/news/spectrum-x-ethernet-networking-xai-colossus#:~:text=The%20supporting%20facility%20and%20state,take%20many%20months%20to%20years.">122 days</a>, the company had a 100k cluster up and running for xAI in Tennessee, with ambitions <a href="https://www.datacenterdynamics.com/en/news/xai-to-double-colossus-compute-capacity-reveals-cluster-uses-nvidia-spectrum-x-ethernet/">to double it in size</a>. You can have a glimpse inside the cluster <a href="https://www.servethehome.com/inside-100000-nvidia-gpu-xai-colossus-cluster-supermicro-helped-build-for-elon-musk/">here</a>.</p><p>NVIDIA also unveiled <a href="https://blogs.nvidia.com/blog/denmark-sovereign-ai-supercomputer/">Denmark&#8217;s first AI supercomputer</a>, Gefion, powered by 1,528 H100s. Gefion is part of the Danish Center for AI Innovation, funded by the Novo Nordisk Foundation and the Export and Investment Fund of Denmark. The new supercomputer is supporting pilots, pilots the large-scale simulation of quantum computer circuits and a multi-modal genomic foundation model for vaccine design.</p><p>A few months ago, we covered how NVIDIA challenger Cerebras was gearing up for an IPO. Despite a spike in good publicity following the launch of a buzzy new inference platform, questions are beginning to mount for the company.&nbsp;</p><p>The publication of the company&#8217;s <a href="https://www.sec.gov/Archives/edgar/data/2021728/000162828024041596/cerebras-sx1.htm">Form S1</a> revealed that almost all of the business&#8217; revenue came from UAE-based G42. This revenue also comes with an option-shaped catch. If G42 makes a single order worth more than $500M but less than $5B before the end of 2025, they will have the right to purchase up to 10% of its value in G42 stock at a discounted rate.</p><p>Whatever happens with the existing challengers, it now looks like Sam Altman&#8217;s planned multi-trillion dollar chip empire is no more. While the company has made progress on its plans with Broadcom and TSMC to build its first custom chip, it looks like its planned <a href="https://www.reuters.com/technology/artificial-intelligence/openai-builds-first-chip-with-broadcom-tsmc-scales-back-foundry-ambition-2024-10-29/">network of foundries has been dropped</a>. Even if it had been possible to raise the vast sums of capital required, the company appears to have concluded that it wasn&#8217;t worth the investment in time.</p><p>Whoever is providing it, all this compute power needs energy. Microsoft&#8217;s <a href="https://www.bbc.co.uk/news/articles/cx25v2d7zexo">move to revive</a> the Three Mile Island nuclear power site looks like the start of a trend. Google unveiled <a href="https://blog.google/outreach-initiatives/sustainability/google-kairos-power-nuclear-energy-agreement/">the world&#8217;s first corporate agreement</a> to buy the output of small nuclear reactors, while nuclear looks <a href="https://www.bloomberg.com/news/articles/2024-10-29/private-equity-s-next-bet-on-artificial-intelligence-is-nuclear-energy">set to become</a> a big theme for private equity. But will this come quickly enough? Three Mile Island won&#8217;t be online until 2028, while on optimistic assumptions, small nuclear reactors are unlikely to reach commercial scale before the 2030s. Expect the AI energy squeeze to get worse before it gets better.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/p/the-state-of-ai-2024-november?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:&quot;button-wrapper&quot;}" data-component-name="ButtonCreateButton"><a class="button primary button-wrapper" href="https://press.airstreet.com/p/the-state-of-ai-2024-november?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><h3><em><strong>&#127981; Big <s>tech</s> start-ups&nbsp;</strong></em></h3><p>Despite its recent internal struggles, OpenAI continues to grow unabated. The company <a href="https://www.bloomberg.com/news/articles/2024-10-02/openai-has-closed-new-funding-round-raising-over-6-5-billion">pulled in $6.6B</a> at an $157B valuation in a funding round led by Thrive Capital. While we&#8217;ve become used to mega-rounds in the GenAI era, this is one of the largest ever rounds for a private company and means OpenAI&#8217;s valuation is up there with the likes of ByteDance and SpaceX.&nbsp;</p><p>OpenAI is now in the process of unpicking the capped-profit model it created when it abandoned its non-profit status in 2019. Under the current model, a non-profit entity oversees a for-profit entity, with early investors returns capped at 100x and later investors at lower numbers. Microsoft, OpenAI&#8217;s main benefactor technically holds no equity, but instead gets the first cut of any profits as they start coming in.</p><p>This model manages both to be cumbersome and to provide little in the way of oversight. As OpenAI transitions to a for-profit company, both sides&#8217; bankers are gearing up to fight over how equity should be distributed. Check out Matt Levine&#8217;s <a href="https://www.bloomberg.com/opinion/articles/2024-10-21/who-owns-openai">analysis</a> of how these negotiations tend to work and Nathan&#8217;s <a href="https://www.ft.com/content/8f192751-598c-4b92-ab8d-c1d368fdfb36">FT op-ed</a> on why OpenAI is right to make the move.</p><p>Outside lawyers&#8217; offices and back in the arena, we see the journey from model to product continue unabated. Anthropic, again continues to lead the way, unveiling its <a href="https://www.anthropic.com/news/3-5-models-and-computer-use">computer use feature</a>, along with a <a href="https://x.com/alexalbert__/status/1852418257195823550">bunch</a> of other new features. This allows developers to direct Claude to look at a screen, click buttons, and type text. It&#8217;s still in beta and Anthropic <a href="https://www.anthropic.com/news/developing-computer-use">acknowledges that it&#8217;s error-prone</a>.</p><div id="youtube2-ODaHJzOyVCQ" class="youtube-wrap" data-attrs="{&quot;videoId&quot;:&quot;ODaHJzOyVCQ&quot;,&quot;startTime&quot;:null,&quot;endTime&quot;:null}" data-component-name="Youtube2ToDOM"><div class="youtube-inner"><iframe src="https://www.youtube-nocookie.com/embed/ODaHJzOyVCQ?rel=0&amp;autoplay=0&amp;showinfo=0&amp;enablejsapi=0" frameborder="0" loading="lazy" gesture="media" allow="autoplay; fullscreen" allowautoplay="true" allowfullscreen="true" width="728" height="409"></iframe></div></div><p>The list of early users Anthropic provides - Asana, Canca, DoorDash, Replit, and The Browser Company - provides a clue as to potential use cases.</p><p>Again, this is likely a case of a frontier model leader probably putting a bunch of start-ups out of business. However, there is an open question about whether or not an API is the right vehicle for a computer use feature at scale. It&#8217;s not hard to see the potential privacy or liability issues stacking up&#8230;</p><p>Computer use came alongside the launch of Claude 3.5 Haiku and an upgraded version of Sonnet. Anthropic&#8217;s own benchmarking shows it edging out GPT-4o across a range of benchmarks, with Haiku a close competitor to 4o-mini. They&#8217;ve also thrown in a new agentic tool use stat for good measure.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!UfON!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79328f54-c771-407c-b18e-7d9463f21122_1600x1188.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!UfON!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79328f54-c771-407c-b18e-7d9463f21122_1600x1188.png 424w, https://substackcdn.com/image/fetch/$s_!UfON!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79328f54-c771-407c-b18e-7d9463f21122_1600x1188.png 848w, https://substackcdn.com/image/fetch/$s_!UfON!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79328f54-c771-407c-b18e-7d9463f21122_1600x1188.png 1272w, https://substackcdn.com/image/fetch/$s_!UfON!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79328f54-c771-407c-b18e-7d9463f21122_1600x1188.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!UfON!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79328f54-c771-407c-b18e-7d9463f21122_1600x1188.png" width="1456" height="1081" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/79328f54-c771-407c-b18e-7d9463f21122_1600x1188.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1081,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!UfON!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79328f54-c771-407c-b18e-7d9463f21122_1600x1188.png 424w, https://substackcdn.com/image/fetch/$s_!UfON!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79328f54-c771-407c-b18e-7d9463f21122_1600x1188.png 848w, https://substackcdn.com/image/fetch/$s_!UfON!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79328f54-c771-407c-b18e-7d9463f21122_1600x1188.png 1272w, https://substackcdn.com/image/fetch/$s_!UfON!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79328f54-c771-407c-b18e-7d9463f21122_1600x1188.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>OpenAI is also getting in on the productizing, with the <a href="https://openai.com/index/introducing-canvas/">launch of Canvas</a>, a writing and coding interface. It doesn&#8217;t take a particularly forensic examination to notice the similarities between Canvas and Claude Artifacts.&nbsp;</p><p>Where the company is differentiating itself is web search. While OpenAI had started integrating links into some queries previously, the <a href="https://openai.com/index/introducing-chatgpt-search/">new search function</a> comes with a jazzier UX and enhanced reasoning capabilities. To accomplish the latter, OpenAI&#8217;s search model is post-trained using new distilled outputs from OpenAI o1-preview. As a nod to the copyright wars, OpenAI is using its newly-minted content partners, as well as existing search engines, as a means of serving up results.&nbsp;</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:&quot;button-wrapper&quot;}" data-component-name="ButtonCreateButton"><a class="button primary button-wrapper" href="https://press.airstreet.com/subscribe?"><span>Subscribe now</span></a></p><h3><strong>&#128300;Research</strong></h3><p><strong><a href="https://www.physicalintelligence.company/blog/pi0">&#960;0: A Vision-Language-Action Flow Model for General Robot Control</a></strong>, <em>Physical Intelligence</em>.</p><p>The authors present &#960;0, a robot foundation model that combines a pre-trained vision-language model with flow matching to generate continuous robot actions. The model is trained on over 10,000 hours of demonstration data across 7 different robot configurations and 68 distinct tasks. By using flow matching rather than discrete action tokens, the model can generate precise, high-frequency control signals needed for dexterous manipulation.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Lrnl!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F708d842f-e804-4c35-8b1b-79b3806d4f72_1600x900.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Lrnl!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F708d842f-e804-4c35-8b1b-79b3806d4f72_1600x900.png 424w, https://substackcdn.com/image/fetch/$s_!Lrnl!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F708d842f-e804-4c35-8b1b-79b3806d4f72_1600x900.png 848w, https://substackcdn.com/image/fetch/$s_!Lrnl!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F708d842f-e804-4c35-8b1b-79b3806d4f72_1600x900.png 1272w, https://substackcdn.com/image/fetch/$s_!Lrnl!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F708d842f-e804-4c35-8b1b-79b3806d4f72_1600x900.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Lrnl!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F708d842f-e804-4c35-8b1b-79b3806d4f72_1600x900.png" width="1456" height="819" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/708d842f-e804-4c35-8b1b-79b3806d4f72_1600x900.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:819,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!Lrnl!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F708d842f-e804-4c35-8b1b-79b3806d4f72_1600x900.png 424w, https://substackcdn.com/image/fetch/$s_!Lrnl!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F708d842f-e804-4c35-8b1b-79b3806d4f72_1600x900.png 848w, https://substackcdn.com/image/fetch/$s_!Lrnl!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F708d842f-e804-4c35-8b1b-79b3806d4f72_1600x900.png 1272w, https://substackcdn.com/image/fetch/$s_!Lrnl!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F708d842f-e804-4c35-8b1b-79b3806d4f72_1600x900.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The training process involves two phases: pre-training on diverse data to build general capabilities, followed by fine-tuning on specific tasks to achieve mastery. This approach allows the model to learn both broad skills and specialized behaviors while maintaining the ability to recover from mistakes. The model can follow natural language commands and work with high-level language models that break complex tasks into simpler steps.</p><p>The researchers evaluated &#960;0 through both zero-shot testing (using only pre-trained capabilities) and fine-tuning experiments on new tasks. The model outperformed existing approaches on dexterous manipulation tasks like folding laundry, assembling boxes, and clearing tables. Zero-shot performance was particularly strong on tasks similar to those in the pre-training data.</p><p>However, the work has clear limitations. The researchers note that they don't yet understand which types of pre-training data are most valuable or how much data is needed for reliable performance. The model's success on novel tasks varies significantly, and it remains unclear whether this approach will generalize to substantially different domains like autonomous driving or legged locomotion.</p><p><strong><a href="https://ai.meta.com/blog/fair-robotics-open-source/?utm_source=twitter&amp;utm_medium=organic_social&amp;utm_content=video&amp;utm_campaign=fair">Advancing embodied AI through progress in touch perception, dexterity, and human-robot interaction</a></strong>, <em>Meta</em>.</p><p>Meta dropped a bunch of new robotics work in a single release. First up, Sparsh, a general-purpose tactile encoder, provides a versatile, vision-based approach to tactile sensing across different tasks and sensor types, eliminating the need for extensive labeled datasets.&nbsp;</p><p>Digit 360, a new fingertip sensor, extends human-like tactile precision to robots. It offers rich multimodal sensory data, capturing intricate surface details and minute forces. This makes it useful in delicate tasks in sectors like healthcare and manufacturing. It&#8217;s bolstered by on-device AI that allows responsive processing, with obvious potential uses in robotics, VR, and potentially even prosthetics.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!YsB1!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F070ac21d-939e-41e1-9072-9e9d72507732_1600x900.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!YsB1!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F070ac21d-939e-41e1-9072-9e9d72507732_1600x900.png 424w, https://substackcdn.com/image/fetch/$s_!YsB1!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F070ac21d-939e-41e1-9072-9e9d72507732_1600x900.png 848w, https://substackcdn.com/image/fetch/$s_!YsB1!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F070ac21d-939e-41e1-9072-9e9d72507732_1600x900.png 1272w, https://substackcdn.com/image/fetch/$s_!YsB1!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F070ac21d-939e-41e1-9072-9e9d72507732_1600x900.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!YsB1!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F070ac21d-939e-41e1-9072-9e9d72507732_1600x900.png" width="1456" height="819" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/070ac21d-939e-41e1-9072-9e9d72507732_1600x900.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:819,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!YsB1!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F070ac21d-939e-41e1-9072-9e9d72507732_1600x900.png 424w, https://substackcdn.com/image/fetch/$s_!YsB1!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F070ac21d-939e-41e1-9072-9e9d72507732_1600x900.png 848w, https://substackcdn.com/image/fetch/$s_!YsB1!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F070ac21d-939e-41e1-9072-9e9d72507732_1600x900.png 1272w, https://substackcdn.com/image/fetch/$s_!YsB1!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F070ac21d-939e-41e1-9072-9e9d72507732_1600x900.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Meta also introduced Digit Plexus, a hardware-software framework that integrates tactile sensors into a unified robotic hand system. This platform coordinates sensory inputs across multiple tactile points, from fingertips to palms, supporting complex hand interactions similar to human touch feedback and motor response.&nbsp;</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!JJ9M!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff139298b-bfb4-4dc8-99dc-e01596520181_1280x720.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!JJ9M!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff139298b-bfb4-4dc8-99dc-e01596520181_1280x720.png 424w, https://substackcdn.com/image/fetch/$s_!JJ9M!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff139298b-bfb4-4dc8-99dc-e01596520181_1280x720.png 848w, https://substackcdn.com/image/fetch/$s_!JJ9M!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff139298b-bfb4-4dc8-99dc-e01596520181_1280x720.png 1272w, https://substackcdn.com/image/fetch/$s_!JJ9M!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff139298b-bfb4-4dc8-99dc-e01596520181_1280x720.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!JJ9M!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff139298b-bfb4-4dc8-99dc-e01596520181_1280x720.png" width="1280" height="720" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f139298b-bfb4-4dc8-99dc-e01596520181_1280x720.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:720,&quot;width&quot;:1280,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!JJ9M!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff139298b-bfb4-4dc8-99dc-e01596520181_1280x720.png 424w, https://substackcdn.com/image/fetch/$s_!JJ9M!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff139298b-bfb4-4dc8-99dc-e01596520181_1280x720.png 848w, https://substackcdn.com/image/fetch/$s_!JJ9M!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff139298b-bfb4-4dc8-99dc-e01596520181_1280x720.png 1272w, https://substackcdn.com/image/fetch/$s_!JJ9M!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff139298b-bfb4-4dc8-99dc-e01596520181_1280x720.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>To smooth human-robot collaboration, Meta also developed PARTNR, a benchmark tool for assessing how AI models perform in collaborative, physical tasks with humans. Built on a simulation platform, PARTNR allows evaluation of large language models in scenarios mimicking household environments.&nbsp;</p><p><strong><a href="https://arxiv.org/abs/2410.23262">EMMA: End-to-End Multimodal Model for Autonomous Driving</a></strong>, <em>Waymo</em>.</p><p>Introduces EMMA, an end-to-end multimodal model designed for autonomous driving, which leverages a multimodal to process raw camera data and directly produce outputs for various driving tasks. EMMA operates within a unified language space, which represents all inputs and outputs (such as navigation instructions, vehicle states, and planned trajectories) as natural language text. This allows EMMA to apply language processing strengths to tasks traditionally segmented into modules, like perception, motion planning, and road graph estimation, effectively merging these functions for comprehensive task handling.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Dt1H!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F762ef9bd-2b87-4de0-a4a2-aa8a93e098e3_814x373.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Dt1H!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F762ef9bd-2b87-4de0-a4a2-aa8a93e098e3_814x373.png 424w, https://substackcdn.com/image/fetch/$s_!Dt1H!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F762ef9bd-2b87-4de0-a4a2-aa8a93e098e3_814x373.png 848w, https://substackcdn.com/image/fetch/$s_!Dt1H!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F762ef9bd-2b87-4de0-a4a2-aa8a93e098e3_814x373.png 1272w, https://substackcdn.com/image/fetch/$s_!Dt1H!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F762ef9bd-2b87-4de0-a4a2-aa8a93e098e3_814x373.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Dt1H!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F762ef9bd-2b87-4de0-a4a2-aa8a93e098e3_814x373.png" width="814" height="373" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/762ef9bd-2b87-4de0-a4a2-aa8a93e098e3_814x373.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:373,&quot;width&quot;:814,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!Dt1H!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F762ef9bd-2b87-4de0-a4a2-aa8a93e098e3_814x373.png 424w, https://substackcdn.com/image/fetch/$s_!Dt1H!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F762ef9bd-2b87-4de0-a4a2-aa8a93e098e3_814x373.png 848w, https://substackcdn.com/image/fetch/$s_!Dt1H!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F762ef9bd-2b87-4de0-a4a2-aa8a93e098e3_814x373.png 1272w, https://substackcdn.com/image/fetch/$s_!Dt1H!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F762ef9bd-2b87-4de0-a4a2-aa8a93e098e3_814x373.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>EMMA achieves high performance on public datasets such as nuScenes and the Waymo Open Motion Dataset, particularly excelling in motion planning and 3D object detection using camera inputs alone. A key feature is its "chain-of-thought" reasoning, which enhances decision-making transparency by prompting the model to explain its decisions sequentially, integrating world knowledge. This approach produces outputs such as future vehicle trajectories and object detection estimates in a readable, interpretable format.</p><p>While promising, EMMA has limitations: it is constrained to camera-only inputs, lacks fusion with 3D sensing modalities like LiDAR, and requires high computational power, which limits real-world deployment. Despite these, the model shows potential as a generalist framework by achieving competitive results across multiple autonomous driving tasks and efficiently sharing learned knowledge across tasks, outperforming task-specific models.</p><p>Also on our radar:</p><ul><li><p><strong><a href="https://www.anthropic.com/research/evaluating-feature-steering">Evaluating feature steering: A case study in mitigating social biases</a></strong>, <em>Anthropic</em>. Investigates "feature steering" - a technique for modifying AI behavior by adjusting interpretable features within Claude 3 Sonnet's neural network. Testing 29 features related to social biases, they found a moderate range where steering could influence outputs without severely degrading performance, and identified a "neutrality" feature that reduced multiple types of bias. However, features often had unexpected "off-target" effects beyond their apparent purpose, and extreme steering impaired model capabilities.</p></li><li><p><strong><a href="https://arxiv.org/abs/2410.06468">Does Spatial Cognition Emerge in Frontier Models?</a>, </strong><em>Apple</em>. Using a new benchmark called SPACE, the researchers test both large-scale abilities (like mapping environments and finding shortcuts) and small-scale skills (like mental rotation and spatial memory). The results reveal that even advanced AI models perform poorly on many spatial cognition tasks that animals handle well, often scoring near chance level on tests.&nbsp;</p></li><li><p><strong><a href="https://arxiv.org/abs/2410.19034">Mixture of Parrots: Experts improve memorization more than reasoning</a>, </strong><em>Harvard. </em>Explores the trade-offs between mixture of expert models and standard dense transformers. They demonstrated that MoEs can effectively leverage additional experts to improve memory-intensive tasks like fact retrieval, but find diminishing returns in reasoning tasks like mathematical problem-solving or graph analysis.&nbsp;</p></li><li><p><strong><a href="https://www.arxiv.org/abs/2410.12491#:~:text=Insights%20from%20the%20Inverse%3A%20Reconstructing%20LLM%20Training%20Goals%20Through%20Inverse%20RL,-Jared%20Joselowitz%2C%20Arjun&amp;text=Large%20language%20models%20(LLMs)%20trained,decision%2Dmaking%20processes%20remain%20opaque.">Insights from the Inverse: Reconstructing LLM Training Goals Through Inverse RL</a></strong>, <em>Imperial College London, Harvard. </em>Presents a novel approach to understanding how LLMs make decisions by using inverse reinforcement learning to reconstruct their implicit reward functions. On 70M and 410M parameter LLMs, they successfully extracted reward models that could predict human preferences with up to 80.40% accuracy. They then used these IRL-derived reward models to fine-tune new LLMs, finding that the 70M model drove improvements on a toxicity benchmark versus standard RLHF, whereas the 410M resulted in worse performance.</p></li><li><p><strong><a href="https://arxiv.org/abs/2410.22071">Distinguishing Ignorance from Error in LLM Hallucinations</a>, </strong><em>Technion, Google Research</em>. Distinguishes between two types of LLM hallucinations: those that occur when the model lacks knowledge (HK-) versus when it hallucinates despite having the correct knowledge (HK+). The researchers developed a method called WACK to systematically capture HK+ across models, using techniques like "bad shots" (showing incorrect examples) and "Alice-Bob" (using subtle persuasion) to induce HK+ hallucinations. They found that hallucination types leave distinct signatures in the models' internal states, different models hallucinate in unique ways even with shared knowledge, and that detecting hallucinations works better when using model-specific datasets rather than generic ones.&nbsp;</p></li><li><p><strong><a href="https://arxiv.org/abs/2410.12851">VibeCheck: Discover and Quantify Qualitative Differences in Large Language Models</a>, </strong><em>UC Berkeley</em>. Proposes VibeCheck, a system that identifies qualitative differences between LLM outputs. It discovers "vibes" (like tone, formatting style, or level of detail) through automated analysis of model outputs and validates each vibe through three criteria: how consistently different judges agree on that vibe, how well it distinguishes between different models' outputs, and whether it predicts user preferences.&nbsp;&nbsp;</p></li></ul><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/p/the-state-of-ai-2024-november?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:&quot;button-wrapper&quot;}" data-component-name="ButtonCreateButton"><a class="button primary button-wrapper" href="https://press.airstreet.com/p/the-state-of-ai-2024-november?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><h3><strong>&#128176;Startups</strong></h3><h4><em><strong>&#128640; </strong></em><strong>Funding highlight reel</strong></h4><p><strong>11x</strong>, developing autonomous digital workers for enterprise, <a href="https://techcrunch.com/2024/09/30/11x-ai-a-developer-of-ai-sales-reps-has-raised-50m-series-b-led-by-a16z-sources-say/">raised</a> a $50M Series B, led by a16z</p><p><strong>Archon Biosciences</strong>, using AI to design novel biomolecules, <a href="https://techcrunch.com/2024/10/30/archon-emerges-from-stealth-with-20m-and-antibody-cages-to-power-up-drug-development/">raised</a> a $20M seed, led by Madrona Ventures</p><p><strong>Basecamp Research</strong>, using the natural world to discover better medicines, <a href="https://tech.eu/2024/10/09/basecamp-research-raises-60m-series-b-to-fuel-ai-advancements-in-life-sciences/">raised</a> a $60M Series B, led by Singular</p><p><strong>CoreWeave</strong>, the GPU compute provider, <a href="https://www.bloomberg.com/news/articles/2024-10-11/coreweave-closes-650-million-credit-facility-for-ai-cloud-computing-push">closed</a> a $650M credit facility, led by Goldman Sachs, JPMorgan, and Morgan Stanley</p><p><strong>EvenUp</strong>, creating AI-based legal case management solutions, <a href="https://www.evenuplaw.com/blog/evenup-announces-135-million-series-d-and-launches-four-new-groundbreaking-products">raised</a> a $135M Series D, led by Bain Capital Ventures&nbsp;</p><p><strong>Galileo</strong>, the AI evaluation and observability platform, <a href="https://galileo.ai/blog/announcing-our-series-b">raised</a> a $45M Series B, led by Scale Venture Partners&nbsp;</p><p><strong>Genie AI</strong>, creating an automatic contract drafting platform, <a href="https://www.businessinsider.in/tech/news/exclusive-legal-tech-startup-genie-ai-just-raised-17-8-million-from-google-and-khosla-ventures-using-this-13-slide-deck/slidelist/114497618.cms#slideid=114497956">raised</a> a $17.8M Series A, led by Khosla Ventures and Google Ventures&nbsp;</p><p><strong>Lightmatter</strong>, producing 3D-stacked photonic chips, <a href="https://techcrunch.com/2024/10/16/lightmatters-400m-d-round-has-ai-hyperscalers-hyped-for-photonic-datacenters/">raised</a> a $400M Series D, led by T. Rowe Price</p><p><strong>Granola</strong>, the AI notetaker, <a href="https://www.granola.ai/blog/series-a">raised</a> a $20M Series A, led by Spark Capital</p><p><strong>Nimble</strong>, building autonomous warehouses for order fulfillment, <a href="https://x.com/simonkalouche/status/1849110090969940176?s=51&amp;t=8YCMEcmVVXRPm8SXTMgdlw">raised</a> a $106M Series C, led by FedEx</p><p><strong>OpenAI</strong>, the frontier AI lab, <a href="https://openai.com/index/scale-the-benefits-of-ai/">raised</a> a $6.6B funding round, led by Thrive Capital</p><p><strong>Oriole Networks</strong>, speeding up data centers for AI, <a href="https://fortune.com/2024/10/21/oriole-networks-series-a-22-milliion-photonics-startups-ai-data-center-energy-demands/">raised</a> a $22M Series A, led by Plural</p><p><strong>Poolside</strong>, the AI for software engineering, <a href="https://sifted.eu/articles/poolside-500m-series-b-news">raised</a> a $500M Series B</p><p><strong>Reality Defender</strong>, a platform for detecting false media content, <a href="https://www.prnewswire.com/news-releases/reality-defender-expands-series-a-to-33-million-to-enhance-ai-detection-capabilities-302283098.html">raised</a> a $33M Series A, led by Illuminate Financial</p><p><strong>Sana</strong>, building an enterprise AI assistant,<strong> </strong><a href="https://www.forbes.com/sites/iainmartin/2024/10/30/this-swedish-startup-raised-55-million-to-build-an-army-of-ai-agents/">raised</a> a $55M venture round, led by New Enterprise Associates</p><p><strong>Sierra</strong>, the AI customer service agent platform, <a href="https://www.cnbc.com/2024/10/28/bret-taylors-ai-startup-sierra-valued-at-4point5-billion-in-funding.html">raised</a> a $175M venture round, led by Greenoaks Capital</p><p><strong>Terray Therapeutics</strong>, working on small molecule drug discovery, raised a $120M Series B, led by NVentures and Bedford Ridge Capital</p><p><strong>Waymo</strong>, the self-driving company, <a href="https://x.com/Waymo/status/1849806827074019654">raised</a> a $5.6B funding round, led by Alphabet</p><p><strong>Wispr</strong>, a voice-based personal computing platform, <a href="https://www.wsj.com/articles/startup-wispr-secures-funding-to-build-wearable-device-that-harnesses-neural-signals-11666605603">raised</a> a $10M seed round, led by TriplePoint Capital and Neo&nbsp;</p><p><strong>Xscape Photonics</strong>, building photonic chips for HPC, <a href="https://techcrunch.com/2024/10/15/xscape-is-building-multicolor-lasers-for-datacenters/">raised</a> a $44M Series A, led by IAG Capital Partners&nbsp;</p><h4><em><strong>&#129309; </strong></em><strong>Exits</strong></h4><p><strong>Applied Intuition</strong>, the company building an autonomy stack for self-driving, <a href="https://www.appliedintuition.com/news/applied-intuition-ghost-autonomy-patents">has acquired</a> the IP portfolio of defunct self-driving start-up Ghost Autonomy</p><p><strong>DeepHealth</strong>, a portfolio of AI-powered healthcare solutions, <a href="https://deephealth.com/news/deephealth-expands-its-ai-powered-cancer-diagnostic-and-screening-portfolio-with-the-acquisition-of-kheiron/">has acquired</a> Kheiron Medical, a cancer diagnostics company</p><p><strong>Thomson Reuters</strong>, the media and content giant, <a href="https://www.thomsonreuters.com/en/press-releases/2024/october/thomson-reuters-acquires-materia-a-specialist-in-agentic-ai-for-the-tax-audit-and-accounting-profession.html?cid=soc&amp;chl=linkedin&amp;postid=18a75956-6dee-4f83-86e9-d2b57da80d22">has acquired</a> Materia, a start-up working on AI agents for tax, audit and accounting</p><div><hr></div><p>Signing off,&nbsp;</p><p>Nathan Benaich and Alex Chalmers on 3 November 2024</p><p><a href="https://www.airstreet.com/?utm_campaign=Your%20guide%20to%20AI&amp;utm_medium=email&amp;utm_source=Substack%20newsletter">Air Street Capital</a> | <a href="https://www.twitter.com/nathanbenaich?utm_campaign=Your%20guide%20to%20AI&amp;utm_medium=email&amp;utm_source=Substack%20newsletter">Twitter</a> | <a href="https://www.linkedin.com/in/nathanbenaich/?utm_campaign=Your%20guide%20to%20AI&amp;utm_medium=email&amp;utm_source=Substack%20newsletter">LinkedIn</a> | <a href="https://www.stateof.ai/?utm_campaign=Your%20guide%20to%20AI&amp;utm_medium=email&amp;utm_source=Substack%20newsletter">State of AI Report</a> | <a href="https://www.raais.co/?utm_campaign=Your%20guide%20to%20AI&amp;utm_medium=email&amp;utm_source=Substack%20newsletter">RAAIS</a> | <a href="https://airstreet.com/events">Events</a></p><p><em>Air Street Capital invests in AI-first entrepreneurs from the very beginning of your company-building journey.</em></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!2HOA!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1431346b-67e1-4e80-8a4a-cdef19006582_1864x1038.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!2HOA!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1431346b-67e1-4e80-8a4a-cdef19006582_1864x1038.png 424w, https://substackcdn.com/image/fetch/$s_!2HOA!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1431346b-67e1-4e80-8a4a-cdef19006582_1864x1038.png 848w, https://substackcdn.com/image/fetch/$s_!2HOA!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1431346b-67e1-4e80-8a4a-cdef19006582_1864x1038.png 1272w, https://substackcdn.com/image/fetch/$s_!2HOA!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1431346b-67e1-4e80-8a4a-cdef19006582_1864x1038.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!2HOA!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1431346b-67e1-4e80-8a4a-cdef19006582_1864x1038.png" width="1456" height="811" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1431346b-67e1-4e80-8a4a-cdef19006582_1864x1038.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:811,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:126808,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!2HOA!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1431346b-67e1-4e80-8a4a-cdef19006582_1864x1038.png 424w, https://substackcdn.com/image/fetch/$s_!2HOA!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1431346b-67e1-4e80-8a4a-cdef19006582_1864x1038.png 848w, https://substackcdn.com/image/fetch/$s_!2HOA!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1431346b-67e1-4e80-8a4a-cdef19006582_1864x1038.png 1272w, https://substackcdn.com/image/fetch/$s_!2HOA!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1431346b-67e1-4e80-8a4a-cdef19006582_1864x1038.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div>]]></content:encoded></item><item><title><![CDATA[State of AI outtakes #1: scaling in defense]]></title><description><![CDATA[The State of AI Report aims to provide a comprehensive overview of everything you need to know across AI research, industry, politics, and safety.]]></description><link>https://press.airstreet.com/p/state-of-ai-outtakes-1-defense</link><guid isPermaLink="false">https://press.airstreet.com/p/state-of-ai-outtakes-1-defense</guid><dc:creator><![CDATA[Air Street Press]]></dc:creator><pubDate>Thu, 24 Oct 2024 15:55:20 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/a14121d7-acb2-4ace-9974-b1e94b3defc2_838x471.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>The State of AI Report aims to provide a comprehensive overview of everything you need to know across AI research, industry, politics, and safety. To ensure the report remains at a manageable length, lots of material doesn&#8217;t make it into the final version. We&#8217;re bringing Air Street Press readers some of the research that didn&#8217;t make the original cut, subsequent follow-up work, and our reflections.</em></p><h3>Introduction</h3><p>Exactly a week after the <a href="http://stateof.ai">State of AI Report</a> landed, we were in Madrid for the European Defense Tech Summit, organized by our friend Eric Slesinger of 201 Ventures. Alex kicked off the day, hosting a panel diving into the challenges founders and investors face in building and scaling defense tech companies in Europe, where he was joined by Jonathan Turner, a partner at Accel (which invested in Helsing&#8217;s Series C) and Lorenz Meier, co-founder and CEO of drone autonomy company Auterion.&nbsp;</p><p>The panel touched on a number of the defense-specific themes that we brought out in this year&#8217;s report, including the pains of defense contracting, the need for courage among investors to bet ahead of government contracts landing, as well as the importance of spending time on the frontline to understand your end user&#8217;s needs.</p><h3>The path to scale</h3><p>In last year&#8217;s report, we covered the record level of investment being pumped into US defense start-ups. Anduril, the clear leader amongst new defense companies, has gone from strength to strength since then. The company has begun to win serious contracts in the US, including <a href="https://www.anduril.com/article/anduril-industries-awarded-air-defense-production-contract-with-department-of-defense/">$250M to deliver</a> advanced air defense capabilities for the Pentagon. </p><p>In August, Anduril delivered the first Tactical Intelligence Targeting Access Node (TITAN) to the US Army, as part of a $178M contract, <a href="https://techcrunch.com/2024/08/12/anduril-reaches-milestone-with-major-defense-hardware-contract/">in partnership</a> with Palantir. Anduril also delivered its first Ghost Submarine prototype to the Australian military <a href="https://breakingdefense.com/2024/04/andurils-aussie-drone-sub-one-year-early-and-on-budget-heads-to-production/">on budget and a year early</a> - unheard of in the world of defense procurement.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!HHfv!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fefa860bc-dc23-41af-b47c-8a225d2bd78e_960x540.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!HHfv!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fefa860bc-dc23-41af-b47c-8a225d2bd78e_960x540.png 424w, https://substackcdn.com/image/fetch/$s_!HHfv!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fefa860bc-dc23-41af-b47c-8a225d2bd78e_960x540.png 848w, https://substackcdn.com/image/fetch/$s_!HHfv!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fefa860bc-dc23-41af-b47c-8a225d2bd78e_960x540.png 1272w, https://substackcdn.com/image/fetch/$s_!HHfv!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fefa860bc-dc23-41af-b47c-8a225d2bd78e_960x540.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!HHfv!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fefa860bc-dc23-41af-b47c-8a225d2bd78e_960x540.png" width="960" height="540" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/efa860bc-dc23-41af-b47c-8a225d2bd78e_960x540.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:540,&quot;width&quot;:960,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!HHfv!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fefa860bc-dc23-41af-b47c-8a225d2bd78e_960x540.png 424w, https://substackcdn.com/image/fetch/$s_!HHfv!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fefa860bc-dc23-41af-b47c-8a225d2bd78e_960x540.png 848w, https://substackcdn.com/image/fetch/$s_!HHfv!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fefa860bc-dc23-41af-b47c-8a225d2bd78e_960x540.png 1272w, https://substackcdn.com/image/fetch/$s_!HHfv!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fefa860bc-dc23-41af-b47c-8a225d2bd78e_960x540.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>But Anduril with its $3.7B warchest from investors is unusual. It&#8217;s too early to judge whether the ecosystem is likely to broaden. There are potential bright spots, like Saronic, a US-based manufacturer of unmanned surface vehicles. But the company, as far as we can tell, is yet to move beyond small pilots to larger contracts.</p><p>These scaling challenges were a primary theme of Alex&#8217;s panel in Madrid. Jonathan from Accel emphasized that one of the reasons Helsing had raised so much so quickly was that the next generation of winners couldn&#8217;t simply wait for government contracting cycles to work themselves out. Only a small handful of winners were likely to prevail and this would require courage from investors.</p><p>He was also bearish on the prospect of M&amp;A coming to early-stage defense companies anytime soon, with national governments wary of foreign acquisitions and the pool of potential acquirers in most European countries being small. This is a dynamic we&#8217;re already seeing play out in France. Preligens, for example, <a href="https://www.reuters.com/markets/deals/safran-buys-ai-firm-preligens-220-million-euros-2024-09-02/">was bought</a> by Safran only after the French government announced that it would bar an overseas buyer. We see M&amp;A as <a href="https://press.airstreet.com/p/defense-exits-european-dynamism">part of the solution</a>, but also believe that defense primes will only spend money on it when there&#8217;s an incentive for them to do so. While procurement remains broken and competition lackluster, they can continue to focus on dividends and buybacks at the expense of R&amp;D.</p><p>There&#8217;s a stark contrast between this sluggishness and the speed with which Ukraine is moving to procure technology for use on the battlefield. Alex Bornyakov, from the country&#8217;s Ministry of Digital Transformation, talked about how contracting cycles had fallen from 2-3 years before the war to 21 days today, while the gap between a manufacturer presenting its equipment for testing to being certified is 3-4 months.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!pAF8!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffa352174-8d4b-424d-ae25-00bd2f426267_1024x768.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!pAF8!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffa352174-8d4b-424d-ae25-00bd2f426267_1024x768.png 424w, https://substackcdn.com/image/fetch/$s_!pAF8!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffa352174-8d4b-424d-ae25-00bd2f426267_1024x768.png 848w, https://substackcdn.com/image/fetch/$s_!pAF8!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffa352174-8d4b-424d-ae25-00bd2f426267_1024x768.png 1272w, https://substackcdn.com/image/fetch/$s_!pAF8!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffa352174-8d4b-424d-ae25-00bd2f426267_1024x768.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!pAF8!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffa352174-8d4b-424d-ae25-00bd2f426267_1024x768.png" width="1024" height="768" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/fa352174-8d4b-424d-ae25-00bd2f426267_1024x768.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:768,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!pAF8!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffa352174-8d4b-424d-ae25-00bd2f426267_1024x768.png 424w, https://substackcdn.com/image/fetch/$s_!pAF8!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffa352174-8d4b-424d-ae25-00bd2f426267_1024x768.png 848w, https://substackcdn.com/image/fetch/$s_!pAF8!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffa352174-8d4b-424d-ae25-00bd2f426267_1024x768.png 1272w, https://substackcdn.com/image/fetch/$s_!pAF8!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffa352174-8d4b-424d-ae25-00bd2f426267_1024x768.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Alex Bornyakov at the European Defense Tech Summit</figcaption></figure></div><p>This is crucial when rapidly-evolving electronic warfare means that equipment usually has a shelf life of about six months on the battlefield.</p><h3>Western start-ups in Ukraine</h3><p>Our other main point around defense concerned the fate of US start-ups, whose enthusiasm for the Ukrainian cause did not always translate into convincing results on the battlefield.&nbsp;</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!-OuX!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42dffd86-e47b-4b77-b8d1-2fce6d442e88_960x540.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!-OuX!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42dffd86-e47b-4b77-b8d1-2fce6d442e88_960x540.png 424w, https://substackcdn.com/image/fetch/$s_!-OuX!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42dffd86-e47b-4b77-b8d1-2fce6d442e88_960x540.png 848w, https://substackcdn.com/image/fetch/$s_!-OuX!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42dffd86-e47b-4b77-b8d1-2fce6d442e88_960x540.png 1272w, https://substackcdn.com/image/fetch/$s_!-OuX!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42dffd86-e47b-4b77-b8d1-2fce6d442e88_960x540.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!-OuX!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42dffd86-e47b-4b77-b8d1-2fce6d442e88_960x540.png" width="960" height="540" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/42dffd86-e47b-4b77-b8d1-2fce6d442e88_960x540.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:540,&quot;width&quot;:960,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!-OuX!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42dffd86-e47b-4b77-b8d1-2fce6d442e88_960x540.png 424w, https://substackcdn.com/image/fetch/$s_!-OuX!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42dffd86-e47b-4b77-b8d1-2fce6d442e88_960x540.png 848w, https://substackcdn.com/image/fetch/$s_!-OuX!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42dffd86-e47b-4b77-b8d1-2fce6d442e88_960x540.png 1272w, https://substackcdn.com/image/fetch/$s_!-OuX!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42dffd86-e47b-4b77-b8d1-2fce6d442e88_960x540.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>It appears we&#8217;re now moving into a new phase. The focus has shifted from attempting to equip the Ukrainian military with drones made overseas, in favor of boosting the domestic Ukrainian industry. At the time of writing, the US is thought to be <a href="https://www.bloomberg.com/news/articles/2024-10-21/zelenskiy-says-us-to-give-ukraine-800m-for-drone-production">preparing</a> an $800M aid package.</p><p>This makes sense for a few reasons.</p><p>Ukrainian teams have a combination of proximity to the frontline and technical ability. Wahid Nawabi, the CEO of US-based Aerovironment, <a href="https://www.defensenews.com/pentagon/2024/10/17/as-ukraine-builds-better-drones-do-american-firms-still-have-a-role/">has acknowledged</a> that <em>&#8220;Ukrainian drone companies in many different domains are going to be a global, legitimate player&#8221;</em>.</p><p>However, this technical talent <a href="https://kyivindependent.com/despite-global-hype-around-ukrainian-weapons-tech-financial-rules-leave-foreign-investors-gun-shy/">often struggles</a> to attract international VC investment; according to Brave1, foreign investment in Ukrainian defense tech start-ups over the past year amounts to a mere $9M. VCs aren&#8217;t used to investing in companies whose facilities can be blown up, especially when the start-up concerned can&#8217;t secure any wartime insurance. The other blocker is the Ukraine Central Bank&#8217;s capital controls, which make transferring money out of the country a logistical nightmare.</p><p>Finally, US-based start-ups simply aren&#8217;t able to match Ukrainian start-ups on price. As well as lower operating costs, Ukrainian companies are able to create &#8216;good enough&#8217; products tailored to the needs of end users.</p><p>Anduril is <a href="https://www.dsca.mil/press-media/major-arms-sales/taipei-economic-and-cultural-representative-office-united-states-34">charging</a> Taiwan close to $1M a unit for its Altius 600 drones, which are propeller-powered and weigh 45 lbs. The costs are also stiff for more established contractors. The Switchblade 600, which the US is purchasing via the Replicator program, comes in at <a href="https://skykam.co.uk/how-much-does-a-switchblade-drone-cost/">over $100k a unit</a>. While these costs are low in the context of US military expenditure, they&#8217;re unworkable for a cash-strapped Ukrainian government.&nbsp;</p><p>Software-only procurement may not be the norm for many countries, but AI-first software companies are gaining some traction in Ukraine by partnering with local manufacturers. While many of these companies don&#8217;t need help building hardware, they often have less in-house AI expertise.&nbsp;</p><p>In Madrid, Lorenz spoke about the team Auterion has built in Ukraine, as well as the speed with which they&#8217;d been able to navigate Ukrainian procurement versus other European countries. This local team, whose software helps drones lock onto targets to counter the effects of electronic warfare, has grown to <a href="https://www.wsj.com/world/drones-in-ukraine-get-smarter-to-dodge-russias-jamming-signals-9ebe3c07">over 20</a> people. Similarly, the company has scored a number of contract wins in the US with the DoD, while remaining stuck in negotiations in Germany, their home market.&nbsp;</p><p>Helsing also appears to be<a href="https://mspu.gov.ua/en/news/ai-in-ukrainian-made-drones-ministry-of-strategic-industries-and-helsing-sign-memorandum-of-understanding"> partnering with local manufacturers</a>, but as ever, their work is shrouded in a degree of mystery (potentially on purpose).</p><h3>Closing thoughts</h3><p>While the defense world looks like a maze to outsiders, the underlying principles are straightforward. It&#8217;s possible to circumvent a broken market, but only if you have the balance sheet to outlast it. Similarly, it's possible to charge the US or European militaries bloated margins if you&#8217;re part of an oligopoly, but don&#8217;t be surprised when a cash-strapped wartime government would rather turn to local innovation to get the job done.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://press.airstreet.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://press.airstreet.com/subscribe?"><span>Subscribe now</span></a></p><p></p>]]></content:encoded></item></channel></rss>