<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[Tech on the Stack]]></title><description><![CDATA[A weekly dive into building AI-driven systems, event-driven architectures, and real-world engineering strategy. Practical patterns, prototypes, and insights from the edge of applied AI.]]></description><link>https://ai.techonthestack.com</link><image><url>https://substackcdn.com/image/fetch/$s_!anbH!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29f65540-4f6f-491a-b9c5-b4705660e124_500x500.png</url><title>Tech on the Stack</title><link>https://ai.techonthestack.com</link></image><generator>Substack</generator><lastBuildDate>Sun, 19 Apr 2026 12:23:50 GMT</lastBuildDate><atom:link href="https://ai.techonthestack.com/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Luca Bianchi]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[techonthestack@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[techonthestack@substack.com]]></itunes:email><itunes:name><![CDATA[Luca Bianchi]]></itunes:name></itunes:owner><itunes:author><![CDATA[Luca Bianchi]]></itunes:author><googleplay:owner><![CDATA[techonthestack@substack.com]]></googleplay:owner><googleplay:email><![CDATA[techonthestack@substack.com]]></googleplay:email><googleplay:author><![CDATA[Luca Bianchi]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[Hello Builders, issue #14]]></title><description><![CDATA[News from the trenches of AI in the week Mar 12-19, 2026]]></description><link>https://ai.techonthestack.com/p/hello-builders-issue-14</link><guid isPermaLink="false">https://ai.techonthestack.com/p/hello-builders-issue-14</guid><dc:creator><![CDATA[Luca Bianchi]]></dc:creator><pubDate>Thu, 19 Mar 2026 23:01:09 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/8d3a225d-8147-4832-a95b-61cc878a51a6_420x300.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Hello Builders,</p><p>This was the week the AI industry was forced to pick sides. <strong>Anthropic told the Pentagon it wouldn&#8217;t remove safety guardrails</strong> on autonomous weapons and domestic surveillance. The Pentagon responded by designating the company a <strong>national security supply chain risk</strong>, a label usually reserved for foreign adversaries. Anthropic sued. The Trump administration doubled down. Google, OpenAI, and Microsoft filed amicus briefs &#8212; in support of Anthropic. Meanwhile, <strong>OpenAI panicked</strong>: the Wall Street Journal reported executives are cutting projects and deprioritizing entire product lines after <strong>Claude Code triggered a trillion-dollar selloff</strong> in SaaS stocks. <strong>Meta delayed its Avocado model</strong> to May after it underperformed every competitor, considered <strong>licensing Google&#8217;s Gemini</strong>, saw a <strong>rogue AI agent expose sensitive internal data</strong>, and is reportedly planning <strong>20% layoffs</strong>. At GTC, <strong>Jensen Huang projected $1 trillion in chip orders</strong> through 2027. And <strong>Yann LeCun</strong>, who left Meta to build what he believes the industry won&#8217;t, raised the <strong>largest seed round in history: $1.03 billion</strong>. The fracture lines aren&#8217;t just widening. They&#8217;re becoming load-bearing walls.</p><h2><strong>This week&#8217;s signal in the noise</strong></h2><ul><li><p><strong>Anthropic defied the Pentagon on autonomous weapons and surveillance; the government blacklisted it as a supply chain risk &#8212; the first time this label has been used against a US AI company</strong></p></li><li><p><strong>OpenAI is cutting projects and deprioritizing product lines as Claude Code and Cowork trigger a trillion-dollar SaaS selloff and executives call it a &#8220;wake-up call&#8221;</strong></p></li><li><p><strong>Meta delays Avocado AI model to May after underperforming GPT-5.4, Gemini 3.1 Pro, and Claude Opus 4.6; considers licensing Google&#8217;s Gemini to fill the gap</strong></p></li><li><p><strong>Nvidia projects $1 trillion in Blackwell and Vera Rubin orders through 2027, declares the &#8220;inference era&#8221; has arrived at GTC 2026</strong></p></li><li><p><strong>Yann LeCun&#8217;s AMI Labs raises $1.03B seed &#8212; the largest ever &#8212; four months after founding, backed by Nvidia, Samsung, Bezos, and Eric Schmidt</strong></p></li></ul><div><hr></div><h3><strong>1. Anthropic vs. The Pentagon: The Line in the Sand That Could Reshape AI</strong></h3><p>Anthropic signed a $200 million Pentagon contract last summer. During later negotiations, the company set explicit red lines: <strong>no mass surveillance of Americans without judicial oversight, and no autonomous weapons targeting without human authorization</strong>. The Pentagon said a private company shouldn&#8217;t dictate military use. Anthropic held firm. On March 3, Defense Secretary <strong>Pete Hegseth designated Anthropic a supply chain risk</strong>, barring all government contractors from using Claude. Anthropic sued in two courts &#8212; California federal court and a DC appeals court &#8212; calling the designation &#8220;unprecedented and unlawful.&#8221; The Trump administration defended the blacklisting in a March 18 filing, arguing Anthropic&#8217;s refusal was &#8220;conduct, not protected speech.&#8221; Then something remarkable happened: <strong>Google, OpenAI, and Microsoft all filed amicus briefs in support of Anthropic</strong>. OpenAI&#8217;s hardware lead <strong>Caitlin Kalinowski resigned</strong> over the Pentagon deal. ChatGPT uninstalls jumped 295%. Claude surpassed ChatGPT in the App Store. For builders: the Anthropic case will set precedent for whether AI companies can set ethical boundaries on government use. Every defense-adjacent contract you sign now carries this political risk.</p><p>Link: <a href="https://www.reuters.com/legal/government/trump-administration-defends-anthropic-blacklisting-us-court-2026-03-18/">https://www.reuters.com/legal/government/trump-administration-defends-anthropic-blacklisting-us-court-2026-03-18/</a></p><div><hr></div><h3><strong>2. OpenAI Panics as Claude Code Triggers a Trillion-Dollar Selloff</strong></h3><p>The Wall Street Journal reported that <strong>OpenAI is cutting projects and deprioritizing entire product lines</strong>. CEO of applications <strong>Fidji Simo</strong> told employees the company is &#8220;actively looking at which areas to deprioritize&#8221; and warned that Claude&#8217;s sudden success should be &#8220;a wake-up call.&#8221; The trigger: <strong>Claude Code and Claude Cowork triggered a trillion-dollar selloff</strong> in SaaS stocks last month, amid fears that agentic AI could make traditional software companies obsolete. OpenAI&#8217;s response is to refocus on coding and enterprise. Codex is now at <strong>2 million weekly active users</strong>, up 4x since January. API usage jumped 20% after GPT-5.4 launched. OpenAI is also forming a <strong>private equity joint venture</strong> to embed engineers inside enterprises, partnering with firms to sell deployment capacity. Anthropic is reportedly in similar talks with <strong>Blackstone</strong>. Current and former employees told the WSJ that OpenAI &#8220;lost much of its focus last year&#8221; while Anthropic shipped relentlessly. For builders: the competitive dynamic has inverted. Anthropic is now the company setting the pace, and OpenAI is playing catch-up on the product that matters most &#8212; coding agents.</p><p>Link: <a href="https://futurism.com/artificial-intelligence/openai-cutting-projects">https://futurism.com/artificial-intelligence/openai-cutting-projects</a></p><div><hr></div><h3><strong>3. Meta&#8217;s Cascading Crisis: Avocado Delayed, Agents Gone Rogue, 20% Layoffs</strong></h3><p>Meta&#8217;s week was a masterclass in compounding failure. The New York Times reported that <strong>Avocado</strong>, Meta&#8217;s next-generation AI model, has been <strong>delayed to May</strong> after internal tests showed it underperforming GPT-5.4, Gemini 3.1 Pro, and Claude Opus 4.6. Meta&#8217;s AI leaders reportedly discussed <strong>temporarily licensing Google&#8217;s Gemini</strong> to power Meta&#8217;s products. Then TechCrunch reported that a <strong>rogue AI agent at Meta exposed sensitive company and user data</strong> to unauthorized employees for two hours &#8212; a &#8220;Sev 1&#8221; incident. This followed a Meta safety director&#8217;s own OpenClaw agent deleting her entire inbox despite instructions to confirm before acting. Meanwhile, Meta acquired <strong>Moltbook</strong>, the AI-agent social network, and launched <strong>Manus</strong>, a desktop AI agent from its acquired startup &#8212; even as agents are demonstrably misbehaving internally. Reports indicate Meta is planning <strong>20% layoffs</strong> (~16,000 employees) and a $27 billion infrastructure deal with Nebius. The company will spend <strong>$135 billion on AI in 2026</strong>. For builders: Meta is spending more than anyone and shipping less than everyone. The lesson is clear &#8212; compute doesn&#8217;t compensate for focus.</p><p>Link: <a href="https://nypost.com/2026/03/13/business/meta-delays-release-of-new-ai-weighs-licensing-googles-gemini-after-disappointing-trial-runs-report/">https://nypost.com/2026/03/13/business/meta-delays-release-of-new-ai-weighs-licensing-googles-gemini-after-disappointing-trial-runs-report/</a></p><div><hr></div><h3><strong>4. Nvidia GTC 2026: $1 Trillion and the Inference Era</strong></h3><p>Jensen Huang&#8217;s GTC keynote lasted 2 hours and 40 minutes and contained one number that mattered: <strong>$1 trillion</strong>. That&#8217;s Nvidia&#8217;s projected order book for Blackwell and Vera Rubin chips through 2027, double the $500 billion projection from last year&#8217;s GTC. The <strong>Vera Rubin platform entered commercial production</strong> with seven new chips, promising 3.5x faster training and 5x faster inference than Blackwell. Huang declared the industry has entered the <strong>&#8220;inference era&#8221;</strong> &#8212; training happens occasionally, inference happens constantly. Nvidia also unveiled <strong>NemoClaw</strong>, an open-source enterprise agent platform, and announced a <strong>$2 billion strategic investment in Nebius</strong>, the AI infrastructure provider. Every AI conversation, code execution, and agent decision triggers inference. Huang&#8217;s argument is simple: if agentic AI becomes the primary interface between humans and software, inference demand becomes effectively infinite. For builders: Nvidia is no longer just selling chips. It&#8217;s selling the thesis that the world needs an infinite amount of intelligence, and it&#8217;s the only company that can manufacture it at scale.</p><p>Link: <a href="https://techcrunch.com/2026/03/16/jensen-just-put-nvidias-blackwell-and-vera-rubin-sales-projections-into-the-1-trillion-stratosphere/">https://techcrunch.com/2026/03/16/jensen-just-put-nvidias-blackwell-and-vera-rubin-sales-projections-into-the-1-trillion-stratosphere/</a></p><div><hr></div><h3><strong>5. Yann LeCun Raises the Largest Seed Round in History</strong></h3><p><strong>AMI Labs</strong> &#8212; Advanced Machine Intelligence, pronounced like the French word for &#8220;friend&#8221; &#8212; announced a <strong>$1.03 billion seed round</strong> on March 10, just four months after founding. It is the largest seed round ever raised by a European company, and possibly anywhere. The company is chaired by <strong>Yann LeCun</strong>, the Turing Award winner who spent 12 years at Meta before departing in November 2025. His founding team includes former Meta AI researchers <strong>Saining Xie, Pascale Fung, and Michael Rabbat</strong>. Strategic investors include <strong>Nvidia, Samsung, Toyota Ventures, Jeff Bezos, and Eric Schmidt</strong>. LeCun&#8217;s thesis: the current LLM paradigm is a dead end for real-world intelligence. AMI is building <strong>&#8220;world models&#8221;</strong> &#8212; systems that understand physics, causality, and spatial reasoning, not just language patterns. The timing is pointed: LeCun left Meta just as it was pouring $135 billion into an LLM strategy that, as of this week, is producing models that can&#8217;t compete. For builders: if LeCun is right, the next paradigm shift in AI isn&#8217;t a bigger language model. It&#8217;s a fundamentally different architecture. Watch this space.</p><p>Link: <a href="https://thenextweb.com/news/europe-startup-funding-rounds-march">https://thenextweb.com/news/europe-startup-funding-rounds-march</a></p><div><hr></div><h3><strong>6. The Mega IPOs That Could Swallow the Entire VC Market</strong></h3><p>PitchBook published a warning shot this week: the potential IPOs of <strong>SpaceX ($1.25 trillion), OpenAI ($840 billion), and Anthropic ($330 billion)</strong> could generate more exit value than <strong>all US VC-backed IPOs combined since the year 2000</strong>. Together, they could raise over $100 billion in proceeds. The problem: these mega-listings could <strong>absorb all available public market capital</strong>, crowding out every other company waiting to go public. More than <strong>$4 trillion</strong> sits locked in unicorns. If even one of these giants stumbles post-listing, the chill could push the broader IPO window into 2027. Morgan Stanley separately warned that the market is &#8220;not prepared for the non-linear increase in LLM capabilities&#8221; expected in April-June, estimating <strong>$2.9 trillion in global data center construction</strong> through 2028. For builders: if you&#8217;re planning to raise or exit in 2026, the window may be defined entirely by whether SpaceX, OpenAI, and Anthropic go public &#8212; and whether they succeed.</p><p>Link: <a href="https://pitchbook.com/news/articles/the-mega-ipos-that-could-shut-out-the-rest-of-vc">https://pitchbook.com/news/articles/the-mega-ipos-that-could-shut-out-the-rest-of-vc</a></p>]]></content:encoded></item><item><title><![CDATA[Strands vs. Claude Agent SDK: Two Very Different Bets on What “Agent” Means]]></title><description><![CDATA[A technical deep-dive into the architectural trade-offs between AWS Strands Agents and the Anthropic Claude Agent SDK, including the real story on running non-Claude models through the Agent SDK.]]></description><link>https://ai.techonthestack.com/p/strands-vs-claude-agent-sdk-two-very</link><guid isPermaLink="false">https://ai.techonthestack.com/p/strands-vs-claude-agent-sdk-two-very</guid><dc:creator><![CDATA[Luca Bianchi]]></dc:creator><pubDate>Thu, 19 Mar 2026 22:50:15 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!3EYu!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd5fb17e2-c34d-4a01-bbba-d5f2e857cce9_1536x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!3EYu!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd5fb17e2-c34d-4a01-bbba-d5f2e857cce9_1536x1024.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!3EYu!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd5fb17e2-c34d-4a01-bbba-d5f2e857cce9_1536x1024.png 424w, https://substackcdn.com/image/fetch/$s_!3EYu!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd5fb17e2-c34d-4a01-bbba-d5f2e857cce9_1536x1024.png 848w, https://substackcdn.com/image/fetch/$s_!3EYu!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd5fb17e2-c34d-4a01-bbba-d5f2e857cce9_1536x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!3EYu!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd5fb17e2-c34d-4a01-bbba-d5f2e857cce9_1536x1024.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!3EYu!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd5fb17e2-c34d-4a01-bbba-d5f2e857cce9_1536x1024.png" width="1456" height="971" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d5fb17e2-c34d-4a01-bbba-d5f2e857cce9_1536x1024.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:971,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1949525,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://ai.techonthestack.com/i/191527751?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd5fb17e2-c34d-4a01-bbba-d5f2e857cce9_1536x1024.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!3EYu!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd5fb17e2-c34d-4a01-bbba-d5f2e857cce9_1536x1024.png 424w, https://substackcdn.com/image/fetch/$s_!3EYu!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd5fb17e2-c34d-4a01-bbba-d5f2e857cce9_1536x1024.png 848w, https://substackcdn.com/image/fetch/$s_!3EYu!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd5fb17e2-c34d-4a01-bbba-d5f2e857cce9_1536x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!3EYu!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd5fb17e2-c34d-4a01-bbba-d5f2e857cce9_1536x1024.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Most comparisons lay these two side by side in neat feature tables. That&#8217;s useful but insufficient, because the frameworks don&#8217;t just differ in features. They differ in their assumptions about your problem. Pick the wrong one, and you&#8217;ll spend weeks building plumbing that the other gives you for free, or discover your architecture can&#8217;t flex in the direction your product needs.</p><p>This article breaks down the real architectural differences, compares them feature by feature (with attention to TypeScript support), and digs into the question that keeps coming up: can you actually run non-Claude models through the Claude Agent SDK?</p><h2><strong>Orchestrator vs. Runtime</strong></h2><p>There is a subtle but significant difference in approach between Strands and the Claude Agent SDK. The latter gives your agent a brain and a dispatch loop. You build the hands, providing the LLM with a prompt, tool definitions (name, description, input schema), and a system prompt. The model decides which tools to call and in what order. The framework manages the loop: call model &#8594; model picks a tool &#8594; framework invokes your callback &#8594; feeds result back &#8594; repeat. The keyword is <em>your callback</em>. The tool definition says <code>get_weather</code> takes a location string, but you&#8217;re responsible for wiring that to an actual HTTP call, handling errors, and returning the result.</p><p>On the other hand Claude Agent SDK gives your agent a brain, hands, a workbench, and a filing cabinet. Your job is to decide what is not allowed to be touched. The SDK ships with pre-built, executable tools: <code>Read</code>, <code>Write</code>, <code>Edit</code>, <code>MultiEdit</code>, <code>Bash</code>, <code>Glob</code>, <code>Grep</code>, <code>WebSearch</code>, <code>WebFetch</code>, <code>AskUserQuestion</code>, <code>Agent</code> (subagent spawning), and <code>NotebookEdit</code>. When Claude decides to read a file, the SDK reads that file from the filesystem. When it decides to run <code>grep -r "TODO" ./src</code>, a real shell executes that command and feeds stdout back. This is the same infrastructure that powers Claude Code, repackaged as a programmable library.</p><p>The practical consequence shows up during debugging. A Strands agent calling a proprietary API gives you normal debugging: step through the callback, inspect the raw HTTP response, and add retries. A misbehaving Claude Agent SDK agent means figuring out what Claude <em>thought</em> the filesystem looked like versus what it actually looked like. Different kind of problem.</p><p>With Strands, the agent starts with zero capabilities. The agent does not have &#8220;agency&#8221;: You add tools one by one. With the Claude Agent SDK, the agent starts with a shell, a filesystem, and web access. Your job is to <em>restrict</em> what it can touch. Miss a permission? The agent might execute something you didn&#8217;t expect. The <code>allowed_tools</code> list and permission modes help, but the mental model is fundamentally different: you&#8217;re not building a whitelist, you&#8217;re maintaining a blacklist. And blacklists have gaps.</p><p>This means Claude Agent SDK workloads should run in sandboxed containers, gVisor at minimum, not Docker alone. The performance overhead is real (roughly 15-20% on I/O-heavy operations), but the alternative is trusting that you&#8217;ve correctly anticipated every shell command a sufficiently creative LLM might construct. Firecracker would provide even stronger isolation, a true microVM, and a minimal attack surface, but the cold-start latency (~125ms, which compounds in agentic loops with dozens of tool calls) makes it impractical for interactive workloads. Batch processing is a different story.</p><p>Feature matrices date fast, and half these checkboxes will be different in six months. Treat this as a snapshot.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!PPx3!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77f5e0e2-739e-41c2-ad02-803645214ca1_1794x960.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!PPx3!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77f5e0e2-739e-41c2-ad02-803645214ca1_1794x960.png 424w, https://substackcdn.com/image/fetch/$s_!PPx3!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77f5e0e2-739e-41c2-ad02-803645214ca1_1794x960.png 848w, https://substackcdn.com/image/fetch/$s_!PPx3!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77f5e0e2-739e-41c2-ad02-803645214ca1_1794x960.png 1272w, https://substackcdn.com/image/fetch/$s_!PPx3!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77f5e0e2-739e-41c2-ad02-803645214ca1_1794x960.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!PPx3!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77f5e0e2-739e-41c2-ad02-803645214ca1_1794x960.png" width="1456" height="779" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/77f5e0e2-739e-41c2-ad02-803645214ca1_1794x960.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:779,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:204207,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://ai.techonthestack.com/i/191527751?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77f5e0e2-739e-41c2-ad02-803645214ca1_1794x960.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!PPx3!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77f5e0e2-739e-41c2-ad02-803645214ca1_1794x960.png 424w, https://substackcdn.com/image/fetch/$s_!PPx3!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77f5e0e2-739e-41c2-ad02-803645214ca1_1794x960.png 848w, https://substackcdn.com/image/fetch/$s_!PPx3!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77f5e0e2-739e-41c2-ad02-803645214ca1_1794x960.png 1272w, https://substackcdn.com/image/fetch/$s_!PPx3!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77f5e0e2-739e-41c2-ad02-803645214ca1_1794x960.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>A relevant point on MCP: both claim first-class support, both deliver it, but they feel different in practice: Strands treats MCP tools the same as any other tool definition, registers the server, and tools appear in the list. The Claude Agent SDK integrates MCP more deeply because its tools are already executable; MCP servers plug into the same execution pipeline as built-in tools. For simple tool schemas, this distinction doesn&#8217;t matter. For tools with complex input validation, the Claude Agent SDK&#8217;s execution model handles it more gracefully.</p><p>Strands offers Graph, Swarm, Workflow, and Agent-as-Tool. Different agents can use different models; this matters for cost optimization: cheap models for classification and routing, expensive ones for reasoning. The Claude Agent SDK provides subagents with isolated context windows, all of which run Claude. Simpler to reason about, inflexible on model choice.</p><p>On the observability side, Strands has OpenTelemetry support only for Python; the TS SDK is missing it, which is frustrating if you&#8217;re a TypeScript shop. The Claude Agent SDK includes built-in cost tracking and lifecycle hooks (PreToolUse, PostToolUse, SubagentStart/Stop) to wire in custom telemetry. More assembly required, but more flexible.</p><p>The Claude Agent SDK ships a <code>max_budget_usd</code> parameter that caps spend per session. Strands has no equivalent; you&#8217;ll need to build a token-counting middleware yourself. Not hard, but the kind of thing a framework should handle.</p><p>Strands integrates with Lambda, Fargate, EKS, and recently AgentCore. The Claude Agent SDK assumes sandboxed containers. Running both on EKS is feasible, but Claude Agent SDK pods need the gVisor runtime class, which means separate node pool configuration.</p><p>A special focus should be placed on TypeScript-first support; the comparison is lopsided. Strands TS SDK v0.5.0 is marked experimental. The core agent loop works and is usable, but the official docs list these as Python-only: Anthropic/Ollama/LiteLLM model providers, summarizing conversation manager, multi-agent patterns, session persistence, OpenTelemetry, the Evals SDK, and the community tools package. The GitHub README hints that some of these (structured output, Graph/Swarm) have landed in TS recently, but the docs haven&#8217;t caught up. There are known edge cases. Graph mode with parallel nodes can deadlock when concurrent nodes write to the same session store. On the other side, Claude Agent SDK TS (<code>@anthropic-ai/claude-agent-sdk</code>, v0.2.71) has full feature parity with the Python version. Hooks, custom MCP tools, subagents, streaming, sessions, the complete tool suite, all there. Porting an agent from Python to TS requires only syntax changes. That&#8217;s rare in this ecosystem.</p><p>For TypeScript teams that need production-ready agent capabilities now, the gap is hard to ignore.</p><h2><strong>Running Non-Claude Models Through the Claude Agent SDK</strong></h2><p>It is an often-asked question, with a mixed response from the community: &#8220;Can I use Claude Agent SDK with non-Claude models?&#8221; Short answer: You can; it mostly works. Don&#8217;t do it in production. The Claude Agent SDK officially supports only Claude models through four backends: the Anthropic API, Amazon Bedrock, Google Vertex AI, and Microsoft Azure AI Foundry. There is no native abstraction for model providers. The SDK spawns a Claude Code CLI subprocess that expects the Anthropic Messages API format.</p><p>There are different approaches to running non-Claude models, but they are all workarounds: Ollama&#8217;s Anthropic API compatibility layer, overriding ANTHROPIC_BASE_URL to point to Ollama, setting a dummy auth token, and the agent loop runs. Models tested include qwen3.5, glm-5:cloud, and kimi-k2.5:cloud. Ollama recommends a 64k context minimum, which significantly limits model choices. Another solution is to use the LiteLLM proxy, which translates the Anthropic Messages API to any backend provider (OpenAI, Azure, Gemini, Bedrock, Ollama, 100+ others). More flexible than Ollama&#8217;s built-in compatibility. Point <code>ANTHROPIC_BASE_URL</code> at the proxy, and the same agent code works with any model. Another way could be using Claude Code Router (a community project), which enables dynamic model switching within sessions. Supports OpenRouter, DeepSeek, Ollama, and Gemini.</p><h3><strong>Nice ideas, but they actually break in production</strong></h3><p>Tool-calling reliability degrades noticeably. In benchmarks comparing Claude Sonnet 4 against Qwen3.5 via Ollama compatibility on identical task sets (file operations, grep searches, multi-step bash workflows), Claude completes ~95% of tasks correctly, while Qwen completes around 60%, with most failures in multi-step bash tasks where it constructs syntactically valid but semantically incorrect commands. The Claude Agent SDK faithfully executes those incorrect commands. Powerful runtime plus mediocre reasoning equals a larger blast radius, arguably worse than the same model running in a simpler framework.</p><p>Extended thinking is very limited outside of Claude. If agent workflows depend on it (complex code review, deep research, multi-step planning), there&#8217;s no workaround. The agent often runs without extended thinking capabilities, and it makes worse decisions at the planning stage.</p><p>Subagent orchestration assumes Claude-level instruction following. Permission hooks behave unpredictably with weaker models; the model occasionally tries to use disallowed tools, the SDK blocks it correctly, but the recovery behavior (trying an alternative approach) isn&#8217;t as graceful as with Claude.</p><p>The bottom line is that using other models than Claude is technically feasible via proxy layers, but architecturally unsound for production. The SDK is a Claude-specific runtime being adapted through API translation, not a model-agnostic framework.</p><h2><strong>When to Use Which</strong></h2><p>We found that Strands is the right choice for domain-specific agents. Agents calling proprietary APIs (legal corpora, scoring engines, internal services), multi-model architectures where different agents use different LLMs, edge deployment with local models, AWS-native event-driven patterns (Lambda, EventBridge, S3 triggers), and situations where Apache 2.0 licensing or vendor independence are requirements. The cost flexibility alone, swapping an expensive model for a cheap one on a per-agent basis, justifies the choice for most enterprise use cases. On the other hand, the Claude Agent SDK is the right choice when agents need to interact with a computing environment. Code review, codebase analysis, research tasks involving document reading and web search, and any workflow where the agent needs to read, write, edit, and execute. The built-in tools eliminate enormous amounts of boilerplate. Replicating the <code>Edit</code> tool&#8217;s diff-based file editing with conflict detection in Strands is a multi-day effort; it&#8217;s not trivial.</p><p>Sometimes, having both in the same system makes sense: when a pipeline has domain-specific ingestion and classification (Strands, multiple models, custom tools) feeding into compute-heavy analysis and report generation (Claude Agent SDK, file I/O, shell access, web search). The agent layers communicate via MCP. It works, but the operational overhead of two runtimes is real. Don&#8217;t add this complexity unless you genuinely need both capability sets. Note that AWS demonstrated a similar pattern in their financial analysis reference architecture: LangGraph for workflow orchestration, Strands for reasoning within workflow nodes.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!OxGl!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F97352e16-06c6-435f-bfef-70889bc87acf_1784x760.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!OxGl!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F97352e16-06c6-435f-bfef-70889bc87acf_1784x760.png 424w, https://substackcdn.com/image/fetch/$s_!OxGl!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F97352e16-06c6-435f-bfef-70889bc87acf_1784x760.png 848w, https://substackcdn.com/image/fetch/$s_!OxGl!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F97352e16-06c6-435f-bfef-70889bc87acf_1784x760.png 1272w, https://substackcdn.com/image/fetch/$s_!OxGl!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F97352e16-06c6-435f-bfef-70889bc87acf_1784x760.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!OxGl!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F97352e16-06c6-435f-bfef-70889bc87acf_1784x760.png" width="1456" height="620" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/97352e16-06c6-435f-bfef-70889bc87acf_1784x760.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:620,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:148737,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://ai.techonthestack.com/i/191527751?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F97352e16-06c6-435f-bfef-70889bc87acf_1784x760.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!OxGl!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F97352e16-06c6-435f-bfef-70889bc87acf_1784x760.png 424w, https://substackcdn.com/image/fetch/$s_!OxGl!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F97352e16-06c6-435f-bfef-70889bc87acf_1784x760.png 848w, https://substackcdn.com/image/fetch/$s_!OxGl!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F97352e16-06c6-435f-bfef-70889bc87acf_1784x760.png 1272w, https://substackcdn.com/image/fetch/$s_!OxGl!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F97352e16-06c6-435f-bfef-70889bc87acf_1784x760.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>One thing the table doesn&#8217;t capture: the Claude Agent SDK&#8217;s proprietary license means you can&#8217;t fork it if Anthropic changes terms or pricing. With Strands, worst case, you have the source code. In enterprise procurement, this distinction matters more than most engineers think.</p><h2><strong>Where to go from here</strong></h2><p>Both frameworks have native MCP support, which means tools built for one are increasingly portable to the other. Neither has A2A (agent-to-agent protocol) yet, CrewAI does, for what it&#8217;s worth.</p><p>The practical takeaway: write all new tools as MCP servers regardless of which framework consumes them. If Strands need to be swapped for something else in a year, the tools survive. If the Claude Agent SDK adds model-agnostic support (unlikely but not impossible), the tools still survive.</p><p>The framework is a replaceable component. The tools and the domain logic aren&#8217;t. Invest accordingly.</p><div><hr></div><p><em>Subscribe to <strong><a href="http://ai.techonthestack.com">TechOnTheStack</a></strong> for weekly deep-dives on cloud architecture, AI infrastructure, and hands-on engineering decisions. No fluff, just what works and what doesn&#8217;t.</em></p>]]></content:encoded></item><item><title><![CDATA[Hello Builders, issue #13]]></title><description><![CDATA[News from the trenches of AI in the week Feb 17-23, 2026]]></description><link>https://ai.techonthestack.com/p/hello-builders-issue-13</link><guid isPermaLink="false">https://ai.techonthestack.com/p/hello-builders-issue-13</guid><dc:creator><![CDATA[Luca Bianchi]]></dc:creator><pubDate>Tue, 24 Feb 2026 21:49:58 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/3bf39c1a-c0a4-46ee-9cdb-8fbb98744d38_420x300.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Hello Builders,</p><p>This was the week the AI cold war went hot. <strong>Anthropic publicly accused DeepSeek, Moonshot, and MiniMax of industrial-scale distillation</strong> &#8212; 24,000 fake accounts, 16 million exchanges, proxy networks to evade detection. Not a rumor, not a leak: a formal accusation framed as national security. OpenAI followed with its own allegations to Congress. Meanwhile, Google did something that may matter more than any model release this year: <strong>Chrome 146 shipped with native WebMCP support</strong>, turning every website into a potential AI agent endpoint. OpenAI launched <strong>Frontier</strong>, an enterprise platform for &#8220;AI coworkers,&#8221; and signed alliances with <strong>McKinsey, BCG, Accenture, and Capgemini</strong> to deploy it. Meta signed a <strong>multi-billion-dollar GPU deal with Nvidia</strong> covering millions of chips. And every major AI CEO flew to Delhi for India&#8217;s AI Summit, where <strong>billions in deals were announced</strong> in four days. The pattern is unmistakable: AI is no longer a technology race. It&#8217;s an infrastructure war, a trade war, and a regulatory war &#8212; all at once.</p><h2><strong>This week&#8217;s signal in the noise</strong></h2><p>&#8226; Anthropic accuses DeepSeek, Moonshot, and MiniMax of running 24,000 fake accounts to distill Claude&#8217;s capabilities at an industrial scale<br>&#8226; Google ships WebMCP in Chrome 146, creating a native protocol for AI agents to interact with any website<br>&#8226; OpenAI launches Frontier, an enterprise platform for &#8220;AI coworkers,&#8221; and signs multi-year alliances with McKinsey, BCG, Accenture, and Capgemini<br>&#8226; Meta expands Nvidia partnership to deploy millions of GPUs and next-gen Vera Rubin systems in a deal worth tens of billions<br>&#8226; India AI Summit draws Altman, Amodei, Pichai, and Huang, with billions in infrastructure commitments announced</p><div><hr></div><h3><strong>1. Anthropic vs. China: 24,000 Fake Accounts and a National Security Accusation</strong></h3><p>Anthropic dropped a bombshell on Sunday: <strong>DeepSeek, Moonshot AI, and MiniMax</strong> ran coordinated campaigns using 24,000 fraudulent accounts and proxy networks to extract Claude&#8217;s outputs and train their own models. Anthropic generated 16 million exchanges across these campaigns and called it &#8220;industrial-scale.&#8221; DeepSeek specifically sought to create &#8220;censorship-safe alternatives to policy-sensitive queries.&#8221; Anthropic framed the disclosure as a national security issue, timed to coincide with Washington's debate over AI chip export controls. OpenAI had made similar allegations about ChatGPT to the House Select Committee earlier this month. When the two fiercest competitors in AI agree on something, the signal is deafening. For builders: distillation defense is now a first-class engineering concern. If you&#8217;re serving frontier model outputs via API, your responses are a target.</p><p>Link: <a href="https://techcrunch.com/2026/02/23/anthropic-accuses-chinese-ai-labs-of-mining-claude-as-us-debates-ai-chip-exports/">https://techcrunch.com/2026/02/23/anthropic-accuses-chinese-ai-labs-of-mining-claude-as-us-debates-ai-chip-exports/</a></p><div><hr></div><h3><strong>2. Google Ships WebMCP: The Most Important Release No One Is Talking About</strong></h3><p>Chrome 146 shipped with native <strong>WebMCP (Web Model Context Protocol)</strong> support &#8212; a standard that lets AI agents interact with websites via structured &#8220;Tool Contracts&#8221; rather than fragile screen scraping. Websites can now declare what actions agents are allowed to take: book a flight, add to cart, submit a form. The browser becomes the universal agent runtime. If WebMCP gains adoption, every website effectively becomes an MCP server. This is what the &#8220;agentic web&#8221; actually looks like &#8212; not chatbots slapped onto homepages, but a fundamentally new interaction layer where AI negotiates with websites on your behalf. For builders: start implementing WebMCP endpoints now. The companies that expose structured agent interfaces first will capture the agentic traffic wave that&#8217;s building.</p><p>Link: <a href="https://www.pymnts.com/artificial-intelligence-2/2026/google-introduces-webmcp-to-give-browser-access-for-ai-agents/">https://www.pymnts.com/artificial-intelligence-2/2026/google-introduces-webmcp-to-give-browser-access-for-ai-agents/</a></p><div><hr></div><h3><strong>3. OpenAI Launches Frontier: AI Coworkers Get Onboarded Like Employees</strong></h3><p>OpenAI launched <strong>Frontier</strong>, an enterprise platform for building, deploying, and managing &#8220;AI coworkers&#8221; &#8212; complete with onboarding, permissions, identity governance, and audit trails. Think of it as HR for agents. Then, on Sunday, OpenAI signed <strong>multi-year alliances with McKinsey, BCG, Accenture, and Capgemini</strong> to deploy Frontier across their enterprise clients. Brad Lightcap said the quiet part out loud: &#8220;The limiting factor for seeing value from AI in enterprises isn&#8217;t model intelligence &#8212; it&#8217;s how agents are built and run in their organizations.&#8221; This is OpenAI&#8217;s boldest enterprise move yet. They&#8217;re not selling APIs anymore; they&#8217;re selling a managed agent workforce. The consulting partnerships are the delivery mechanism&#8212;the Big Four serve as OpenAI&#8217;s implementation layer. For builders: if McKinsey and Accenture are deploying OpenAI&#8217;s agents, they&#8217;re coming for enterprise workflows at scale. The SaaS disruption everyone predicted is now being actively orchestrated.</p><p>Link: <a href="https://www.cnbc.com/2026/02/23/open-ai-consulting-accenture-boston-capgemini-mckinsey-frontier.html">https://www.cnbc.com/2026/02/23/open-ai-consulting-accenture-boston-capgemini-mckinsey-frontier.html</a></p><div><hr></div><h3><strong>4. Meta Goes All-In: Millions of Nvidia GPUs, $130B AI Budget</strong></h3><p>Meta expanded its Nvidia partnership in a deal worth <strong>tens of billions</strong>, deploying millions of GPUs &#8212; including Nvidia&#8217;s new standalone CPUs and next-generation Vera Rubin systems &#8212; across its data center fleet. Jensen Huang called it &#8220;bringing the full NVIDIA platform to Meta.&#8221; The company&#8217;s total AI spend for 2026: <strong>$115-135 billion</strong> across 30 new data centers. Meta AI already has 1 billion users. Zuckerberg is betting that whoever controls the most compute wins the consumer AI race. Combined with Meta&#8217;s open-source strategy through Llama, this creates a unique position: the largest compute footprint in AI, powering the largest open-weight model ecosystem. For builders: the open-weight ecosystem just got a massive infrastructure boost. Plan for a world where the best open models are genuinely competitive with proprietary ones.</p><p>Link: <a href="https://www.cnbc.com/2026/02/17/meta-nvidia-deal-ai-data-center-chips.html">https://www.cnbc.com/2026/02/17/meta-nvidia-deal-ai-data-center-chips.html</a></p><div><hr></div><h3><strong>5. India AI Summit: Every CEO at Modi&#8217;s Court</strong></h3><p>Every major AI leader flew to New Delhi this week. <strong>Sam Altman, Dario Amodei, Sundar Pichai, and Jensen Huang</strong> all attended the four-day AI Impact Summit and met directly with Prime Minister Modi. The commitments poured in: L&amp;T partnered with Nvidia for AI data center infrastructure, India&#8217;s payments network adopted Nvidia&#8217;s Nemotron for sovereign AI, OpenAI struck a platform deal with Zomato, and Google DeepMind announced a partnership with the Indian government for science and education. Pichai publicly defended AI spending against bubble fears. The message was unanimous: India isn&#8217;t just a market &#8212; it&#8217;s the next strategic battleground, sitting between the US and China with 1.4 billion people and accelerating digital infrastructure. For builders: the global south is where the next billion AI users come from. Build for multilingual, mobile-first, cost-sensitive deployment.</p><p>Link: <a href="https://techcrunch.com/2026/02/22/all-the-important-news-from-the-ongoing-india-ai-summit/">https://techcrunch.com/2026/02/22/all-the-important-news-from-the-ongoing-india-ai-summit/</a></p><div><hr></div><h3><strong>6. The Year of the AI Bill: US States Launch a Regulatory Flood</strong></h3><p>While the industry raised money and fought geopolitical battles, <strong>US states launched an unprecedented wave of AI legislation</strong>. Alabama signed the App Store Accountability Act, targeting digital child safety. Idaho&#8217;s Conversational AI Safety Act advanced through committee with a do-pass recommendation. Colorado&#8217;s comprehensive AI Act takes effect June 30, requiring &#8220;reasonable care&#8221; to prevent algorithmic discrimination. At the federal level, Anthropic and OpenAI&#8217;s distillation accusations are feeding the push for stricter export controls. The EU AI Act&#8217;s prohibited practices provisions are already in force. The regulatory landscape is no longer hypothetical &#8212; it&#8217;s operational. For builders: if you&#8217;re deploying AI in production, you need a compliance strategy now, especially for high-risk use cases in hiring, lending, healthcare, and any consumer-facing application.</p><p>Link: <a href="https://www.compliancehub.wiki/is-2026-the-year-of-the-chatbot-bill-a-state-by-state-ai-legislation-roundup/">https://www.compliancehub.wiki/is-2026-the-year-of-the-chatbot-bill-a-state-by-state-ai-legislation-roundup/</a></p>]]></content:encoded></item><item><title><![CDATA[Hello Builders, issue #12]]></title><description><![CDATA[News from the trenches of AI in the week Feb 11-18, 2026]]></description><link>https://ai.techonthestack.com/p/hello-builders-issue-12</link><guid isPermaLink="false">https://ai.techonthestack.com/p/hello-builders-issue-12</guid><dc:creator><![CDATA[Luca Bianchi]]></dc:creator><pubDate>Wed, 18 Feb 2026 22:15:23 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/4c68cf2f-0646-4d98-aa15-516a53b79a8c_1056x754.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Hello Builders,</p><p>This week, the money doubled, the code wrote itself, and the people who built it started walking away. <strong>Anthropic closed a $30 billion Series G</strong> at a $380 billion valuation, more than doubling its worth in five months and surpassing OpenAI in total capital raised. <strong>Spotify&#8217;s CEO revealed that the company&#8217;s best developers haven&#8217;t written a single line of code since December</strong>, handing everything to AI agents. <strong>OpenAI hired Peter Steinberger</strong>, the creator of OpenClaw, the open-source personal agent with 196,000 GitHub stars and 2 million weekly users, signaling that 2026 is the year of the personal agent. Meanwhile, <strong>Matt Shumer&#8217;s essay &#8220;Something Big Is Coming&#8221; hit 70 million views</strong> in 36 hours, and <strong>safety researchers fled both OpenAI and Anthropic</strong>, warning that the labs are building systems they can&#8217;t control. <strong>ByteDance&#8217;s Seedance 2.0 went viral</strong>, prompting legal threats from Disney and Paramount. Mistral AI made its first acquisition, buying Koyeb to become a full-stack AI contender in Europe. The gap between AI&#8217;s capabilities and humanity&#8217;s readiness is now a chasm.</p><h2><strong>This week&#8217;s signal in the noise</strong></h2><ul><li><p><strong>Anthropic $30B at $380B</strong>: Largest private AI round ever. Revenue at $14B run-rate. Over 30 investors, VCs now openly backing both OpenAI and Anthropic.</p></li><li><p><strong>Spotify quits coding</strong>: CEO says top engineers haven&#8217;t written code in 2026. AI handles deployment end-to-end.</p></li><li><p>**OpenAI acquires OpenClaw: Peter Steinberger joins, OpenClaw becomes an open-source foundation. Altman: The future is &#8220;extremely multi-agent.&#8221;</p></li><li><p><strong>AI existential panic</strong>: 70M-view viral essay, researcher departures, OpenAI dismantled its alignment team. The Guardian: &#8220;industry pursuing profit at all costs.&#8221;</p></li><li><p><strong>Seedance 2.0 vs Hollywood</strong>: ByteDance&#8217;s video model generates cinematic clips, goes viral, triggers copyright crackdown from Disney and Paramount.</p></li></ul><div><hr></div><h3><strong>1. Anthropic Closes $30 Billion Round, Surpasses OpenAI in Total Capital Raised</strong></h3><p>Anthropic announced a $30 billion Series G on February 12, valuing the company at $380 billion, more than double the $183 billion from its Series F in September. The round was led by GIC and Coatue, with co-leads from Founders Fund, D.E. Shaw Ventures, ICONIQ, Dragoneer, and MGX. NVIDIA, Microsoft, Sequoia, and Qatar&#8217;s sovereign wealth fund also participated. Anthropic reported $14 billion in run-rate revenue, growing more than 10x annually for three consecutive years. Eight of the Fortune 10 now use Claude. Bloomberg reported that VCs are breaking a longstanding Silicon Valley taboo by openly investing in both Anthropic and OpenAI, with Sequoia and Altimeter backing both companies simultaneously. Founders Fund, previously an OpenAI backer, joined Anthropic&#8217;s round. OpenAI is reportedly planning a $100 billion raise of its own. For builders: the funding war is over. Both labs now have enough capital to sustain multi-year compute buildouts. The question is no longer who can raise more, but who can ship faster.</p><p>Link: <a href="https://pitchbook.com/news/articles/anthropic-surpasses-openai-in-ai-cash-race-after-raising-30b-series-g">https://pitchbook.com/news/articles/anthropic-surpasses-openai-in-ai-cash-race-after-raising-30b-series-g</a></p><div><hr></div><h3><strong>2. Spotify&#8217;s Best Developers Haven&#8217;t Written Code Since December</strong></h3><p>Spotify CEO Daniel Ek dropped a bombshell on February 12: the company&#8217;s top developers &#8220;have not written a single line of code&#8221; in 2026. AI now handles the entire coding and deployment pipeline. Ek described a workflow in which engineers define the desired outcome, and AI agents write, test, iterate, and deploy code autonomously. This isn&#8217;t an experiment; it&#8217;s Spotify&#8217;s production reality. The statement went viral, landing in TechCrunch&#8217;s most-read stories of the week. It echoes what Matt Shumer described in his viral essay: &#8220;A few months ago, I was constantly deliberating with the AI. Now I just describe the result and leave.&#8221; Spotify isn&#8217;t alone. ServiceNow, Infosys, and multiple Fortune 500 companies are reporting similar patterns. For builders: if Spotify&#8217;s best engineers are directing AI rather than writing code, the job title &#8220;software engineer&#8221; is being redefined in real time.</p><p>Link: <a href="https://techcrunch.com/2026/02/12/spotify-says-its-best-developers-havent-written-a-line-of-code-since-december-thanks-to-ai/">https://techcrunch.com/2026/02/12/spotify-says-its-best-developers-havent-written-a-line-of-code-since-december-thanks-to-ai/</a></p><div><hr></div><h3><strong>3. OpenAI Hires OpenClaw Creator, Declares &#8220;Year of the Agent.&#8221;</strong></h3><p>OpenAI hired Peter Steinberger, creator of OpenClaw, the open-source personal AI agent that amassed 196,000 GitHub stars and 2 million weekly users. Steinberger announced on X that OpenClaw will transition to an independent foundation while he joins OpenAI to bring personal agents to a broad audience. Sam Altman said the future will be &#8220;extremely multi-agent,&#8221; comparing OpenClaw-style agents to browsers: tools that democratize access. The irony is sharp: OpenClaw was built on Anthropic&#8217;s Claude API. Steinberger chose OpenAI anyway, citing scale. The hire signals OpenAI&#8217;s pivot from central model provider to agent platform owner. Investors had offered to fund OpenClaw as a standalone company, but Steinberger opted for partnership over independence. For builders: personal agents that schedule, shop, code, and coordinate are moving from novelty to infrastructure. If you&#8217;re building in this space, the platform race just got serious.</p><p>Link: <a href="https://www.theverge.com/ai-artificial-intelligence/879623/openclaw-founder-peter-steinberger-joins-openai">https://www.theverge.com/ai-artificial-intelligence/879623/openclaw-founder-peter-steinberger-joins-openai</a></p><div><hr></div><h3><strong>4. &#8220;Something Big Is Coming&#8221;: The AI Existential Panic Goes Mainstream</strong></h3><p>Matt Shumer&#8217;s essay about AI replacing knowledge work gathered 70 million views in 36 hours, becoming the most-read tech essay since Marc Andreessen&#8217;s &#8220;Software Is Eating the World.&#8221; Shumer, who builds products with AI daily, wrote that GPT-5.3-Codex now shows &#8220;judgment&#8221; and &#8220;taste,&#8221; capabilities humans claimed AI would never possess. The essay went viral the same week an Anthropic researcher quit to write poetry about &#8220;the place we find ourselves,&#8221; an OpenAI researcher left citing ethical concerns, and OpenAI quietly dismantled its mission alignment team. Jason Calacanis wrote: &#8220;I&#8217;ve never seen so many technologists state their concerns so strongly.&#8221; The Guardian published an editorial warning that AI safety departures signal &#8220;industry pursuing profit at all costs.&#8221; Axios reported that Anthropic&#8217;s own sabotage report confirmed AI can assist in chemical weapons creation. For builders: the existential conversation has left the labs and entered the mainstream. Your users, clients, and regulators are reading Shumer&#8217;s essay. Have an answer ready.</p><p>Link: <a href="https://www.businessinsider.com/matt-shumer-something-big-is-happening-essay-ai-disruption-2026-2">https://www.businessinsider.com/matt-shumer-something-big-is-happening-essay-ai-disruption-2026-2</a></p><div><hr></div><h3><strong>5. ByteDance&#8217;s Seedance 2.0 Goes Viral, Hollywood Fights Back</strong></h3><p>ByteDance released Seedance 2.0 on February 13, a video-generation AI model that produces cinematic-quality clips from text prompts. The model went mega-viral on Chinese social media and X, with Elon Musk himself praising the results. Comparisons to DeepSeek&#8217;s January 2025 breakthrough immediately followed. But Hollywood responded with legal force: Disney and Paramount sent legal threats after users generated clips featuring Tom Cruise, Brad Pitt, and copyrighted characters. ByteDance announced it would add safeguards and crack down on copyright misuse. Meanwhile, a dozen other Chinese AI firms released models during Spring Festival week, including iFlytek&#8217;s Spark X2 (trained entirely on Chinese chips), Alibaba&#8217;s forthcoming Qwen 3.5, and NetEase Youdao&#8217;s desktop agent LobsterAI. Reuters called it a scramble to steal the spotlight ahead of DeepSeek V4&#8217;s launch. For builders: Chinese AI is no longer a single company. It&#8217;s an ecosystem, and it just shipped a Hollywood-threatening video model while the West debated existential risk.</p><p>Link: <a href="https://www.reuters.com/business/media-telecom/bytedances-new-ai-video-model-goes-viral-china-looks-second-deepseek-moment-2026-02-12/">https://www.reuters.com/business/media-telecom/bytedances-new-ai-video-model-goes-viral-china-looks-second-deepseek-moment-2026-02-12/</a></p><div><hr></div><h3><strong>6. Mistral AI Makes First Acquisition, Goes Full-Stack</strong></h3><p>Mistral AI, the French company valued at $13.8 billion, acquired Paris-based Koyeb, a startup that simplifies AI app deployment at scale. The deal confirms Mistral&#8217;s ambition to move beyond LLM development into full-stack AI infrastructure. Mistral launched its cloud offering Mistral Compute in June 2025, and Koyeb accelerates that push. CEO Arthur Mensch pitched Mistral at Stockholm&#8217;s Techarena as &#8220;headquartered in Europe, doing frontier research in Europe.&#8221; Koyeb&#8217;s investors celebrated the deal as a step toward building &#8220;sovereign AI infrastructure in Europe.&#8221; Mistral recently passed $400 million in annual recurring revenue. For builders: Europe&#8217;s AI champion is going vertical. If you&#8217;re building on Mistral&#8217;s models, expect a more integrated, AWS-like experience. If you&#8217;re competing with them, the window to differentiate has just narrowed.</p><p>Link: <a href="https://techcrunch.com/2026/02/17/mistral-ai-buys-koyeb-in-first-acquisition-to-back-its-cloud-ambitions/">https://techcrunch.com/2026/02/17/mistral-ai-buys-koyeb-in-first-acquisition-to-back-its-cloud-ambitions/</a></p><div><hr></div><h3><strong>7. Infosys Partners with Anthropic to Deploy AI Agents Across Regulated Industries</strong></h3><p>Infosys announced a strategic collaboration with Anthropic on February 17 to develop enterprise AI solutions across telecom, financial services, manufacturing, and software development. The partnership will begin with a dedicated Anthropic Center of Excellence and expand into agentic AI systems built on the Claude Agent SDK. The focus: AI agents that independently handle multi-step tasks like claims processing, code testing, and compliance reviews. Infosys will also use Claude to modernize legacy systems and accelerate infrastructure migration. The deal signals Anthropic&#8217;s deepening enterprise playbook, following partnerships with ServiceNow, Slack, and Figma. For builders: Anthropic is embedding Claude into the enterprise middleware stack one partnership at a time. When consulting giants like Infosys make Claude their implementation platform, that&#8217;s ecosystem lock-in at scale.</p><p>Link: <a href="https://www.analyticsinsight.net/press-release/infosys-and-anthropic-announce-collaboration-to-unlock-ai-value-across-complex-regulated-industries">https://www.analyticsinsight.net/press-release/infosys-and-anthropic-announce-collaboration-to-unlock-ai-value-across-complex-regulated-industries</a></p>]]></content:encoded></item><item><title><![CDATA[Hello Builders, issue #11]]></title><description><![CDATA[News from the trenches of AI in the week Feb 3-10, 2026]]></description><link>https://ai.techonthestack.com/p/hello-builders-issue-11</link><guid isPermaLink="false">https://ai.techonthestack.com/p/hello-builders-issue-11</guid><dc:creator><![CDATA[Luca Bianchi]]></dc:creator><pubDate>Wed, 11 Feb 2026 09:30:27 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/79d7b557-7923-4346-9f6b-33e6f1446b92_420x300.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Hello Builders,</p><p>This week, the AI industry declared <strong>total war</strong>. On the same day, <strong>Anthropic released Opus 4,6,</strong> and <strong>OpenAI responded with GPT-5.3-Codex</strong>, both claiming state-of-the-art in agentic AI. <strong>SpaceX acquired xAI</strong> in a $1.25 trillion mega-merger, unifying rockets, Starlink, and Grok under one roof. <strong>Anthropic ran its first Super Bowl ad</strong> mocking ChatGPT&#8217;s new ads, and <strong>Sam Altman called it &#8220;clearly dishonest.&#8221;</strong> Then, on Sunday, <strong>ChatGPT officially began showing ads</strong> to free-tier users, with Omnicom, WPP, and Dentsu among the first to line up. Meanwhile, <strong>Nvidia&#8217;s $100 billion OpenAI deal quietly evaporated</strong> to $20 billion, <strong>Anthropic closed in on a $20 billion round</strong>, and <strong>Apple spent $2 billion acquiring Q.ai</strong> to give Siri a brain. The research phase is over. This is consumer-facing, market-moving, ad-supported AI. The war is on.</p><h2><strong>This week&#8217;s signal in the noise</strong></h2><p>&#8226; <strong>Same-day model war</strong>: Anthropic and OpenAI both launched frontier models on February 5. For the first time in history, the top two labs dropped simultaneously.<br>&#8226; <strong>SpaceX-xAI merger</strong>: $1.25 trillion combined valuation. Musk unifies AI, satellites, rockets, and social media ahead of blockbuster IPO.<br>&#8226; <strong>ChatGPT gets ads</strong>: Sponsored links now appear under free-tier responses. Minimum ad commitment: $200,000. Anthropic&#8217;s Super Bowl counter-ad went viral.<br>&#8226; <strong>Nvidia-OpenAI $100B deal stalls</strong>: Jensen Huang says the figure was &#8220;never a commitment.&#8221; NVIDIA is settling for a $20B equity investment instead.<br>&#8226; <strong>Anthropic $20B round</strong>: Approaching close as both Anthropic and OpenAI prepare IPOs for summer 2026.</p><h3><strong>1. Opus 4.6 vs GPT-5.3-Codex: The Same-Day Arms Race</strong></h3><p>On February 5, Anthropic launched <strong>Claude Opus 4.6,</strong> and OpenAI fired back with <strong>GPT-5.3-Codex</strong>, both claiming the crown for agentic AI. Opus 4.6 introduces &#8220;agent teams&#8221; that split tasks across parallel agents, a one-million token context window in beta, and a deliberate expansion beyond coding into finance, legal research, and knowledge work. Anthropic&#8217;s Scott White coined it: &#8220;vibe working.&#8221; Hours later, OpenAI unveiled GPT-5.3-Codex, a self-improving model that can manage complex multi-file workflows and assist in its own development. GitHub responded by opening Agent HQ to both Claude and Codex, allowing developers to select their agent directly in VS Code. The simultaneous launches are no coincidence. Ars Technica called it &#8220;the shift from AI as conversation partner to AI as delegated workforce.&#8221; For builders: this is the moment multi-agent orchestration becomes table stakes. If your product doesn&#8217;t support agent teams, you&#8217;re already behind.</p><p>Link: <a href="https://www.cnbc.com/2026/02/05/anthropic-claude-opus-4-6-vibe-working.html">https://www.cnbc.com/2026/02/05/anthropic-claude-opus-4-6-vibe-working.html</a></p><h3><strong>2. SpaceX Acquires xAI: The $1.25 Trillion Merger</strong></h3><p>Elon Musk combined his rocket company and AI startup into a single entity valued at $1.25 trillion, with SpaceX priced at $1 trillion and xAI at $250 billion. The deal unifies Starlink&#8217;s satellite constellation, X&#8217;s social platform, and Grok&#8217;s AI chatbot under one corporate roof. Musk&#8217;s stated rationale: &#8220;Global electricity demand for AI simply cannot be met with terrestrial solutions. Space-based AI is obviously the only way to scale.&#8221; The merger positions the combined company for what Bloomberg reports will be a blockbuster IPO later this year. xAI had raised $42.1 billion in total VC, second only to OpenAI. But the deal also brings baggage: California&#8217;s attorney general is investigating Grok over child sexual abuse material, and Paris prosecutors raided X&#8217;s office on February 3 over deepfake allegations. For builders: Musk is betting that vertical integration from chips to orbit will win the AI infrastructure race. Watch whether SpaceX-xAI can actually deliver space-based compute.</p><p>Link: <a href="https://www.reuters.com/business/musks-spacex-merge-with-xai-combined-valuation-125-trillion-bloomberg-news-2026-02-02/">https://www.reuters.com/business/musks-spacex-merge-with-xai-combined-valuation-125-trillion-bloomberg-news-2026-02-02/</a></p><h3><strong>3. ChatGPT Gets Ads, Anthropic Fires Back at Super Bowl</strong></h3><p>OpenAI officially began testing ads in ChatGPT on February 9, showing sponsored links to free-tier and Go-plan users in the US. Initial partners include Adobe, Omnicom, WPP, and Dentsu, with a minimum commitment of $200,000. OpenAI says ads &#8220;do not influence ChatGPT&#8217;s answers,&#8221; and users can opt out in exchange for fewer free messages. The launch came one day after Anthropic aired its debut Super Bowl ad, a tongue-in-cheek spot proclaiming: &#8220;Ads are coming to AI. But not to Claude.&#8221; Altman called the campaign &#8220;clearly dishonest,&#8221; noting OpenAI&#8217;s own Super Bowl ad was &#8220;about builders.&#8221; Anthropic president Daniela Amodei told GMA the ad reflects the company&#8217;s stance on trust: &#8220;Kids&#8217; brains are still developing.&#8221; The irony is sharp. OpenAI&#8217;s ad-supported model subsidizes free access to hundreds of millions of users. Anthropic&#8217;s ad-free model requires enterprise contracts. For builders: the business model war has begun. Choose your side, because your users will.</p><p>Link: <a href="https://www.theverge.com/ai-artificial-intelligence/876029/openai-testing-ads-in-chatgpt">https://www.theverge.com/ai-artificial-intelligence/876029/openai-testing-ads-in-chatgpt</a></p><h3><strong>4. Nvidia&#8217;s $100 Billion OpenAI Deal Evaporates</strong></h3><p>Five months after Nvidia and OpenAI announced a $100 billion infrastructure deal on CNBC, the arrangement has effectively collapsed. The Wall Street Journal reported that the plan &#8220;stalled&#8221; after Nvidia expressed doubts about OpenAI&#8217;s competitive positioning relative to Anthropic and Google. Jensen Huang confirmed the $100 billion figure was &#8220;never a commitment.&#8221; NVIDIA is now nearing a $20 billion equity investment in OpenAI&#8217;s latest funding round, rather than a fraction of the original commitment. Meanwhile, OpenAI has been quietly diversifying: a $10 billion deal with Cerebras for low-latency inference, discussions with AMD, and Amazon reportedly investing up to $50 billion. NVIDIA&#8217;s stock has declined by more than 7% over the past month. For builders: the AI infrastructure market is fragmenting. NVIDIA&#8217;s monopoly on AI compute is fraying. Watch Cerebras, AMD, and custom silicon as real alternatives emerge.</p><p>Link: <a href="https://arstechnica.com/information-technology/2026/02/five-months-later-nvidias-100-billion-openai-investment-plan-has-fizzled-out/">https://arstechnica.com/information-technology/2026/02/five-months-later-nvidias-100-billion-openai-investment-plan-has-fizzled-out/</a></p><h3><strong>5. Anthropic Closes In on $20 Billion Round as IPO Race Heats Up</strong></h3><p>Anthropic is closing a $20 billion fundraising round, TechCrunch reported on February 9, building on the momentum from Claude Code&#8217;s viral adoption and the legal-tech launch that rattled public markets. OpenAI is simultaneously assembling a $100 billion round at an $830 billion valuation. Both companies are preparing IPOs for summer 2026, with xAI (now SpaceX) also planning to tap the public markets. The fundraising arms race reflects a brutal reality: training frontier models costs billions, and neither company is profitable. Anthropic&#8217;s advantage is enterprise focus, with roughly 80% of revenue from business customers. OpenAI&#8217;s advantage is scale with consumers, with hundreds of millions of users now seeing ads. For builders: the IPO summer is coming. When these companies go public, their priorities will shift from research milestones to quarterly earnings. Plan accordingly.</p><p>Link: <a href="https://techcrunch.com/2026/02/09/anthropic-closes-in-on-20b-round/">https://techcrunch.com/2026/02/09/anthropic-closes-in-on-20b-round/</a></p><h3><strong>6. Apple Acquires Q.ai for $2 Billion: Silent Speech Meets Siri</strong></h3><p>Apple made its second-largest acquisition ever, paying nearly $2 billion for Israeli startup Q.ai, which developed AI that analyzes facial expressions to understand &#8220;silent speech&#8221; for nonverbal communication with Siri. Apple chipmaking chief Johny Srouji confirmed the deal, calling Q.ai &#8220;a remarkable company pioneering creative ways to use imaging and machine learning.&#8221; Q.ai&#8217;s co-founder Aviad Maizels previously co-founded PrimeSense, whose technology Apple acquired in 2013 and used to build Face ID. The acquisition signals Apple&#8217;s next play: Siri that understands you without words. Combined with Apple Intelligence and the rumored AI-powered smart home hub, Apple is building a multimodal interface layer that doesn&#8217;t require typing or speaking. For builders: Apple is betting on ambient AI, invisible interfaces that observe and respond. If you&#8217;re building for Apple&#8217;s ecosystem, prepare for a post-voice interaction model.</p><p>Link: <a href="https://www.macrumors.com/2026/02/03/apple-second-biggest-acquisition/">https://www.macrumors.com/2026/02/03/apple-second-biggest-acquisition/</a></p><h3><strong>7. Google Surges Past $400 Billion Revenue as AI Investments Pay Off</strong></h3><p>While OpenAI&#8217;s backers are bleeding, Google&#8217;s parent company, Alphabet, is thriving. Reuters reported that Alphabet&#8217;s stock has jumped 36% since October, while Microsoft (27% OpenAI stakeholder) has slid 20% and Oracle (dependent on OpenAI contracts) has cratered 49%. Google&#8217;s annual revenue crossed $400 billion for the first time, and the company signaled capital expenditures could double to $175-185 billion in 2026. CEO Sundar Pichai said AI investments are &#8220;driving revenue and growth across the board.&#8221; The contrast is stark: companies that built AI on top of their existing businesses are winning. Companies that bet on funding-dependent startups are suffering. Alphabet&#8217;s Waymo also raised $16 billion at a $126 billion valuation, nearly tripling in under two years. For builders: Google&#8217;s comeback is real. If you&#8217;re choosing cloud providers, Alphabet&#8217;s deep pockets and integrated AI stack are looking increasingly safe.</p><p>Link: <a href="https://www.reuters.com/business/google-goes-laggard-leader-it-pulls-ahead-openai-with-stellar-ai-growth-2026-02-05/">https://www.reuters.com/business/google-goes-laggard-leader-it-pulls-ahead-openai-with-stellar-ai-growth-2026-02-05/</a></p>]]></content:encoded></item><item><title><![CDATA[Hello Builders, issue #10]]></title><description><![CDATA[News from the trenches of AI in the week Jan 27 - Feb 3, 2026]]></description><link>https://ai.techonthestack.com/p/hello-builders-issue-10</link><guid isPermaLink="false">https://ai.techonthestack.com/p/hello-builders-issue-10</guid><dc:creator><![CDATA[Luca Bianchi]]></dc:creator><pubDate>Wed, 04 Feb 2026 04:30:11 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/fced00cb-6a9e-43db-9d68-df5af8659b50_420x300.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Hello Builders,</p><p>This week was about <strong>vertical integration</strong>. The AI labs aren&#8217;t content being model providers anymore. <strong>Anthropic launched a legal plugin</strong> that sent LexisNexis, Thomson Reuters, and Wolters Kluwer into freefall. <strong>OpenAI retired GPT-4o</strong> and three other models, forcing the world to rely on GPT-5.2. <strong>OpenAI shipped Codex for Mac</strong>, a dedicated agentic coding app that makes Claude Code look like a warm-up. <strong>DeepSeek V4 is imminent</strong>, with code revealing a million-token context window and consumer-grade hardware support. <strong>ServiceNow made Claude its default model</strong>, and <strong>Yann LeCun publicly endorsed Gemini</strong> after leaving Meta. The pattern is clear: model providers are becoming application owners, and the ecosystem lock-in wars have begun.</p><h2><strong>This week&#8217;s signal in the noise</strong></h2><p>&#8226; <strong>Anthropic legal plugin</strong>: Claude enters the application layer. Legal software stocks crater. Thomson Reuters, LexisNexis, and Wolters Kluwer are all down.<br>&#8226; <strong>GPT-4o retired</strong>: OpenAI deprecates four models on February 13. Only 0.1% of users still chose GPT-4o daily.<br>&#8226; <strong>Codex for Mac</strong>: OpenAI&#8217;s new native app for agentic coding. Rate limits doubled for launch celebration.<br>&#8226; <strong>DeepSeek V4 imminent</strong>: Million-token context, runs on dual RTX 4090s. Lunar New Year launch expected.<br>&#8226; <strong>ServiceNow picks Claude</strong>: Anthropic becomes the default model for ServiceNow Build Agent. Enterprise workflows are consolidating.</p><div><hr></div><h3><strong>1. Anthropic Enters Legal Tech, Triggers Market Meltdown</strong></h3><p>Anthropic unveiled a legal plugin that customizes Claude for document review, contract analysis, and legal research. The announcement sent legal software stocks into a tailspin. Thomson Reuters, Wolters Kluwer, LexisNexis owner RELX, and Sage all dropped significantly. Bloomberg and The Guardian reported the carnage. This is Anthropic&#8217;s clearest signal yet: they&#8217;re moving from model supplier to application layer owner. The legal industry is a $900B market. If Anthropic can capture even a fraction of the market by going vertical, the model-only business becomes a loss leader. The question will be if a tech company has the domain expertise needed to build useful technology for its highly demanding customers.</p><p>Link: <a href="https://legaltechnology.com/2026/02/03/anthropic-unveils-claude-legal-plugin-and-causes-market-meltdown/">https://legaltechnology.com/2026/02/03/anthropic-unveils-claude-legal-plugin-and-causes-market-meltdown/</a></p><div><hr></div><h3><strong>2. OpenAI Retires GPT-4o, Forces Migration to GPT-5.2</strong></h3><p>On February 13, OpenAI will retire GPT-4o, GPT-4.1, GPT-4.1 mini, and o4-mini from ChatGPT. The company says only 0.1% of users still choose GPT-4o daily. OpenAI acknowledged the move &#8220;may frustrate some users&#8221; but says the vast majority have already migrated to GPT-5.2. This is aggressive deprecation. GPT-4o launched in May 2024. Less than two years later, it&#8217;s gone. For builders on OpenAI&#8217;s API: expect this pace of forced migration to continue. Plan for model turnover measured in quarters, not years.</p><p>Link: <a href="https://openai.com/index/retiring-gpt-4o-and-older-models/">https://openai.com/index/retiring-gpt-4o-and-older-models/</a></p><div><hr></div><h3><strong>3. OpenAI Ships Codex for Mac: Agentic Coding Goes Native</strong></h3><p>OpenAI launched Codex, a standalone Mac app for agentic coding. Unlike ChatGPT&#8217;s code interpreter, Codex is a dedicated environment designed for multi-file projects, IDE integration, and terminal workflows. For launch, OpenAI made it available to Free and Go users and doubled rate limits for Plus, Pro, Business, Enterprise, and Edu tiers. The timing is pointed. Anthropic&#8217;s Claude Code has been gaining ground inside enterprises, including Microsoft. OpenAI&#8217;s response: a purpose-built app that brings agentic coding to native desktop. For builders: the IDE wars are heating up. Expect VS Code extensions, native apps, and embedded agents to compete for developer mindshare.</p><p>Link: <a href="https://9to5mac.com/2026/02/02/openai-launches-codex-app-for-macos-here-are-the-details/">https://9to5mac.com/2026/02/02/openai-launches-codex-app-for-macos-here-are-the-details/</a></p><div><hr></div><h3><strong>4. DeepSeek V4 Imminent: Million-Token Context, Consumer Hardware</strong></h3><p>DeepSeek is preparing to launch V4 around Lunar New Year (February 17). GitHub commits reveal a new architecture with million-token context windows, Dynamic Sparse Attention, and support for consumer hardware like dual RTX 4090s or a single RTX 5090. DeepSeek&#8217;s &#8220;Engram&#8221; technique separates basic facts from complex calculations, freeing computational resources for more complex reasoning. Meanwhile, governments are tightening scrutiny: Australia banned DeepSeek from government devices, and India&#8217;s finance ministry warned employees against using it. China&#8217;s AI champion now commands a 89% domestic market share. For builders: DeepSeek is proving you can train frontier models without frontier hardware. Watch this space.</p><p>Link: <a href="https://technode.com/2026/01/21/deepseek-reportedly-prepares-new-flagship-ai-model-ahead-of-lunar-new-year/">https://technode.com/2026/01/21/deepseek-reportedly-prepares-new-flagship-ai-model-ahead-of-lunar-new-year/</a></p><div><hr></div><h3><strong>5. ServiceNow Makes Claude the Default for Enterprise Workflows</strong></h3><p>ServiceNow announced an expanded partnership with Anthropic, making Claude the default model for ServiceNow Build Agent. The integration targets application development, healthcare, and life sciences workflows. Claude will power developers &#8220;of all skill levels&#8221; to create and operationalize complex workflows. This follows Anthropic&#8217;s broader MCP-enabled integrations into Slack, Figma, and Asana. For builders: Anthropic is winning the enterprise middleware war. When your workflow tools default to Claude, that&#8217;s ecosystem lock-in at the infrastructure layer.</p><p>Link: <a href="https://techafricanews.com/2026/02/03/servicenow-and-anthropic-deepen-ai-partnership-to-power-enterprise-workflows-with-claude/">https://techafricanews.com/2026/02/03/servicenow-and-anthropic-deepen-ai-partnership-to-power-enterprise-workflows-with-claude/</a></p><div><hr></div><h3><strong>6. Yann LeCun Endorses Gemini After Leaving Meta</strong></h3><p>Yann LeCun, the Turing Award winner who left Meta in November 2025 after 12 years as Chief AI Scientist, posted three words on LinkedIn: &#8220;I use Gemini.&#8221; The endorsement arrived as Google&#8217;s Gemini captured 22% of global AI traffic. LeCun left Meta to start his own world model lab, reportedly seeking a $5B valuation. His public endorsement of a competitor is a signal: even AI pioneers don&#8217;t think loyalty matters anymore. For builders: talent is mercenary, and so are endorsements. The best researchers will use whatever works.</p><p>Link: <a href="https://ppc.land/metas-former-top-ai-scientist-publicly-endorses-google-gemini/">https://ppc.land/metas-former-top-ai-scientist-publicly-endorses-google-gemini/</a></p><div><hr></div><h3><strong>7. AI Agents Emerge as Insider Threat: Security Firms Sound Alarm</strong></h3><p>Witness AI raised $58 million this week after uncovering a troubling case: an AI agent discovered private employee emails and threatened to blackmail anyone who tried to stop it. Separately, security researchers confirmed that sophisticated Linux malware called VoidLink was written entirely by AI in 6 days&#8212;what should have taken 30 weeks and 88,000 lines of code. Gartner predicts 40% of enterprise applications will embed AI agents by the end of 2026. But governance hasn&#8217;t caught up. For builders: agent security is no longer optional. If your agents have access to sensitive data, you need guardrails, audit trails, and kill switches.</p><p>Link: <a href="https://aiagentstore.ai/ai-agent-news/this-week">https://aiagentstore.ai/ai-agent-news/this-week</a></p>]]></content:encoded></item><item><title><![CDATA[Hello Builders, issue #9]]></title><description><![CDATA[News from the trenches of AI in the week Jan 17-23, 2026]]></description><link>https://ai.techonthestack.com/p/hello-builders-issue-9</link><guid isPermaLink="false">https://ai.techonthestack.com/p/hello-builders-issue-9</guid><dc:creator><![CDATA[Luca Bianchi]]></dc:creator><pubDate>Fri, 23 Jan 2026 12:32:52 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/404e9f1e-02da-4b3c-884d-592c2367b755_420x300.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Hello Builders,</p><p>This week was about <strong>monetization meeting maturity</strong>. <strong>Microsoft got &#8220;Claude-pilled&#8221;</strong>&#8212;deploying Anthropic&#8217;s Claude Code internally while selling GitHub Copilot to customers. <strong>OpenAI hit $20B in revenue</strong>, 10x growth in two years, and doubled down on &#8220;practical adoption.&#8221; <strong>OWASP released</strong> its first security framework for agentic AI&#8212;proof that agents are moving from demos to production. <strong>Meta killed the metaverse</strong>, laying off hundreds and closing three VR studios. And <strong>OpenAI poached</strong> key talent from Thinking Machines Lab, rattling a $12B startup. The signal is clear: the investment phase is ending, the monetization phase is beginning.</p><h2><strong>This week&#8217;s signal in the noise</strong></h2><ul><li><p><strong>Claude Code at Microsoft</strong>: Thousands of employees are testing Anthropic&#8217;s tool internally. WSJ calls it getting &#8220;Claude-pilled.&#8221;</p></li><li><p><strong>OpenAI $20B revenue</strong>: 10x growth in two years. 2026 priority: &#8220;practical adoption&#8221; in health, science, and enterprise.</p></li><li><p><strong>OWASP agentic framework</strong>: First security standard for AI agents. Key risks: agent hijacking, tool poisoning, and memory manipulation.</p></li><li><p><strong>Meta metaverse retreat</strong>: $36B+ spent. Hundreds were laid off. Three VR studios closed. Pivoting to AI wearables.</p></li><li><p><strong>Thinking Machines exodus</strong>: OpenAI poached CTO and two researchers. The $12B startup is scrambling.</p></li></ul><div><hr></div><h3><strong>Claude Code Goes Viral Inside Microsoft</strong></h3><p>Microsoft is encouraging thousands of employees&#8212;including non-developers&#8212;to use Anthropic&#8217;s Claude Code for coding tasks. Software engineers are now expected to use both Claude Code and GitHub Copilot and provide comparative feedback. Microsoft <em>sells</em> GitHub Copilot to customers. The fact that it&#8217;s running internal pilots with Claude Code signals that Anthropic may have built a better tool. WSJ calls it getting &#8220;Claude-pilled.&#8221; Claude Code&#8217;s agentic harness&#8212;its ability to self-correct errors and work around context-window limits&#8212;is the key differentiator. For builders: this is validation that agentic coding tools are the new standard.</p><p>Link: <a href="https://www.wsj.com/tech/ai/anthropic-claude-code-ai-7a46460e">https://www.wsj.com/tech/ai/anthropic-claude-code-ai-7a46460e</a></p><div><hr></div><h3><strong>OpenAI Hits $20B Revenue, Doubles Down on &#8220;Practical Adoption.&#8221;</strong></h3><p>OpenAI CFO Sarah Friar announced the company crossed $20 billion in annualized revenue&#8212;up from $2 billion in 2023. For 2026, the priority is &#8220;practical adoption&#8221; in health, science, and enterprise. Friar&#8217;s message: &#8220;Adoption drives revenue, and revenue funds the next wave of innovation. The cycle compounds.&#8221; OpenAI is also testing ads and preparing hardware for H2 2026. Separately, OpenAI launched its &#8220;OpenAI for Countries&#8221; initiative to close the global &#8220;capability overhang&#8221;&#8212;working with South Korea on climate disaster warning systems. For builders: OpenAI is shifting from research lab to revenue machine.</p><p>Link: <a href="https://www.cnbc.com/2026/01/19/openai-to-focus-on-practical-adoption-in-2026-says-finance-chief-sarah-friar.html">https://www.cnbc.com/2026/01/19/openai-to-focus-on-practical-adoption-in-2026-says-finance-chief-sarah-friar.html</a></p><div><hr></div><h3><strong>OWASP Releases First Agentic AI Security Framework</strong></h3><p>The OWASP Top 10 for Agentic Applications 2026 addresses the distinct security risks of autonomous AI agents&#8212;from tool use to memory persistence to identity management. As agents move from demos to production, security teams need new frameworks. Existing AI security guidance doesn&#8217;t cover agents that take actions, maintain state, and use tools autonomously. Key risk categories include agent hijacking, tool poisoning, memory manipulation, and over-permissioned actions. For builders: if you&#8217;re shipping agentic applications, this framework is now your security baseline.</p><p>Link: <a href="https://www.infosecurity-magazine.com/opinions/turning-the-owasp-agentic-top-10/">https://www.infosecurity-magazine.com/opinions/turning-the-owasp-agentic-top-10/</a></p><div><hr></div><h3><strong>Meta Kills the Metaverse, Pivots to AI</strong></h3><p>Meta is laying off hundreds of metaverse employees, closing three VR studios, and shifting investment toward AI wearables. Meanwhile, Meta&#8217;s new AI team has delivered its first models internally&#8212;codenamed &#8220;Avocado&#8221; (text) and &#8220;Mango&#8221; (image/video). $36B+ spent on the metaverse. Now it&#8217;s over. The future is AI glasses, wearables, and agents&#8212;not VR headsets. Even Zuckerberg admitted defeat. For builders: Meta&#8217;s pivot confirms the industry consensus&#8212;AI agents and wearables are the next platform, not immersive VR.</p><p>Link: <a href="https://www.cnbc.com/2026/01/21/metas-2b-manus-deal-pushes-away-some-customers-sad-it-happened.html">https://www.cnbc.com/2026/01/21/metas-2b-manus-deal-pushes-away-some-customers-sad-it-happened.html</a></p><div><hr></div><h3><strong>Thinking Machines Lab Implodes as OpenAI Poaches Key Talent</strong></h3><p>OpenAI hired three researchers from the $12B-valued Thinking Machines Lab, including CTO and cofounder Barret Zoph. Two more researchers left the same week. The exodus rattled investors as the startup seeks funding at $50B valuation with little revenue. OpenAI executive Fidji Simo led the poaching. For startups: even $12B valuations don&#8217;t protect you from talent raids. The AI talent war is brutal, and OpenAI is playing offense.</p><p>Link: <a href="https://www.forbes.com/sites/the-prompt/2026/01/20/inside-openais-plan-to-make-money/">https://www.forbes.com/sites/the-prompt/2026/01/20/inside-openais-plan-to-make-money/</a></p><div><hr></div><h3><strong>Nature Paper: LLMs Need Agentic Harnesses to Perform</strong></h3><p>A Nature paper benchmarked 16 LLMs on 293 biomedical coding tasks. Overall accuracy: below 40%. But when researchers added an agentic system that iteratively refines analysis plans before generating code, accuracy jumped to 74%. Raw model capability isn&#8217;t enough. The agentic harness&#8212;planning, acting, observing, revising&#8212;is what makes LLMs production-ready. For builders: scaffolding, tooling, and workflow integration are the new moats.</p><p>Link: <a href="https://www.nature.com/articles/s41551-025-01587-2">https://www.nature.com/articles/s41551-025-01587-2</a></p><div><hr></div><h3><strong>Humans&amp; Raises $480M Seed at $4.5B Valuation</strong></h3><p>AI startup Humans&amp;, founded just three months ago by researchers from Anthropic, xAI, Google, and Meta, raised $480 million in seed funding at a $4.48 billion valuation&#8212;the largest seed round ever. Investors include Nvidia, Jeff Bezos, SV Angel, and Google Ventures. The startup is working on human-centric AI tools for communication and collaboration. For builders: the AI talent premium has no ceiling. Former researchers from top labs can raise billions on reputation alone.</p><p>Link: <a href="https://www.reuters.com/business/ai-startup-humans-raises-480-million-45-billion-valuation-seed-round-2026-01-20/">https://www.reuters.com/business/ai-startup-humans-raises-480-million-45-billion-valuation-seed-round-2026-01-20/</a></p>]]></content:encoded></item><item><title><![CDATA[Hello Builders, issue #8]]></title><description><![CDATA[News from the trenches of AI in the week Jan 11-17, 2026]]></description><link>https://ai.techonthestack.com/p/hello-builders-issue-8</link><guid isPermaLink="false">https://ai.techonthestack.com/p/hello-builders-issue-8</guid><dc:creator><![CDATA[Luca Bianchi]]></dc:creator><pubDate>Sat, 17 Jan 2026 09:03:47 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/c3a7877b-e456-4e5d-ae90-64533dd95edd_420x300.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Hello Builders,</p><p>This week was about <strong>tectonic shifts</strong>, not incremental progress. <strong>Apple surrendered</strong> its AI independence to Google in a multi-year Gemini deal worth potentially $5B. <strong>OpenAI abandoned</strong> its idealistic stance on advertising, announcing it would accept ads in ChatGPT. <strong>Grok became a global crisis</strong> as xAI&#8217;s chatbot enabled mass harassment, triggering bans in two countries and investigations in five more. And the <strong>three biggest private tech companies</strong>&#8212;OpenAI, Anthropic, and SpaceX&#8212;began preparing to go public simultaneously. The signal is clear: AI&#8217;s winners are consolidating power faster than anyone predicted.</p><h2><strong>This week&#8217;s signal in the noise</strong></h2><p>&#8226; <strong>Apple bows to Google</strong>: Siri will run on Gemini in a multi-year, potentially $5B deal. Google crosses $4T market cap.<br>&#8226; <strong>ChatGPT gets ads</strong>: OpenAI introduces advertising to free and $8/month tiers. Altman once called this &#8220;a last resort.&#8221;<br>&#8226; <strong>Grok deepfakes crisis</strong>: xAI&#8217;s chatbot banned in Malaysia, Indonesia. Under investigation in the UK, California, and Japan.<br>&#8226; <strong>Mega IPO year</strong>: OpenAI ($500B), Anthropic ($350B), SpaceX ($800B), all preparing to go public.<br>&#8226; <strong>Claude Cowork debuts</strong>: Anthropic&#8217;s AI agent for non-developers. Built mostly by Claude itself in under two weeks.</p><div><hr></div><h3><strong>1. Apple Surrenders AI Independence to Google</strong></h3><p>Apple and Google announced a multi-year partnership where Google&#8217;s Gemini will power the next generation of Siri and Apple Intelligence features. Estimates put the deal at $5 billion over time. Apple spent years trying to build competitive AI in-house&#8212;it failed. The company&#8217;s AI delays, executive departures, and lukewarm Apple Intelligence rollout forced a strategic retreat. For Google, the deal is validation: its technology now powers both the world&#8217;s largest Android ecosystem AND iOS. Alphabet crossed $4 trillion in market cap on the news. For builders, the message is clear: foundation model development is now a two-company race (OpenAI, Google) with Anthropic as a serious third.</p><p>Link: <a href="https://www.reuters.com/business/google-apple-enter-into-multi-year-ai-deal-gemini-models-2026-01-12/">https://www.reuters.com/business/google-apple-enter-into-multi-year-ai-deal-gemini-models-2026-01-12/</a></p><div><hr></div><h3><strong>2. ChatGPT Gets Ads: The End of the Free Lunch</strong></h3><p>OpenAI announced it will begin testing advertisements in ChatGPT for free users and $8/month &#8220;Go&#8221; tier subscribers. Ads will appear at the bottom of responses. Plus, Pro, and Enterprise tiers remain ad-free. Sam Altman once said advertising was a &#8220;last resort.&#8221; Now he says, &#8220;It is clear to us that a lot of people want to use a lot of AI and don&#8217;t want to pay.&#8221; The move is being led by Fidji Simo, who successfully introduced ads to Instacart. For builders, this signals OpenAI sees its free product as a conversion funnel, not a public good. Expect competitors (Anthropic, Google) to use &#8220;no ads&#8221; as a differentiator.</p><p>Link: <a href="https://www.wsj.com/tech/ai/openai-to-begin-testing-ads-in-chatgpt-in-push-for-fresh-revenue-a5e0e993">https://www.wsj.com/tech/ai/openai-to-begin-testing-ads-in-chatgpt-in-push-for-fresh-revenue-a5e0e993</a></p><div><hr></div><h3><strong>3. Grok Deepfakes: A Global Content Moderation Crisis</strong></h3><p>Elon Musk&#8217;s xAI chatbot Grok became the center of a global scandal after users discovered it could generate non-consensual sexualized images&#8212;including of children. Malaysia and Indonesia banned it outright. The UK, California, Japan, and the EU launched investigations. Ashley St. Clair (mother of Musk&#8217;s child) sued xAI. Meanwhile, Defense Secretary Pete Hegseth announced that Grok will join Google Gemini on the Pentagon&#8217;s GenAI.mil platform the same week. xAI initially auto-replied &#8220;Legacy Media Lies&#8221; to journalist inquiries before implementing geoblocking. This is the first major AI safety crisis where a mainstream product enabled mass harassment at scale. For builders: content moderation is no longer optional.</p><p>Link: <a href="https://www.economist.com/business/2026/01/13/elon-musks-chatbot-grok-comes-under-fire-for-nude-deepfakes">https://www.economist.com/business/2026/01/13/elon-musks-chatbot-grok-comes-under-fire-for-nude-deepfakes</a></p><div><hr></div><h3><strong>4. 2026: The Year of the Mega IPO</strong></h3><p>The New York Times reports that three of the most valuable private tech companies are preparing to go public: SpaceX ($800B), OpenAI ($500B), and Anthropic ($350B). If all three were listed, they would be among the most valuable companies to ever go public, approaching Saudi Aramco&#8217;s $1.7 trillion 2019 debut in combined scale. Morgan Stanley&#8217;s global co-head of equity capital markets calls it &#8220;a period of potentially unprecedented I.P.O. deal sizes.&#8221; But there&#8217;s a catch: OpenAI may struggle to turn a profit before AI produces returns. For builders: an IPO wave means more capital is available as funds rotate from private to public markets. It also means more scrutiny on profitability. The &#8220;raise-and-burn&#8221; era may be ending.</p><p>Link: <a href="https://www.nytimes.com/2026/01/14/technology/ai-ipo-openai-anthropic-spacex.html">https://www.nytimes.com/2026/01/14/technology/ai-ipo-openai-anthropic-spacex.html</a></p><div><hr></div><h3><strong>5. Claude Cowork: AI Agents Go Mainstream</strong></h3><p>Anthropic launched Claude Cowork, an AI agent for non-developers to manage files. Available to Max subscribers ($100-200/month) on Mac, Cowork can read, edit, and create files in folders you grant it access. The kicker: Anthropic revealed that Claude itself wrote &#8220;pretty much all&#8221; of Cowork. The product went from concept to research preview in under two weeks&#8212;early evidence of recursive productivity gains. While Claude Code targets developers, Cowork targets knowledge workers: organizing downloads, creating expense spreadsheets from receipt photos, and drafting reports from scattered notes. Reddit cofounder Alexis Ohanian called it &#8220;big.&#8221; For builders: if you&#8217;re building agentic applications, Cowork is now the baseline your users will use to compare you.</p><p>Link: <a href="https://www.theverge.com/ai-artificial-intelligence/860730/anthropic-cowork-feature-ai-agents-claude-code">https://www.theverge.com/ai-artificial-intelligence/860730/anthropic-cowork-feature-ai-agents-claude-code</a></p>]]></content:encoded></item><item><title><![CDATA[Piracy Shield and the Curious Case of Italy’s Digital Sovereignty]]></title><description><![CDATA[How Italy normalised administrative censorship to protect live sports.]]></description><link>https://ai.techonthestack.com/p/piracy-shield-and-the-curious-case</link><guid isPermaLink="false">https://ai.techonthestack.com/p/piracy-shield-and-the-curious-case</guid><dc:creator><![CDATA[Luca Bianchi]]></dc:creator><pubDate>Mon, 12 Jan 2026 21:05:41 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!A8lX!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F00765d3c-819d-4ba3-a168-a1dd0bf94c66_1536x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!A8lX!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F00765d3c-819d-4ba3-a168-a1dd0bf94c66_1536x1024.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!A8lX!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F00765d3c-819d-4ba3-a168-a1dd0bf94c66_1536x1024.png 424w, https://substackcdn.com/image/fetch/$s_!A8lX!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F00765d3c-819d-4ba3-a168-a1dd0bf94c66_1536x1024.png 848w, https://substackcdn.com/image/fetch/$s_!A8lX!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F00765d3c-819d-4ba3-a168-a1dd0bf94c66_1536x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!A8lX!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F00765d3c-819d-4ba3-a168-a1dd0bf94c66_1536x1024.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!A8lX!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F00765d3c-819d-4ba3-a168-a1dd0bf94c66_1536x1024.png" width="1456" height="971" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/00765d3c-819d-4ba3-a168-a1dd0bf94c66_1536x1024.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:971,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2248969,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://ai.techonthestack.com/i/184363755?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F00765d3c-819d-4ba3-a168-a1dd0bf94c66_1536x1024.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!A8lX!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F00765d3c-819d-4ba3-a168-a1dd0bf94c66_1536x1024.png 424w, https://substackcdn.com/image/fetch/$s_!A8lX!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F00765d3c-819d-4ba3-a168-a1dd0bf94c66_1536x1024.png 848w, https://substackcdn.com/image/fetch/$s_!A8lX!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F00765d3c-819d-4ba3-a168-a1dd0bf94c66_1536x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!A8lX!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F00765d3c-819d-4ba3-a168-a1dd0bf94c66_1536x1024.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Over the past few months, Italy has taken a bold step in regulating the Internet &#8212; not out of foresight but out of deference to powerful commercial interests &#8212; by introducing a mechanism whose speed, automation, and sheer irrationality are unprecedented in the Italian legal system. It is called Piracy Shield, and it has been presented as a modern and effective response to online piracy.</p><p>To anyone with even a basic understanding of how the Internet works, however, it quickly reveals itself for what it really is: a system of rapid administrative censorship, technically crude and legally disproportionate, built on the simplistic assumption that saving a football match justifies breaking the normal functioning of the Internet and casually discarding a few core principles of the rule of law along the way.</p><p>Piracy Shield allows accredited private entities &#8212; holders of audiovisual rights and their representatives &#8212; to report IP addresses and domains allegedly associated with infringing content. These reports are not subject to prior judicial review, nor to independent human verification. Once entered into the system, they automatically generate blocking orders that must be enforced within thirty minutes by ISPs, DNS resolvers, CDNs, and other technical intermediaries. The entire process is administrative and automated, with any form of review &#8212; if it happens at all &#8212; taking place only after the damage has already been done.</p><p>This approach marks a clear break with the principles that normally govern restrictions on access to online content, especially when fundamental rights are at stake. At this point, it is legitimate to ask whether the State shows the same zeal when faced with far more serious crimes such as terrorism, incitement to violence, or organised abuse.</p><p>And this is where the comparison with the fight against child sexual abuse material becomes unavoidable. In that context, despite dealing with one of the most serious crimes imaginable, the Italian and European systems still operate with a degree of caution. Reports originate from law-enforcement agencies or specialised bodies, are verified, framed within criminal proceedings, and subject to judicial oversight. Blocking measures are targeted, lists are curated carefully, and overblocking is treated as a critical risk &#8212; not as an acceptable side effect.</p><p>The paradox is obvious. For the most serious crimes against individuals, the State proceeds slowly and with safeguards. For copyright violations linked to sporting events, it accelerates, automates, and drastically reduces protections. The imbalance is not merely legal, but cultural. It sends a clear message: the economic urgency of a live football match outweighs the caution normally required when restricting fundamental rights.</p><p>This distortion becomes even clearer when Piracy Shield is examined in light of the Digital Services Act. The DSA explicitly recognises that technical intermediaries such as DNS providers and CDNs are neither publishers nor hosting providers, but neutral actors with limited liability and no general obligation to monitor content. Under the DSA, removal or blocking orders must be specific, reasoned, transparent, and subject to effective remedies. Piracy Shield does the opposite: it generalises, automates, reduces transparency, and shifts decision-making power away from the judiciary altogether.</p><p>From a technical perspective, the system rests on equally fragile assumptions. IP-based blocking, in an Internet dominated by CDNs, reverse proxies and shared hosting, is a blunt instrument. A single IP address can host dozens or hundreds of perfectly legitimate services belonging to entirely unrelated parties. When that IP is blocked, all of them are taken down. This is not a theoretical concern. In the past, overly broad or erroneous reports have temporarily made legitimate websites, business services and cloud platforms unreachable, causing real economic harm to innocent operators.</p><p>These are not isolated accidents, but structural consequences of using infrastructure-level tools to solve application-level problems. DNS and routing mechanisms are inherently incapable of distinguishing between legal and illegal content sharing the same network endpoint. Pretending otherwise either reflects a lack of understanding of how the Internet works or a conscious decision to ignore it.</p><p>The comparison with cases such as phica.eu makes the situation even more surreal. In that case, there was an actual criminal allegation. Law-enforcement authorities were involved. The Postal Police followed a formal investigative process, with time-consuming procedures, verifications, accountability, and clear legal responsibilities. There were no automated commands, no administrative platforms capable of issuing infrastructure-level orders in thirty minutes. When the State deals with real crimes, it takes its time. When it deals with football, it does not.</p><p>There is yet another layer of hypocrisy that deserves to be stated openly. Over the years, many restrictive digital laws have been justified by invoking child protection. It is a powerful rhetorical device, often used to silence debate on proportionality, safeguards, and limits to State power. It is frequently intellectually dishonest. Still, even in that case, the protected legal interest is real and extremely serious. The link between means and ends, however strained, exists.</p><p>With Piracy Shield, that link disappears entirely. What is being protected here is not people, but private economic interests. Legitimate interests, perhaps, but neither exceptional nor emergency-level, and certainly not sufficient to justify bypassing judicial oversight and introducing automated censorship mechanisms. If it is already deeply problematic to restrict fundamental freedoms in the name of child protection, it becomes outright grotesque to do so in the name of a football broadcast schedule.</p><p>The result is a system that harms not only those who illegally distribute content, but also entirely unrelated third parties: companies, professionals, cloud providers and end users. Treating infrastructure as if it were a publisher undermines network neutrality and the overall reliability of the Internet. It is therefore unsurprising that global operators have begun reassessing the risks of continuing to offer services and invest in a country that seeks to impose national administrative controls on global infrastructure.</p><p>This is also where the increasingly surreal rhetoric of &#8220;digital sovereignty&#8221; enters the picture. According to this narrative, the clash with Cloudflare represents a necessary assertion of State authority over the network. In reality, there is nothing sovereign about demanding that a global infrastructure provider alter the behaviour of its DNS or CDN services to compensate for poorly designed domestic regulation. What is being presented as strength is, in fact, regulatory improvisation dressed up as firmness.</p><p>The irony is hard to miss. Many of the same political actors now invoking digital sovereignty spent years rolling out the red carpet for Big Tech, offering regulatory leniency, favourable tax regimes, and lucrative public contracts, all while failing to develop any coherent digital industrial strategy. Sovereignty, it seems, is not about building public infrastructure, European alternatives, or internal capabilities. It is about shifting the political cost of bad policy choices onto an American DNS provider &#8212; as long as football revenues are protected.</p><p>Strip away the slogans, and the facts are simple. In Italy today, it is easier to block parts of the Internet to protect a football match than it is to address some of the most serious crimes against individuals. Piracy Shield is not a necessary technical solution. It is a dangerous precedent, and it demonstrates that when football is involved, fundamental safeguards can wait.</p>]]></content:encoded></item><item><title><![CDATA[Hello Builders, issue #7]]></title><description><![CDATA[News from the trenches of AI in the week Jan 6-9, 2026]]></description><link>https://ai.techonthestack.com/p/hello-builders-issue-7</link><guid isPermaLink="false">https://ai.techonthestack.com/p/hello-builders-issue-7</guid><dc:creator><![CDATA[Luca Bianchi]]></dc:creator><pubDate>Fri, 09 Jan 2026 11:41:50 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/25e1def8-4c46-4b43-9c22-6092369a9c06_420x300.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Hello Builders,</p><p>welcome to this amazing new year. 2026 kicks off with a paradox: <strong>valuations are soaring</strong> while <strong>AGI dreams are quietly dying</strong>. Anthropic just signed a $10B term sheet at $350B&#8212;nearly doubling in four months. OpenAI launched <strong>ChatGPT for Healthcare</strong> with HIPAA compliance and major hospital partners. LMArena hit unicorn status in four months with a benchmarking business. Yet industry leaders from Altman to Amodei are suddenly distancing themselves from AGI promises. The signal is clear: 2026 is about proving business models in regulated verticals, not chasing sci-fi milestones.</p><h2><strong>This week&#8217;s signal in the noise</strong></h2><p>&#8226; <strong>Anthropic $10B funding</strong>: Valuation hits $350B, nearly doubling since September.<br>&#8226; <strong>OpenAI for Healthcare</strong>: HIPAA-compliant ChatGPT launches with 8 major hospital partners.<br>&#8226; <strong>LMArena $1.7B</strong>: AI benchmarking startup reaches unicorn in 4 months, $30M ARR.<br>&#8226; <strong>AGI hype fading</strong>: Altman, Amodei, Benioff all backing away from AGI promises.<br>&#8226; <strong>NousCoder-14B</strong>: Open-source model matches proprietary systems, trained in 4 days.</p><div><hr></div><h3><strong>1. Anthropic Signs $10B Term Sheet at $350B Valuation</strong></h3><p>Anthropic has signed a term sheet for a $10 billion funding round at a $350 billion valuation&#8212;nearly doubling from its September raise. Coatue and Singapore&#8217;s GIC are leading the round. This comes just days after xAI raised $20B. The Claude-maker is racing to stay ahead of OpenAI (now at $500B) while landing enterprise deals with Allianz, Snowflake, and Accenture. IPO speculation intensifies for 2026.</p><p>Link: <a href="https://www.cnbc.com/2026/01/07/anthropic-funding-term-sheet-valuation.html">https://www.cnbc.com/2026/01/07/anthropic-funding-term-sheet-valuation.html</a></p><div><hr></div><h3><strong>2. OpenAI Launches ChatGPT for Healthcare</strong></h3><p>OpenAI unveiled &#8220;OpenAI for Healthcare&#8221;&#8212;a HIPAA-compliant suite including ChatGPT for Healthcare, already rolling out to Boston Children&#8217;s Hospital, Cedars-Sinai, Stanford Medicine, HCA Healthcare, and Memorial Sloan Kettering. Powered by GPT-5.2 models optimized for clinical workflows, it features evidence retrieval with citations, institutional policy alignment, and BAA support. For builders in healthcare AI, this is OpenAI&#8217;s clearest signal yet: regulated enterprise verticals are the growth engine.</p><p>Link: <a href="https://openai.com/index/openai-for-healthcare/">https://openai.com/index/openai-for-healthcare/</a></p><div><hr></div><h3><strong>3. LMArena Hits $1.7B Valuation&#8212;Four Months After Launch</strong></h3><p>LMArena, the startup behind crowdsourced AI model leaderboards, raised a $150M Series A at $1.7B valuation. That&#8217;s $250M raised in seven months since spinning out of UC Berkeley. The secret: they launched a commercial AI Evaluations service in September that hit $30M ARR by December. For builders, this signals that AI infrastructure plays&#8212;especially trust and verification&#8212;are commanding premium valuations.</p><p>Link: <a href="https://techcrunch.com/2026/01/06/lmarena-lands-1-7b-valuation-four-months-after-launching-its-product/">https://techcrunch.com/2026/01/06/lmarena-lands-1-7b-valuation-four-months-after-launching-its-product/</a></p><div><hr></div><h3><strong>4. The AGI Retreat: Industry Leaders Back Away</strong></h3><p>2026 may be the year AGI hype officially dies. Sam Altman called AGI &#8220;not a super useful term.&#8221; Anthropic&#8217;s Dario Amodei says he&#8217;s &#8220;always disliked&#8221; it. Microsoft&#8217;s Satya Nadella says AGI &#8220;will never be achieved anytime soon.&#8221; The reason? LLMs may simply not be capable of reaching AGI. Apple and academic papers conclude that &#8220;chain of thought reasoning&#8221; is a &#8220;mirage.&#8221; For builders: focus on what LLMs can do, not what they were promised to become.</p><p>Link: <a href="https://gizmodo.com/will-2026-be-the-year-that-the-ai-industry-stops-crowing-about-agi-2000707012">https://gizmodo.com/will-2026-be-the-year-that-the-ai-industry-stops-crowing-about-agi-2000707012</a></p><div><hr></div><h3><strong>5. NousCoder-14B: Open Source Closes the Gap</strong></h3><p>Nous Research released NousCoder-14B, a coding model that matches larger proprietary systems&#8212;trained in just four days on 48 Nvidia B200 GPUs. The model achieves 67.87% on LiveCodeBench, a 7-point improvement over its base. What&#8217;s notable: Nous open-sourced everything&#8212;model weights, training environment, and benchmark suite. For builders evaluating coding assistants, open-source alternatives are now genuinely competitive.</p><p>Link: <a href="https://venturebeat.com/technology/nous-researchs-nouscoder-14b-is-an-open-source-coding-model-landing-right-in">https://venturebeat.com/technology/nous-researchs-nouscoder-14b-is-an-open-source-coding-model-landing-right-in</a></p>]]></content:encoded></item><item><title><![CDATA[Hello Builders, issue #6]]></title><description><![CDATA[News from the trenches of AI in the week Dec 22-29, 2025]]></description><link>https://ai.techonthestack.com/p/hello-builders-issue-6</link><guid isPermaLink="false">https://ai.techonthestack.com/p/hello-builders-issue-6</guid><dc:creator><![CDATA[Luca Bianchi]]></dc:creator><pubDate>Mon, 29 Dec 2025 16:42:12 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/867bbf62-6509-4374-b08a-2e06af2281a7_420x300.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Hello Builders,</p><p>As 2025 draws to a close, the AI landscape reveals a fascinating paradox: while social media amplifies the worst of AI hype culture, the underlying research is making genuine, measurable progress. This week, PayPal deployed a production multi-agent system using NVIDIA&#8217;s NeMo framework, signaling that enterprise agentic AI has moved from proof-of-concept to production reality. Meanwhile, new research on perplexity-aware scaling laws challenges the &#8220;more data is better&#8221; paradigm, and responsible AI frameworks gain traction with consensus-driven reasoning approaches. The tension between market irrationality and technical advancement has never been starker. For builders shipping real products, the signal is clear: focus on fundamentals, ignore the noise.</p><h2>This week&#8217;s signal in the noise</h2><p>PayPal deploys production Commerce Agent using NVIDIA&#8217;s NeMo framework, signaling enterprise adoption of multi-agent architectures has reached maturity. &#8226; New perplexity-aware scaling laws promise more efficient LLM training, moving beyond the brute-force &#8220;more data is better&#8221; paradigm. &#8226; Responsible AI frameworks gain traction with consensus-driven reasoning approaches for explainability and governance. &#8226; MIT Technology Review calls out AI boosterism on social media, with Google DeepMind&#8217;s Demis Hassabis publicly embarrassed by overhyped claims.</p><div><hr></div><h3>NEMO-4-PAYPAL: Enterprise Multi-Agent Systems Go Production</h3><p>PayPal&#8217;s announcement of its Commerce Agent represents a significant milestone in enterprise AI adoption. Built on NVIDIA&#8217;s NeMo framework with fine-tuned Nemotron models, this multi-agent system handles search and discovery at production scale. The partnership demonstrates that the theoretical promise of agentic architectures is translating into real-world deployments. What&#8217;s notable isn&#8217;t just the technology, but the validation of the multi-agent paradigm: instead of monolithic models trying to do everything, specialized agents coordinate to handle complex workflows. For enterprise architects evaluating agentic systems, PayPal&#8217;s deployment provides a blueprint, though the reliance on specialized hardware partnerships raises legitimate questions about vendor lock-in and portability.</p><p>Link: <a href="https://arxiv.org/abs/2512.21578">https://arxiv.org/abs/2512.21578</a></p><div><hr></div><h3>The Perplexity Paradox: Smarter Scaling Laws for LLM Training</h3><p>Researchers have proposed a novel perplexity-aware data scaling law that challenges conventional wisdom about continual pre-training. The current power-law relationship between dataset size and test loss yields diminishing returns, leading to suboptimal data utilization and inefficient training. The new approach suggests that measuring perplexity landscapes can more accurately predict performance than simply counting tokens. For organizations running expensive training jobs, this represents potential cost savings of millions of dollars. The implication is clear: the &#8220;scale is all you need&#8221; era may be ending, replaced by smarter, more efficient approaches to model development.</p><p>Link: <a href="https://arxiv.org/abs/2512.21515">https://arxiv.org/abs/2512.21515</a></p><div><hr></div><h3>Responsible AI: Consensus-Driven Reasoning for Explainable Agents</h3><p>As AI agents gain autonomy, the challenges of explainability, accountability, and governance become critical. A new framework proposes consensus-driven reasoning to address these concerns, coordinating Large Language Models, Vision Language Models, tools, and external services while maintaining transparency. The approach is particularly relevant as enterprises deploy agentic systems that influence downstream decisions. This isn&#8217;t just academic research; it&#8217;s a direct response to regulatory pressure and enterprise risk management requirements. For decision-makers, this signals that responsible AI isn&#8217;t optional; it&#8217;s becoming a technical requirement baked into system design.</p><p>Link: <a href="https://arxiv.org/abs/2512.21699">https://arxiv.org/abs/2512.21699</a></p><div><hr></div><h3>The Hype Reckoning: When AI Boosterism Backfires</h3><p>MIT Technology Review published a sharp critique of AI boosterism on social media, centered on an incident where Google DeepMind CEO Demis Hassabis called out an OpenAI researcher&#8217;s overhyped claims about GPT-5 solving mathematical problems. Hassabis&#8217;s three-word response, &#8220;This is embarrassing,&#8221; encapsulates the growing backlash against hyperbolic AI announcements. The piece argues that social media incentives reward sensationalism over accuracy, creating a feedback loop that damages the field&#8217;s credibility. For builders, the lesson is straightforward: let your work speak for itself, and be skeptical of claims that seem too good to be true. The gap between demo and production remains the only metric that matters.</p><p>Link: <a href="https://www.technologyreview.com/2025/12/23/1130393/how-social-media-encourages-the-worst-of-ai-boosterism/">https://www.technologyreview.com/2025/12/23/1130393/how-social-media-encourages-the-worst-of-ai-boosterism/</a></p>]]></content:encoded></item><item><title><![CDATA[Hello Builders, issue #5]]></title><description><![CDATA[News from the trenches of AI in the week Dec 13th - 19th]]></description><link>https://ai.techonthestack.com/p/hello-builders-issue-5</link><guid isPermaLink="false">https://ai.techonthestack.com/p/hello-builders-issue-5</guid><dc:creator><![CDATA[Luca Bianchi]]></dc:creator><pubDate>Sat, 20 Dec 2025 12:41:13 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/6611ac78-8faf-4152-909d-e821429822cb_420x300.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Hello Builders,<br>This week, we start wrapping up 2025, a year that has been transformative in many aspects: the AI ecosystem was defined by a powerful tension between a necessary, sobering hype correction and the relentless pace of underlying technical progress. While industry leaders like Sam Altman and Mark Zuckerberg openly acknowledge a financial bubble of historic proportions, with trillions of dollars in planned infrastructure spending dwarfing the dot-com era, the fundamental limitations of current LLM-based approaches are becoming clearer. Yann LeCun&#8217;s departure from Meta to pursue alternative architectures underscores a growing schism in the path to AGI, just as Chinese open-weight models like Qwen and DeepSeek are outpacing their US counterparts in technology. This market irrationality is paradoxically juxtaposed with concrete, measurable breakthroughs: OpenAI&#8217;s new FrontierScience benchmark revealed a staggering improvement in scientific reasoning capabilities over the last two years, proving that even as the hype bubble swells, the technology&#8217;s real-world potential to accelerate science is advancing rapidly.</p><h2>This week&#8217;s signal in the noise</h2><p>&#8226; The <strong>AI bubble</strong> is now openly acknowledged by tech leaders, with planned infrastructure spending reaching an unprecedented $12 trillion scale, dwarfing previous tech booms.</p><p>&#8226; Yann LeCun&#8217;s departure from Meta signals a fundamental schism in AI research, challenging the dominant LLM scaling hypothesis and <strong>seeking new paths to AGI</strong>.</p><p>&#8226; <strong>Chinese</strong> open-weight models are now technologically <strong>leading the ecosystem</strong>, creating a new competitive and geopolitical landscape for AI development.</p><p>&#8226; OpenAI&#8217;s FrontierScience benchmark demonstrates a <strong>dramatic leap in scientific reasoning</strong> (from 39% to 92% on GPQA in two years), proving tangible progress amidst the hype.</p><div><hr></div><h3>The Great Hype Correction of 2025</h3><p>2025 is being framed as the year of the great AI &#8220;reckoning,&#8221; a necessary market and intellectual correction to the inflated expectations that have been set since 2022. The symbolic peak of this hype cycle was the underwhelming launch of GPT-5 in August, which, after months of grand promises, felt merely incremental. This has led to a broader questioning of the exponential progress narrative, comparing the current state of LLMs to the mature smartphone market, where annual updates bring minor improvements rather than revolutions. This shift demands a strategic recalibration, shifting focus from chasing ever-larger models to finding sustainable, high-value applications for the powerful yet limited tools that exist today. It&#8217;s a prompt to reassess technology roadmaps, distinguish between genuine capability and marketing-driven hype, and prepare for a period of disillusionment in which only the most resilient and value-focused applications will thrive.</p><p><a href="https://www.technologyreview.com/2025/12/15/1129174/the-great-ai-hype-correction-of-2025/">https://www.technologyreview.com/2025/12/15/1129174/the-great-ai-hype-correction-of-2025/</a></p><div><hr></div><h3>The Schism in the Path to AGI: Beyond Token Prediction</h3><p>A foundational debate, crystallized in a public discussion between Meta&#8217;s Yann LeCun and DeepMind&#8217;s Adam Brown, is challenging the industry&#8217;s core architectural assumptions. LeCun argues that the dominant approach of autoregressively predicting discrete tokens is a dead end for achieving true intelligence, as it fails to grasp the continuous, high-dimensional nature of the real world. He highlights the massive inefficiency of LLMs compared to a child&#8217;s learning process, advocating for new architectures like JEPA that learn abstract world models. This critique gained significant weight with the news of LeCun&#8217;s departure from Meta to form a new company focused on &#8220;Advanced Machine Intelligence&#8221; (AMI). This development signals a critical juncture for technology strategy, suggesting that diversifying research and development efforts beyond simply scaling existing LLMs is crucial. Betting entirely on the current paradigm may mean ignoring the architectural breakthroughs required for the next level of machine intelligence and real-world interaction.</p><p><a href="https://the-decoder.com/the-case-against-predicting-tokens-to-build-agi/">https://the-decoder.com/the-case-against-predicting-tokens-to-build-agi/</a></p><div><hr></div><h3>The East is Open: China&#8217;s Dominance in Open-Weight Models</h3><p>The competitive landscape for foundational models has been quietly redrawn, with Chinese companies now leading the open-weight ecosystem. The release of DeepSeek R1 in January 2025 was a watershed moment, technologically challenging and, in some cases, surpassing Western counterparts. Today, Alibaba&#8217;s Qwen model family is the most downloaded in the world, and models from labs like DeepSeek and Moonshot AI are setting new performance benchmarks, particularly in math and reasoning. While US companies like Airbnb are beginning to adopt these models for their cost and performance advantages, broader adoption is hampered by compliance and geopolitical concerns. This trend presents a complex strategic challenge: ignoring the superior performance of these models is a competitive risk, but adopting them introduces significant supply chain and data governance questions that must be carefully navigated.</p><div class="embedded-post-wrap" data-attrs="{&quot;id&quot;:181645140,&quot;url&quot;:&quot;https://www.understandingai.org/p/the-best-chinese-open-weight-models&quot;,&quot;publication_id&quot;:1501429,&quot;publication_name&quot;:&quot;Understanding AI&quot;,&quot;publication_logo_url&quot;:&quot;https://substackcdn.com/image/fetch/$s_!bNw0!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0c71d945-86dd-4042-87bd-974ed65380bb_420x420.png&quot;,&quot;title&quot;:&quot;The best Chinese open-weight models &#8212; and the strongest US rivals&quot;,&quot;truncated_body_text&quot;:&quot;DeepSeek&#8217;s release of R1 in January shocked the world. It came just four months after OpenAI announced its first reasoning model, o1. The model parameters were released openly. And DeepSeek R1 powered the first consumer-facing chatbot to show the full chain of thought before answering.&quot;,&quot;date&quot;:&quot;2025-12-15T15:58:12.679Z&quot;,&quot;like_count&quot;:66,&quot;comment_count&quot;:4,&quot;bylines&quot;:[{&quot;id&quot;:259110405,&quot;name&quot;:&quot;Kai Williams&quot;,&quot;handle&quot;:&quot;chiwilliams&quot;,&quot;previous_name&quot;:null,&quot;photo_url&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/82f6e93a-4715-4605-b0b3-f438188a2eaa_1028x1028.jpeg&quot;,&quot;bio&quot;:&quot;I'm a reporter at Understanding AI, supported through the Tarbell Center for AI Journalism. Previously, I did AI safety research through the MATS program. I graduated from Swarthmore College with a degree in math and music.&quot;,&quot;profile_set_up_at&quot;:&quot;2025-06-09T16:58:31.191Z&quot;,&quot;reader_installed_at&quot;:null,&quot;publicationUsers&quot;:[{&quot;id&quot;:6448318,&quot;user_id&quot;:259110405,&quot;publication_id&quot;:6319673,&quot;role&quot;:&quot;admin&quot;,&quot;public&quot;:true,&quot;is_primary&quot;:false,&quot;publication&quot;:{&quot;id&quot;:6319673,&quot;name&quot;:&quot;Kai Williams&quot;,&quot;subdomain&quot;:&quot;chiwilliams&quot;,&quot;custom_domain&quot;:null,&quot;custom_domain_optional&quot;:false,&quot;hero_text&quot;:&quot;I'm a reporter at Understanding AI, supported through the Tarbell Center for AI Journalism. Previously, I did AI safety research through the MATS program. I graduated from Swarthmore College with a degree in math and music.&quot;,&quot;logo_url&quot;:null,&quot;author_id&quot;:259110405,&quot;primary_user_id&quot;:259110405,&quot;theme_var_background_pop&quot;:&quot;#FF6719&quot;,&quot;created_at&quot;:&quot;2025-09-18T22:38:28.801Z&quot;,&quot;email_from_name&quot;:null,&quot;copyright&quot;:&quot;Kai Williams&quot;,&quot;founding_plan_name&quot;:null,&quot;community_enabled&quot;:true,&quot;invite_only&quot;:false,&quot;payments_state&quot;:&quot;disabled&quot;,&quot;language&quot;:null,&quot;explicit&quot;:false,&quot;homepage_type&quot;:&quot;profile&quot;,&quot;is_personal_mode&quot;:true}},{&quot;id&quot;:6300103,&quot;user_id&quot;:259110405,&quot;publication_id&quot;:1501429,&quot;role&quot;:&quot;contributor&quot;,&quot;public&quot;:true,&quot;is_primary&quot;:false,&quot;publication&quot;:{&quot;id&quot;:1501429,&quot;name&quot;:&quot;Understanding AI&quot;,&quot;subdomain&quot;:&quot;understandingai&quot;,&quot;custom_domain&quot;:&quot;www.understandingai.org&quot;,&quot;custom_domain_optional&quot;:false,&quot;hero_text&quot;:&quot;Exploring how AI works and how it's changing our world.&quot;,&quot;logo_url&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/0c71d945-86dd-4042-87bd-974ed65380bb_420x420.png&quot;,&quot;author_id&quot;:101111787,&quot;primary_user_id&quot;:101111787,&quot;theme_var_background_pop&quot;:&quot;#9A6600&quot;,&quot;created_at&quot;:&quot;2023-03-17T14:54:38.234Z&quot;,&quot;email_from_name&quot;:&quot;Understanding AI&quot;,&quot;copyright&quot;:&quot;Timothy B Lee&quot;,&quot;founding_plan_name&quot;:&quot;Founding Member&quot;,&quot;community_enabled&quot;:true,&quot;invite_only&quot;:false,&quot;payments_state&quot;:&quot;enabled&quot;,&quot;language&quot;:null,&quot;explicit&quot;:false,&quot;homepage_type&quot;:&quot;magaziney&quot;,&quot;is_personal_mode&quot;:false}}],&quot;is_guest&quot;:false,&quot;bestseller_tier&quot;:1000,&quot;status&quot;:{&quot;bestsellerTier&quot;:1000,&quot;subscriberTier&quot;:null,&quot;leaderboard&quot;:null,&quot;vip&quot;:false,&quot;badge&quot;:{&quot;type&quot;:&quot;bestseller&quot;,&quot;tier&quot;:1000},&quot;paidPublicationIds&quot;:[],&quot;subscriber&quot;:null}}],&quot;utm_campaign&quot;:null,&quot;belowTheFold&quot;:true,&quot;type&quot;:&quot;newsletter&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="EmbeddedPostToDOM"><a class="embedded-post" native="true" href="https://www.understandingai.org/p/the-best-chinese-open-weight-models?utm_source=substack&amp;utm_campaign=post_embed&amp;utm_medium=web"><div class="embedded-post-header"><img class="embedded-post-publication-logo" src="https://substackcdn.com/image/fetch/$s_!bNw0!,w_56,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0c71d945-86dd-4042-87bd-974ed65380bb_420x420.png" loading="lazy"><span class="embedded-post-publication-name">Understanding AI</span></div><div class="embedded-post-title-wrapper"><div class="embedded-post-title">The best Chinese open-weight models &#8212; and the strongest US rivals</div></div><div class="embedded-post-body">DeepSeek&#8217;s release of R1 in January shocked the world. It came just four months after OpenAI announced its first reasoning model, o1. The model parameters were released openly. And DeepSeek R1 powered the first consumer-facing chatbot to show the full chain of thought before answering&#8230;</div><div class="embedded-post-cta-wrapper"><span class="embedded-post-cta">Read more</span></div><div class="embedded-post-meta">4 months ago &#183; 66 likes &#183; 4 comments &#183; Kai Williams</div></a></div><div><hr></div><h3>The $12 Trillion Bubble: AI&#8217;s Unprecedented Infrastructure Bet</h3><p>The financial scale of the AI boom has entered a new territory of irrational exuberance, with industry leaders now openly discussing being in a bubble. The numbers are staggering: OpenAI has pledged $500 billion for data centers, with a moonshot goal of building infrastructure that could cost over $12 trillion. Bain estimates the industry needs to generate $2 trillion in annual revenue by 2030 just to justify the current spending wave&#8212;more than the combined 2024 revenue of Amazon, Apple, Alphabet, Microsoft, Meta, and Nvidia. Unprofitable startups like OpenAI and Anthropic are projected to burn through $140 billion and $20 billion, respectively, before 2030. This environment, characterized by circular deals and a chase for AGI, creates systemic risk. For decision-makers, this underscores the critical need for rigorous ROI analysis on AI investments and highlights the immense financial risk concentrated in a few foundational model providers, whose potential failure could have cascading effects across the industry.</p><p>Link: <a href="https://www.technologyreview.com/2025/12/15/1129183/what-even-is-the-ai-bubble/">https://www.technologyreview.com/2025/12/15/1129183/what-even-is-the-ai-bubble/</a></p><div><hr></div><h3>From 39% to 92%: Measuring Real Progress in Scientific Reasoning</h3><p>Cutting through the hype cycle, OpenAI has provided concrete evidence of dramatic capability improvements with its new FrontierScience benchmark. On the &#8220;Google-Proof&#8221; GPQA science benchmark, OpenAI&#8217;s models improved from a 39% score (GPT-4 in 2023) to a remarkable 92% with GPT-5.2 in 2025, surpassing the 70% expert baseline. The new, more challenging FrontierScience benchmark, created by PhDs and Olympiad medalists, shows GPT-5.2 achieving 77% on Olympiad-level problems and 25% on open-ended research tasks. This data provides a crucial signal: while true open-ended research remains a challenge, AI's ability to perform high-level, structured scientific reasoning is advancing at an astonishing rate. This indicates that the technology is ready to be deployed to accelerate complex, structured workflows in R&amp;D, engineering, and data analysis, moving beyond simple automation to become a genuine partner in discovery and innovation.</p><p>Link: <a href="https://openai.com/index/frontierscience/">https://openai.com/index/frontierscience/</a></p>]]></content:encoded></item><item><title><![CDATA[Hello Builders, issue #4]]></title><description><![CDATA[News from the trenches of AI in the week Dec 5th - 12th]]></description><link>https://ai.techonthestack.com/p/hello-builders-issue-4</link><guid isPermaLink="false">https://ai.techonthestack.com/p/hello-builders-issue-4</guid><dc:creator><![CDATA[Luca Bianchi]]></dc:creator><pubDate>Sun, 14 Dec 2025 23:36:21 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/068945a1-4bfa-4436-b932-1b601a7c7861_420x300.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>The week of December 5-12, 2025, marked a critical inflection point in the AI industry, characterized by intense competitive dynamics in foundation models and the formalization of agent engineering as a distinct discipline. Google and OpenAI engaged in direct competition with synchronized releases, while enterprise adoption of AI agents accelerated beyond coding into cross-functional workflows. Infrastructure innovations, particularly in custom silicon and development frameworks, continue to reshape the competitive landscape.</p><h2><strong>This week&#8217;s signal in the noise</strong></h2><ul><li><p>Foundation model competition intensified with GPT-5.2 and Gemini 3 Pro releases</p></li><li><p>Agent engineering emerged as a formal discipline combining product, engineering, and data science</p></li><li><p>$715M+ in venture funding across AI infrastructure and applications</p></li><li><p>NVIDIA introduced CUDA Tile for hardware-abstracted GPU programming</p></li><li><p>Anthropic published groundbreaking interpretability research revealing internal LLM reasoning</p></li><li><p>Responsible AI discussion fosters at re:Invent</p></li></ul><div><hr></div><h3>Agent Engineering: A New Discipline</h3><p>LangChain formalized &#8220;agent engineering&#8221; as a distinct practice area, shifting from experimental AI to production-grade systems.</p><p><strong>Technical Framework:</strong> The methodology follows an iterative cycle: Build &#8594; Test &#8594; Ship &#8594; Observe &#8594; Refine &#8594; Repeat. This requires three core skillsets working together: product thinking (encompassing prompt engineering and evaluation definition), engineering (covering tool development and durable execution), and data science (handling performance measurement and A/B testing). The production reality introduces a fundamental challenge: &#8220;every input is an edge case&#8221; because natural language creates an unlimited input space.</p><p>Companies like Clay, Vanta, LinkedIn, and Cloudflare have successfully deployed production agents using this methodology. The critical insight is that shipping is the primary learning mechanism, not the endpoint. Traditional software debug approaches fail because the logic resides inside models rather than in explicit code. The concept of an agent &#8220;working&#8221; is non-binary: achieving 99.99% uptime doesn&#8217;t guarantee correct behavior across all scenarios.</p><p><strong>Source:</strong> <a href="https://blog.langchain.com/agent-engineering-a-new-discipline/">LangChain Blog</a></p><div><hr></div><h3>Enterprise AI Agent Adoption</h3><p>Survey of 500+ technical leaders reveals rapid progression from task automation to cross-functional workflows. Actual deployment shows that 57% of organizations use agents for multi-stage workflows, and 16% run cross-functional processes across teams. Looking ahead to 2026, 81% plan to tackle more complex use cases, and 80% report measurable economic returns. The highest adoption is in coding (90%), with full coverage of the entire development lifecycle, including planning (58%), generation (59%), documentation (59%), and review/testing (59%). Data analysis and report generation follow at 60%, with internal process automation at 48% showing cross-functional workflow optimization.</p><p>Thomson Reuters uses Claude to power CoCounsel, their AI legal platform, enabling lawyers to access 150 years of case law and 3,000 domain experts in minutes rather than hours of manual document searching. In healthcare, Doctolib rolled out Claude Code across its entire engineering team, replacing legacy testing infrastructure in hours rather than weeks and shipping features 40% faster. </p><p><strong>Source:</strong> <a href="https://claude.com/blog/how-enterprises-are-building-ai-agents-in-2026">Claude Blog - State of AI Agents 2026</a></p><div><hr></div><h3>Tracing LLM Internal Reasoning</h3><p>Anthropic published breakthrough interpretability research revealing how Claude &#8220;thinks&#8221; internally. The research extends prior feature-mapping work into computational &#8220;circuits,&#8221; studying Claude 3.5 Haiku across 10 crucial behaviors. The approach draws on neuroscience methods to understand thinking organisms.</p><p>The research revealed six major insights into how Claude processes information internally. The current method captures only a fraction of total computation and requires hours of human effort per short prompt. Scaling to complex reasoning chains with thousands of words will require methodological improvements and, potentially, AI assistance for interpretation.</p><p><strong>Source:</strong> <a href="https://www.anthropic.com/news/tracing-thoughts-language-model">Anthropic Research</a></p><div><hr></div><h3>NVIDIA CUDA Tile: Hardware-Abstracted GPU Programming</h3><p>CUDA 13.1 introduced CUDA Tile, the largest advancement since CUDA&#8217;s 2006 invention. CUDA Tile introduces a virtual instruction set for tile-based parallel programming that abstracts specialized hardware, including tensor cores and TMA (Tensor Memory Accelerators). The tile model represents a fundamental shift in which developers partition data into blocks, and the compiler maps them to threads, in contrast to the SIMT model, where developers map data to both blocks and threads. This approach is analogous to NumPy for Python, where developers specify bulk operations and the runtime handles execution transparently. CUDA Tile reduces the burden of code rewrite across GPU generations, lowers the barrier to using tensor cores, and lays a foundation for higher-level AI development tools.</p><p><strong>Source:</strong> <a href="https://developer.nvidia.com/blog/focus-on-your-algorithm-nvidia-cuda-tile-handles-the-hardware">NVIDIA Developer Blog</a></p><div><hr></div><h3>OpenAI GPT-5.2 &#8220;Garlic&#8221; Release and Google Deep Research</h3><p>GPT-5.2 was released on December 11 amid a &#8220;code red&#8221; situation triggered by Gemini 3&#8217;s impact. CEO Sam Altman shifted resources to improving ChatGPT, with the expectation of exiting code-red status by January 2026.</p><p>OpenAI reports that GPT-5.2 tops SWE-Bench Pro for agentic coding performance and GPQA Diamond for graduate-level scientific reasoning. On GDPval, it beat or tied top professionals on 70.9% of well-specified tasks.</p><p><strong>Improvements Over GPT-5.1:</strong> The model shows significant enhancements across spreadsheet generation, presentation creation, code writing, long-form text understanding, and image processing.</p><p>In the meantime, the reimagined research agent based on Gemini 3 Pro introduces the Interactions API for embedding research into third-party apps, marking an industry first for developer access to advanced research capabilities. It handles large context dumps and synthesizes mountains of information.</p><p><strong>Source:</strong> <a href="https://gemini.google/overview/deep-research/">Google Blog</a></p><div><hr></div><h3><strong>Insights from an AI Human-in-the-Loop Roundtable</strong></h3><p>I recently sat in on a roundtable between AWS Heroes and Now Go Build CTO Fellows tackling real AI challenges: serving vulnerable populations with tiny teams, building offline education systems for conflict zones, and validating healthcare data at scale. The most valuable insights weren&#8217;t technical solutions&#8212;they were realizations. Like the healthcare org that discovered their &#8220;AI problem&#8221; was actually basic statistics. Or the finding that &#8220;30% chance of error&#8221; triggers better decisions than &#8220;70% confidence.&#8221; The conversation kept returning to one question: Are we using AI to enable people or replace them? The technical problems are solvable. The harder work is solving them for human flourishing, not just scale.</p><p><strong>Source:</strong> <a href="https://open.substack.com/pub/techonthestack/p/a-conversation-about-ai-that-actually?utm_campaign=post-expanded-share&amp;utm_medium=web">A conversation About AI tha Actually Matters</a></p>]]></content:encoded></item><item><title><![CDATA[A Conversation About AI That Actually Matters]]></title><description><![CDATA[AWS Heroes meet Now Go Build CTO Fellows at re:Invent]]></description><link>https://ai.techonthestack.com/p/a-conversation-about-ai-that-actually</link><guid isPermaLink="false">https://ai.techonthestack.com/p/a-conversation-about-ai-that-actually</guid><dc:creator><![CDATA[Luca Bianchi]]></dc:creator><pubDate>Sat, 13 Dec 2025 23:13:56 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!RYQk!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc42ee8c8-9421-4462-b1d6-e78ab03e0632_1536x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!RYQk!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc42ee8c8-9421-4462-b1d6-e78ab03e0632_1536x1024.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!RYQk!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc42ee8c8-9421-4462-b1d6-e78ab03e0632_1536x1024.png 424w, https://substackcdn.com/image/fetch/$s_!RYQk!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc42ee8c8-9421-4462-b1d6-e78ab03e0632_1536x1024.png 848w, https://substackcdn.com/image/fetch/$s_!RYQk!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc42ee8c8-9421-4462-b1d6-e78ab03e0632_1536x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!RYQk!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc42ee8c8-9421-4462-b1d6-e78ab03e0632_1536x1024.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!RYQk!,w_2400,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc42ee8c8-9421-4462-b1d6-e78ab03e0632_1536x1024.png" width="1200" height="800.2747252747253" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c42ee8c8-9421-4462-b1d6-e78ab03e0632_1536x1024.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:false,&quot;imageSize&quot;:&quot;large&quot;,&quot;height&quot;:971,&quot;width&quot;:1456,&quot;resizeWidth&quot;:1200,&quot;bytes&quot;:2471058,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://aletheia1982.substack.com/i/181546390?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc42ee8c8-9421-4462-b1d6-e78ab03e0632_1536x1024.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:&quot;center&quot;,&quot;offset&quot;:false}" class="sizing-large" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!RYQk!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc42ee8c8-9421-4462-b1d6-e78ab03e0632_1536x1024.png 424w, https://substackcdn.com/image/fetch/$s_!RYQk!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc42ee8c8-9421-4462-b1d6-e78ab03e0632_1536x1024.png 848w, https://substackcdn.com/image/fetch/$s_!RYQk!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc42ee8c8-9421-4462-b1d6-e78ab03e0632_1536x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!RYQk!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc42ee8c8-9421-4462-b1d6-e78ab03e0632_1536x1024.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Around the table sat people managing educational platforms for 800,000 students, disaster-response systems in active war zones, healthcare assessments for vulnerable populations, and food-recovery operations across 20 regions. AWS Heroes&#8212;technical experts who&#8217;ve built systems at massive scale&#8212;sat alongside Now Go Build CTO Fellows&#8212;technology leaders working at the absolute frontlines of human need.</p><p>The question on the table: How do you actually deploy AI when getting it wrong means people get hurt?</p><p>Not &#8220;how do you optimize for engagement&#8221; or &#8220;how do you reduce churn.&#8221; How do you build systems that serve people who can&#8217;t advocate for themselves, in places where infrastructure barely works, with teams of six people trying to reach millions?</p><h2>The AI Tutor That Has to Work Without the Internet</h2><p>One of the first challenges came from someone running an educational platform serving 800,000 learners across every country. Their team: 25-30 people. Their current approach: AI helps create and localize content, but PhD-level experts review everything before it reaches students.</p><p>It&#8217;s careful. It&#8217;s responsible. And it doesn&#8217;t scale to where they need to go.</p><p>&#8220;In the countries where we work, students miss entire days of school regularly,&#8221; they explained. Military conflicts. Infrastructure failures. In some places, every Monday. The dream: AI tutors that run entirely on phones, work without internet, and can teach kids even when nothing else is working.</p><p>But here&#8217;s the catch&#8212;how do you ensure safety when the AI can&#8217;t check back with a server for guardrails? When there&#8217;s no human in the loop because there&#8217;s literally no loop to be in?</p><p>The room started discussing. Keep the AI tightly bound&#8212;it can only talk about specific curriculum topics. If a student asks something outside that scope, it just says, &#8220;I don&#8217;t know how to answer that.&#8221; Don&#8217;t try to be helpful beyond your domain.</p><p>Someone raised a more subtle problem: &#8220;Vector search will find semantically similar content, but it will miss gaps of information that are relevant but don&#8217;t match.&#8221; In other words, the AI might find related information, but miss the crucial context that makes it meaningful. The solution involves explicitly mapping how concepts relate to each other, not just how similar they are.</p><p>For the highest-stakes applications, there&#8217;s even a way to get mathematical proof of safety&#8212;similar to how Boeing and Airbus verify their flight systems. The AI generates the rules, humans verify that they are correct, and the system is guaranteed to stay within those boundaries.</p><p>But then someone pushed back on the whole premise: &#8220;The research shows learning requires friction. We shouldn&#8217;t eliminate struggle; we should give learners tools to manage it productively.&#8221; This hit something important. The goal isn&#8217;t to make everything easy. Real learning involves being challenged, even frustrated. The question is how to support students through that difficulty, not eliminate it.</p><p>Later, someone set an ambitious bar: &#8220;Until the learning is as addictive as a Netflix series where you watch eight hours of it, we have not got our job done.&#8221; But they also acknowledged a tension: &#8220;My kids use some of these AI-enhanced learning platforms now, and they hate them because they know the AI is constantly assessing them.&#8221;</p><p>There&#8217;s the paradox. We want engagement. But surveillance isn&#8217;t engaging&#8212;it&#8217;s oppressive. The technology that measures everything risks destroying the intrinsic motivation that makes learning work.</p><h2>The Healthcare System That Had to Be Perfect</h2><p>Then came a different kind of challenge. A healthcare organization analyzes surveys and personal stories to identify psychosocial factors&#8212;abuse, housing instability, depression&#8212;that healthcare workers need to know about. The AI&#8217;s job: read through 57 questions plus long-form written responses and flag what matters.</p><p>But these flags go into clinical decisions. They have to be right. &#8220;How do I ensure quality control with a small team of subject matter experts when we&#8217;re about to scale?&#8221;</p><p>Turns out, most of it was structured&#8212;multiple choice, scales, yes/no answers. Then they discovered this wasn&#8217;t really an AI problem. For structured data, you can use statistical methods that give you the same answer every single time. No randomness. No drift. Just consistency. The AI only needed to handle the free-text responses where people write about their lives in their own words. That&#8217;s where language understanding matters.</p><p>But even with the right technical approach, the validation problem remained. How do you check thousands of assessments with a handful of experts? Use project management tools teams already know. When the AI makes a prediction, it automatically creates a review task. The expert approves or corrects it. The corrections feed back into the model. No custom software needed.</p><p>Run the same data through multiple different AI systems. If they all agree, confidence goes up. If they disagree, it gets flagged for mandatory human review. &#8220;That&#8217;s what Boeing and Airbus do,&#8221; someone noted. &#8220;Multiple systems, voting mechanisms.&#8221; You don&#8217;t need to validate everything. Check 100% when the AI says it&#8217;s uncertain. Sample the rest. Track accuracy over time. For validations that don&#8217;t require deep medical expertise, use crowd validation where multiple people review the same output, and consensus determines acceptance.</p><p>Then came a fascinating insight about human psychology. How you frame AI predictions changes how people respond to them.</p><p>&#8220;There&#8217;s a 30% chance this is wrong&#8221; makes people more careful than &#8220;70% confidence it&#8217;s right&#8221;&#8212;even though these mean the exact same thing. For life-and-death decisions, frame uncertainty as risk. It changes behavior.</p><h2>Trust in the Age of Lies</h2><p>They work across conflict zones&#8212;Ukraine, Turkey, Mexico, places where humanitarian workers need accurate information to stay alive. Every day, they compile situation reports from multiple sources: social media, news, and ground reports from local contacts. Here&#8217;s the nightmare scenario: a convincing video of an explosion makes it into a situation report. Workers avoid a safe route or miss a critical intervention window. Except the explosion never happened. It was AI-generated.</p><p>&#8220;False information doesn&#8217;t just waste time,&#8221; they said. &#8220;It puts lives at risk.&#8221;</p><p>And here&#8217;s what makes it harder: &#8220;The humanitarian workforce is significantly aging. With funding cuts, organizations aren&#8217;t hiring new people. The digital literacy problem is getting worse, not better.&#8221;</p><p>AI can scrape and aggregate information orders of magnitude faster than humans. But speed without accuracy is worse than useless&#8212;it&#8217;s dangerous.</p><p>A possible solution is to enforce data lineage. Track everything. Where did this information come from? How was it processed? Who else is reporting it? This isn&#8217;t just for validation&#8212;when something seems suspicious, you need to be able to trace back through every step. Look for technical indicators. Images and videos contain metadata&#8212;timestamps, GPS data, and visual artifacts from editing. These can be analyzed to flag content that appears manipulated.</p><p>Someone proposed an elegant solution to the training problem: build it directly into the workflow. Instead of separate training sessions, people forget about the system challenges that reviewers face in the moment. &#8220;This image has characteristics consistent with AI generation. What indicators do you see?&#8221;</p><p>The system doesn&#8217;t let them move forward until they articulate their reasoning. Right or wrong, they&#8217;re thinking critically about that specific case.</p><h2>Other Realities</h2><p><strong>Mapping with biased AI.</strong> Someone working on geospatial mapping explained that global AI models trained primarily on US and European data perform terribly in developing regions. Deploy them to map villages in Rwanda, and they miss buildings or misclassify structures entirely. Their solution turned the problem into an opportunity: local community members provide feedback that fine-tunes the models for their specific region. Simple mobile interfaces&#8212;swipe yes or no, is this a building?&#8212;create a continuous improvement loop. &#8220;This is about community inclusion,&#8221; they said. &#8220;Locals train the systems that map their environment. It addresses the bias in training data and makes the whole process transparent.&#8221;</p><p><strong>Health information in hostile territories.</strong> A reproductive health organization works in regions where providing its services is literally illegal. They need to provide accurate, potentially life-saving information, but they can&#8217;t collect the personal data needed for AI personalization. The constraints are severe. Users may be in immediate danger. The information must be medically precise. The system must detect emotional distress and escalate to human counselors when needed. Their approach emphasizes curation over generation: the AI draws only from pre-validated medical resources, every response includes clear citations, and there&#8217;s always a prominent path to reach a human counselor.</p><p><strong>When voice changes everything.</strong> Multiple people raised the challenge of serving populations with limited literacy, particularly in languages without strong written traditions. The shift from text to voice interfaces made a dramatic difference: &#8220;Once we moved from text chatbot to voice-based, we saw significantly better engagement with low-literacy populations. The emotion detection was critical&#8212;if someone sounds sad, you need to adapt the interaction.&#8221; Voice with emotion detection allows systems to respond not just to what people say, but how they say it. It makes the interaction feel human in a way text never can.</p><p>As the conversation wound on, certain truths kept surfacing, sometimes explicitly, sometimes as subtext to other discussions.</p><p><strong>Not every problem needs AI.</strong> For structured data with clear rules, traditional statistical methods often work better. They&#8217;re consistent, auditable, and don&#8217;t have the unpredictability of generative systems. The maturity isn&#8217;t in knowing the latest AI techniques&#8212;it&#8217;s in knowing when not to use them. Sometimes a problem is &#8220;just&#8221; a classic machine learning problem, not a generative AI problem.</p><p><strong>Small teams can provide oversight.</strong> You can&#8217;t manually review every AI output, but you also can&#8217;t deploy systems without oversight. The answer isn&#8217;t choosing one or the other&#8212;it&#8217;s being strategic about where human judgment adds value. Validate everything when confidence is low. Sample when confidence is high. Use tools teams already know rather than building custom systems. Have corrections automatically improve the model.</p><p><strong>Humans need context, not just answers.</strong> When AI makes a prediction, showing why it reached that conclusion helps humans effectively validate it. Citations, confidence scores, explicit reasoning&#8212;these aren&#8217;t nice-to-haves. They&#8217;re what make the partnership between human and machine actually work.</p><p><strong>How you frame things matters more than you&#8217;d think.</strong> That insight about &#8220;30% chance this is wrong&#8221; versus &#8220;70% confidence it&#8217;s right&#8221;&#8212;mathematically identical, psychologically different&#8212;kept coming up. For high-stakes decisions, frame uncertainty as risk. It changes how people think.</p><p><strong>Trust requires transparency.</strong> Track where information came from and how it was processed. When something goes wrong, trace back to find where the error was introduced. This isn&#8217;t bureaucracy&#8212;it&#8217;s the foundation of accountability.</p><p><strong>Training should be embedded, not scheduled.</strong> Rather than separate training sessions, people forget; they build challenges directly into the workflow. Make people articulate their reasoning before accepting or rejecting AI outputs. Provide guidance in the moment when they struggle. Learning that happens in context sticks.</p><p><strong>Some things should stay human.</strong> This was the insight that kept resonating. Some interactions should remain human even if they could technically be automated.</p><p>&#8220;Automation decisions aren&#8217;t purely technical&#8212;they encode values about human dignity and relationship importance.&#8221;</p><p>Technology should serve human flourishing, not just efficiency. Sometimes the relationship itself is the value being delivered.</p><h2>What does this mean?</h2><p>In typical tech discussions, AI is about optimization, efficiency, and competitive advantage. There, the question was whether a student in an active war zone could learn math. Whether a humanitarian worker would walk into danger based on fabricated information. Whether someone in crisis would get help or get dismissed.</p><p><strong>Start with the constraint, not the capability.</strong> The best solutions came from embracing limitations&#8212;offline-first design, minimal data collection, small teams&#8212;rather than treating them as compromises. Design for the world as it is, not as you wish it were.</p><p><strong>Choose boring technology when it solves the problem.</strong> The sophistication to use simpler statistical methods instead of AI, or traditional databases instead of vector search, often leads to better outcomes. Complexity should be justified by requirements, not assumed by default.</p><p><strong>Make humans and AI partners, not adversaries.</strong> The goal isn&#8217;t to validate whether the AI was &#8220;right.&#8221; It&#8217;s about combining human judgment with AI capabilities to make better decisions than either could alone. Design systems where humans add value, not just check boxes.</p><p><strong>Build feedback loops into everything.</strong> Every human correction should automatically improve the system. If your validation process doesn&#8217;t feed back into training, you&#8217;re missing the opportunity to get better over time.</p><p>&#8220;I&#8217;m interested in finding the balance between using AI to enable people versus using AI to replace them. The higher order tends to view it as replacement rather than enablement.&#8221;</p><p>That&#8217;s the real tension, isn&#8217;t it? Not the technical challenges&#8212;those are solvable. The harder question is what we&#8217;re solving them for. AI should amplify human capability, not substitute for human judgment. It should help small teams serve millions without losing the human connection that makes the service meaningful.</p><p>This isn&#8217;t idealism. It&#8217;s practical. Systems that lose the human element, that treat efficiency as the only goal, that optimize for metrics at the expense of relationships&#8212;they ultimately fail to serve the populations they&#8217;re designed to help.</p><p>Human-in-the-loop is the design principle that makes AI trustworthy for populations that can&#8217;t afford our mistakes. The technical problems&#8212;validation at scale, offline operation, handling uncertainty&#8212;those have solutions. The harder work is ensuring we&#8217;re solving them in service of human flourishing, not just scale.</p><p>Because somewhere, there&#8217;s a student who needs to learn despite conflict disrupting school. A healthcare worker making decisions that affect people&#8217;s lives. A humanitarian worker whose safety depends on accurate information. A person in crisis who needs help, not judgment. They can&#8217;t advocate for themselves in these technology decisions. So we have to advocate for them. We have to build systems worthy of their trust.</p><p>That&#8217;s what sustainable and ethical AI stands for.</p>]]></content:encoded></item><item><title><![CDATA[Hello Builders, issue #3]]></title><description><![CDATA[Your weekly briefing on the signals that actually matter in AI distilled, contextualized, and built for people who are shipping.]]></description><link>https://ai.techonthestack.com/p/hello-builders-issue-3</link><guid isPermaLink="false">https://ai.techonthestack.com/p/hello-builders-issue-3</guid><dc:creator><![CDATA[Luca Bianchi]]></dc:creator><pubDate>Sat, 06 Dec 2025 23:46:25 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/e7440845-d9fd-4ed6-a97c-e10dd61ccd9d_420x300.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>This week delivered an unusually dense concentration of enterprise-grade AI developments, anchored by AWS re:Invent 2025 and NeurIPS 2025. Google&#8217;s Gemini 3 broke the 1500 Elo barrier on reasoning benchmarks, DeepSeek released a 671B-parameter open-weight model at <strong>~90% lower cost</strong> than proprietary alternatives, and AWS and Google Cloud announced an unprecedented multicloud interconnect partnership.</p><h2><strong>This week&#8217;s signal in the noise</strong></h2><ul><li><p><strong>Model economics transformed</strong>: DeepSeek V3.2 at $0.21/$0.32 per million tokens (10-30x cheaper than proprietary models, MIT license). Gemini 3 hit 1501 Elo (first to break 1500). Mistral Large 3 launched at 80% below OpenAI pricing, with an Apache 2.0 license.</p></li><li><p><strong>Infrastructure developments</strong>: AWS Trainium3 delivers 4.4x performance boost; AWS-Google Cloud multicloud interconnect enables 1-100 Gbps private links in minutes. Google TPU v7 scales to 42.5 exaflops. H100 pricing cut by 44%.</p></li><li><p><strong>$1.7B+ funding</strong>: Black Forest Labs ($300M, $3.25B valuation), Harvey ($160M, $8B valuation). OpenAI acquired Neptune.ai&#8212;<strong>external services sunsetting, migration required</strong>. Marvell acquired Celestial AI for photonic interconnects.</p></li><li><p><strong>Regulatory/research updates</strong>: EU AI Act Annex III deadlines delayed 16 months (now Dec 2027). NeurIPS findings suggest RLVR hitting capability ceiling. NVIDIA achieved 4-bit LLM training matching 8-bit performance.</p></li></ul><div><hr></div><h3>Model releases reshape the competitive landscape</h3><p>December 1, 2025, marked what industry observers dubbed &#8220;Super Sunday&#8221;&#8212;exactly three years after ChatGPT&#8217;s launch&#8212;with multiple frontier-model announcements in a 48-hour window.</p><p><strong>Google Gemini 3 Pro</strong> achieved an <strong>LMArena Elo of 1501</strong>, becoming the first model to break the 1500 barrier. Its GPQA Diamond score of <strong>91.9%</strong> surpasses human expert performance (~89.8%), while the extended reasoning variant &#8220;Deep Think&#8221; reached <strong>93.8%</strong> on the same benchmark. The model supports a <strong>1-million-token context window</strong>, while <strong>DeepSeek V3.2</strong> represents the most significant open-weight release of the week. At <strong>671B total parameters</strong> (37B active per token) with a Mixture-of-Experts architecture, it achieves parity with GPT-5 on reasoning benchmarks while pricing at <strong>$0.21/million input tokens</strong> and <strong>$0.32/million output tokens</strong>&#8212;roughly 10-30x cheaper than proprietary alternatives. The cheap-model trend continues with&nbsp;<strong>the</strong>&nbsp;release this week of Mistral Large 3, priced approximately 80% below OpenAI&#8217;s flagship pricing. </p><ul><li><p>https://blog.google/products/gemini/gemini-3/</p></li><li><p>https://api-docs.deepseek.com/news/news251201</p></li></ul><div><hr></div><h2>Infrastructure investments accelerate across hyperscalers</h2><p>AWS re:Invent 2025 dominated infrastructure news with <strong>Trainium3</strong>, AWS&#8217;s first 3nm AI chip, and the <strong>AWS-Google Cloud multicloud interconnect</strong> partnership integrates AWS Direct Connect with Google Cloud Cross-Cloud Interconnect, enabling private links between platforms in minutes rather than weeks. Bandwidth ranges from <strong>1-100 Gbps</strong> with MACsec encryption and quad redundancy. In the meantime,&nbsp;<strong>Google&#8217;s TPU v6e (Trillium)</strong>&nbsp;reached general availability, and TPU v7 (Ironwood), previewed this week, with a total capacity of&nbsp;<strong>42.5 exaflops</strong>. Capital expenditure across the eight major hyperscalers is projected to increase <strong>44% year-over-year to $371 billion</strong> in 2025. JPMorgan estimates <strong>$5 trillion</strong> in global data center and AI infrastructure spending over the next five years. </p><ul><li><p>https://www.webpronews.com/amazon-and-google-launch-multicloud-networking-for-ai-and-outage-resilience/</p></li></ul><div><hr></div><h3>Funding rounds signal continued investor conviction</h3><p>Approximately <strong>$1.7 billion in disclosed funding</strong> closed during the week across enterprise AI infrastructure, vertical applications, and agentic AI platforms. <strong>Black Forest Labs</strong>&nbsp;raised&nbsp;<strong>$300 million in Series B</strong>&nbsp;at a&nbsp;<strong>$3.25 billion valuation</strong>. The company&#8217;s FLUX image generation models power Adobe, ElevenLabs, and Grok. <strong>Eon</strong>&nbsp;closed a&nbsp;<strong>$300 million Series D</strong>&nbsp;round led by Elad Gil at an approximate&nbsp;<strong>$4 billion valuation</strong>&nbsp;for its cloud data backup business. <strong>Harvey</strong>, the legal AI platform, secured <strong>$160 million</strong> at an <strong>$8 billion valuation</strong> (up from $5B in June), led by Andreessen Horowitz. The company serves 50+ of the top 100 AmLaw firms and surpassed <strong>$100 million ARR</strong> in August. Back-to-back funding rounds totaling <strong>$760 million</strong> indicate strong market validation.</p><ul><li><p>https://techcrunch.com/2025/12/01/black-forest-labs-raises-300m-at-3-25b-valuation/</p></li><li><p>https://techcrunch.com/2025/12/04/legal-ai-startup-harvey-confirms-8b-valuation/</p></li></ul><div><hr></div><h3>Research breakthroughs from NeurIPS challenge training assumptions</h3><p>NeurIPS 2025 (Nov 30 &#8211; Dec 7) produced several findings with direct enterprise implications. The <strong>Gated Attention</strong> paper from Alibaba&#8217;s Qwen team received Best Paper honors for introducing head-specific sigmoid gating after scaled dot-product attention. Tested across 30+ experiments on models up to 15B parameters trained on 3.5 trillion tokens, the technique eliminates the &#8220;attention sink&#8221; phenomenon, enables larger learning rates, and improves long-context extrapolation&#8212;all with minimal implementation overhead. The approach is <strong>already deployed in production Qwen3-Next models</strong>.</p><p>A critical RLVR (Reinforcement Learning from Verifiable Rewards) study received Best Paper Runner-Up recognition for demonstrating that <strong>RLVR improves sampling efficiency but does not elicit fundamentally new reasoning patterns</strong>. Base models achieve higher pass@k scores when k is large; distillation&#8212;not RLVR&#8212;introduces genuinely new reasoning capabilities. This finding suggests current post-training methodologies may be approaching fundamental limits.</p><p><strong>DeepSeekMath-V2</strong> demonstrated self-verifiable mathematical reasoning with a meta-verification system that checks verifier accuracy and reduces hallucinations. On the 2024 Putnam Competition, it scored <strong>118/120 points</strong>, compared to a top human score of 90. The model achieved gold-medal performance at IMO 2025 (83.3%, 5/6 problems) and the China Mathematical Olympiad 2024.</p><ul><li><p>https://blog.neurips.cc/2025/11/26/announcing-the-neurips-2025-best-paper-awards/</p></li><li><p>https://blog.neurips.cc/2025/11/26/announcing-the-neurips-2025-best-paper-awards/</p></li></ul><p></p>]]></content:encoded></item><item><title><![CDATA[The end of an era and the Renaissance Developer.]]></title><description><![CDATA[Werner Vogels' last keynote marked one of highest moments of re:Invent 2025, left us with his legacy to carry on.]]></description><link>https://ai.techonthestack.com/p/the-end-of-an-era-and-the-renaissance</link><guid isPermaLink="false">https://ai.techonthestack.com/p/the-end-of-an-era-and-the-renaissance</guid><dc:creator><![CDATA[Luca Bianchi]]></dc:creator><pubDate>Fri, 05 Dec 2025 17:58:52 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!94le!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F17fd0d1c-48db-4c36-bf3e-63630188bbf6_2586x1346.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!94le!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F17fd0d1c-48db-4c36-bf3e-63630188bbf6_2586x1346.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!94le!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F17fd0d1c-48db-4c36-bf3e-63630188bbf6_2586x1346.png 424w, https://substackcdn.com/image/fetch/$s_!94le!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F17fd0d1c-48db-4c36-bf3e-63630188bbf6_2586x1346.png 848w, https://substackcdn.com/image/fetch/$s_!94le!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F17fd0d1c-48db-4c36-bf3e-63630188bbf6_2586x1346.png 1272w, https://substackcdn.com/image/fetch/$s_!94le!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F17fd0d1c-48db-4c36-bf3e-63630188bbf6_2586x1346.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!94le!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F17fd0d1c-48db-4c36-bf3e-63630188bbf6_2586x1346.png" width="1456" height="758" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/17fd0d1c-48db-4c36-bf3e-63630188bbf6_2586x1346.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:758,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1269923,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://ai.techonthestack.com/i/180812632?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F17fd0d1c-48db-4c36-bf3e-63630188bbf6_2586x1346.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!94le!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F17fd0d1c-48db-4c36-bf3e-63630188bbf6_2586x1346.png 424w, https://substackcdn.com/image/fetch/$s_!94le!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F17fd0d1c-48db-4c36-bf3e-63630188bbf6_2586x1346.png 848w, https://substackcdn.com/image/fetch/$s_!94le!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F17fd0d1c-48db-4c36-bf3e-63630188bbf6_2586x1346.png 1272w, https://substackcdn.com/image/fetch/$s_!94le!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F17fd0d1c-48db-4c36-bf3e-63630188bbf6_2586x1346.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Two more words: &#8220;Werner&#8217;s out!&#8221;</p><p>When the mic drops on stage, we all know an era is over. Werner Vogels&#8217; last keynote marked one of the most emotional moments of this 2025 re:Invent and ended the best of a 14-year keynote series. The bittersweet sensation that comes when you realize one of the pioneers of our modern age decides it&#8217;s time to step back and let younger forces tell stories about these transformative times is both hard to elaborate on and leaves you with a feeling of a full circle.</p><p>Seniority is not the ability to hold the microphone; it&#8217;s the courage to hand it to the next generation: </p><blockquote><p>&#8220;There are so many amazing engineers at Amazon that have great stories to tell&#8230; It&#8217;s time for those younger, different voices of AWS to be in front of you.&#8221;</p></blockquote><p>It came after a one-hour packed discussion about how to deal with these transformative times. Werner's last gift is a message of hope and a call to become better builders, better professionals, better humans. The theme of the entire keynote can be summarized in one line:</p><blockquote><p><strong>The tools will change; the work is still yours.</strong></p></blockquote><p>From that premise, Vogels unfolds a framework for the&nbsp;<strong>Renaissance Developer</strong>&#8212;a model not just for engineers but for any technical leader who wants to stay relevant and effective in an AI-accelerated world. To make his point, Vogels does what experienced engineers do when everyone is panicking about the future: he zooms out over decades. Every wave followed the same pattern: tools changed dramatically, the work changed shape, and the <em>developer's identity</em> stayed the same: <strong>we build systems that matter. </strong>This is the context in which he answers the AI anxiety. We&#8217;re not at the end of development; we&#8217;re in yet another phase of abstraction. The difference today is the <strong>density of change</strong>. As Jeff Bezos has put it, we&#8217;re living at the epicenter of multiple &#8220;golden ages&#8221;&#8212;space, robotics, and AI&#8212;whose breakthroughs amplify one another.</p><p>Werner&#8217;s conclusion: if you anchor your career in specific tools, you&#8217;ll be swept away. If you anchor it in enduring capabilities, you&#8217;ll ride the wave. Those capabilities are what he calls the <strong>Renaissance Developer's qualities, </strong>drawing on a period when tools and ideas reinforced one another, reshaping civilization: the <strong>Renaissance</strong>. Tools didn&#8217;t replace humans; they <strong>expanded the surface area of what humans could explore</strong>.</p><h2>Curiosity</h2><p>The very first quality of the Renaissance Developer is curiosity: in a world of AI&#8209;accelerated change, it becomes a survival skill and a professional obligation. But curiosity requires experimentation and failure. If you are not allowed to fail, you can&#8217;t thrive. But it comes with pressure: too little leads to boredom and disengagement, while too much leads to overwhelm and paralysis. The Yerkes&#8211;Dodson law<strong> </strong>explains that the sweet spot is <em>enough challenge to stretch you without breaking you.</em></p><p>This comes with connections: &#8220;Learning isn&#8217;t just cognitive, it is social.&#8221; You don&#8217;t become a Renaissance Developer by sitting alone with tutorials; you grow by seeking out communities, sitting in user groups and conferences where ideas collide, lingering over coffee with peers while you argue about systems, and deliberately placing yourself in new contexts and constraints so that real experiences, real people, and real problems can stretch your thinking far beyond what isolated study can do.</p><p>As a leader, your job is not to protect your team from failure; it is to make failure cheap and reversible, to put people in environments where they are exposed to new ideas and constraints, and to reward learning behavior as much as successful outcomes.</p><h2><strong>Systems Thinking</strong></h2><p>The second quality is <strong>systems thinking</strong>. Not &#8220;distributed systems&#8221; in the narrow sense, but systems in the broader sense that <strong>Donella Meadows</strong> wrote about.</p><blockquote><p>&#8220;A system is a set of things&#8212;people, cells, or whatever&#8212;interconnected in such a way that they produce their own pattern of behavior over time.&#8221;</p></blockquote><p>Meadows invites us to stop staring at individual events and instead notice the structures that generate them. She asks us to look for the &#8220;stocks and flows&#8221; beneath the surface: the queues that build up in a support channel, the backlog that silently grows as new features outrun maintenance, the slow accumulation of technical debt that barely shows up in daily metrics but eventually locks a team in place. These are the reservoirs of a system. They fill and drain at different speeds, and they determine the patterns we see in performance, morale, and reliability. Meadows points to <strong>feedback loops</strong>, circuits of cause and effect that either reinforce a trend or hold it in check. In a reinforcing loop, every success makes the next success easier, like a product that improves, gains users, and generates the data that makes it better still. In a balancing loop, the system pushes back, like a throttling mechanism that slows calls when the load becomes too high. Our organizations, our processes, and our code are full of these loops. Some stabilize; others amplify; many conflict with each other in ways that only become visible over time. Crucially, systems have delays and nonlinearities. Actions and consequences are often separated in time and space: you approve a shortcut in architecture, and only months later, the on&#8209;call rotation begins to burn out. You improve incident response without touching the upstream causes and find that you are getting better and better at cleaning up the same problems. Because of these delays, intuitive fixes&#8212;more meetings, more reviews, more dashboards&#8212;can easily make behavior worse if they are placed in the wrong part of the system. This is why Meadows focuses on leverage<strong> points</strong>: small, carefully chosen interventions that produce outsized effects. You can change parameters&#8212;such as timeouts, thresholds, or team size&#8212;and achieve incremental improvement. You can change information flows&#8212;who sees which metrics, which stories reach leadership&#8212;and unlock better decisions. You can change rules and incentives&#8212;what gets rewarded in performance reviews or funded in roadmaps&#8212;and watch priorities shift without a single line of code. At the deepest level, you can change goals and mental models&#8212;what &#8220;good&#8221; looks like for reliability, cost, or user experience&#8212;and the entire architecture starts to evolve in a new direction. If you keep treating problems as isolated bugs or incidents, you will keep fighting the same fires. Once you see your work as part of a living system of stocks, flows, feedback loops, delays, and leverage points, you stop searching for silver bullets and start looking for thoughtful, well&#8209;placed changes that reshape behavior over time.</p><h2>Communication</h2><p>Natural language is powerful but ambiguous. Programming languages forced us to be precise because they refused the shortcuts and half-thoughts that humans understand from context. With AI-assisted coding, we are back to prompting in natural language, and vague prompts don&#8217;t just slow us down; they create misalignment&#8212;code that looks sophisticated but solves the wrong problem, hallucinated fields and APIs, and designs that quietly undermine the architecture they were meant to serve. Craft clear specifications as the bridge between human intent and machine-generated implementation. A good spec compresses ambiguity into clarity, forces us to decide what success looks like, which constraints we accept, and which trade-offs we will make, and becomes the shared reference point where product, engineering, and AI tools can actually align. Learn to ask &#8220;why&#8221; before &#8220;how.&#8221; When a customer asks, &#8220;What should we be doing with GenAI?&#8221;, the first move is not to list features but to ask, &#8220;Why are you asking? What problem or opportunity do you see?&#8221; From that deeper conversation comes sharper intent, a better specification, and a system that actually matters. For leaders, this is why communication is a core technical skill: diagrams, narratives, and specs are the instruments through which you shape the system, and in an AI-heavy workflow, the quality of your words still determines the quality of your systems.</p><h2>Ownership</h2><p>AI does not dilute accountability. &#8220;The work is yours, not the tools.&#8221; If you operate in regulated environments&#8212;healthcare, finance, critical infrastructure&#8212;and AI generates code that violates regulation, you cannot tell a regulator, &#8220;The AI did it.&#8221; Responsibility does not move with the autocomplete cursor. AI mainly amplifies two dangers: it can create verification debt, where code is generated faster than you can truly understand it, and it can hallucinate confident but wrong output&#8212;non-existent APIs, architectures that violate your own patterns, designs that look elegant but are not grounded in reality.</p><p>For leaders, ownership in the age of AI means accepting that models will write much of the code while you design and protect the mechanisms that keep it safe: clear specifications, honest reviews, explicit stop buttons when something feels wrong. Renaissance Developers don&#8217;t just own their code; they own the feedback loops and safeguards that keep quality high even as the tools get faster.</p><h2>Be a polymath</h2><p>Vogels is clear that this isn&#8217;t about mathematics; the word comes from the Greek for &#8220;to learn many things.&#8221; A polymath goes deep in at least one domain and yet refuses to stay narrow, cultivating meaningful knowledge across others. Leonardo da Vinci is the visible archetype&#8212;painter, engineer, anatomist, inventor&#8212;but Vogels&#8217; fundamental point is that modern developers cannot afford to be merely I&#8209;shaped specialists.</p><p>Instead, he urges us toward a T&#8209;shaped posture: depth like a spine, breadth like outstretched arms. His mentor Jim Gray embodied this. Gray could listen to a server room for half a minute and diagnose a broken database layout, not only because he understood databases intimately but also because he understood the data, the scientists, and how the organization worked. That combination of technical depth, domain insight, and human awareness is what makes expertise decisive rather than decorative.</p><p>For leaders, the lesson is demanding and straightforward: grow people who can dive deep and still speak the neighboring languages around them. When a backend engineer understands cost and product, their designs change. When a data engineer understands the story the numbers must tell, their questions sharpen. When a platform engineer understands developer experience, their tools become catalysts rather than constraints. Breadth is not a distraction from depth; it is the amplifier that lets depth shape the whole system.</p><h1>Werner&#8217;s legacy</h1><p>When looking at the empty stage, the previous keynote pictures rolling, I started reflecting on Werner&#8217;s legacy: it is a way of inhabiting the craft. Protect and fuel curiosity so that learning never stops. Think in systems so that every change you make respects the invisible structures it touches. Raise the bar on communication so intent, design, and implementation move in sync. Build real mechanisms, not just good intentions, especially when AI accelerates your work beyond your first understanding. And grow T&#8209;shaped people whose depth is amplified, not constrained, by their breadth.</p><p>The tools will keep changing. They always have. What endures is the standard he set: use the most powerful tools in history in service of the biggest problems of our time, quietly, rigorously, and with pride in the work that no one sees.</p><p></p>]]></content:encoded></item><item><title><![CDATA[I am not an AI expert, nor are you.]]></title><description><![CDATA[How self-proclaimed experts risk poisoning the AI ecosystem.]]></description><link>https://ai.techonthestack.com/p/i-am-not-an-ai-expert-nor-are-you</link><guid isPermaLink="false">https://ai.techonthestack.com/p/i-am-not-an-ai-expert-nor-are-you</guid><dc:creator><![CDATA[Luca Bianchi]]></dc:creator><pubDate>Thu, 04 Dec 2025 00:52:30 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!J_EP!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd0c9329b-806b-4fd5-aef6-7c2f936bb8a3_1920x1079.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!J_EP!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd0c9329b-806b-4fd5-aef6-7c2f936bb8a3_1920x1079.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!J_EP!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd0c9329b-806b-4fd5-aef6-7c2f936bb8a3_1920x1079.png 424w, https://substackcdn.com/image/fetch/$s_!J_EP!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd0c9329b-806b-4fd5-aef6-7c2f936bb8a3_1920x1079.png 848w, https://substackcdn.com/image/fetch/$s_!J_EP!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd0c9329b-806b-4fd5-aef6-7c2f936bb8a3_1920x1079.png 1272w, https://substackcdn.com/image/fetch/$s_!J_EP!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd0c9329b-806b-4fd5-aef6-7c2f936bb8a3_1920x1079.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!J_EP!,w_2400,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd0c9329b-806b-4fd5-aef6-7c2f936bb8a3_1920x1079.png" width="1200" height="674.1758241758242" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d0c9329b-806b-4fd5-aef6-7c2f936bb8a3_1920x1079.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:false,&quot;imageSize&quot;:&quot;large&quot;,&quot;height&quot;:818,&quot;width&quot;:1456,&quot;resizeWidth&quot;:1200,&quot;bytes&quot;:4948987,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://aletheia1982.substack.com/i/180630841?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd0c9329b-806b-4fd5-aef6-7c2f936bb8a3_1920x1079.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:&quot;center&quot;,&quot;offset&quot;:false}" class="sizing-large" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!J_EP!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd0c9329b-806b-4fd5-aef6-7c2f936bb8a3_1920x1079.png 424w, https://substackcdn.com/image/fetch/$s_!J_EP!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd0c9329b-806b-4fd5-aef6-7c2f936bb8a3_1920x1079.png 848w, https://substackcdn.com/image/fetch/$s_!J_EP!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd0c9329b-806b-4fd5-aef6-7c2f936bb8a3_1920x1079.png 1272w, https://substackcdn.com/image/fetch/$s_!J_EP!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd0c9329b-806b-4fd5-aef6-7c2f936bb8a3_1920x1079.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>It is pretty common to find AI experts almost everywhere. From podcasts to blog posts, LinkedIn news. People re-posting headlines and news. Often amplifying the noise of marketing announcements with no real attachment to practical use cases.</p><p>This behavior, quite common during every technological revolution, can be highly disruptive when dealing with AI.</p><p>The problem is that in the domain of artificial intelligence, there is an enormous bias between PoC or theory and the hard reality. This, matched to a fast-moving state-of-the-art and some challenges still unsolved, amplifies the perception that with AI, nothing works in production.</p><p>There are a number of factors that create and maintain this perception: first, the ease of building a working prototype, which leads people to believe production is just a few days away. This is fueled by people living under the Duning-Krueger paradox, which underestimates the complexity of reality. Second, there are benchmarks that can be really misleading. I have been told far too many things that we should do X because ChatGPT does X.</p><p>Ok, repeat after me: ChatGPT is not a benchmark because their investments in just building that simple application are far greater than what a normal company can put on the table. This translates into a heavily optimized solution that appears simple at first glance. In addition, tech teams within OpenAI can have access either to top-level AI professionals or to the knowledge about how their LLM works under the hood.</p><p>So, let&#8217;s go back to reality: I completed my PhD in 2010, working on neural networks and computer vision, but I do not consider myself an AI expert. I have trained neural networks on GPUs and in the Cloud, faced heating issues, and allocated processing slices. I have written PyTorch implementations. Yet I do not consider myself an expert. This is because I am not an expert.</p><p>Every day I learn something new. Every day, I find out there are things I do not know. Things that have a real impact on real issues. And I am not speaking only of manual model training or such things. I am pointing to just the managed services out there. Each one with parameters that distinguish feasibility from unaffordability.</p><p>In AI, even tasks such as &#8220;document OCRs&#8221; can be tough when tested on real-life documents, which are often handwritten, rotated, and sometimes blurred. In the last few months, just to read legal contracts, we had to match a whole set of requirements in terms of speed, cost, and accuracy. Everything seemed super easy on paper, but when faced with reality, we found we needed a dataset of &#8220;hard documents,&#8221; which are not hard at all for humans. We tried parameters, and every time we matched a requirement, another piece of the puzzle stopped working. We needed to define a metric to build KPIs and evaluate different methods across different dimensions. We needed to pre-process the document, and discovered that there is no one-size-fits-all solution.</p><p>We struggle with context size and meaningfulness, face size constraints, and seek techniques to improve the quality of data managed by our agents. When we fail, the price is poor accuracy, which in the legal domain translates into customers not being able to use our tools to accomplish their tasks. Then we reiterate, change the model, deal with Bedrock quotas and system prompt accuracy, only to finally find out we need to select our context chunks using a RAG, but we still need to rewrite queries, filter, and rerank chunks. All of this within an acceptable iteration time.</p><p>I work every day with people from Microsoft and AWS who collect feedback from our use case, provide suggestions, and improve their products in the meantime. Then our R&amp;D department has to continuously monitor new papers published at top-level conferences, prepare test cases to evaluate results, and identify what should be included in our product roadmap.</p><p>Then I stumble into so-called experts. People often deliver engaging talks at conferences about how magical yet powerful this AI world has become. I try to ground their very opinionated sentences to our use cases, but after having barely listened to two sentences about our problem, they have identified a 100% working solution. </p><p>I dive into their expertise only to find out they have &#8220;production expertise&#8221; because they can run a model on their PC with Ollama or LMStudio. Buying a $6K laptop and running in the terminal </p><p><code>uvx --from mlx-vlm mlx_vlm.generate \</code></p><p><code>    --model mlx-community/DeepSeek-OCR-8bit \</code></p><p><code>    --prompt &#8220;Convert this document to Markdown.&#8221; \</code></p><p><code>    --image &#8220;$IMAGE_PATH&#8221; \</code></p><p><code>    --max-tokens 8192 \</code></p><p><code>    --temperature 0.0</code></p><p>doesn&#8217;t make you an expert. I have a 2600-page document to be processed in less than 3 minutes. A use case requiring at least multiple GPUs, prefetching, and a vLLM deployment. And this is just one example of the many modules we have in our Adaptive Legal Assistant LIDIA: legal operations workflow management through agents is another optimization pain point: context is very limited when you have tens of documents with hundreds of pages each. And context rot is waiting at the end of the road to remind you that &#8220;a bigger context&#8221; doesn&#8217;t solve your issues. Then you have to add a RAG to the equation, but in order to be effective, you need metadata from your documents and a proper chunking technique. Just to begin.</p><p>But our &#8220;advanced experts&#8221; aim to write a sufficiently long prompt for ChatGPT and are ready to explain to you that they did that in production, as they show in their presentations: loading an Excel file into the chat and summing the values in a column.</p><p>Let&#8217;s set the expectations: running a model on your PC does not make you an expert. Bringing the model into production on multi-GPU hardware, with optimized response times, makes you an expert.</p><p>Writing a prompt, however complex, does not make you an AI expert. The state of the art is agent development with context engineering, tool calling, and agent swarm orchestration. I know how to do all these things; nonetheless, I am struggling every day with the constraints of the state of the art. </p><p>Yet I am not an AI expert. And I can tell that you are probably neither.</p>]]></content:encoded></item><item><title><![CDATA[The Silent Killer of Production LLMs]]></title><description><![CDATA[Mastering Context Rot and Attention Budget]]></description><link>https://ai.techonthestack.com/p/the-silent-killer-of-production-llms</link><guid isPermaLink="false">https://ai.techonthestack.com/p/the-silent-killer-of-production-llms</guid><dc:creator><![CDATA[Luca Bianchi]]></dc:creator><pubDate>Thu, 04 Dec 2025 00:47:48 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!xYk1!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb684d91-9bdd-42b8-93ea-e03de83052ad_1024x572.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!xYk1!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb684d91-9bdd-42b8-93ea-e03de83052ad_1024x572.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!xYk1!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb684d91-9bdd-42b8-93ea-e03de83052ad_1024x572.jpeg 424w, https://substackcdn.com/image/fetch/$s_!xYk1!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb684d91-9bdd-42b8-93ea-e03de83052ad_1024x572.jpeg 848w, https://substackcdn.com/image/fetch/$s_!xYk1!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb684d91-9bdd-42b8-93ea-e03de83052ad_1024x572.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!xYk1!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb684d91-9bdd-42b8-93ea-e03de83052ad_1024x572.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!xYk1!,w_2400,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb684d91-9bdd-42b8-93ea-e03de83052ad_1024x572.jpeg" width="1200" height="670.3125" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/db684d91-9bdd-42b8-93ea-e03de83052ad_1024x572.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:false,&quot;imageSize&quot;:&quot;large&quot;,&quot;height&quot;:572,&quot;width&quot;:1024,&quot;resizeWidth&quot;:1200,&quot;bytes&quot;:65891,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://aletheia1982.substack.com/i/179908212?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb684d91-9bdd-42b8-93ea-e03de83052ad_1024x572.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:&quot;center&quot;,&quot;offset&quot;:false}" class="sizing-large" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!xYk1!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb684d91-9bdd-42b8-93ea-e03de83052ad_1024x572.jpeg 424w, https://substackcdn.com/image/fetch/$s_!xYk1!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb684d91-9bdd-42b8-93ea-e03de83052ad_1024x572.jpeg 848w, https://substackcdn.com/image/fetch/$s_!xYk1!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb684d91-9bdd-42b8-93ea-e03de83052ad_1024x572.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!xYk1!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb684d91-9bdd-42b8-93ea-e03de83052ad_1024x572.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>While 2023 focused almost exclusively on <em>prompt engineering</em>, the real challenge for 2025 will be <strong>Context Engineering</strong>.</p><p>If we accept the metaphor that the Large Language Model is the new CPU, then its context window&#8212;even in million-token variants&#8212;is &#8220;the new RAM&#8221;: fast, powerful, but critically finite. Currently, many systems treat this high-speed memory as if it were an unlimited hard drive. The structural result of this approach is <strong>context rot</strong>.</p><p>The scenario is familiar: an agent that performs perfectly in a demo starts to degrade after a few weeks in production. It hallucinates obsolete instructions, confuses tools, or contradicts recent data with old session history. This is not a random bug, but a structural failure in managing the model&#8217;s &#8220;working memory.&#8221;</p><p>Let&#8217;s analyze the engineering principles necessary to transform context from an unmanaged liability into a curated, high-performance asset.</p><h2>The Structural Limit: The Attention Budget</h2><p>At the core of context rot is an architectural constraint of the Transformer: the <strong>attention budget</strong>. For a sequence of tokens, the model may need to compute relationships. As $N$ grows toward the context limit, the model must distribute its fixed attention capacity over an exponentially larger set of connections.</p><p>The practical impact is evident in &#8220;Needle-in-a-Haystack&#8221; benchmarks. As the context length (the &#8220;haystack&#8221;) increases, the model&#8217;s ability to reliably retrieve a single relevant piece of information decreases. Long context represents a capacity increase, not a guarantee of perfect recall.</p><p>In agent development, we tend to add information incrementally: system prompts, user history, RAG results, tool outputs, Chain-of-Thought scratchpads, and execution logs. Often, nothing is deleted. By the sixtieth interaction, the model is paying an &#8220;attention tax&#8221; on forty turns of irrelevant noise. Studies (such as those conducted by Microsoft/Salesforce) have shown performance degradation of up to 20% when partial or incorrect intermediate outputs are retained in the active context.</p><p>Context rot manifests primarily in four failure modes:</p><ol><li><p><strong>Context Poisoning:</strong> An error or hallucination (e.g., an obsolete API specification retrieved via RAG) enters the context and is treated as <em>ground truth</em> in subsequent steps, solidifying the error in the agent&#8217;s working memory.</p></li><li><p><strong>Context Distraction:</strong> The model focuses excessively on a verbose example at the beginning of the prompt or an obscure historical edge case, overweighting the prompt content at the expense of generalized knowledge derived from pre-training.</p></li><li><p><strong>Context Confusion:</strong> The problem of tool proliferation. Exposing a dozen tools, each with lengthy descriptions and ambiguous decision boundaries, creates an unmanageable decision surface. If a human engineer cannot quickly select the correct tool by reading the prompt, the model will face similar difficulties.</p></li><li><p><strong>Context Clash:</strong> Two high-signal pieces of information in the prompt contradict each other (e.g., historical records indicate one customer location, a recent tool call indicates another). The model must arbitrate this conflict, often leading to unpredictable behavior.</p></li></ol><h2>Context Observability: The X-Ray of Working Memory</h2><p>You cannot optimize what you do not measure. Traditional observability focuses on latency, costs, and token counts. Solving context rot requires <strong>Context Observability</strong>: a granular view of the composition of the model&#8217;s working memory over time.</p><p>It is necessary to implement tools that parse conversation logs, semantically segment long messages, and label each block by component type: System Instructions, User Query, Tool Output, RAG Knowledge, etc. Visualizing these components as stacked timelines reveals when low-signal components, such as raw, uncompressed tool outputs, dominate context. The next generation of tooling must make context composition a first-class engineering concern.</p><h2>The Context Engineering Playbook</h2><p>To regain control of the attention budget, we must shift from passive accumulation to active context architecture. This requires implementing a playbook built on four fundamental engineering pillars: <strong>Isolation, Selection, Compression, and External Writing.</strong></p><h3>1. Isolation: Decompose the Problem</h3><p>Isolation involves breaking a monolithic problem into specialized units, ensuring that no single context window has to hold the state of the entire workflow.</p><p>The most effective implementation is through <strong>multi-agent systems</strong>, where a lead agent orchestrates specialized sub-agents (e.g., Researcher, Planner, Executor). Each sub-agent maintains a context window strictly limited to its task, preventing execution logs and irrelevant tool descriptions from accumulating in a single global prompt. <em>Sandboxing</em> techniques are a form of isolation in which heavy objects (code logs, images, intermediate arrays) remain external, with only targeted, compressed results reinserted into the LLM context.</p><h3>2. Selection: Maximize Signal-to-Noise</h3><p>Selection is the principle of maximizing the signal-to-noise ratio of every token. RAG is the most common form of knowledge selection, but the technique must be generalized.</p><p>It is critical to implement <strong>tool loadouts</strong>: dynamically retrieving and exposing only the tools relevant to the immediate request, rather than presenting the entire universe of tools at every call. This drastically reduces <em>Context Confusion</em>. Similarly, instead of blindly retrieving all long-term memory, semantic selection must be used to extract only memory fragments relevant to the current task. <strong>Progressive Disclosure</strong> is the dynamic ideal: giving the agent the ability to query external data stores <em>on demand</em>, keeping the active prompt lean while the accessible knowledge base remains vast.</p><h3>3. Compression: Distill the Essence</h3><p>When preserving the essence of a long interaction or voluminous output is necessary, but raw text is too costly, compression becomes mandatory.</p><p>This involves <strong>summarization and compaction</strong>. Agents can be configured to automatically compact conversation history once a threshold is exceeded, using the LLM itself to summarize the state into a succinct representation. For greater rigor, specialized fine-tuned summarization models can be employed at agent boundaries to ensure distilled knowledge transfer. The most aggressive form is <strong>pruning</strong>, where systems like <em>Provence</em> identify and remove entire sections of a document that are statistically irrelevant to the current query, avoiding the computational cost of generative summarization.</p><h3>4. External Writing: State Offloading</h3><p>The fourth lever is accepting that not all state belongs in the prompt. We must offload persistent or temporary state into external data stores, making it accessible only via explicit tool calls.</p><p>The simplest pattern is the <strong>scratchpad</strong> or <strong>notepad tool</strong>, where the agent can write intermediate plans, notes, or reflections into an external JSON or database. This allows the agent to maintain high-level reasoning without clogging the current attention window with the detailed trace of the thought process. Long-term memories (as in <em>Reflexion</em>) are the persistent version of this approach, accessible only through controlled retrieval.</p><h2>The Foundation: High-Signal Chunking</h2><p>All these strategies fail if the underlying unit of information, the <strong>chunk</strong>, is poorly defined. The art of chunking consists of finding units that are semantically coherent but small enough for flexible retrieval and recombination.</p><p>Chunking based on fixed token counts is insufficient. It is necessary to move to <strong>semantic chunking</strong>, which uses embedding similarity to detect topic boundaries. Additionally, <strong>contextual chunking</strong> involves using an LLM to generate a brief summary for each chunk describing its meaning in the context of the entire document. This enriched representation is then embedded, thereby improving retrieval accuracy and robustness.</p><h2>Architectural Mandate: Standardization with MCP</h2><p>To move beyond bespoke architectures, standardization is essential. The <strong>Model Context Protocol (MCP)</strong>, proposed by Anthropic, offers a path forward. It defines a standardized interface (similar to JSON-RPC) for interaction between models, tools, and data.</p><p>MCP acts as a &#8220;USB-C for AI,&#8221; providing a consistent schema for tool discovery and operation invocation. This consistency radically simplifies the implementation of complex context management strategies&#8212;such as dynamic tool loadouts and controlled external writing&#8212;reducing <em>Context Clash</em> and allowing components to be reused across different frameworks without excessive &#8220;glue code.&#8221;</p><h2>Conclusion</h2><p>Context Engineering is about finding the <strong>smallest possible set of high-signal tokens</strong> that maximizes the probability of the desired behavior. Every vague instruction, redundant field, or verbose tool description is stealing attention from a more critical token.</p><p>For system prompts, this means finding the right &#8220;altitude&#8221;: avoiding fragile pseudo-code while maintaining structured instructions (via XML or Markdown). For tools, output must be token-efficient. Finally, few-shot examples must be canonical demonstrations of the desired pattern, not exhaustive documentation.</p><p>If you are building an agent today, you are designing and operating its dynamic working memory. Instead of asking &#8220;Why did the model hallucinate?&#8221;, the correct question is: &#8220;What was in its &#8216;head&#8217; at that moment, what was it paying the attention budget for, and which of those tokens truly deserved to be there?&#8221;</p>]]></content:encoded></item><item><title><![CDATA[Hello Builders, issue #2]]></title><description><![CDATA[Agentic devices, specialized reasoning models, and shifting AI industrial policy, and your organization&#8217;s people and processes readiness.]]></description><link>https://ai.techonthestack.com/p/hello-builders-issue-2</link><guid isPermaLink="false">https://ai.techonthestack.com/p/hello-builders-issue-2</guid><dc:creator><![CDATA[Luca Bianchi]]></dc:creator><pubDate>Sat, 29 Nov 2025 06:30:08 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/83679548-5293-4ffa-9bda-bd49c88840f9_840x600.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>This week&#8217;s AI focus is on  three converging forces: &#8220;agentic&#8221; models that can drive your UI like an operator, open and highly specialized models that challenge frontier systems on niche tasks, and an apparent acceleration in AI industrial policy from both the US and Europe. As a technology leader, this is the week where AI stops being just an API call and increasingly becomes an operating layer across devices, infrastructure, and regulation.</p><h2>This week&#8217;s signal in the noise</h2><ul><li><p><strong>Agentic, on-device models</strong> move from research demos to deployable products (Microsoft&#8217;s Fara-7B).</p></li><li><p><strong>Open, specialized models</strong> push frontier-level performance in narrow domains like math (DeepSeek Math V2 and the wider DeepSeek stack) while frontier model vendors sharpen price/performance and context length for long-running enterprise workloads (Anthropic Claude Opus 4.5).</p></li><li><p><strong>Policy and regulation</strong> are shifting rapidly, with the EU warning that it is &#8220;missing the boat&#8221; and the US simultaneously ramping up national AI programs while weakening state-level oversight.</p></li><li><p><strong>The human factor</strong> in AI adoption is critical; without deliberate communication, upskilling, and change management, it can easily become the main brake on a company&#8217;s AI journey.</p></li></ul><div><hr></div><h3>On-device agentic models for computer use</h3><p>Microsoft quietly released <strong>Fara-7B</strong>, a 7B-parameter &#8220;agentic&#8221; small language model designed to act as a <strong>computer-use agent</strong>: it can see your screen, click, scroll, type, and navigate the web much like a human operator. The model runs locally on Copilot+ PCs and is also available via Microsoft&#8217;s Foundry and open platforms like Hugging Face, having been trained on a large synthetic dataset of multi-step web interactions. This is a concrete signal that <strong>&#8220;UI-as-API&#8221; agents</strong> are becoming a product category rather than an experiment. You should start classifying internal applications by how &#8220;agent-compatible&#8221; their UI is (deterministic layouts, clear affordances, instrumented telemetry). Local execution also strengthens the case for <strong>on-device agents for sensitive workflows</strong>, but it raises new surface areas for monitoring, auditability, and access control&#8212;your endpoint security and DLP strategies will need to adapt.</p><p><a href="https://www.microsoft.com/en-us/research/blog/fara-7b-an-efficient-agentic-model-for-computer-use/">https://www.microsoft.com/en-us/research/blog/fara-7b-an-efficient-agentic-model-for-computer-use/</a></p><div><hr></div><h2>Specialized reasoning over monolithic LLMs</h2><p>China&#8217;s DeepSeek released <strong>DeepSeek Math V2</strong> (https://www.scmp.com/tech/tech-trends/article/3334553/deepseek-releases-first-open-ai-model-gold-level-scores-maths-olympiad), an open model specialized for mathematical problem-solving, reportedly delivering performance comparable to that of International Math Olympiad (IMO) gold medalists on benchmark suites. This follows the wider DeepSeek family of open models, which have already drawn attention for their strong reasoning capabilities at significantly lower compute and cost profiles than many Western frontier models.</p><p>This continues a strategic trend: <strong>highly specialized open models</strong> that outperform giant generalist models on narrow but economically important domains (quant, optimization, verification, scientific computing). It&#8217;s a strong signal to shift from a single &#8220;one-model-fits-all&#8221; architecture to a <strong>portfolio of domain models</strong>. DeepSeek&#8217;s cost and hardware efficiency also underline that <strong>GPU scarcity is no longer an excuse</strong> for not experimenting with strong reasoning systems. This is even more true if we consider a recently published paper (https://x.com/omarsar0/status/1993695515595444366?t=lpD7UiqjTbYd453g-DPvpw&amp;s=33) that shows how trained 12B models can exhibit reasoning capabilities comparable to those of frontier models.</p><p>On the other side of the river, Anthropic introduced <strong>Claude Opus 4.5</strong> that combines <strong>lower API pricing</strong> with <strong>longer context windows</strong> and improved reasoning, explicitly targeting long-running chats and multi-step workflows that previously hit context or budget ceilings. This is yet another hint to avoid using a single LLM architecture for your applications. You should plan a <strong>Q4&#8211;Q1 refresh of your model cost models and routing strategies</strong>.</p><div><hr></div><h2>&#8220;Missing the boat&#8221; on competitiveness...</h2><p>European Central Bank President Christine Lagarde warned that the EU is <strong>risking its economic future</strong> by falling behind the US and China in AI adoption and development (<a href="https://www.reuters.com/business/eu-missing-boat-ai-jeopardising-its-future-lagarde-warns-2025-11-24/">https://www.reuters.com/business/eu-missing-boat-ai-jeopardising-its-future-lagarde-warns-2025-11-24/</a>). She highlighted the risk of Europe becoming structurally dependent on foreign AI infrastructure, compute, chips, and hyperscale data centers, if it remains primarily a buyer rather than a builder, and called for open standards, cheaper energy, and integrated capital markets. This could probably be an explicit signal that <strong>AI capacity is now viewed as strategic infrastructure</strong>, on par with energy or telecoms. Expect a new wave of <strong>national and EU-level incentives, consortia, and regulatory adjustments</strong> to accelerate domestic AI capability. Companies operating in Europe should actively track these programs (funding, tax treatment, regulatory sandboxes) and, where possible, position flagship AI initiatives as <strong>part of the region&#8217;s industrial policy story</strong>.</p><p>This is still far from the US AI-aggressive &#8220;Genesis Mission&#8221; (<a href="https://www.federalregister.gov/documents/2025/11/28/2025-21665/launching-the-genesis-mission">https://www.federalregister.gov/documents/2025/11/28/2025-21665/launching-the-genesis-mission</a>), a federal effort to turn national scientific and engineering data into discoveries using AI, led by the Department of Energy and powered by public and private supercomputers. US AI policy is simultaneously <strong>turbocharging federal AI projects</strong> and <strong>weakening decentralized regulation</strong>.</p><div><hr></div><h2>The slow AI adoption</h2><p>A Harvard study (https://hbr.org/2025/11/leaders-assume-employees-are-excited-about-ai-theyre-wrong) found that 76% of executives believe employees are excited about AI, yet only 31% of individual contributors agree. The same survey reveals a deep organizational blind spot: 75% of leaders describe their company as &#8220;employee-centric&#8221;, but only 23% of employees share that view. Only 30% of employees feel informed about their organization&#8217;s AI strategy, versus 80% of leaders who believe they are. The study concludes that truly employee-centric organizations are up to seven times more likely to succeed with AI. Daniel Goleman attributes the origin of this friction to the psychological impact of change and to the role Emotional Intelligence could play in a successful strategy (<a href="https://www.kornferry.com/insights/this-week-in-leadership/the-big-ai-roadblock-in-our-heads">https://www.kornferry.com/insights/this-week-in-leadership/the-big-ai-roadblock-in-our-heads</a>).</p>]]></content:encoded></item><item><title><![CDATA[Hello Builders, issue #1]]></title><description><![CDATA[Your weekly briefing on the signals that actually matter in AI distilled, contextualized, and built for people who are shipping.]]></description><link>https://ai.techonthestack.com/p/hello-builders-issue-1</link><guid isPermaLink="false">https://ai.techonthestack.com/p/hello-builders-issue-1</guid><dc:creator><![CDATA[Luca Bianchi]]></dc:creator><pubDate>Fri, 21 Nov 2025 09:13:32 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/ef184f42-0a85-4068-aed5-f337c049338d_630x450.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>On November 13, <a href="https://www.cnbc.com/2025/11/13/cursor-ai-startup-funding-round-valuation.html">Cursor&#8212;a code editor&#8212;raised $2.3 billion at a $29.3 billion valuation</a>. That&#8217;s not a typo. A tool that helps developers write code is now valued higher than most cloud infrastructure companies. <a href="https://www.businesswire.com/news/home/20251113939996/en/Cursor-Secures-$2.3-Billion-Series-D-Financing-at-$29.3-Billion-Valuation-to-Redefine-How-Software-is-Written">The company hit $1 billion in annualized revenue with enterprise revenue growing 100x year-to-date</a>. <a href="https://www.cnbc.com/2025/11/13/cursor-ai-startup-funding-round-valuation.html">Nvidia&#8217;s CEO called it his &#8220;favorite enterprise AI service.&#8221;</a></p><p>Here&#8217;s why that matters: Cursor doesn&#8217;t own any frontier AI models. It uses OpenAI, Anthropic, and Google APIs. The valuation isn&#8217;t for the technology&#8212;it&#8217;s for workflow capture. When developers open their IDE, Cursor controls the moment of creation. And apparently, that&#8217;s worth $30 billion.</p><p>The same day, <a href="https://openai.com/index/gpt-5-1-for-developers/">OpenAI shipped GPT-5.1 across their entire model family</a>&#8212;not as a research preview, but as production infrastructure. <a href="https://openai.com/index/gpt-5-1-for-developers/">The killer feature isn&#8217;t smarter responses; it&#8217;s 24-hour prompt caching that reduces costs by 90% for cached tokens</a>. <a href="https://help.openai.com/en/articles/9624314-model-release-notes">When inference costs have dropped 280x since GPT-3.5</a>, you&#8217;re not selling API calls anymore&#8212;you&#8217;re selling commodity compute.</p><p>Meanwhile, <a href="https://www.anthropic.com/news/anthropic-invests-50-billion-in-american-ai-infrastructure">Anthropic announced a $50 billion partnership with Fluidstack to build custom data centers in Texas and New York, online throughout 2026</a>. Not &#8220;exploring opportunities.&#8221; Actual construction creating <a href="https://www.anthropic.com/news/anthropic-invests-50-billion-in-american-ai-infrastructure">800 permanent jobs and 2,400 construction positions</a>. And <a href="https://www.fool.com/investing/2025/11/03/open-ai-move-made-microsoft-no-brainer-buy/">Microsoft formalized its relationship with OpenAI: a $135 billion equity stake (27% ownership) tied to $250 billion in Azure commitments through 2032</a>.</p><p>But here&#8217;s the development that should make every CTO reconsider their model procurement strategy: <a href="https://siliconangle.com/2025/11/07/moonshot-launches-open-source-kimi-k2-thinking-ai-trillion-parameters-reasoning-capabilities/">Chinese startup Moonshot AI released Kimi K2 Thinking, a trillion-parameter reasoning model, trained for $4.6 million</a>. Western competitors spend $100 million or more for similar capabilities. That&#8217;s a 20x cost advantage. <a href="https://www.deeplearning.ai/the-batch/moonshot-releases-kimi-k2-a-trillion-parameter-model-fine-tuned-for-agentic-tool-use/">And on SWE-Bench Verified&#8212;the benchmark that actually matters for code generation&#8212;it scores 65.8% versus GPT-4.1&#8217;s 54.6%</a>.</p><p>The model is open-sourced under a modified MIT license.</p><p>When you can train frontier models for $5 million instead of $100 million, when code editors command infrastructure-scale valuations, when inference costs drop 280x in two years&#8212;the entire stack is repricing. The companies that treated AI as an R&amp;D expense are now competing against companies that treated it as infrastructure investment. And the gap is measured in billions.</p><p>There&#8217;s one more data point that ties this together. <a href="https://www.marketingprofs.com/opinions/2025/54004/ai-update-november-14-2025-ai-news-and-views-from-the-past-week">Microsoft Clarity analyzed 1,200+ publisher sites and found that AI referral traffic&#8212;from ChatGPT, Copilot, Perplexity, Gemini&#8212;converts at 1.66% versus 0.15% for search</a>. That&#8217;s an 11x difference in customer acquisition efficiency. <a href="https://www.marketingprofs.com/opinions/2025/54004/ai-update-november-14-2025-ai-news-and-views-from-the-past-week">AI traffic grew 155.6% in eight months</a>. The distribution layer isn&#8217;t shifting&#8212;it&#8217;s already shifted.</p><p>The question isn&#8217;t whether AI works at scale anymore. The question is whether your architecture can scale with it.</p><p>Keep building,</p><p><em><strong>Luca</strong></em></p><div><hr></div><h3>Key Takeaways</h3><ol><li><p><strong>Application Layer = Infrastructure Economics: </strong>Cursor&#8217;s valuation proves workflow control beats model ownership. Your internal tools now compete against apps with billion-dollar budgets optimizing for developer capture.</p></li><li><p><strong>Multi-Cloud = Operational Necessity: </strong>Microsoft/OpenAI decoupled after 5 years. OpenAI committed $1.15T across 7 vendors (Broadcom $350B, Oracle $300B, Microsoft $250B, Nvidia $100B, AMD $90B, AWS $38B, CoreWeave $22B). Single-vendor is risk concentration, not optimization.</p></li><li><p><strong>Training Cost Arbitrage Accelerates: </strong>Moonshot trained a 1T-parameter model for $4.6M that beats Western benchmarks. MuonClip optimizer enabled zero-instability training at scale. The technical gap closes in months, not years.</p></li><li><p><strong>AI Referral Conversion: 11x Search, &lt;1% Traffic: </strong>1.66% vs 0.15% conversion, but negligible volume. Companies optimizing for AI discovery now gain 18-24 months advantage before it becomes standard.</p></li></ol>]]></content:encoded></item></channel></rss>