<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>AI on Matt Suiche</title><link>https://www.msuiche.com/categories/ai/</link><description>Recent content in AI on Matt Suiche</description><generator>Hugo</generator><language>en-us</language><lastBuildDate>Tue, 21 Apr 2026 00:00:00 +0000</lastBuildDate><atom:link href="https://www.msuiche.com/categories/ai/index.xml" rel="self" type="application/rss+xml"/><item><title>Legacy Security Is the Real Enterprise AI Bottleneck</title><link>https://www.msuiche.com/posts/legacy-security-is-the-real-enterprise-ai-bottleneck/</link><pubDate>Tue, 21 Apr 2026 00:00:00 +0000</pubDate><guid>https://www.msuiche.com/posts/legacy-security-is-the-real-enterprise-ai-bottleneck/</guid><description>&lt;p&gt;High quality data is expensive to collect, clean, and maintain. Poor security makes all of it free. To someone else.&lt;/p&gt;
&lt;p&gt;As software collapses toward zero marginal cost, that sentence stops being a cybersecurity truism and starts being a business model observation. Data is the last asset with durable value in an AI-native stack. The only thing that keeps that value is the discipline most AI-native companies are treating as optional.&lt;/p&gt;</description></item><item><title>From La Fontaine to Lego: Characters as Ideological Delivery Systems</title><link>https://www.msuiche.com/posts/from-la-fontaine-to-lego-characters-as-ideological-delivery-systems/</link><pubDate>Thu, 09 Apr 2026 00:00:00 +0000</pubDate><guid>https://www.msuiche.com/posts/from-la-fontaine-to-lego-characters-as-ideological-delivery-systems/</guid><description>&lt;p&gt;&lt;em&gt;Cute characters as ideological delivery systems, and how AI accelerated the propaganda playbook.&lt;/em&gt;&lt;/p&gt;
&lt;blockquote class="twitter-tweet"&gt;&lt;p lang="en" dir="ltr"&gt;Kind of crazy that the big propaganda medium to come out of AI wasn&amp;#39;t deepfakes but LEGO men and Persian cats &lt;a href="https://t.co/PEzEiamwJg"&gt;https://t.co/PEzEiamwJg&lt;/a&gt;&lt;/p&gt;&amp;mdash; Tracy Alloway (@tracyalloway) &lt;a href="https://twitter.com/tracyalloway/status/2042206704478122448?ref_src=twsrc%5Etfw"&gt;April 9, 2026&lt;/a&gt;&lt;/blockquote&gt;
&lt;script async src="https://platform.twitter.com/widgets.js" charset="utf-8"&gt;&lt;/script&gt;


&lt;p&gt;Tracy Alloway nailed it: &amp;ldquo;Kind of crazy that the big propaganda medium to come out of AI wasn&amp;rsquo;t deepfakes but LEGO men and Persian cats.&amp;rdquo;&lt;/p&gt;
&lt;p&gt;Everyone was bracing for deepfakes. The national security community spent years warning about synthetic video of world leaders saying things they never said, doctored footage designed to deceive at the pixel level. Instead, what showed up was Lego minifigures of Trump and Netanyahu set to AI-generated rap tracks, produced by an Iran-based group calling themselves the &lt;a href="https://www.rawstory.com/slopaganda-wars-the-us-and-iran-are-flooding-the-zone-with-viral-ai-generated-noise/" target="_blank" rel="noopener"&gt;&amp;ldquo;Explosive News Team&amp;rdquo;&lt;/a&gt;. And it wasn&amp;rsquo;t just Iran. Chinese state media CCTV joined in with its own GenAI animal fable: &lt;a href="https://www.realclearpolitics.com/video/2026/03/27/chinese_state_tv_shares_viral_ai_cartoons_on_iran_war_white_eagle_vs_persian_cat.html" target="_blank" rel="noopener"&gt;&amp;ldquo;The White Eagle and Persian Cat&amp;rdquo;&lt;/a&gt;, a stop-motion style animation where a White Eagle Alliance dominates trade by forcing other animals to use its currency. Not trying to fool anyone into thinking the footage was real. Just trying to be catchy, shareable, and memetically sticky.&lt;/p&gt;</description></item><item><title>Local Models Within Reach: Everything That Changed in Eight Months</title><link>https://www.msuiche.com/posts/local-models-within-reach-everything-that-changed-in-eight-months/</link><pubDate>Sun, 05 Apr 2026 00:00:00 +0000</pubDate><guid>https://www.msuiche.com/posts/local-models-within-reach-everything-that-changed-in-eight-months/</guid><description>&lt;p&gt;Eight months ago I published &lt;a href="https://www.msuiche.com/posts/building-agents-for-small-language-models-a-deep-dive-into-lightweight-ai/"&gt;Building Agents for Small Language Models&lt;/a&gt;, a set of hard-won notes from shipping agents on 270M–32B parameter models. At the time, running useful local models meant embracing constraints: small context windows, CPU-only fallbacks, broken UTF-8 streams, and reasoning that fell apart past two steps.&lt;/p&gt;
&lt;p&gt;I stand by that post. But the ground has shifted fast. What was a set of careful workarounds in August 2025 is starting to look like the default architecture for a large class of workloads. Local models are no longer the constrained sibling of cloud APIs — for many agent use cases, they are the better answer. Here is what has changed.&lt;/p&gt;</description></item><item><title>When Machines Pay Machines: The Economics of Agentic AI</title><link>https://www.msuiche.com/posts/when-machines-pay-machines-the-economics-of-agentic-ai/</link><pubDate>Mon, 15 Dec 2025 00:00:00 +0200</pubDate><guid>https://www.msuiche.com/posts/when-machines-pay-machines-the-economics-of-agentic-ai/</guid><description>&lt;p&gt;The internet was built with a missing piece. In 1994, when the HTTP specification reserved status code 402 for &amp;ldquo;Payment Required,&amp;rdquo; the architects knew money would eventually flow as freely as data. Three decades later, that vision is finally materializing—not because humans demanded it, but because AI agents need it.&lt;/p&gt;
&lt;h2 id="the-402-awakening"&gt;The 402 Awakening &lt;a href="#the-402-awakening" class="anchor"&gt;🔗&lt;/a&gt;&lt;/h2&gt;&lt;p&gt;HTTP 402 sat dormant for years, a placeholder for a future nobody could quite figure out. Credit cards required human intervention. PayPal needed accounts. Stripe demanded integration. None of these worked for a world where software talks to software at millisecond intervals.&lt;/p&gt;</description></item><item><title>The Hidden Math Bug That Makes AI Unpredictable</title><link>https://www.msuiche.com/posts/the-hidden-math-bug-that-makes-ai-unpredictable/</link><pubDate>Sun, 14 Sep 2025 00:00:00 +0200</pubDate><guid>https://www.msuiche.com/posts/the-hidden-math-bug-that-makes-ai-unpredictable/</guid><description>&lt;p&gt;This &lt;a href="https://x.com/awnihannun/status/1966953027451118012" target="_blank" rel="noopener"&gt;tweet from Awni Hannun&lt;/a&gt; demonstrates in one line of MLX code the nondeterminism phenomenon detailed in &lt;a href="https://thinkingmachines.ai/blog/defeating-nondeterminism-in-llm-inference/" target="_blank" rel="noopener"&gt;Thinking Machines&amp;rsquo; research&lt;/a&gt;. We will explore the PyTorch equivalent that reveals a fundamental issue in AI systems, because I&amp;rsquo;ve found that tweet extremely helpful to understand what the original blogpost was about.&lt;/p&gt;
&lt;blockquote class="twitter-tweet"&gt;&lt;p lang="en" dir="ltr"&gt;Here&amp;#39;s a one-line code summary in MLX of the &lt;a href="https://twitter.com/thinkymachines?ref_src=twsrc%5Etfw"&gt;@thinkymachines&lt;/a&gt; blog post on non-determinism in LLM inference.&lt;br&gt;&lt;br&gt;I&amp;#39;d guess the difference is larger the lower the precision, as you get larger affects from non-associativity of FP math.&lt;br&gt;&lt;br&gt;Interestingly, that implies that training at low… &lt;a href="https://t.co/jYcDK9GiLn"&gt;pic.twitter.com/jYcDK9GiLn&lt;/a&gt;&lt;/p&gt;</description></item><item><title>Building Agents for Small Language Models: A Deep Dive into Lightweight AI</title><link>https://www.msuiche.com/posts/building-agents-for-small-language-models-a-deep-dive-into-lightweight-ai/</link><pubDate>Wed, 27 Aug 2025 00:00:00 +0000</pubDate><guid>https://www.msuiche.com/posts/building-agents-for-small-language-models-a-deep-dive-into-lightweight-ai/</guid><description>&lt;p&gt;The landscape of AI agents has been dominated by large language models (LLMs) like GPT-4 and Claude, but a new frontier is opening up: lightweight, open-source, locally-deployable agents that can run on consumer hardware. This post shares internal notes and discoveries from my journey building agents for small language models (SLMs) – models ranging from 270M to 32B parameters that run efficiently on CPUs or modest GPUs. These are lessons learned from hands-on experimentation, debugging, and optimizing inference pipelines.&lt;/p&gt;</description></item></channel></rss>