<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>MoE on Matt Suiche</title><link>https://www.msuiche.com/tags/moe/</link><description>Recent content in MoE on Matt Suiche</description><generator>Hugo</generator><language>en-us</language><lastBuildDate>Sun, 05 Apr 2026 00:00:00 +0000</lastBuildDate><atom:link href="https://www.msuiche.com/tags/moe/index.xml" rel="self" type="application/rss+xml"/><item><title>Local Models Within Reach: Everything That Changed in Eight Months</title><link>https://www.msuiche.com/posts/local-models-within-reach-everything-that-changed-in-eight-months/</link><pubDate>Sun, 05 Apr 2026 00:00:00 +0000</pubDate><guid>https://www.msuiche.com/posts/local-models-within-reach-everything-that-changed-in-eight-months/</guid><description>&lt;p&gt;Eight months ago I published &lt;a href="https://www.msuiche.com/posts/building-agents-for-small-language-models-a-deep-dive-into-lightweight-ai/"&gt;Building Agents for Small Language Models&lt;/a&gt;, a set of hard-won notes from shipping agents on 270M–32B parameter models. At the time, running useful local models meant embracing constraints: small context windows, CPU-only fallbacks, broken UTF-8 streams, and reasoning that fell apart past two steps.&lt;/p&gt;
&lt;p&gt;I stand by that post. But the ground has shifted fast. What was a set of careful workarounds in August 2025 is starting to look like the default architecture for a large class of workloads. Local models are no longer the constrained sibling of cloud APIs — for many agent use cases, they are the better answer. Here is what has changed.&lt;/p&gt;</description></item></channel></rss>