<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Engineering on Matt Suiche</title><link>https://www.msuiche.com/categories/engineering/</link><description>Recent content in Engineering on Matt Suiche</description><generator>Hugo</generator><language>en-us</language><lastBuildDate>Mon, 13 Apr 2026 00:00:00 +0000</lastBuildDate><atom:link href="https://www.msuiche.com/categories/engineering/index.xml" rel="self" type="application/rss+xml"/><item><title>Seeing Sound: Generative Techno and DSP in Pure NumPy</title><link>https://www.msuiche.com/posts/seeing-sound-generative-techno-and-dsp-in-pure-numpy/</link><pubDate>Mon, 13 Apr 2026 00:00:00 +0000</pubDate><guid>https://www.msuiche.com/posts/seeing-sound-generative-techno-and-dsp-in-pure-numpy/</guid><description>&lt;p&gt;This post is a bit of a grab bag — personal notes dumped here so I can pick up the thread later. The main goal: achieving generative EDM/techno music. Everything else — DSP, frequency bands, oscillators, filters — is machinery toward that end.&lt;/p&gt;
&lt;p&gt;Especially now with AI/GenAI, this feels achievable: create bangers with a few Python scripts, provide generative sound experiences that are unique each time. Not generating samples from prompts — actually synthesizing sound from first principles.&lt;/p&gt;</description></item><item><title>Local Models Within Reach: Everything That Changed in Eight Months</title><link>https://www.msuiche.com/posts/local-models-within-reach-everything-that-changed-in-eight-months/</link><pubDate>Sun, 05 Apr 2026 00:00:00 +0000</pubDate><guid>https://www.msuiche.com/posts/local-models-within-reach-everything-that-changed-in-eight-months/</guid><description>&lt;p&gt;Eight months ago I published &lt;a href="https://www.msuiche.com/posts/building-agents-for-small-language-models-a-deep-dive-into-lightweight-ai/"&gt;Building Agents for Small Language Models&lt;/a&gt;, a set of hard-won notes from shipping agents on 270M–32B parameter models. At the time, running useful local models meant embracing constraints: small context windows, CPU-only fallbacks, broken UTF-8 streams, and reasoning that fell apart past two steps.&lt;/p&gt;
&lt;p&gt;I stand by that post. But the ground has shifted fast. What was a set of careful workarounds in August 2025 is starting to look like the default architecture for a large class of workloads. Local models are no longer the constrained sibling of cloud APIs — for many agent use cases, they are the better answer. Here is what has changed.&lt;/p&gt;</description></item><item><title>Building Agents for Small Language Models: A Deep Dive into Lightweight AI</title><link>https://www.msuiche.com/posts/building-agents-for-small-language-models-a-deep-dive-into-lightweight-ai/</link><pubDate>Wed, 27 Aug 2025 00:00:00 +0000</pubDate><guid>https://www.msuiche.com/posts/building-agents-for-small-language-models-a-deep-dive-into-lightweight-ai/</guid><description>&lt;p&gt;The landscape of AI agents has been dominated by large language models (LLMs) like GPT-4 and Claude, but a new frontier is opening up: lightweight, open-source, locally-deployable agents that can run on consumer hardware. This post shares internal notes and discoveries from my journey building agents for small language models (SLMs) – models ranging from 270M to 32B parameters that run efficiently on CPUs or modest GPUs. These are lessons learned from hands-on experimentation, debugging, and optimizing inference pipelines.&lt;/p&gt;</description></item></channel></rss>