<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>GPU Programming on Matt Suiche</title><link>https://www.msuiche.com/categories/gpu-programming/</link><description>Recent content in GPU Programming on Matt Suiche</description><generator>Hugo</generator><language>en-us</language><lastBuildDate>Fri, 17 Oct 2025 00:00:00 +0000</lastBuildDate><atom:link href="https://www.msuiche.com/categories/gpu-programming/index.xml" rel="self" type="application/rss+xml"/><item><title>Porting CUDA FFT to Mojo: Achieving Bit-Exact Precision</title><link>https://www.msuiche.com/posts/porting-cuda-fft-to-mojo-achieving-bit-exact-precision/</link><pubDate>Fri, 17 Oct 2025 00:00:00 +0000</pubDate><guid>https://www.msuiche.com/posts/porting-cuda-fft-to-mojo-achieving-bit-exact-precision/</guid><description>&lt;p&gt;Porting a CUDA Fast Fourier Transform (FFT) implementation to Mojo for the &lt;a href="https://leetgpu.com/challenges/fast-fourier-transform" target="_blank" rel="noopener"&gt;LeetGPU Fast Fourier Transform challenge&lt;/a&gt; presented an unexpected challenge: achieving bit-exact precision matching between CUDA&amp;rsquo;s &lt;code&gt;sinf()&lt;/code&gt;/&lt;code&gt;cosf()&lt;/code&gt; functions and their Mojo equivalents. This required PTX assembly analysis, cross-platform testing, and ultimately upgrading to Float64 precision for deterministic results.&lt;/p&gt;
&lt;h2 id="challenge-constraints"&gt;Challenge Constraints &lt;a href="#challenge-constraints" class="anchor"&gt;🔗&lt;/a&gt;&lt;/h2&gt;&lt;ul&gt;
&lt;li&gt;N range: $1 \leq N \leq 262,144$ (power-of-2 FFT sizes)&lt;/li&gt;
&lt;li&gt;Data type: All values are 32-bit floating point numbers&lt;/li&gt;
&lt;li&gt;Accuracy requirements: Absolute error $\leq 10^{-3}$, Relative error $\leq 10^{-3}$&lt;/li&gt;
&lt;li&gt;Array format: Input and output arrays have length $2N$ (interleaved real/imaginary)&lt;/li&gt;
&lt;/ul&gt;
&lt;h2 id="initial-problem-accuracy-mismatch"&gt;Initial Problem: Accuracy Mismatch &lt;a href="#initial-problem-accuracy-mismatch" class="anchor"&gt;🔗&lt;/a&gt;&lt;/h2&gt;&lt;p&gt;The initial Mojo FFT implementation failed correctness tests with a maximum absolute difference of 0.023 compared to the reference CUDA implementation. For a coding challenge requiring exact equality, this was unacceptable.&lt;/p&gt;</description></item><item><title>AMD GPU Support in Triton Gluon Framework</title><link>https://www.msuiche.com/posts/amd-gpu-support-in-triton-gluon-framework/</link><pubDate>Wed, 15 Oct 2025 02:00:00 +0000</pubDate><guid>https://www.msuiche.com/posts/amd-gpu-support-in-triton-gluon-framework/</guid><description>&lt;h2 id="introduction"&gt;Introduction &lt;a href="#introduction" class="anchor"&gt;🔗&lt;/a&gt;&lt;/h2&gt;&lt;p&gt;This document analyzes AMD GPU support implementation in Triton&amp;rsquo;s Gluon framework, examining architecture-specific optimizations, performance characteristics, and implementation details relative to NVIDIA GPU support.&lt;/p&gt;
&lt;p&gt;For background on Gluon and its motivation as a lower-level alternative to Triton, see my previous post: &lt;a href="https://www.msuiche.com/posts/gluon-when-triton-isnt-low-level-enough/" target="_blank" rel="noopener"&gt;&amp;ldquo;Gluon: When Triton Isn&amp;rsquo;t Low-Level Enough&amp;rdquo;&lt;/a&gt;.&lt;/p&gt;
&lt;h2 id="background-gpu-programming-architecture-landscape"&gt;Background: GPU Programming Architecture Landscape &lt;a href="#background-gpu-programming-architecture-landscape" class="anchor"&gt;🔗&lt;/a&gt;&lt;/h2&gt;&lt;p&gt;The GPU programming ecosystem has evolved with distinct architectural approaches between NVIDIA and AMD, creating implementation challenges for cross-platform frameworks.&lt;/p&gt;</description></item><item><title>Gluon: When Triton Isn't Low-Level Enough</title><link>https://www.msuiche.com/posts/gluon-when-triton-isnt-low-level-enough/</link><pubDate>Tue, 23 Sep 2025 00:00:00 +0000</pubDate><guid>https://www.msuiche.com/posts/gluon-when-triton-isnt-low-level-enough/</guid><description>&lt;h1 id="my-journey-from-pytorch-to-gluon"&gt;My Journey from PyTorch to Gluon &lt;a href="#my-journey-from-pytorch-to-gluon" class="anchor"&gt;🔗&lt;/a&gt;&lt;/h1&gt;&lt;p&gt;After spending the last month diving into PyTorch, learning Triton, understanding CUDA, and even peeking at PTX/SASS assembly, I&amp;rsquo;ve come to a surprising realization: I&amp;rsquo;ve yet to meet anyone who&amp;rsquo;s actually writing &lt;a href="https://siboehm.com/articles/22/CUDA-MMM" target="_blank" rel="noopener"&gt;raw CUDA code in production anymore&lt;/a&gt;. Everyone I&amp;rsquo;ve talked to – from ML engineers at startups to researchers at big tech companies – seems to have converged on Triton as their go-to solution for custom GPU kernels. And honestly? The &lt;a href="https://www.gpumode.com/v2/leaderboard/496?tab=rankings" target="_blank" rel="noopener"&gt;fused kernels performance they&amp;rsquo;re getting is impressive enough&lt;/a&gt; that I understand why.&lt;/p&gt;</description></item></channel></rss>