Finetuning LLMs on Strix Halo – Full, LoRA, and QLoRA on Gemma-3, Qwen-3, and GPT-OSS-20B



Date: 12/24/2025

Watch the Video

Fine-Tuning LLMs on AMD Strix Halo: Local Powerhouse Tutorial

Loved this deep dive into fine-tuning large language models (LLMs) right on your desk with the Framework Desktop powered by AMD Ryzen AI Max “Strix Halo.” The creator shares a ready-to-go Jupyter toolbox for seamless setup, letting you tweak Gemma-3, Qwen, and GPT-OSS 20B models using full-parameter training, LoRA, or QLoRA (with 8/4-bit quantization).

Covers everything from dataset prep and unified memory tricks to checkpointing and exporting for inference. Perfect for AI tinkerers in no-code/low-code spaces looking to harness edge hardware without cloud hassle.

Timestamps for quick jumps: Intro (00:00), Toolbox Install (04:09), Gemma-3 Deep Dive (10:49+), LoRA/QLoRA (30:00+), and GPT-OSS (43:08).

Grab the toolbox and notebooks here: GitHub Repo
More resources: Framework Desktop, Gemma Models, Full Playlist

Watch now and level up your local AI game! 🚀