Local AI anywhere, for everyone — LLM inference, chat UI, voice, agents, workflows, RAG, and image generation. No cloud, no subscriptions.
-
Updated
Apr 28, 2026 - Rust
Local AI anywhere, for everyone — LLM inference, chat UI, voice, agents, workflows, RAG, and image generation. No cloud, no subscriptions.
This is a mirror of the Strix Halo HomeLab wiki, to browse the wiki click on the link below
[ARCHIVED 2026-04-20 — stampby retired; see bong-water-water-bong] no longer actively developed
Experimental support for many TTS/STT LLMs wrapped in a Wyoming API for consumption via Homeassistant
Sixunited AXB35 EC control & monitoring for Windows
The definitive Strix Halo LLM guide — 65 t/s on a $2,999 mini PC. Live benchmarks, tested optimizations, and everything that doesn't work.
A comprehensive guide to running Linux (Omarchy/Arch) on the 2025 ASUS ROG Flow Z13 (AMD Strix Halo). Includes CachyOS Kernel setup, Tablet Mode fixes, and Power Management for the Ryzen AI Max
vLLM + Qwen3.6-27B (BF16) OpenAI-compatible inference server on AMD Strix Halo (Ryzen AI Max+ 395, gfx1151). Vision input, 256K context, /v1/responses with separated reasoning, via TheRock ROCm.
Tools and documentation related to the AMD Strix-Halo AGU family (Ryzen AI Max 395) of systems. Tested on GMKtec EVO-2
llama.cpp + Qwen3.6-27B (Q8_0 GGUF) OpenAI-compatible inference server on AMD Strix Halo (Ryzen AI Max+ 395, gfx1151). 256K context, ~7.5 t/s decode via TheRock ROCm Docker.
Local, ternary-weight LLM inference on AMD Strix Halo. Rust above the kernels, HIP below, zero Python at runtime. https://discord.gg/EhQgmNePg
Simple installer script which take a download (if newer) and installs it globally. Sets Vulkan support
ComfyUI on AMD Strix Halo (RDNA 3.5 / gfx1151) via Docker. Ubuntu Rolling + UV-managed Python 3.12 + ROCm preview wheels. Solves the silent CPU fallback Debian/Python 3.13 images hit on gfx1151.
Ansible playbook to configure AMD Strix Halo machines (e.g. Framework Desktop or GMKtec EVO-X2) as local AI inference servers running Fedora 43. Sets up llama.cpp with llama-swap and Open WebUI and downloads GGUF models. With NGINX reverse proxy and TLS via ACME or self-signed certificate.
Claude Code skill for AMD Strix Halo (Ryzen AI MAX+ 395) ML setup. Handles PyTorch installation (official wheels don't work with gfx1151), GTT memory config, and environment setup. Enables 30B parameter models.
llama.cpp setup on dedicated AMD Strix Halo machine
Talos-O (Omni): A sovereign, embodied agentic organism forged on AMD Strix Halo. Integrating the Chimera Kernel (Linux 7.0), Zero-Copy Introspection, and the Phronesis Engine. Built from First Principles.
Stable Diffusion image generation on AMD Ryzen AI NPUs for Linux
Local LLM benchmarks on AMD Strix Halo — 26+ models tested across RADV, AMDVLK, and ROCm with llama.cpp
Add a description, image, and links to the strix-halo topic page so that developers can more easily learn about it.
To associate your repository with the strix-halo topic, visit your repo's landing page and select "manage topics."