Author
Alex Voss
Hardware Reviewer · The AI Desk
Independent hardware reviewer focused on local AI inference — GPUs, Mini PCs, and the software stack that ties them together. Writing about what actually runs, at what speed, and whether it's worth the money.
Background
Alex has been building and benchmarking local AI hardware since 2022, when running a 7B model locally still required a dedicated workstation. The goal then — and now — is the same: find out what ordinary consumer hardware can actually do, without relying on manufacturer benchmarks or cloud-hosted demos.
Background in software infrastructure led to a deep interest in the systems layer: memory hierarchies, driver stacks, compute scheduling. When LLMs started demanding the same understanding of hardware that distributed systems engineers apply to servers, the overlap became obvious.
Every review on The AI Desk follows a fixed benchmark protocol — same models, same quantization, same tools as our readers use. No manufacturer samples. No cherry-picked runs. Results are cross-referenced against community benchmarks from r/LocalLLaMA and llama.cpp before publishing.
Areas of Expertise
- →
GPU Architecture
CUDA, ROCm, Metal — how memory hierarchy and compute clusters translate to real inference throughput.
- →
LLM Inference
Tokens-per-second measurement methodology using Ollama and llama.cpp. Quantization tradeoffs (Q4_K_M vs Q8_0 vs fp16).
- →
Memory Bandwidth
Why GB/s matters more than FLOPS for LLM workloads. VRAM sizing for common model families.
- →
Apple Silicon
Unified memory architecture, Metal Performance Shaders, and how M-series chips compare to discrete GPUs for AI.
- →
Mini PC Ecosystem
AMD Ryzen AI, Intel Core Ultra, eGPU compatibility, and thermal limits for always-on inference servers.
- →
Image Generation
Stable Diffusion XL, FLUX, and ComfyUI pipeline benchmarks. VRAM minimums per resolution.
Review Methodology
All benchmarks use a fixed protocol: Ollama latest stable, Llama 3.1 8B Q4_K_M, 500-token seed prompt, 3 runs with median reported. Image generation tested via ComfyUI with SDXL and FLUX.1-dev at 1024×1024. Full details on the methodology page →
Recent Articles
- →
Apple Mac Mini M4 Pro for Local LLM Review 2026: 70B Models at 30W
April 29, 2026
- →
Apple Silicon vs NVIDIA for Local AI 2026: M4 Pro vs RTX 5070
April 29, 2026
- →
Best AI PC Under $1,000 in 2026
April 28, 2026
- →
How to Run Stable Diffusion Locally in 2026: Complete Setup Guide
April 28, 2026
- →
Best GPU for Stable Diffusion and FLUX in 2026
April 26, 2026
- →
Best Local AI Hardware Under $500 in 2026
April 26, 2026
Products Reviewed
The AI Desk catalog covers GPUs, Mini PCs, and AI accessories — 27 products benchmarked to date. Every product page includes tokens-per-second data, VRAM or unified memory specs, power draw, and a verdict on the specific use case it best serves.
Browse all products →