Author

Alex Voss

Hardware Reviewer · The AI Desk

Independent hardware reviewer focused on local AI inference — GPUs, Mini PCs, and the software stack that ties them together. Writing about what actually runs, at what speed, and whether it's worth the money.

Background

Alex has been building and benchmarking local AI hardware since 2022, when running a 7B model locally still required a dedicated workstation. The goal then — and now — is the same: find out what ordinary consumer hardware can actually do, without relying on manufacturer benchmarks or cloud-hosted demos.

Background in software infrastructure led to a deep interest in the systems layer: memory hierarchies, driver stacks, compute scheduling. When LLMs started demanding the same understanding of hardware that distributed systems engineers apply to servers, the overlap became obvious.

Every review on The AI Desk follows a fixed benchmark protocol — same models, same quantization, same tools as our readers use. No manufacturer samples. No cherry-picked runs. Results are cross-referenced against community benchmarks from r/LocalLLaMA and llama.cpp before publishing.

Areas of Expertise

  • GPU Architecture

    CUDA, ROCm, Metal — how memory hierarchy and compute clusters translate to real inference throughput.

  • LLM Inference

    Tokens-per-second measurement methodology using Ollama and llama.cpp. Quantization tradeoffs (Q4_K_M vs Q8_0 vs fp16).

  • Memory Bandwidth

    Why GB/s matters more than FLOPS for LLM workloads. VRAM sizing for common model families.

  • Apple Silicon

    Unified memory architecture, Metal Performance Shaders, and how M-series chips compare to discrete GPUs for AI.

  • Mini PC Ecosystem

    AMD Ryzen AI, Intel Core Ultra, eGPU compatibility, and thermal limits for always-on inference servers.

  • Image Generation

    Stable Diffusion XL, FLUX, and ComfyUI pipeline benchmarks. VRAM minimums per resolution.

Review Methodology

All benchmarks use a fixed protocol: Ollama latest stable, Llama 3.1 8B Q4_K_M, 500-token seed prompt, 3 runs with median reported. Image generation tested via ComfyUI with SDXL and FLUX.1-dev at 1024×1024. Full details on the methodology page →

Recent Articles

All articles →

Products Reviewed

The AI Desk catalog covers GPUs, Mini PCs, and AI accessories — 27 products benchmarked to date. Every product page includes tokens-per-second data, VRAM or unified memory specs, power draw, and a verdict on the specific use case it best serves.

Browse all products →