What is CUDA?
NVIDIA's proprietary parallel computing platform. Industry standard for AI/ML. Nearly every AI framework (PyTorch, Ollama, ComfyUI) supports CUDA natively and first.
Full Explanation
CUDA (Compute Unified Device Architecture) is NVIDIA's proprietary programming model for GPU-accelerated computing. Introduced in 2007, it became the de-facto standard for AI research and production. Every major deep learning framework — PyTorch, TensorFlow, JAX — treats CUDA as the primary accelerator target. For local AI, this means NVIDIA GPUs have near-universal software compatibility: Ollama, ComfyUI, llama.cpp, LM Studio, and Automatic1111 all work out of the box with CUDA.
Why It Matters for Local AI
CUDA's maturity is a practical advantage. AMD ROCm works well on Linux but has inconsistent Windows support and frequent compatibility issues with bleeding-edge tools. If you want everything to just work without debugging driver issues, CUDA on Windows is the path of least resistance.
Hardware Relevant to CUDA
gpu · Check Price on Amazon · 12 GB VRAM · 672 GB/s
gpu · Check Price on Amazon · 12 GB VRAM · 672 GB/s
Related Terms
ROCm→
AMD's open-source GPU compute platform — AMD's answer to NVIDIA CUDA. Required for GPU-accelerated AI on AMD cards. Mature on Linux; less reliable on Windows.
Tensor Cores→
Specialized hardware units on NVIDIA GPUs designed for matrix multiplication — the core math operation in neural networks. 5th-gen Tensor Cores (Blackwell) are significantly faster than 4th-gen (Ada Lovelace) for AI inference.
Ollama→
Free open-source tool for running LLMs locally on macOS, Linux, and Windows. Download a model with a single command. No cloud account required. Supports Llama, Mistral, Qwen, Phi, and more.
ComfyUI→
The node-based GUI for Stable Diffusion and Flux image generation. Industry standard for advanced AI image workflows. Requires a CUDA GPU for practical speeds; AMD ROCm on Linux works.