GPU
How to Run Karpathy's Autoresearch on Your Local GPU
Set up Karpathy's autoresearch on your GPU to run 100+ ML experiments overnight. Works on RTX 3090/4090 as-is, scales down to 6GB cards with tweaks.
Local AI Upscaling: Make Blurry Images Sharp Without the Cloud
Upscayl, Real-ESRGAN, chaiNNer, and ComfyUI can upscale your photos for free on your own hardware. No subscriptions, no uploads, no per-image fees. Even a GTX 1060 works. Here's how to pick the right tool and start.
Running OpenClaw on 4GB, 6GB, and 8GB GPUs: What Actually Works
OpenClaw on low VRAM GPUs: 4GB is rough, 6GB is marginal, 8GB is where it starts working. Model picks, quantization tricks, partial offload, and when to just use a cloud API instead.
Intel Arc B580 for Local LLMs: 12GB VRAM at $250, With Caveats
The Arc B580 gives you 12GB VRAM for $250, but Intel's AI software stack needs work. Real tok/s benchmarks, setup paths, and honest comparison with RTX 3060.
Docker for Local AI: The Complete Setup Guide for Ollama, Open WebUI, and GPU Passthrough
Run Ollama and Open WebUI in Docker with GPU passthrough. Five copy-paste compose files for NVIDIA, AMD, multi-GPU, and CPU-only setups, plus the Mac gotcha most guides skip.
WSL2 + Ollama on Windows: Complete Setup Guide (GPU Passthrough Included)
Install Ollama in WSL2 with full GPU acceleration in 20 minutes. GPU passthrough, Open WebUI, Docker Compose, VPN fixes, and the gotchas that will waste your afternoon.
RTX 5060 Ti Review for Local AI — The New Budget King
Real benchmarks for the RTX 5060 Ti 16GB running local LLMs. Qwen 3.5 35B at 44 tok/s, 100K context for ~$430. Compared against RTX 3060, 3090, and 4060 Ti.
Used Server GPUs for Local AI: Tesla P40, V100, A100, and the eBay Goldmine
A Tesla P40 has 24GB VRAM for $175. A V100 has 32GB for $350. Server GPUs offer insane VRAM per dollar for local AI — if you can handle the quirks. Full breakdown with prices, benchmarks, and cooling fixes.
Best Qwen 3.5 Setup: Which Model Fits Your GPU (Complete Cheat Sheet)
Pick the right Qwen 3.5 model for your hardware. Covers 0.8B through 397B with VRAM requirements, quant recommendations, and benchmarks for every GPU tier.
Used Tesla P40 for Local AI: The $200 Budget Beast
24GB VRAM for $150-$200 on eBay. Pascal architecture, no display output, passive cooling. Full benchmarks, setup guide, and honest comparison to the RTX 3060 and 3090.
RTX 5090 for Local AI: Worth the Upgrade?
32GB GDDR7, 1,792 GB/s bandwidth, 67% faster than 4090 — but $3,500+ street price. Full benchmarks, value analysis, and who should actually buy one.
RTX 4090 vs Used RTX 3090 for Local AI: Which to Buy in 2026
Both have 24GB VRAM. One costs 2-3x more. RTX 4090 vs used RTX 3090 — real benchmarks, real prices, and who should buy which for local LLM inference and image generation.
ROCm Not Detecting GPU: AMD Troubleshooting Guide
AMD GPU not detected in ROCm? Check supported GPUs, fix rocminfo errors, HSA_OVERRIDE hack for unsupported cards, and Ollama/llama.cpp ROCm build fixes.
Ollama Not Using GPU: Complete Fix Guide
Ollama running on CPU instead of GPU? Diagnose with ollama ps and nvidia-smi, then fix CUDA drivers, ROCm setup, VRAM limits, and Docker GPU passthrough.
Best LLM Speed Trick: ExLlamaV2 vs llama.cpp Benchmarks (50-85% Faster)
Head-to-head speed benchmarks on RTX 3090 and 4090. ExLlamaV2 generates tokens 50-85% faster than llama.cpp on NVIDIA GPUs. Full comparison with setup guides for both.
RTX 3060 vs 3060 Ti vs 3070 for Local AI
The RTX 3060 has 12GB VRAM, the 3060 Ti and 3070 only have 8GB. For LLMs, the cheapest card wins — it runs 14B models the others can't fit. Speeds, prices, and when the 3070 still makes sense.
Mixtral 8x7B & 8x22B VRAM Requirements
Mixtral 8x7B and 8x22B VRAM requirements at every quantization level. Exact numbers from Q2 to FP16, GPU recommendations, and KV cache impact explained.
RTX 3090 vs 4070 Ti Super for Local LLMs
Head-to-head comparison of the RTX 3090 and RTX 4070 Ti Super for running LLMs locally. Covers VRAM, speed, power, price, and which to buy for your use case.
Best Used GPUs for Local AI: 2026 Buying Guide
RTX 3090 at $700-850 for 24GB, RTX 3060 12GB at $170-220, RTX 3080 at $350-400. Tier rankings, fair prices, what to avoid (skip the 8GB 3070), and where to buy safely.
Best GPU Under $500 for Local AI (2026 Picks)
Find the best GPU under $500 for running local AI in 2026. RTX 4060 Ti 16GB, used RTX 3080, RTX 3060 12GB, and RX 7700 XT compared with real benchmarks.
Best GPU Under $300 for Local AI (2026 Picks)
Find the best GPU under $300 for local AI. We compare the RTX 3060 12GB, RX 7600, and Intel Arc B580 with VRAM analysis, LLM benchmarks, and real pricing.
Ollama Troubleshooting Guide: Every Common Problem and Fix
GPU not detected? Running at 1/30th speed on CPU? OOM crashes mid-generation? Every common Ollama error with exact diagnostic commands and fixes for Mac, Windows, and Linux. Updated for v0.17.7 and Qwen 3.5.
What Can You Actually Run on 16GB VRAM?
13B-14B models hit 22-53 tok/s at Q4-Q6, Flux runs at FP8, and 20B models squeeze in with short context. Where 16GB beats 12GB, where it trails 24GB, and the best cards at this tier.
Mac vs PC for Local AI: Which Should You Choose?
RTX 3090 runs 7B-14B models 2-3x faster than M4 Pro. M4 Max with 128GB loads 70B models a PC can't touch. Real benchmarks, prices, and which platform fits your use case.
What Can You Actually Run on 24GB VRAM?
32B models at 25-38 tok/s, 70B at Q3 with limited context, Flux at full FP16, and LoRA fine-tuning. RTX 3090 at $700 vs 4090 at $1,800—every model that fits and which GPU to buy.
What Can You Actually Run on 8GB VRAM?
7B-8B models hit 35-42 tok/s at Q4, SD 1.5 runs great, SDXL is tight but doable. Nothing above 13B fits. Every model that works on RTX 4060 and 3060 Ti, plus the best upgrade path.
What Can You Actually Run on 12GB VRAM?
14B models at Q4 hit 25-32 tok/s, 7B-8B run at near-lossless Q6-Q8, and SDXL generates without workarounds. Every model that fits on an RTX 3060 12GB and the best upgrade path.
Used RTX 3090 Buying Guide for Local AI
24GB VRAM for $650-750—half the cost of an RTX 4090 with the same capacity. Fair prices, eBay red flags, PSU requirements (850W minimum), and how to test before your return window closes.
NVIDIA GPU Prices Are Rising: What to Do Now
GPU prices are spiking due to GDDR7 shortages and AI datacenter demand. Here's what's happening, which cards are affected, and strategies for local AI builders.
Best VRAM Cheat Sheet for Local LLMs: Every Model, Every Quant
Exact VRAM for Qwen 3.5, Llama, Mistral, and DeepSeek at Q3 through FP16. Lookup tables for 7B, 9B, 13B, 27B, 32B, 70B, and 120B models with real measurements and GPU recommendations. Updated March 2026.
RTX 5060 Ti 16GB Killed? Local AI Alternatives
The RTX 5060 Ti 16GB faces production cuts from GDDR7 shortages. See what is really happening and explore the best alternative GPUs for local AI in 2026.
GPU Buying Guide for Local AI: Pick the Right Card
The complete GPU buying guide for local AI. Covers RTX 3060 through 4090 with VRAM analysis, performance benchmarks, prices, and used vs new buying advice.