Used Hardware
Used Server GPUs for Local AI: Tesla P40, V100, A100, and the eBay Goldmine
A Tesla P40 has 24GB VRAM for $175. A V100 has 32GB for $350. Server GPUs offer insane VRAM per dollar for local AI — if you can handle the quirks. Full breakdown with prices, benchmarks, and cooling fixes.
Best Used GPUs for Local AI: 2026 Buying Guide
RTX 3090 at $700-850 for 24GB, RTX 3060 12GB at $170-220, RTX 3080 at $350-400. Tier rankings, fair prices, what to avoid (skip the 8GB 3070), and where to buy safely.
Used RTX 3090 Buying Guide for Local AI
24GB VRAM for $650-750—half the cost of an RTX 4090 with the same capacity. Fair prices, eBay red flags, PSU requirements (850W minimum), and how to test before your return window closes.