Setup-Guide
Docker for Local AI: The Complete Setup Guide for Ollama, Open WebUI, and GPU Passthrough
Run Ollama and Open WebUI in Docker with GPU passthrough. Five copy-paste compose files for NVIDIA, AMD, multi-GPU, and CPU-only setups, plus the Mac gotcha most guides skip.
Best 8GB GPU Model: How to Set Up Qwen 3.5 9B (Step by Step)
Qwen 3.5 9B fits in 6.6GB and beats models 3x its size. Complete setup with Ollama, benchmarks, and real-world testing on RTX 3060 and 4060.
WSL2 + Ollama on Windows: Complete Setup Guide (GPU Passthrough Included)
Install Ollama in WSL2 with full GPU acceleration in 20 minutes. GPU passthrough, Open WebUI, Docker Compose, VPN fixes, and the gotchas that will waste your afternoon.
WSL2 for Local AI: The Complete Windows Setup Guide
Install WSL2, configure GPU passthrough, set up Ollama and llama.cpp with CUDA, and optimize memory for LLM inference. Step-by-step for Windows 11.