Multimodal
Qwen 3.5 Small Models: The 9B Beats Last-Gen 30B — Here's What Matters for Local AI
Alibaba's Qwen 3.5 drops 4 small models (0.8B to 9B) — all natively multimodal, 262K context, Apache 2.0. The 9B beats Qwen3-30B on reasoning and destroys GPT-5-Nano on vision. VRAM tables and what to run.
Best 8GB GPU Model: How to Set Up Qwen 3.5 9B (Step by Step)
Qwen 3.5 9B fits in 6.6GB and beats models 3x its size. Complete setup with Ollama, benchmarks, and real-world testing on RTX 3060 and 4060.
DeepSeek V4: Everything We Know Before It Drops
DeepSeek V4 launches next week with native image and video generation, 1M context, and rumored 1T MoE params with only 32B active. Here's what local AI builders need to know and how to prepare.
Qwen2.5-VL Not Loading in LM Studio? Fix mmproj and Vision Errors
Fix every Qwen2.5-VL error in LM Studio: missing mmproj, 'model type not supported', no eye icon, vision crashes. Exact fixes with file paths.
Llama 4 Guide: Running Scout and Maverick Locally
Complete Llama 4 guide for local AI — Scout (109B MoE, 17B active) and Maverick (400B). VRAM requirements, Ollama setup, benchmarks, and honest hardware reality check.
Run Qwen2.5-VL Vision in LM Studio (Setup)
Get Qwen2.5-VL running in LM Studio in 5 minutes. Covers the mmproj file most people miss, correct download links, and how to analyze images and PDFs locally.
Best Vision Models You Can Run Locally: Every Model, Every GPU Tier (2026)
Qwen3-VL 8B replaced Qwen2.5-VL as the best local vision model. Full VRAM table, Ollama commands, speed benchmarks, and setup for every GPU from 4GB to 48GB+. Updated March 2026.