
AIOpenCode
1 min read
Deploying Qwen3-Coder-30B-A3B on 8GB GPU with Docker
A 30B model on an 8GB GPU sounds impossible, but quantization and llama.cpp make it work. This guide shows how to run it with Docker and use it in OpenCode.
Browse all articles tagged with 2026. Found 2 articles covering this topic.

A 30B model on an 8GB GPU sounds impossible, but quantization and llama.cpp make it work. This guide shows how to run it with Docker and use it in OpenCode.
Comparing how Gemini CLI, Copilot CLI, and OpenCode CLI approach generating Playwright E2E tests for a React authentication flow. Real code examples and workflow trade-offs for 2026.