OwnRig
FLOW
7 workflows

AI Builder Workflows

How much hardware do you actually need? It depends on your workflow. Pick the setup that matches how you build with AI.

Basic Coding Assistant

basic

Single model, code completion

Run a single local coding model for code completion and chat. The entry-level builder setup — replace API-dependent code completion with a local 7-8B model.

CursorContinueLM StudioOllama
Concurrent VRAM: 5.8 GB
Models: 1
API cost saved: $30/mo
Break-even: 25 months

Recommended Build

Budget AI Desktop

$753

Required Models

Llama 3.1 8B Instruct(code completion and chat)

Power Coding Workflow

power

Dedicated coding model + embeddings

Run a dedicated 32B coding model with embeddings for local RAG. The sweet spot for developers who want serious local AI coding without API dependency.

CursorContinueAiderOllamaOpen WebUI
Concurrent VRAM: 18.9 GB
Models: 2
API cost saved: $80/mo
Break-even: 12 months

Recommended Build

AI Builder Workstation

$2,902

Required Models

Qwen 2.5 Coder 32B Instruct(code completion)concurrentnomic-embed-text v1.5(embeddings for rag)concurrent

Full AI Builder

full

Concurrent coding + reasoning + embeddings

The complete local AI development stack: concurrent coding model + reasoning model + embeddings. Switch between QwQ for architecture decisions and Qwen Coder for implementation, with local RAG always available.

CursorAiderOpen WebUIAnythingLLMCodex CLI
Concurrent VRAM: 18.9 GB
Models: 4
API cost saved: $150/mo
Break-even: 8 months

Recommended Build

AI Builder Workstation

$2,902

Required Models

Qwen 2.5 Coder 32B Instruct(code completion)QwQ 32B Preview(reasoning and architecture)nomic-embed-text v1.5(embeddings for rag)concurrentLlama 3.1 8B Instruct(fast chat and drafting)

Mac AI Builder

mac

Apple Silicon unified memory

The silent, unified-memory approach: Apple Silicon with enough memory to run coding + reasoning + embeddings concurrently. No fan noise, no separate GPU. The premium option for builders who value silence and simplicity.

CursorContinueLM StudioMLXOllama
Concurrent VRAM: 27.3 GB
Models: 3
API cost saved: $100/mo
Break-even: 35 months

Recommended Build

Mac Studio AI Builder

$3,999

Required Models

Qwen 2.5 Coder 32B Instruct(code completion)concurrentLlama 3.1 8B Instruct(fast chat)concurrentnomic-embed-text v1.5(embeddings for rag)concurrent

Home AI Server

server

Always-on, multi-user household AI

Always-on local AI server for a household or small team. Runs Ollama + Open WebUI accessible from any device on the network. Serves chat, coding assistance, document Q&A, and transcription to multiple simultaneous users — with zero API costs and complete data privacy.

OllamaOpen WebUIAnythingLLMWhisperLibreChat
Concurrent VRAM: 7 GB
Models: 3
API cost saved: $80/mo
Break-even: 10 months

Recommended Build

Budget Home AI Server

$1,162

Required Models

Llama 3.1 8B Instruct(fast chat for household)concurrentnomic-embed-text v1.5(document search and rag)concurrentWhisper Large V3(voice transcription)

Model Fine-Tuning & Training

training

Fine-tune language models locally with QLoRA, LoRA, and full fine-tuning. Train custom adapters for domain-specific tasks without sending proprietary data to third-party APIs. VRAM requirements scale with model size and method: QLoRA fine-tuning a 7B model fits in 16GB, while full fine-tuning of 32B models needs 48GB+. System RAM matters — gradient checkpointing and dataset loading use 2-4x the model's VRAM in system memory.

AxolotlUnslothHugging Face TRLLLaMA FactoryPEFT
Concurrent VRAM: 10 GB
Models: 3
API cost saved: $200/mo
Break-even: 10 months

Recommended Build

Mid-Range AI Workstation

$1,228

Required Models

Llama 3.1 8B Instruct(qlora fine tuning target)Mistral 7B Instruct v0.3(qlora fine tuning target)Qwen 2.5 Coder 7B Instruct(code fine tuning target)

AI Image & Video Generation

creative

Local image and video generation with FLUX.1, SDXL, and Stable Diffusion 3. Run ComfyUI workflows with LoRAs, ControlNet, and upscaling — no cloud credits, no content filters, no rate limits. VRAM is the bottleneck: FLUX.1 at full quality needs 24GB, but Q4 quantization brings it to 8GB GPUs.

ComfyUIFooocusAutomatic1111InvokeAIdiffusers
Concurrent VRAM: 13 GB
Models: 3
API cost saved: $60/mo
Break-even: 15 months

Recommended Build

Mid-Range AI Workstation

$1,228

Required Models

FLUX.1 Dev(primary image generation)Stable Diffusion XL 1.0(image generation with loras)Stable Diffusion 3 Medium(fast image generation)