OwnRig

Home AI Server

server

Always-on local AI server for a household or small team. Runs Ollama + Open WebUI accessible from any device on the network. Serves chat, coding assistance, document Q&A, and transcription to multiple simultaneous users — with zero API costs and complete data privacy.

OllamaOpen WebUIAnythingLLMWhisperLibreChat

Concurrent VRAM

7 GB

Peak VRAM

10 GB

Min Bandwidth

250 GB/s

Models Required

3

VRAM Breakdown

How the 7 GB concurrent VRAM is used.

Always Running (Concurrent)

Llama 3.1 8B Instruct(fast chat for household)
6.5 GB

Q5_K_M · 8.03B

nomic-embed-text v1.5(document search and rag)
512 MB

FP16 · 137M

Switched (Loaded As Needed)

These share VRAM with the largest concurrent model — only one runs at a time.

Whisper Large V3(voice transcription)
3 GB

FP16

Local vs API Costs

Typical Monthly API Cost

$80/mo

Break-Even Point

10 months

Annual Savings After Break-Even

~$768/yr

Based on a 3-person household using ChatGPT Plus ($20/mo each = $60/mo) plus occasional API calls for document processing (~$20/mo). Budget Home AI Server at $1,162 including electricity (~$15/mo for the budget tier at 250W average). Break-even includes electricity. After break-even, savings are $65+/month indefinitely. Privacy benefit: no family conversations, documents, or voice recordings leave your network.

Recommended Builds

Pre-configured builds that can run the Home AI Server workflow.

Prefer a Mac? Apple Silicon with unified memory can run this workflow too. See the Mac AI Builder workflow →

Get a personalized recommendation for this workflow →

Author: Ada. Last updated: 2026-03-14.