$1,892

| Category | Component | Price | Rationale | Buy |
|---|---|---|---|---|
| gpu | $799 | The undisputed price-performance king for home AI servers. 24GB VRAM at 936 GB/s bandwidth runs 32B models at Q4 comfortably. Used market price ~$700-900 — roughly half the cost of a new RTX 4090 with identical VRAM. TDP is 350W, so power cost is higher than budget tier (~$35/month at 24/7), but the VRAM-per-dollar ratio is unbeatable. Source: r/LocalLLaMA used market surveys, eBay sold listings. | Buy | |
| cpu | $229 | 65W TDP 8-core. Handles multiple Ollama instances, Open WebUI, and RAG pipelines simultaneously. The non-X variant runs 20W cooler for identical server workloads. 8 cores give headroom for concurrent request handling. | Buy | |
| motherboard | ASRockASRock B650M PG Riptide WiFi | $149 | B650 micro-ATX with 2.5GbE LAN, two M.2 slots, and 4 SATA ports. Strong VRMs handle the 7700 without issue. BIOS supports Wake-on-LAN and PBO configuration for server-optimized power profiles. | Buy |
| ram | 64GB DDR5-5600 (2x32GB) | $159 | 64GB is essential for a multi-user server: Ollama caches models in system RAM, Open WebUI runs its database, and RAG indexes live in memory. When 2-3 household members hit the server simultaneously, RAM prevents swapping. Expandable to 128GB. | Buy |
| storage | $289 | 1TB NVMe for OS and active models — holds 15-20 quantized models ready to serve. Dual 4TB IronWolf HDDs for RAID 1 mirrored data storage: documents, media, backups. IronWolf drives are rated 180TB/year for 24/7 NAS workloads. RAID 1 protects against single drive failure. | Buy | |
| psu | $99 | 750W handles the 3090's 350W TDP with system headroom. At ~500W typical load, the PSU operates at its efficiency peak. Fully modular. 10-year warranty matches the intended server lifespan. | Buy | |
| case | Fractal DesignFractal Design Node 804 | $119 | Dual-chamber micro-ATX case with 8 HDD bays — a proper NAS chassis. GPU and CPU in the front chamber, drives in the rear. 310mm GPU clearance fits most 3090 models. Two 120mm and one 140mm fans included. Cube form factor is stable on shelves. | Buy |
| cooler | NoctuaNoctua NH-U12S redux | $49 | Quiet 120mm tower cooler. Handles the 65W 7700 at whisper-quiet levels. The Node 804 has 160mm cooler clearance — this fits at 158mm. Noctua's fans are rated for 150,000+ hours MTBF. | Buy |
| Total | $1,892 | |||
Search links — prices and availability vary by retailer.
Prices and availability vary. Inspect hardware before purchasing.
AI models tested on this build's hardware.
| Model | Quant | Speed |
|---|---|---|
| Qwen 2.5 Coder 32B Instruct | Q4_K_M | 22 tok/s |
| DeepSeek R1 Distill Qwen 32B | Q4_K_M | 21 tok/s |
| QwQ 32B Preview | Q4_K_M | 21 tok/s |
| Llama 3.1 8B Instruct | Q8_0 | 80 tok/s |
| Gemma 2 27B Instruct | Q4_K_M | 20 tok/s |
| Phi-4 14B | Q5_K_M | 35 tok/s |
| Mistral 7B Instruct v0.3 | Q8_0 | 75 tok/s |
| nomic-embed-text v1.5 | FP16 | — |
| Whisper Large V3 | FP16 | — |
Always-on local AI server for a household or small team. Runs Ollama + Open WebUI accessible from any device on the network. Serves chat, coding assistance, document Q&A, and transcription to multiple simultaneous users — with zero API costs and complete data privacy.
7 GB of 24 GB
17 GB headroom for additional workloads
If you're paying ~$80/month for cloud API access, this build pays for itself in 10 months.
Based on a 3-person household using ChatGPT Plus ($20/mo each = $60/mo) plus occasional API calls for document processing (~$20/mo). Budget Home AI Server at $1,162 including electricity (~$15/mo for the budget tier at 250W average). Break-even includes electricity. After break-even, savings are $65+/month indefinitely. Privacy benefit: no family conversations, documents, or voice recordings leave your network.
Add a second used RTX 3090 (~$800) via an ATX motherboard swap for 48GB total VRAM — enables Llama 3.1 70B at Q3. Both 3090s support NVLink ($80 bridge) for combined memory bandwidth. Or upgrade storage to 4x 8TB IronWolf in RAID 10 for 16TB usable NAS capacity.
Last updated: 2026-03-14.