OwnRig

Code Llama 34B Instruct on Apple M3 Pro (18GB Unified)

Apple M3 Pro (18GB Unified) cannot run Code Llama 34B Instruct at any quantization level. The 18 GB of VRAM is insufficient.

Model Size

33.7B

Device VRAM

18 GB

Bandwidth

150 GB/s

Quantizations Tested

1

Performance by Quantization

Each row shows Code Llama 34B Instruct performance at a different quality level on Apple M3 Pro (18GB Unified).

QuantizationSpeedTTFTFits in VRAMRatingConfidence
Q3_K_M✗ OffloadNot Viableestimated

Notes

Q3_K_M

34B exceeds 14GB effective. Doesn't fit.

About Code Llama 34B Instruct

Code Llama 34B Instruct (33.7B) is a coding, ai coding model. Meta's dedicated 34B coding model. Still competitive for code generation but being surpassed by newer models like Qwen 2.5 Coder 32B. Shorter context window (16K) is a limitation for large codebases.

View all Code Llama 34B Instruct hardware options →

About Apple M3 Pro (18GB Unified)

Apple M3 Pro (18GB Unified) has 18 GB at 150 GB/s. Available in MacBook Pro 14", MacBook Pro 16".

See all models Apple M3 Pro (18GB Unified) can run →

Source: MLX performance estimates (2026-03-15)

Data last updated: 2026-03-01