跳到主要内容
本文尚未翻译

本文暂时显示英文原文,中文翻译正在进行中。翻译完成后将自动更新。

原文链接:English Version

Run Local LLMs on Mac

This guide walks you through running a local LLM server on macOS with an OpenAI-compatible API. You get full privacy, zero API costs, and surprisingly good performance on Apple Silicon.

We cover two backends:

BackendInstallBest atFormat
llama.cppbrew install llama.cppFastest time-to-first-token, quantized KV cache for low memoryGGUF
omlxomlx.aiFastest token generation, native Metal optimizationMLX (safetensors)

Both expose an OpenAI-compatible /v1/chat/completions endpoint. Hermes works with either one — just point it at http://localhost:8080 or http://localhost:8000.

Apple Silicon only

This guide targets Macs with Apple Silicon (M1 and later). Intel Macs will work with llama.cpp but without GPU acceleration — expect significantly slower performance.


Choosing a model

For getting started, we recommend Qwen3.5-9B — it's a strong reasoning model that fits comfortably in 8GB+ of unified memory with quantization.

VariantSize on diskRAM needed (128K context)Backend
Qwen3.5-9B-Q4_K_M (GGUF)5.3 GB~10 GB with quantized KV cachellama.cpp
Qwen3.5-9B-mlx-lm-mxfp4 (MLX)~5 GB~12 GBomlx

Memory rule of thumb: model size + KV cache. A 9B Q4 model is ~5 GB. The KV cache at 128K context with Q4 quantization adds ~4-5 GB. With default (f16) KV cache, that balloons to ~16 GB. The quantized KV cache flags in llama.cpp are the key trick for memory-constrained systems.

For larger models (27B, 35B), you'll need 32 GB+ of unified memory. The 9B is the sweet spot for 8-16 GB machines.


Option A: llama.cpp

llama.cpp is the most portable local LLM runtime. On macOS it uses Metal for GPU acceleration out of the box.

Install

brew install llama.cpp

This gives you the llama-server command globally.

Download the model

You need a GGUF-format model. The easiest source is Hugging Face via the huggingface-cli:

brew install huggingface-cli

Then download:

huggingface-cli download unsloth/Qwen3.5-9B-GGUF Qwen3.5-9B-Q4_K_M.gguf --local-dir ~/models
Gated models

Some models on Hugging Face require authentication. Run huggingface-cli login first if you get a 401 or 404 error.

Start the server

llama-server -m ~/models/Qwen3.5-9B-Q4_K_M.gguf \
-ngl 99 \
-c 131072 \
-np 1 \
-fa on \
--cache-type-k q4_0 \
--cache-type-v q4_0 \
--host 0.0.0.0

Here's what each flag does:

FlagPurpose
-ngl 99Offload all layers to GPU (Metal). Use a high number to ensure nothing stays on CPU.
-c 131072Context window size (128K tokens). Reduce this if you're low on memory.
-np 1Number of parallel slots. Keep at 1 for single-user use — more slots split your memory budget.
-fa onFlash attention. Reduces memory usage and speeds up long-context inference.
--cache-type-k q4_0Quantize the key cache to 4-bit. This is the big memory saver.
--cache-type-v q4_0Quantize the value cache to 4-bit. Together with the above, this cuts KV cache memory by ~75% vs f16.
--host 0.0.0.0Listen on all interfaces. Use 127.0.0.1 if you don't need network access.

The server is ready when you see:

main: server is listening on http://0.0.0.0:8080
srv update_slots: all slots are idle

Memory optimization for constrained systems

The --cache-type-k q4_0 --cache-type-v q4_0 flags are the most important optimization for systems with limited memory. Here's the impact at 128K context:

KV cache typeKV cache memory (128K ctx, 9B model)
f16 (default)~16 GB
q8_0~8 GB
q4_0~4 GB

On an 8 GB Mac, use q4_0 KV cache and reduce context to -c 32768 (32K). On 16 GB, you can comfortably do 128K context. On 32 GB+, you can run larger models or multiple parallel slots.

If you're still running out of memory, reduce context size first (-c), then try a smaller quantization (Q3_K_M instead of Q4_K_M).

Test it

curl -s http://localhost:8080/v1/chat/completions \
-H "Content-Type: application/json" \
-d '{
"model": "Qwen3.5-9B-Q4_K_M.gguf",
"messages": [{"role": "user", "content": "Hello!"}],
"max_tokens": 50
}' | jq .choices[0].message.content

Get the model name

If you forget the model name, query the models endpoint:

curl -s http://localhost:8080/v1/models | jq '.data[].id'

Option B: MLX via omlx

omlx is a macOS-native app that manages and serves MLX models. MLX is Apple's own machine learning framework, optimized specifically for Apple Silicon's unified memory architecture.

Install

Download and install from omlx.ai. It provides a GUI for model management and a built-in server.

Download the model

Use the omlx app to browse and download models. Search for Qwen3.5-9B-mlx-lm-mxfp4 and download it. Models are stored locally (typically in ~/.omlx/models/).

Start the server

omlx serves models on http://127.0.0.1:8000 by default. Start serving from the app UI, or use the CLI if available.

Test it

curl -s http://127.0.0.1:8000/v1/chat/completions \
-H "Content-Type: application/json" \
-d '{
"model": "Qwen3.5-9B-mlx-lm-mxfp4",
"messages": [{"role": "user", "content": "Hello!"}],
"max_tokens": 50
}' | jq .choices[0].message.content

List available models

omlx can serve multiple models simultaneously:

curl -s http://127.0.0.1:8000/v1/models | jq '.data[].id'

Benchmarks: llama.cpp vs MLX

Both backends tested on the same machine (Apple M5 Max, 128 GB unified memory) running the same model (Qwen3.5-9B) at comparable quantization levels (Q4_K_M for GGUF, mxfp4 for MLX). Five diverse prompts, three runs each, backends tested sequentially to avoid resource contention.

Results

Metricllama.cpp (Q4_K_M)MLX (mxfp4)Winner
TTFT (avg)67 ms289 msllama.cpp (4.3x faster)
TTFT (p50)66 ms286 msllama.cpp (4.3x faster)
Generation (avg)70 tok/s96 tok/sMLX (37% faster)
Generation (p50)70 tok/s96 tok/sMLX (37% faster)
Total time (512 tokens)7.3s5.5sMLX (25% faster)

What this means

  • llama.cpp excels at prompt processing — its flash attention + quantized KV cache pipeline gets you the first token in ~66ms. If you're building interactive applications where perceived responsiveness matters (chatbots, autocomplete), this is a meaningful advantage.

  • MLX generates tokens ~37% faster once it gets going. For batch workloads, long-form generation, or any task where total completion time matters more than initial latency, MLX finishes sooner.

  • Both backends are extremely consistent — variance across runs was negligible. You can rely on these numbers.

Which one should you pick?

Use caseRecommendation
Interactive chat, low-latency toolsllama.cpp
Long-form generation, bulk processingMLX (omlx)
Memory-constrained (8-16 GB)llama.cpp (quantized KV cache is unmatched)
Serving multiple models simultaneouslyomlx (built-in multi-model support)
Maximum compatibility (Linux too)llama.cpp

Connect to Hermes

Once your local server is running:

hermes model

Select Custom endpoint and follow the prompts. It will ask for the base URL and model name — use the values from whichever backend you set up above.