Techalicious Academy / 2026-02-24-openclaw-ollama

(Visit our meetup for more great tutorials)

INSTALLING OLLAMA

Ollama is the local model server. It handles downloading models, loading them into memory, running inference on Apple Silicon's GPU via Metal, and exposing an API that other tools (like OpenClaw) can talk to. Think of it as the engine. OpenClaw is the dashboard.

Two Ways To Install

Option A: Direct download (recommended)

Go to https://ollama.com/download/mac and download the DMG. Open it,
drag the app to your Applications folder, and launch it. On first run,
it installs the command-line tool at /usr/local/bin/ollama.

That's it. Done.

Option B: Homebrew

brew install ollama

This puts the binary at /opt/homebrew/bin/ollama on Apple Silicon Macs.
Same result, different path.

Either way works fine. The direct download is simpler for people who don't have Homebrew set up.

Verify The Installation

Open a terminal and run:

ollama --version

You should see a version number. If you get "command not found", the CLI didn't install properly. Try launching the Ollama app from Applications first, then reopen your terminal.

ollama list

This should return an empty table since you haven't pulled any models yet. If it returns an error about connecting to the server, make sure the Ollama app is running (check your menu bar for the Ollama icon).

Confirming Metal Acceleration

This is important. Ollama must be running natively on Apple Silicon to use Metal GPU acceleration. If it's running under Rosetta 2 (the x86 emulator), Metal is completely disabled and everything will be painfully slow.

Verify the binary architecture:

file $(which ollama)

You want to see "arm64" in the output. If you see "x86_64", something went wrong. Reinstall from the official download.

Also verify the path makes sense:

which ollama

Expected paths:

/usr/local/bin/ollama (direct install)
/opt/homebrew/bin/ollama (Homebrew)

If it's pointing somewhere unexpected, you might have a stale installation.

A Note About Docker

You might be tempted to run Ollama in Docker. Don't. Docker on macOS runs inside a Linux VM, and that VM has no access to Apple Silicon's GPU. Your model would run on CPU only, which means maybe 1-2 tokens per second instead of 10-40. It's unusable.

Always run Ollama natively on macOS for Metal acceleration.

Starting Ollama

If you installed via the DMG, just launch the app. You'll see a small llama icon in your menu bar. The server starts automatically.

If you installed via Homebrew and want to start it manually:

ollama serve

This starts the server in the foreground. You'll see log output. Open a new terminal tab for running commands.

To verify the server is running:

curl http://localhost:11434/

This should return: "Ollama is running"

If it does, you're good. The model server is up, Metal is active, and we're ready to pull a model.