Skip to main content

Ollama

Ollama runs open-source AI models locally on your machine. Free, private, no API key needed.

Setup

  1. Install Ollama: ollama.com/download
  2. Pull a model:
    ollama pull llama3.1
    
  3. Pawz auto-detects Ollama at localhost:11434
That’s it — no API key or configuration needed.

Configuration

SettingDefaultNotes
Base URLhttp://localhost:11434/v1Change if Ollama runs on another machine
API key(none)Not needed for local Ollama
Default modelSet to your preferred model
ModelSizeBest for
llama3.1:8b4.7 GBGeneral use, fast
llama3.1:70b40 GBHigh quality, needs GPU
gemma2:9b5.4 GBGood balance
mistral:7b4.1 GBFast, European
codellama:13b7.4 GBCode generation
deepseek-coder-v28.9 GBCode generation

Embedding model

Pawz uses nomic-embed-text for the memory system:
ollama pull nomic-embed-text
Pawz auto-pulls this model if it’s missing.

Remote Ollama

To use Ollama on another machine:
  1. On the remote machine, set OLLAMA_HOST=0.0.0.0:11434
  2. In Pawz, change the base URL to http://<remote-ip>:11434/v1

Tips

  • Ollama uses your GPU automatically if available
  • Models are downloaded to ~/.ollama/models/
  • Run ollama list to see installed models
  • Run ollama rm <model> to delete a model
  • Ollama models use the : syntax (e.g., llama3.1:8b) — Pawz auto-routes these