Skip to main content
BrowserOS works great with local models for Chat Mode. Run models completely offline — your data never leaves your machine.

Context Length

Ollama defaults to 4,096 tokens of context — this is too low for BrowserOS. Below 15K tokens, the context overflows and the agent gets stuck in a loop constantly trying to recover. Only Chat Mode will work at low context lengths. Set at least 15,000–20,000 tokens for local models to function properly.
Set context length when starting Ollama:
OLLAMA_CONTEXT_LENGTH=20000 ollama serve
Increasing context length uses more VRAM. Run ollama ps to check your current allocation. See the Ollama context length docs for more details.

Setup

The easiest way to run models locally.
1

Install Ollama

Download from ollama.com and install it.
2

Pull a model

ollama pull qwen/qwen3-4b
3

Start Ollama with higher context

OLLAMA_CONTEXT_LENGTH=20000 ollama serve
4

Configure in BrowserOS

  1. Go to chrome://browseros/settings
  2. Click USE on the Ollama card
  3. Set Model ID to qwen/qwen3-4b
  4. Set Context Window to 20000
  5. Click Save Ollama in BrowserOS

Pick a model based on your available RAM/VRAM. Smaller models are faster but less capable.

Lightweight (under 5 GB)

Good for machines with 8 GB RAM. Fast responses, suitable for simple chat tasks.
ModelPublisherParamsQuantSize
qwen/qwen3-4bQwen4B4bit2.28 GB
mistralai/ministral-3-3bMistral3BQ4_K_M2.99 GB
deepseek-r1-distill-qwen-7blmstudio-community7BQ4_K_M4.68 GB
deepseek-r1-distill-llama-8blmstudio-community8BQ4_K_M4.92 GB

Mid-range (10–15 GB)

Needs 16+ GB RAM. Better reasoning, handles longer conversations well.
ModelPublisherParamsQuantSize
openai/gpt-oss-20bOpenAI20BMXFP412.11 GB
mistralai/magistral-smallMistral23.6B4bit13.28 GB
mistralai/devstral-small-2-2512Mistral24B4bit14.12 GB

Heavy (60+ GB)

For workstations with 64+ GB RAM. Closest to cloud model quality.
ModelPublisherParamsQuantSize
openai/gpt-oss-120bOpenAI120BMXFP463.39 GB
Start with qwen/qwen3-4b if you’re unsure — it’s small, fast, and surprisingly capable for its size.