Ollama Client is a powerful, privacy-first Chrome extension that lets you chat with locally hosted LLMs using Ollama โ no cloud, no tracking. Itโs lightweight, open source, and designed for fast, offline-friendly AI conversations.
โ Works with any Chromium-based browser: Chrome, Brave, Edge, Opera, Chromium, and Arc.
๐ฆ Firefox support available via temporary addon installation (manual permissions setup required).
๐ Model Search & Pull โ Pull models directly in the UI (with progress indicator)
๐ฆ Model Version Display โ View and compare model versions easily
declarativeNetRequest
, storage
, sidePanel
)๐ Chrome Web Store
brew install ollama # macOS
ollama serve # starts at http://localhost:11434
More info: https://ollama.com
ollama pull gemma3:1b
Other options: mistral
, llama3:8b
, codellama
, etc.
Set your:
http://localhost:11434
gemma:2b
)Advanced parameters like system prompts and stop sequences are available per model.
Want to contribute or customize? You can run and modify the Ollama Client extension locally using Plasmo.
git clone https://github.com/Shishir435/ollama-client.git
cd ollama-client
Using pnpm (recommended):
pnpm install
Or with npm:
npm install
Start development mode with hot reload:
pnpm dev
Or with npm:
npm run dev
This launches the Plasmo dev server and gives instructions for loading the unpacked extension in Chrome:
chrome://extensions
dist/
folder generated by Plasmopnpm dev --target=firefox
Load as a temporary extension.
pnpm build
Output will be in the build/
or dist/
folder depending on your Plasmo version.
src/
: Core logic and componentsbackground.ts
: API bridge + streamingsidepanel.tsx
: Main chat UIoptions.tsx
: Settings pagecontent.ts
: Summarizer / Readabilitylib/
: Utility functionshooks/
, features/
, context/
: Modular structure for maintainabilitypackage.json
System Specs | Suggested Models |
---|---|
๐ป 8GB RAM (no GPU) | gemma:2b , mistral:7b-q4 |
๐ป 16GB RAM (no GPU) | gemma:3b-q4 , mistral |
๐ฎ 16GB+ with GPU (6GB VRAM) | llama3:8b-q4 , gemma:3b |
๐ฅ RTX 3090+ or Apple M3 Max | llama3:70b , mixtral |
๐ฆ Prefer quantized models (q4_0
, q5_1
, etc.) for better performance.
Explore: Ollama Model Library
Ollama Client is a Chrome Manifest V3 extension. To use in Firefox:
about:debugging
manifest.json
from the extension folderIf you find Ollama Client helpful, please consider:
#OllamaClient
)Built with โค๏ธ by @Shishir435