This extension allows you to chat with websites using your local Ollama models.
- Chat with websites using Ollama models
- Customize the model, server URL, and system prompt
- Save and load settings
- Automatically loads current website content for contextual chat
- Chat directly about the webpage you're viewing without any manual setup
- Clone the repository
- Load the extension in Chrome by navigating to
chrome://extensions/, enabling "Developer mode", and clicking "Load unpacked". - Pin the extension to the toolbar
OLLAMA_KEEP_ALIVE=1h OLLAMA_ORIGINS=* OLLAMA_HOST=0.0.0.0:11434
using Ollama OSX app:
launchctl setenv OLLAMA_KEEP_ALIVE "1h"
launchctl setenv OLLAMA_ORIGINS "*"
launchctl setenv OLLAMA_HOST "0.0.0.0:11434"
