Privacy-first extension for local AI chat with multi-provider support. Connect Ollama, LM Studio, or llama.cpp and run conversations directly in your browser. Includes enhanced content extraction, semantic retrieval, and advanced text-to-speech workflows with local-first data handling.
*You control the endpoint. If you configure a remote provider, your data will go to that provider.
Everything you need for productive AI conversations, all running with local-first defaults and user-controlled endpoints
Use local models through Ollama, LM Studio, or llama.cpp. Route models by provider while keeping a single chat experience.
Organize your conversations with multiple chat sessions, all saved locally using IndexedDB.
Advanced content extraction with lazy loading support, site-specific configuration, and intelligent fallback strategies for modern web pages.
Built-in speech synthesis with searchable voice selection, customizable rate and pitch controls, and seamless cross-browser compatibility for an accessible AI chat experience.
Pre-built templates for common tasks to boost your productivity.
Upload PDFs, text, and DOCX files, automatically chunk and embed them for semantic search and RAG.
Full control over AI behavior with comprehensive settings and options.
Local-first by default with full endpoint control. If you point to a remote provider, prompts are sent to that provider.
Leveraging the latest tools and frameworks for optimal performance and developer experience
Simple setup process to get you chatting with AI models locally
Run Ollama, LM Studio, or llama.cpp with a local API endpoint. Configure the matching provider URL in extension settings.
Install the Ollama Client extension from the Chrome Web Store with a single click.
Open the side panel, select a model, and start chatting on any
website.
⚠️ CORS issue? Follow the
setup guide .