-
🔌 Multiple providers, plug in your API keys (stored entirely locally) and you're good to go
- Local models (through Ollama)
- OpenRouter (which lets you use ALL models across many providers: OpenAI, Anthropic, OSS, 50+ others)
- OpenAI
- Anthropic
- Mistral
- Groq
-
🛠️ Tool use
- Check out
server/toolfns/toolfns.go
. You only need to write functions. The function comment is the description the model receives, so it knows what to use. Click theSync
button in the web UI to refresh your tools.
- Check out
-
🖼️ Multimodal input: upload, paste, or share links to images
-
🎨 Image generation using DALL-E 3
-
📝 Multi-shot prompting. Also edit, delete, regenerate messages, whatever. The world is your oyster
-
⚡ Pre-filled responses (where supported by provider)
-
🌐 Support for all available models across all providers
-
🔄 Change model mid-conversation
-
🔐 Sync chats and keys across devices, end-to-end encrypted. Self-hosted, or use our hosted instance.
-
🔗 Conversation sharing (if you choose to share, your conversation has to be stored on an external server for the share link to be made available. Self-hosted share options coming soon. No, I will not view any of your stuff.)
-
🌿 Branching conversation history (like the left-right ChatGPT arrows that you can click to go back to a previous response)
- Completely private and transparent. All your conversation history and keys are stored entirely locally, and kept only in your browser, on your device.
If you don't want to use tools, you don't need to install anything. A hosted instance is available at: https://llum.chat
If you want to use tools, proceed below.
The llum tool server is available prebuilt as a single binary. Download prebuilt package from the releases page.
Download the binary for your platform, then run it, which will start the tool server:
./llum-darwin-amd64
Tool server running at http://localhost:8081
Go back to https://llum.chat, head over to Settings -> Tool calling, and click the "Refresh tools" button. You should be good to go!
- Clone the repository
- Install and start the client:
npm i && npm run dev
. The client will be accessible at http://localhost:5173 - Install and start the server:
cd server && go generate ./... && go build && ./server -password foobar
. The server will be accessible at http://localhost:8081. You can plug this into the server address in the chat UI along with the password you selected.