Image and video analyzer for Home Assistant using multimodal LLMs
🌟 Features · 📖 Resources · ⬇️ Installation · 🚧 Roadmap · 🪲 How to report Bugs · ☕ Support
LLM Vision is a Home Assistant integration to analyze images, videos, live camera feeds and frigate events using the vision capabilities of multimodal LLMs. Supported providers are OpenAI, Anthropic, Google Gemini, Groq, LocalAI, Ollama and any OpenAI compatible API.
- Compatible with OpenAI, Anthropic Claude, Google Gemini, Groq, LocalAI, Ollama and custom OpenAI compatible APIs
- Analyzes images and video files, live camera feeds and Frigate events
- Remembers Frigate events and camera motion events so you can ask about them later
- Seamlessly updates sensors based on image input
With the easy to use blueprint, you'll get important notifications intelligently summarized by AI from either Frigate or cameras in Home Assistant. LLM Vision can also remember events, so you can ask about them later. LLM Vision needs to be installed to use the blueprint.
Learn how to install the blueprint
Check the docs for detailed instructions on how to set up LLM Vision and each of the supported providers, get inspiration from examples or join the discussion on the Home Assistant Community.
- Search for
LLM Vision
in Home Assistant Settings/Devices & services - Select your provider
- Follow the instructions to add your AI providers.
Detailed instruction on how to set up LLM Vision and each of the supported providers are available here: https://llm-vision.gitbook.io/getting-started/
To enable debugging, add the following to your configuration.yaml
:
logger:
logs:
custom_components.llmvision: debug
Note
These are planned features and ideas. They are subject to change and may not be implemented in the order listed or at all.
- HACS: Include in HACS default repository
For features added in previous versions, check the changelogs in the release notes.
Important
Bugs: If you encounter any bugs and have followed the instructions carefully, feel free to file a bug report.
Feature Requests: If you have an idea for a feature, create a feature request.
You can support this project by starring this GitHub repository. If you want, you can also buy me a coffee here: