/ha-llmvision

Let Home Assistant see!

Primary LanguagePythonApache License 2.0Apache-2.0

Issues

Image and video analyzer for Home Assistant using multimodal LLMs

🌟 Features · 📖 Resources · ⬇️ Installation · 🚧 Roadmap · 🪲 How to report Bugs




LLM Vision is a Home Assistant integration to analyze images, videos and camera feeds using the vision capabilities of multimodal LLMs.
Supported providers are OpenAI, Anthropic, Google Gemini, LocalAI, Ollama and any OpenAI compatible API.

Features

  • Compatible with OpenAI, Anthropic Claude, Google Gemini, LocalAI, Ollama and custom OpenAI compatible APIs
  • Takes images and video from camera entities as input
  • Takes local image and video files as input
  • Images can be downscaled for faster processing

Resources

Check the docs for detailed instructions on how to set up LLM Vision and each of the supported providers, get inspiration from examples or join the discussion on the Home Assistant Community.

Installation

Open a repository inside the Home Assistant Community Store.

  1. Search for LLM Vision in Home Assistant Settings/Devices & services
  2. Select your provider
  3. Follow the instructions to add your AI providers.

Detailed instruction on how to set up LLM Vision and each of the supported providers are available here: https://llm-vision.gitbook.io/getting-started/

Debugging

To enable debugging, add the following to your configuration.yaml:

logger:
  logs:
    custom_components.llmvision: debug

Roadmap

Note

These are planned features and ideas. They are subject to change and may not be implemented in the order listed or at all.

  1. New Provider: NVIDIA ChatRTX
  2. HACS: Include in HACS default
  3. Animation Support: Support for animated GIFs
  4. New Provider: Custom (OpenAI API compatible) Providers
  5. Feature: HTTPS support for LocalAI and Ollama
  6. Feature: Support for video files
  7. Feature: Analyze Frigate Recordings using frigate's event_id

How to report a bug or request a feature

Important

Bugs: If you encounter any bugs and have followed the instructions carefully, feel free to file a bug report.
Feature Requests: If you have an idea for a feature, create a feature request.


Create new Issue 

Support

You can support this project by starring this GitHub repository. If you want, you can also buy me a coffee here: