Leverage the power of multimodal LLMs to automate your smart home.
Endless possibilities for automations
Stay in the loop
Analyzes camera feeds and Frigate events to keep you updated.
Just ask
LLM Vision builds a detailed timeline of events around your home, so you know what happened and when. You can even ask it about a specific event.
Optimized
Lightweight local preprocessing automatically picks the right frames to analyze.
Supports most AI Providers
Integrate any AI provider you already use, or host your own LLM for maximum privacy.
Easy to use
Get started with LLM Vision with just a few simple lines of code.
How it works
LLM Vision is a Home Assistant integration that can analyze images, videos, live camera feeds and Frigate events using the vision capabilities of multimodal LLMs.
The integration exposes 4 services:
image_analyzer
Analyzes one or multiple images from different camera or image entities or image files.
video_analyzer
Extracts and analyzes frames from a video file or frigate event.
stream_analyzer
Captures and analyzes frames from camera entities.
data_analyzer
Analyzes and updates a Home Assistant entity with data extracted from an image based on your prompt.
Frames are fetched through Home Assistant's Camera API. When using video- or stream analyzer, frames are preprocessed locally and only frames with enough movement are sent to the provider. This ensures low latency as well as lower token usage.
Events
LLM Vision can remember events so you can ask about them later. A conversation agent such as Extended OpenAI Conversation is required.
Person seen
08:52 AM
White SUV seen
08:55 AM
Courier seen
10:11 AM
Data Analyzer
LLM Vision can extract data from images and update Home Assistant entities.
number
and
input_number
text
and
input_text
select
and
input_select
input_boolean
*LLM Vision is not endorsed by or associated with any of the listed providers. All trademarks are property of their respective owners.
Examples
LLM Vision is a Home Assistant integration that can analyze images, videos, live camera feeds and frigate events using the vision capabilities of multimodal LLMs.
You can install LLM Vision via HACS using the GitHub repository URL. For more details, visit the LLM Vision documentation.
While the integration itself is free, depending on your provider you'll pay per query. For more information on pricing, visit Provider Comparison.