You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
The stream_analyzer action in LLMVision fails when using Ollama as the provider. While the image_analyzer action works correctly with Ollama, the stream_analyzer does not function due to a limitation in the vision model used by Ollama. Specifically, the vision model only supports a single image per message, which causes the stream_analyzer to fail when it attempts to send multiple frames in a single request.
Reproduction steps
Set up LLMVision with Ollama as the provider, following the documentation.
Configure the stream_analyzer to analyze frames from a camera entity.
Trigger the stream_analyzer action.
Tried with multiple models
gemma3:12b
llama3.2-vision:11b
llava-phi3
Debug logs
Home Assistant
2025-04-01 12:02:32.110 DEBUG (MainThread) [custom_components.llmvision.memory] Memory([], [], 0)
2025-04-01 12:02:32.111 INFO (MainThread) [custom_components.llmvision.providers] Request data: {'model': 'llama3.2-vision:11b', 'messages': [{'role': 'user', 'content': [{'type': 'text', 'text': 'porch_fluent frame 1:'}, {'type': 'image_url', 'image_url': {'url': '<long_string>'}}, {'type': 'text', 'text': 'porch_fluent frame 0:'}, {'type': 'image_url', 'image_url': {'url': '<long_string>'}}, {'type': 'text', 'text': 'The attached images are frames from a live camera feed. Describe what you see'}]}], 'max_tokens': 100, 'temperature': 0.2}
2025-04-01 12:02:32.111 INFO (MainThread) [custom_components.llmvision.providers] Posting to http://192.168.x.x:3000/api/chat/completions
2025-04-01 12:02:32.241 INFO (MainThread) [custom_components.llmvision.providers] [INFO] Full Response: {"detail":"500: Ollama: 500, message='Internal Server Error', url='http://localhost:11434/api/chat'"}
2025-04-01 12:02:32.242 ERROR (MainThread) [homeassistant.helpers.script.websocket_api_script] websocket_api script: Error executing script. Error for call_service at pos 1: Unknown error
2025-04-01 12:02:32.245 ERROR (MainThread) [homeassistant.components.websocket_api.http.connection] [546287706048] Unknown error
Ollama OpenWebUI Docker
time=2025-04-01T01:25:52.536Z level=INFO source=server.go:624 msg="llama runner started in 21.82 seconds"
[GIN] 2025/04/01 - 01:25:52 | 500 | 24.39373541s | 127.0.0.1 | POST "/api/chat"
time=2025-04-01T01:25:52.536Z level=ERROR source=routes.go:1516 msg="chat prompt error" error="vision model only supports a single image per message"
2025-04-01 01:25:52.537 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 192.168.x.x:46240 - "POST /api/chat/completions HTTP/1.1" 400 - {}
The text was updated successfully, but these errors were encountered:
To me it looks like your actually using Open WebUI as provider, which then uses Ollama as the backend. Also your logs mention that "vision model only supports a single image per message". This could be a limitation with Open WebUI. Can you try using Ollama directly?
Uh oh!
There was an error while loading. Please reload this page.
Checklist
Describe the issue
The
stream_analyzer
action in LLMVision fails when using Ollama as the provider. While theimage_analyzer
action works correctly with Ollama, thestream_analyzer
does not function due to a limitation in the vision model used by Ollama. Specifically, the vision model only supports a single image per message, which causes thestream_analyzer
to fail when it attempts to send multiple frames in a single request.Reproduction steps
Tried with multiple models
Debug logs
The text was updated successfully, but these errors were encountered: