Update model explanation for genai (#20186)

This commit is contained in:
Nicolas Mowen 2025-09-23 07:30:42 -06:00 committed by GitHub
parent 7f7eefef7f
commit 2bc92cce81
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194

View File

@ -27,13 +27,20 @@ Parallel requests also come with some caveats. You will need to set `OLLAMA_NUM_
You must use a vision capable model with Frigate. Current model variants can be found [in their model library](https://ollama.com/library). Note that Frigate will not automatically download the model you specify in your config, Ollama will try to download the model but it may take longer than the timeout, it is recommended to pull the model beforehand by running `ollama pull your_model` on your Ollama server/Docker container. Note that the model specified in Frigate's config must match the downloaded model tag. You must use a vision capable model with Frigate. Current model variants can be found [in their model library](https://ollama.com/library). Note that Frigate will not automatically download the model you specify in your config, Ollama will try to download the model but it may take longer than the timeout, it is recommended to pull the model beforehand by running `ollama pull your_model` on your Ollama server/Docker container. Note that the model specified in Frigate's config must match the downloaded model tag.
:::tip
Each model is available in multiple parameter sizes (3b, 4b, 8b, etc.). Larger sizes are more capable of complex tasks and understanding of situations, but requires more memory and computational resources. It is recommended to try multiple models and experiment to see which performs best.
:::
The following models are recommended: The following models are recommended:
| Model | Size | Notes | | Model | Notes |
| ----------------- | ------ | ----------------------------------------------------------- | | ----------------- | ----------------------------------------------------------- |
| `gemma3:4b` | 3.3 GB | Strong frame-to-frame understanding, slower inference times | | `Intern3.5VL` | Relatively fast with good vision comprehension
| `qwen2.5vl:3b` | 3.2 GB | Fast but capable model with good vision comprehension | | `gemma3` | Strong frame-to-frame understanding, slower inference times |
| `llava-phi3:3.8b` | 2.9 GB | Lightweight and fast model with vision comprehension | | `qwen2.5vl` | Fast but capable model with good vision comprehension |
| `llava-phi3` | Lightweight and fast model with vision comprehension |
:::note :::note
@ -124,4 +131,4 @@ genai:
provider: azure_openai provider: azure_openai
base_url: https://example-endpoint.openai.azure.com/openai/deployments/gpt-4o/chat/completions?api-version=2023-03-15-preview base_url: https://example-endpoint.openai.azure.com/openai/deployments/gpt-4o/chat/completions?api-version=2023-03-15-preview
api_key: "{FRIGATE_OPENAI_API_KEY}" api_key: "{FRIGATE_OPENAI_API_KEY}"
``` ```