mirror of
https://github.com/blakeblackshear/frigate.git
synced 2025-05-17 01:16:06 +02:00
Make note of multi-model on GPU support (#13813)
* Make note of multi-gpu support * fix typo
This commit is contained in:
parent
e44a9e8921
commit
5e0d8fe4c7
@ -140,6 +140,22 @@ The OpenVINO device to be used is specified using the `"device"` attribute accor
|
|||||||
|
|
||||||
OpenVINO is supported on 6th Gen Intel platforms (Skylake) and newer. It will also run on AMD CPUs despite having no official support for it. A supported Intel platform is required to use the `GPU` device with OpenVINO. For detailed system requirements, see [OpenVINO System Requirements](https://docs.openvino.ai/2024/about-openvino/release-notes-openvino/system-requirements.html)
|
OpenVINO is supported on 6th Gen Intel platforms (Skylake) and newer. It will also run on AMD CPUs despite having no official support for it. A supported Intel platform is required to use the `GPU` device with OpenVINO. For detailed system requirements, see [OpenVINO System Requirements](https://docs.openvino.ai/2024/about-openvino/release-notes-openvino/system-requirements.html)
|
||||||
|
|
||||||
|
:::tip
|
||||||
|
|
||||||
|
When using many cameras one detector may not be enough to keep up. Multiple detectors can be defined assuming GPU resources are available. An example configuration would be:
|
||||||
|
|
||||||
|
```yaml
|
||||||
|
detectors:
|
||||||
|
ov_0:
|
||||||
|
type: openvino
|
||||||
|
device: GPU
|
||||||
|
ov_1:
|
||||||
|
type: openvino
|
||||||
|
device: GPU
|
||||||
|
```
|
||||||
|
|
||||||
|
:::
|
||||||
|
|
||||||
### Supported Models
|
### Supported Models
|
||||||
|
|
||||||
#### SSDLite MobileNet v2
|
#### SSDLite MobileNet v2
|
||||||
@ -298,7 +314,21 @@ model:
|
|||||||
|
|
||||||
## ONNX
|
## ONNX
|
||||||
|
|
||||||
ONNX is an open format for building machine learning models, these models can run on a wide variety of hardware. Frigate supports running ONNX models on CPU, OpenVINO, and TensorRT.
|
ONNX is an open format for building machine learning models, Frigate supports running ONNX models on CPU, OpenVINO, and TensorRT. On startup Frigate will automatically try to use a GPU if one is available.
|
||||||
|
|
||||||
|
:::tip
|
||||||
|
|
||||||
|
When using many cameras one detector may not be enough to keep up. Multiple detectors can be defined assuming GPU resources are available. An example configuration would be:
|
||||||
|
|
||||||
|
```yaml
|
||||||
|
detectors:
|
||||||
|
onnx_0:
|
||||||
|
type: onnx
|
||||||
|
onnx_1:
|
||||||
|
type: onnx
|
||||||
|
```
|
||||||
|
|
||||||
|
:::
|
||||||
|
|
||||||
### Supported Models
|
### Supported Models
|
||||||
|
|
||||||
|
Loading…
Reference in New Issue
Block a user