mirror of
https://github.com/blakeblackshear/frigate.git
synced 2026-03-07 02:18:07 +01:00
* use react-jsonschema-form for UI config * don't use properties wrapper when generating config i18n json * configure for full i18n support * section fields * add descriptions to all fields for i18n * motion i18n * fix nullable fields * sanitize internal fields * add switches widgets and use friendly names * fix nullable schema entries * ensure update_topic is added to api calls this needs further backend implementation to work correctly * add global sections, camera config overrides, and reset button * i18n * add reset logic to global config view * tweaks * fix sections and live validation * fix validation for schema objects that can be null * generic and custom per-field validation * improve generic error validation messages * remove show advanced fields switch * tweaks * use shadcn theme * fix array field template * i18n tweaks * remove collapsible around root section * deep merge schema for advanced fields * add array field item template and fix ffmpeg section * add missing i18n keys * tweaks * comment out api call for testing * add config groups as a separate i18n namespace * add descriptions to all pydantic fields * make titles more concise * new titles as i18n * update i18n config generation script to use json schema * tweaks * tweaks * rebase * clean up * form tweaks * add wildcards and fix object filter fields * add field template for additionalproperties schema objects * improve typing * add section description from schema and clarify global vs camera level descriptions * separate and consolidate global and camera i18n namespaces * clean up now obsolete namespaces * tweaks * refactor sections and overrides * add ability to render components before and after fields * fix titles * chore(sections): remove legacy single-section components replaced by template * refactor configs to use individual files with a template * fix review description * apply hidden fields after ui schema * move util * remove unused i18n * clean up error messages * fix fast refresh * add custom validation and use it for ffmpeg input roles * update nav tree * remove unused * re-add override and modified indicators * mark pending changes and add confirmation dialog for resets * fix red unsaved dot * tweaks * add docs links, readonly keys, and restart required per field * add special case and comments for global motion section * add section form special cases * combine review sections * tweaks * add audio labels endpoint * add audio label switches and input to filter list * fix type * remove key from config when resetting to default/global * don't show description for new key/val fields * tweaks * spacing tweaks * add activity indicator and scrollbar tweaks * add docs to filter fields * wording changes * fix global ffmpeg section * add review classification zones to review form * add backend endpoint and frontend widget for ffmpeg presets and manual args * improve wording * hide descriptions for additional properties arrays * add warning log about incorrectly nested model config * spacing and language tweaks * fix i18n keys * networking section docs and description * small wording tweaks * add layout grid field * refactor with shared utilities * field order * add individual detectors to schema add detector titles and descriptions (docstrings in pydantic are used for descriptions) and add i18n keys to globals * clean up detectors section and i18n * don't save model config back to yaml when saving detectors * add full detectors config to api model dump works around the way we use detector plugins so we can have the full detector config for the frontend * add restart button to toast when restart is required * add ui option to remove inner cards * fix buttons * section tweaks * don't zoom into text on mobile * make buttons sticky at bottom of sections * small tweaks * highlight label of changed fields * add null to enum list when unwrapping * refactor to shared utils and add save all button * add undo all button * add RJSF to dictionary * consolidate utils * preserve form data when changing cameras * add mono fonts * add popover to show what fields will be saved * fix mobile menu not re-rendering with unsaved dots * tweaks * fix logger and env vars config section saving use escaped periods in keys to retain them in the config file (eg "frigate.embeddings") * add timezone widget * role map field with validation * fix validation for model section * add another hidden field * add footer message for required restart * use rjsf for notifications view * fix config saving * add replace rules field * default column layout and add field sizing * clean up field template * refactor profile settings to match rjsf forms * tweaks * refactor frigate+ view and make tweaks to sections * show frigate+ model info in detection model settings when using a frigate+ model * update restartRequired for all fields * fix restart fields * tweaks and add ability enable disabled cameras more backend changes required * require restart when enabling camera that is disabled in config * disable save when form is invalid * refactor ffmpeg section for readability * change label * clean up camera inputs fields * misc tweaks to ffmpeg section - add raw paths endpoint to ensure credentials get saved - restart required tooltip * maintenance settings tweaks * don't mutate with lodash * fix description re-rendering for nullable object fields * hide reindex field * update rjsf * add frigate+ description to settings pane * disable save all when any section is invalid * show translated field name in validation error pane * clean up * remove unused * fix genai merge * fix genai
235 lines
8.7 KiB
Python
235 lines
8.7 KiB
Python
import logging
|
|
|
|
import numpy as np
|
|
import openvino as ov
|
|
from pydantic import ConfigDict, Field
|
|
from typing_extensions import Literal
|
|
|
|
from frigate.detectors.detection_api import DetectionApi
|
|
from frigate.detectors.detection_runners import OpenVINOModelRunner
|
|
from frigate.detectors.detector_config import BaseDetectorConfig, ModelTypeEnum
|
|
from frigate.util.model import (
|
|
post_process_dfine,
|
|
post_process_rfdetr,
|
|
post_process_yolo,
|
|
)
|
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
DETECTOR_KEY = "openvino"
|
|
|
|
|
|
class OvDetectorConfig(BaseDetectorConfig):
|
|
"""OpenVINO detector for AMD and Intel CPUs, Intel GPUs and Intel VPU hardware."""
|
|
|
|
model_config = ConfigDict(
|
|
title="OpenVINO",
|
|
)
|
|
|
|
type: Literal[DETECTOR_KEY]
|
|
device: str = Field(
|
|
default=None,
|
|
title="Device Type",
|
|
description="The device to use for OpenVINO inference (e.g. 'CPU', 'GPU', 'NPU').",
|
|
)
|
|
|
|
|
|
class OvDetector(DetectionApi):
|
|
type_key = DETECTOR_KEY
|
|
supported_models = [
|
|
ModelTypeEnum.dfine,
|
|
ModelTypeEnum.rfdetr,
|
|
ModelTypeEnum.ssd,
|
|
ModelTypeEnum.yolonas,
|
|
ModelTypeEnum.yologeneric,
|
|
ModelTypeEnum.yolox,
|
|
]
|
|
|
|
def __init__(self, detector_config: OvDetectorConfig):
|
|
super().__init__(detector_config)
|
|
self.ov_model_type = detector_config.model.model_type
|
|
|
|
self.h = detector_config.model.height
|
|
self.w = detector_config.model.width
|
|
|
|
self.runner = OpenVINOModelRunner(
|
|
model_path=detector_config.model.path,
|
|
device=detector_config.device,
|
|
model_type=detector_config.model.model_type,
|
|
)
|
|
|
|
# For dfine models, also pre-allocate target sizes tensor
|
|
if self.ov_model_type == ModelTypeEnum.dfine:
|
|
self.target_sizes_tensor = ov.Tensor(
|
|
np.array([[self.h, self.w]], dtype=np.int64)
|
|
)
|
|
|
|
self.model_invalid = False
|
|
|
|
if self.ov_model_type not in self.supported_models:
|
|
logger.error(
|
|
f"OpenVino detector does not support {self.ov_model_type} models."
|
|
)
|
|
self.model_invalid = True
|
|
|
|
if self.ov_model_type == ModelTypeEnum.ssd:
|
|
model_inputs = self.runner.compiled_model.inputs
|
|
model_outputs = self.runner.compiled_model.outputs
|
|
|
|
if len(model_inputs) != 1:
|
|
logger.error(
|
|
f"SSD models must only have 1 input. Found {len(model_inputs)}."
|
|
)
|
|
self.model_invalid = True
|
|
if len(model_outputs) != 1:
|
|
logger.error(
|
|
f"SSD models must only have 1 output. Found {len(model_outputs)}."
|
|
)
|
|
self.model_invalid = True
|
|
|
|
output_shape = model_outputs[0].get_shape()
|
|
if output_shape[0] != 1 or output_shape[1] != 1 or output_shape[3] != 7:
|
|
logger.error(f"SSD model output doesn't match. Found {output_shape}.")
|
|
self.model_invalid = True
|
|
|
|
if self.ov_model_type == ModelTypeEnum.yolonas:
|
|
model_inputs = self.runner.compiled_model.inputs
|
|
model_outputs = self.runner.compiled_model.outputs
|
|
|
|
if len(model_inputs) != 1:
|
|
logger.error(
|
|
f"YoloNAS models must only have 1 input. Found {len(model_inputs)}."
|
|
)
|
|
self.model_invalid = True
|
|
if len(model_outputs) != 1:
|
|
logger.error(
|
|
f"YoloNAS models must be exported in flat format and only have 1 output. Found {len(model_outputs)}."
|
|
)
|
|
self.model_invalid = True
|
|
output_shape = model_outputs[0].partial_shape
|
|
if output_shape[-1] != 7:
|
|
logger.error(
|
|
f"YoloNAS models must be exported in flat format. Model output doesn't match. Found {output_shape}."
|
|
)
|
|
self.model_invalid = True
|
|
|
|
if self.ov_model_type == ModelTypeEnum.yolox:
|
|
self.output_indexes = 0
|
|
while True:
|
|
try:
|
|
tensor_shape = self.runner.compiled_model.output(
|
|
self.output_indexes
|
|
).shape
|
|
logger.info(
|
|
f"Model Output-{self.output_indexes} Shape: {tensor_shape}"
|
|
)
|
|
self.output_indexes += 1
|
|
except Exception:
|
|
logger.info(f"Model has {self.output_indexes} Output Tensors")
|
|
break
|
|
self.num_classes = tensor_shape[2] - 5
|
|
logger.info(f"YOLOX model has {self.num_classes} classes")
|
|
self.calculate_grids_strides()
|
|
|
|
## Takes in class ID, confidence score, and array of [x, y, w, h] that describes detection position,
|
|
## returns an array that's easily passable back to Frigate.
|
|
def process_yolo(self, class_id, conf, pos):
|
|
return [
|
|
class_id, # class ID
|
|
conf, # confidence score
|
|
(pos[1] - (pos[3] / 2)) / self.h, # y_min
|
|
(pos[0] - (pos[2] / 2)) / self.w, # x_min
|
|
(pos[1] + (pos[3] / 2)) / self.h, # y_max
|
|
(pos[0] + (pos[2] / 2)) / self.w, # x_max
|
|
]
|
|
|
|
def detect_raw(self, tensor_input):
|
|
if self.model_invalid:
|
|
return np.zeros((20, 6), np.float32)
|
|
|
|
if self.ov_model_type == ModelTypeEnum.dfine:
|
|
# Use named inputs for dfine models
|
|
inputs = {
|
|
"images": tensor_input,
|
|
"orig_target_sizes": np.array([[self.h, self.w]], dtype=np.int64),
|
|
}
|
|
outputs = self.runner.run(inputs)
|
|
tensor_output = (
|
|
outputs[0],
|
|
outputs[1],
|
|
outputs[2],
|
|
)
|
|
return post_process_dfine(tensor_output, self.w, self.h)
|
|
|
|
# Run inference using the runner
|
|
input_name = self.runner.get_input_names()[0]
|
|
outputs = self.runner.run({input_name: tensor_input})
|
|
|
|
detections = np.zeros((20, 6), np.float32)
|
|
|
|
if self.ov_model_type == ModelTypeEnum.rfdetr:
|
|
return post_process_rfdetr(outputs)
|
|
elif self.ov_model_type == ModelTypeEnum.ssd:
|
|
results = outputs[0][0][0]
|
|
|
|
for i, (_, class_id, score, xmin, ymin, xmax, ymax) in enumerate(results):
|
|
if i == 20:
|
|
break
|
|
detections[i] = [
|
|
class_id,
|
|
float(score),
|
|
ymin,
|
|
xmin,
|
|
ymax,
|
|
xmax,
|
|
]
|
|
return detections
|
|
elif self.ov_model_type == ModelTypeEnum.yolonas:
|
|
predictions = outputs[0]
|
|
|
|
for i, prediction in enumerate(predictions):
|
|
if i == 20:
|
|
break
|
|
(_, x_min, y_min, x_max, y_max, confidence, class_id) = prediction
|
|
# when running in GPU mode, empty predictions in the output have class_id of -1
|
|
if class_id < 0:
|
|
break
|
|
detections[i] = [
|
|
class_id,
|
|
confidence,
|
|
y_min / self.h,
|
|
x_min / self.w,
|
|
y_max / self.h,
|
|
x_max / self.w,
|
|
]
|
|
return detections
|
|
elif self.ov_model_type == ModelTypeEnum.yologeneric:
|
|
return post_process_yolo(outputs, self.w, self.h)
|
|
elif self.ov_model_type == ModelTypeEnum.yolox:
|
|
# [x, y, h, w, box_score, class_no_1, ..., class_no_80],
|
|
results = outputs[0]
|
|
results[..., :2] = (results[..., :2] + self.grids) * self.expanded_strides
|
|
results[..., 2:4] = np.exp(results[..., 2:4]) * self.expanded_strides
|
|
image_pred = results[0, ...]
|
|
|
|
class_conf = np.max(
|
|
image_pred[:, 5 : 5 + self.num_classes], axis=1, keepdims=True
|
|
)
|
|
class_pred = np.argmax(image_pred[:, 5 : 5 + self.num_classes], axis=1)
|
|
class_pred = np.expand_dims(class_pred, axis=1)
|
|
|
|
conf_mask = (image_pred[:, 4] * class_conf.squeeze() >= 0.3).squeeze()
|
|
# Detections ordered as (x1, y1, x2, y2, obj_conf, class_conf, class_pred)
|
|
detections = np.concatenate(
|
|
(image_pred[:, :5], class_conf, class_pred), axis=1
|
|
)
|
|
detections = detections[conf_mask]
|
|
|
|
ordered = detections[detections[:, 5].argsort()[::-1]][:20]
|
|
|
|
for i, object_detected in enumerate(ordered):
|
|
detections[i] = self.process_yolo(
|
|
object_detected[6], object_detected[5], object_detected[:4]
|
|
)
|
|
return detections
|