mirror of
https://github.com/blakeblackshear/frigate.git
synced 2026-03-07 02:18:07 +01:00
* use react-jsonschema-form for UI config * don't use properties wrapper when generating config i18n json * configure for full i18n support * section fields * add descriptions to all fields for i18n * motion i18n * fix nullable fields * sanitize internal fields * add switches widgets and use friendly names * fix nullable schema entries * ensure update_topic is added to api calls this needs further backend implementation to work correctly * add global sections, camera config overrides, and reset button * i18n * add reset logic to global config view * tweaks * fix sections and live validation * fix validation for schema objects that can be null * generic and custom per-field validation * improve generic error validation messages * remove show advanced fields switch * tweaks * use shadcn theme * fix array field template * i18n tweaks * remove collapsible around root section * deep merge schema for advanced fields * add array field item template and fix ffmpeg section * add missing i18n keys * tweaks * comment out api call for testing * add config groups as a separate i18n namespace * add descriptions to all pydantic fields * make titles more concise * new titles as i18n * update i18n config generation script to use json schema * tweaks * tweaks * rebase * clean up * form tweaks * add wildcards and fix object filter fields * add field template for additionalproperties schema objects * improve typing * add section description from schema and clarify global vs camera level descriptions * separate and consolidate global and camera i18n namespaces * clean up now obsolete namespaces * tweaks * refactor sections and overrides * add ability to render components before and after fields * fix titles * chore(sections): remove legacy single-section components replaced by template * refactor configs to use individual files with a template * fix review description * apply hidden fields after ui schema * move util * remove unused i18n * clean up error messages * fix fast refresh * add custom validation and use it for ffmpeg input roles * update nav tree * remove unused * re-add override and modified indicators * mark pending changes and add confirmation dialog for resets * fix red unsaved dot * tweaks * add docs links, readonly keys, and restart required per field * add special case and comments for global motion section * add section form special cases * combine review sections * tweaks * add audio labels endpoint * add audio label switches and input to filter list * fix type * remove key from config when resetting to default/global * don't show description for new key/val fields * tweaks * spacing tweaks * add activity indicator and scrollbar tweaks * add docs to filter fields * wording changes * fix global ffmpeg section * add review classification zones to review form * add backend endpoint and frontend widget for ffmpeg presets and manual args * improve wording * hide descriptions for additional properties arrays * add warning log about incorrectly nested model config * spacing and language tweaks * fix i18n keys * networking section docs and description * small wording tweaks * add layout grid field * refactor with shared utilities * field order * add individual detectors to schema add detector titles and descriptions (docstrings in pydantic are used for descriptions) and add i18n keys to globals * clean up detectors section and i18n * don't save model config back to yaml when saving detectors * add full detectors config to api model dump works around the way we use detector plugins so we can have the full detector config for the frontend * add restart button to toast when restart is required * add ui option to remove inner cards * fix buttons * section tweaks * don't zoom into text on mobile * make buttons sticky at bottom of sections * small tweaks * highlight label of changed fields * add null to enum list when unwrapping * refactor to shared utils and add save all button * add undo all button * add RJSF to dictionary * consolidate utils * preserve form data when changing cameras * add mono fonts * add popover to show what fields will be saved * fix mobile menu not re-rendering with unsaved dots * tweaks * fix logger and env vars config section saving use escaped periods in keys to retain them in the config file (eg "frigate.embeddings") * add timezone widget * role map field with validation * fix validation for model section * add another hidden field * add footer message for required restart * use rjsf for notifications view * fix config saving * add replace rules field * default column layout and add field sizing * clean up field template * refactor profile settings to match rjsf forms * tweaks * refactor frigate+ view and make tweaks to sections * show frigate+ model info in detection model settings when using a frigate+ model * update restartRequired for all fields * fix restart fields * tweaks and add ability enable disabled cameras more backend changes required * require restart when enabling camera that is disabled in config * disable save when form is invalid * refactor ffmpeg section for readability * change label * clean up camera inputs fields * misc tweaks to ffmpeg section - add raw paths endpoint to ensure credentials get saved - restart required tooltip * maintenance settings tweaks * don't mutate with lodash * fix description re-rendering for nullable object fields * hide reindex field * update rjsf * add frigate+ description to settings pane * disable save all when any section is invalid * show translated field name in validation error pane * clean up * remove unused * fix genai merge * fix genai
158 lines
5.3 KiB
Python
158 lines
5.3 KiB
Python
import logging
|
|
import queue
|
|
|
|
import numpy as np
|
|
from pydantic import ConfigDict, Field
|
|
from typing_extensions import Literal
|
|
|
|
from frigate.detectors.detection_api import DetectionApi
|
|
from frigate.detectors.detector_config import BaseDetectorConfig
|
|
|
|
logger = logging.getLogger(__name__)
|
|
DETECTOR_KEY = "degirum"
|
|
|
|
|
|
### DETECTOR CONFIG ###
|
|
class DGDetectorConfig(BaseDetectorConfig):
|
|
"""DeGirum detector for running models via DeGirum cloud or local inference services."""
|
|
|
|
model_config = ConfigDict(
|
|
title="DeGirum",
|
|
)
|
|
|
|
type: Literal[DETECTOR_KEY]
|
|
location: str = Field(
|
|
default=None,
|
|
title="Inference Location",
|
|
description="Location of the DeGirim inference engine (e.g. '@cloud', '127.0.0.1').",
|
|
)
|
|
zoo: str = Field(
|
|
default=None,
|
|
title="Model Zoo",
|
|
description="Path or URL to the DeGirum model zoo.",
|
|
)
|
|
token: str = Field(
|
|
default=None,
|
|
title="DeGirum Cloud Token",
|
|
description="Token for DeGirum Cloud access.",
|
|
)
|
|
|
|
|
|
### ACTUAL DETECTOR ###
|
|
class DGDetector(DetectionApi):
|
|
type_key = DETECTOR_KEY
|
|
|
|
def __init__(self, detector_config: DGDetectorConfig):
|
|
try:
|
|
import degirum as dg
|
|
except ModuleNotFoundError:
|
|
raise ImportError("Unable to import DeGirum detector.")
|
|
|
|
self._queue = queue.Queue()
|
|
self._zoo = dg.connect(
|
|
detector_config.location, detector_config.zoo, detector_config.token
|
|
)
|
|
|
|
logger.debug(f"Models in zoo: {self._zoo.list_models()}")
|
|
|
|
self.dg_model = self._zoo.load_model(
|
|
detector_config.model.path,
|
|
)
|
|
|
|
# Setting input image format to raw reduces preprocessing time
|
|
self.dg_model.input_image_format = "RAW"
|
|
|
|
# Prioritize the most powerful hardware available
|
|
self.select_best_device_type()
|
|
# Frigate handles pre processing as long as these are all set
|
|
input_shape = self.dg_model.input_shape[0]
|
|
self.model_height = input_shape[1]
|
|
self.model_width = input_shape[2]
|
|
|
|
# Passing in dummy frame so initial connection latency happens in
|
|
# init function and not during actual prediction
|
|
frame = np.zeros(
|
|
(detector_config.model.width, detector_config.model.height, 3),
|
|
dtype=np.uint8,
|
|
)
|
|
# Pass in frame to overcome first frame latency
|
|
self.dg_model(frame)
|
|
self.prediction = self.prediction_generator()
|
|
|
|
def select_best_device_type(self):
|
|
"""
|
|
Helper function that selects fastest hardware available per model runtime
|
|
"""
|
|
types = self.dg_model.supported_device_types
|
|
|
|
device_map = {
|
|
"OPENVINO": ["GPU", "NPU", "CPU"],
|
|
"HAILORT": ["HAILO8L", "HAILO8"],
|
|
"N2X": ["ORCA1", "CPU"],
|
|
"ONNX": ["VITIS_NPU", "CPU"],
|
|
"RKNN": ["RK3566", "RK3568", "RK3588"],
|
|
"TENSORRT": ["DLA", "GPU", "DLA_ONLY"],
|
|
"TFLITE": ["ARMNN", "EDGETPU", "CPU"],
|
|
}
|
|
|
|
runtime = types[0].split("/")[0]
|
|
# Just create an array of format {runtime}/{hardware} for every hardware
|
|
# in the value for appropriate key in device_map
|
|
self.dg_model.device_type = [
|
|
f"{runtime}/{hardware}" for hardware in device_map[runtime]
|
|
]
|
|
|
|
def prediction_generator(self):
|
|
"""
|
|
Generator for all incoming frames. By using this generator, we don't have to keep
|
|
reconnecting our websocket on every "predict" call.
|
|
"""
|
|
logger.debug("Prediction generator was called")
|
|
with self.dg_model as model:
|
|
while 1:
|
|
logger.info(f"q size before calling get: {self._queue.qsize()}")
|
|
data = self._queue.get(block=True)
|
|
logger.info(f"q size after calling get: {self._queue.qsize()}")
|
|
logger.debug(
|
|
f"Data we're passing into model predict: {data}, shape of data: {data.shape}"
|
|
)
|
|
result = model.predict(data)
|
|
logger.debug(f"Prediction result: {result}")
|
|
yield result
|
|
|
|
def detect_raw(self, tensor_input):
|
|
# Reshaping tensor to work with pysdk
|
|
truncated_input = tensor_input.reshape(tensor_input.shape[1:])
|
|
logger.debug(f"Detect raw was called for tensor input: {tensor_input}")
|
|
|
|
# add tensor_input to input queue
|
|
self._queue.put(truncated_input)
|
|
logger.debug(f"Queue size after adding truncated input: {self._queue.qsize()}")
|
|
|
|
# define empty detection result
|
|
detections = np.zeros((20, 6), np.float32)
|
|
# grab prediction
|
|
res = next(self.prediction)
|
|
|
|
# If we have an empty prediction, return immediately
|
|
if len(res.results) == 0 or len(res.results[0]) == 0:
|
|
return detections
|
|
|
|
i = 0
|
|
for result in res.results:
|
|
if i >= 20:
|
|
break
|
|
|
|
detections[i] = [
|
|
result["category_id"],
|
|
float(result["score"]),
|
|
result["bbox"][1] / self.model_height,
|
|
result["bbox"][0] / self.model_width,
|
|
result["bbox"][3] / self.model_height,
|
|
result["bbox"][2] / self.model_width,
|
|
]
|
|
i += 1
|
|
|
|
logger.debug(f"Detections output: {detections}")
|
|
return detections
|