2019-02-26 03:27:02 +01:00
|
|
|
import datetime
|
2020-11-04 13:31:25 +01:00
|
|
|
import itertools
|
2020-11-04 04:26:39 +01:00
|
|
|
import logging
|
2019-03-30 02:49:27 +01:00
|
|
|
import multiprocessing as mp
|
2020-11-04 13:31:25 +01:00
|
|
|
import queue
|
2019-06-02 14:29:50 +02:00
|
|
|
import subprocess as sp
|
2020-11-29 23:19:59 +01:00
|
|
|
import signal
|
2020-11-04 13:31:25 +01:00
|
|
|
import threading
|
|
|
|
import time
|
2019-12-14 22:18:21 +01:00
|
|
|
from collections import defaultdict
|
2021-01-03 20:41:02 +01:00
|
|
|
from setproctitle import setproctitle
|
2020-11-04 13:31:25 +01:00
|
|
|
from typing import Dict, List
|
|
|
|
|
2021-02-17 14:23:32 +01:00
|
|
|
from cv2 import cv2
|
2020-11-04 13:31:25 +01:00
|
|
|
import numpy as np
|
|
|
|
|
2020-11-03 15:15:58 +01:00
|
|
|
from frigate.config import CameraConfig
|
2020-02-16 04:07:54 +01:00
|
|
|
from frigate.edgetpu import RemoteObjectDetector
|
2020-12-04 13:59:03 +01:00
|
|
|
from frigate.log import LogPipe
|
2020-02-16 04:07:54 +01:00
|
|
|
from frigate.motion import MotionDetector
|
2020-11-04 13:31:25 +01:00
|
|
|
from frigate.objects import ObjectTracker
|
2021-02-17 14:23:32 +01:00
|
|
|
from frigate.util import (
|
|
|
|
EventsPerSecond,
|
|
|
|
FrameManager,
|
|
|
|
SharedMemoryFrameManager,
|
|
|
|
calculate_region,
|
|
|
|
clipped,
|
|
|
|
listen,
|
|
|
|
yuv_region_2_rgb,
|
|
|
|
)
|
2019-02-26 03:27:02 +01:00
|
|
|
|
2020-11-04 04:26:39 +01:00
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
|
2021-02-17 14:23:32 +01:00
|
|
|
|
2021-01-15 14:52:28 +01:00
|
|
|
def filtered(obj, objects_to_track, object_filters):
|
2020-02-16 04:07:54 +01:00
|
|
|
object_name = obj[0]
|
|
|
|
|
|
|
|
if not object_name in objects_to_track:
|
|
|
|
return True
|
2021-02-17 14:23:32 +01:00
|
|
|
|
2020-02-16 04:07:54 +01:00
|
|
|
if object_name in object_filters:
|
|
|
|
obj_settings = object_filters[object_name]
|
|
|
|
|
|
|
|
# if the min area is larger than the
|
|
|
|
# detected object, don't add it to detected objects
|
2020-11-03 15:15:58 +01:00
|
|
|
if obj_settings.min_area > obj[3]:
|
2020-02-16 04:07:54 +01:00
|
|
|
return True
|
2021-02-17 14:23:32 +01:00
|
|
|
|
2020-02-16 04:07:54 +01:00
|
|
|
# if the detected object is larger than the
|
|
|
|
# max area, don't add it to detected objects
|
2020-11-03 15:15:58 +01:00
|
|
|
if obj_settings.max_area < obj[3]:
|
2020-02-16 04:07:54 +01:00
|
|
|
return True
|
|
|
|
|
2020-09-07 19:17:42 +02:00
|
|
|
# if the score is lower than the min_score, skip
|
2020-11-03 15:15:58 +01:00
|
|
|
if obj_settings.min_score > obj[1]:
|
2020-02-16 04:07:54 +01:00
|
|
|
return True
|
2021-02-17 14:23:32 +01:00
|
|
|
|
2021-01-15 14:52:28 +01:00
|
|
|
if not obj_settings.mask is None:
|
|
|
|
# compute the coordinates of the object and make sure
|
|
|
|
# the location isnt outside the bounds of the image (can happen from rounding)
|
2021-02-17 14:23:32 +01:00
|
|
|
y_location = min(int(obj[2][3]), len(obj_settings.mask) - 1)
|
|
|
|
x_location = min(
|
|
|
|
int((obj[2][2] - obj[2][0]) / 2.0) + obj[2][0],
|
|
|
|
len(obj_settings.mask[0]) - 1,
|
|
|
|
)
|
2021-01-15 14:52:28 +01:00
|
|
|
|
|
|
|
# if the object is in a masked location, don't add it to detected objects
|
|
|
|
if obj_settings.mask[y_location][x_location] == 0:
|
|
|
|
return True
|
2021-02-17 14:23:32 +01:00
|
|
|
|
2020-09-07 19:17:42 +02:00
|
|
|
return False
|
2019-12-15 14:25:40 +01:00
|
|
|
|
2021-02-17 14:23:32 +01:00
|
|
|
|
2020-12-09 14:18:53 +01:00
|
|
|
def create_tensor_input(frame, model_shape, region):
|
2020-10-11 04:28:12 +02:00
|
|
|
cropped_frame = yuv_region_2_rgb(frame, region)
|
2019-03-30 02:49:27 +01:00
|
|
|
|
2020-02-16 04:07:54 +01:00
|
|
|
# Resize to 300x300 if needed
|
2020-12-09 14:18:53 +01:00
|
|
|
if cropped_frame.shape != (model_shape[0], model_shape[1], 3):
|
2021-02-17 14:23:32 +01:00
|
|
|
cropped_frame = cv2.resize(
|
|
|
|
cropped_frame, dsize=model_shape, interpolation=cv2.INTER_LINEAR
|
|
|
|
)
|
|
|
|
|
2020-12-09 14:18:53 +01:00
|
|
|
# Expand dimensions since the model expects images to have shape: [1, height, width, 3]
|
2020-02-16 04:07:54 +01:00
|
|
|
return np.expand_dims(cropped_frame, axis=0)
|
|
|
|
|
2021-02-17 14:23:32 +01:00
|
|
|
|
2020-12-04 13:59:03 +01:00
|
|
|
def stop_ffmpeg(ffmpeg_process, logger):
|
2020-11-29 23:19:59 +01:00
|
|
|
logger.info("Terminating the existing ffmpeg process...")
|
|
|
|
ffmpeg_process.terminate()
|
|
|
|
try:
|
|
|
|
logger.info("Waiting for ffmpeg to exit gracefully...")
|
|
|
|
ffmpeg_process.communicate(timeout=30)
|
|
|
|
except sp.TimeoutExpired:
|
|
|
|
logger.info("FFmpeg didnt exit. Force killing...")
|
|
|
|
ffmpeg_process.kill()
|
|
|
|
ffmpeg_process.communicate()
|
|
|
|
ffmpeg_process = None
|
|
|
|
|
2021-02-17 14:23:32 +01:00
|
|
|
|
|
|
|
def start_or_restart_ffmpeg(
|
|
|
|
ffmpeg_cmd, logger, logpipe: LogPipe, frame_size=None, ffmpeg_process=None
|
|
|
|
):
|
Use dataclasses for config handling
Use config data classes to eliminate some of the boilerplate associated
with setting up the configuration. In particular, using dataclasses
removes a lot of the boilerplate around assigning properties to the
object and allows these to be easily immutable by freezing them. In the
case of simple, non-nested dataclasses, this also provides more
convenient `asdict` helpers.
To set this up, where previously the objects would be parsed from the
config via the `__init__` method, create a `build` classmethod that does
this and calls the dataclass initializer.
Some of the objects are mutated at runtime, in particular some of the
zones are mutated to set the color (this might be able to be refactored
out) and some of the camera functionality can be enabled/disabled. Some
of the configs with `enabled` properties don't seem to have mqtt hooks
to be able to toggle this, in particular, the clips, snapshots, and
detect can be toggled but rtmp and record configs do not, but all of
these configs are still not frozen in case there is some other
functionality I am missing.
There are a couple other minor fixes here, one that was introduced
by me recently where `max_seconds` was not defined, the other to
properly `get()` the message payload when handling publishing mqtt
messages sent via websocket.
2021-05-23 00:28:15 +02:00
|
|
|
if ffmpeg_process is not None:
|
2020-12-04 13:59:03 +01:00
|
|
|
stop_ffmpeg(ffmpeg_process, logger)
|
2020-02-27 02:02:12 +01:00
|
|
|
|
2020-11-29 22:55:53 +01:00
|
|
|
if frame_size is None:
|
2021-02-17 14:23:32 +01:00
|
|
|
process = sp.Popen(
|
|
|
|
ffmpeg_cmd,
|
|
|
|
stdout=sp.DEVNULL,
|
|
|
|
stderr=logpipe,
|
|
|
|
stdin=sp.DEVNULL,
|
|
|
|
start_new_session=True,
|
|
|
|
)
|
2020-11-29 22:55:53 +01:00
|
|
|
else:
|
2021-02-17 14:23:32 +01:00
|
|
|
process = sp.Popen(
|
|
|
|
ffmpeg_cmd,
|
|
|
|
stdout=sp.PIPE,
|
|
|
|
stderr=logpipe,
|
|
|
|
stdin=sp.DEVNULL,
|
|
|
|
bufsize=frame_size * 10,
|
|
|
|
start_new_session=True,
|
|
|
|
)
|
2020-03-10 03:12:19 +01:00
|
|
|
return process
|
2020-02-27 02:02:12 +01:00
|
|
|
|
2021-02-17 14:23:32 +01:00
|
|
|
|
|
|
|
def capture_frames(
|
|
|
|
ffmpeg_process,
|
|
|
|
camera_name,
|
|
|
|
frame_shape,
|
|
|
|
frame_manager: FrameManager,
|
|
|
|
frame_queue,
|
|
|
|
fps: mp.Value,
|
|
|
|
skipped_fps: mp.Value,
|
|
|
|
current_frame: mp.Value,
|
|
|
|
):
|
2020-08-22 14:05:20 +02:00
|
|
|
|
2020-11-03 15:15:58 +01:00
|
|
|
frame_size = frame_shape[0] * frame_shape[1]
|
2020-10-25 16:05:21 +01:00
|
|
|
frame_rate = EventsPerSecond()
|
2020-10-26 13:59:22 +01:00
|
|
|
frame_rate.start()
|
2020-10-25 16:05:21 +01:00
|
|
|
skipped_eps = EventsPerSecond()
|
|
|
|
skipped_eps.start()
|
2020-08-22 14:05:20 +02:00
|
|
|
while True:
|
2020-10-25 16:05:21 +01:00
|
|
|
fps.value = frame_rate.eps()
|
|
|
|
skipped_fps = skipped_eps.eps()
|
2020-08-22 14:05:20 +02:00
|
|
|
|
2020-09-07 19:17:42 +02:00
|
|
|
current_frame.value = datetime.datetime.now().timestamp()
|
2020-10-24 18:36:04 +02:00
|
|
|
frame_name = f"{camera_name}{current_frame.value}"
|
|
|
|
frame_buffer = frame_manager.create(frame_name, frame_size)
|
|
|
|
try:
|
2020-12-12 16:12:15 +01:00
|
|
|
frame_buffer[:] = ffmpeg_process.stdout.read(frame_size)
|
|
|
|
except Exception as e:
|
|
|
|
logger.info(f"{camera_name}: ffmpeg sent a broken frame. {e}")
|
|
|
|
|
|
|
|
if ffmpeg_process.poll() != None:
|
2021-02-17 14:23:32 +01:00
|
|
|
logger.info(
|
|
|
|
f"{camera_name}: ffmpeg process is not running. exiting capture thread..."
|
|
|
|
)
|
2020-12-12 16:12:15 +01:00
|
|
|
frame_manager.delete(frame_name)
|
|
|
|
break
|
|
|
|
continue
|
2020-08-22 14:05:20 +02:00
|
|
|
|
2020-10-25 16:05:21 +01:00
|
|
|
frame_rate.update()
|
2020-08-22 14:05:20 +02:00
|
|
|
|
2020-10-10 13:57:43 +02:00
|
|
|
# if the queue is full, skip this frame
|
|
|
|
if frame_queue.full():
|
2020-10-25 16:05:21 +01:00
|
|
|
skipped_eps.update()
|
2020-10-24 18:36:04 +02:00
|
|
|
frame_manager.delete(frame_name)
|
2020-08-22 14:05:20 +02:00
|
|
|
continue
|
|
|
|
|
2020-10-24 18:36:04 +02:00
|
|
|
# close the frame
|
|
|
|
frame_manager.close(frame_name)
|
2020-09-22 04:02:00 +02:00
|
|
|
|
2020-08-22 14:05:20 +02:00
|
|
|
# add to the queue
|
2020-09-07 19:17:42 +02:00
|
|
|
frame_queue.put(current_frame.value)
|
2020-08-22 14:05:20 +02:00
|
|
|
|
2021-02-17 14:23:32 +01:00
|
|
|
|
2020-10-25 16:05:21 +01:00
|
|
|
class CameraWatchdog(threading.Thread):
|
2021-02-17 14:23:32 +01:00
|
|
|
def __init__(
|
|
|
|
self, camera_name, config, frame_queue, camera_fps, ffmpeg_pid, stop_event
|
|
|
|
):
|
2020-10-25 16:05:21 +01:00
|
|
|
threading.Thread.__init__(self)
|
2020-12-04 13:59:03 +01:00
|
|
|
self.logger = logging.getLogger(f"watchdog.{camera_name}")
|
2020-11-04 13:28:07 +01:00
|
|
|
self.camera_name = camera_name
|
2020-10-25 16:05:21 +01:00
|
|
|
self.config = config
|
|
|
|
self.capture_thread = None
|
2020-11-29 22:55:53 +01:00
|
|
|
self.ffmpeg_detect_process = None
|
2020-12-04 13:59:03 +01:00
|
|
|
self.logpipe = LogPipe(f"ffmpeg.{self.camera_name}.detect", logging.ERROR)
|
2020-11-29 22:55:53 +01:00
|
|
|
self.ffmpeg_other_processes = []
|
2020-10-25 16:05:21 +01:00
|
|
|
self.camera_fps = camera_fps
|
2020-10-26 13:59:05 +01:00
|
|
|
self.ffmpeg_pid = ffmpeg_pid
|
2020-10-25 16:05:21 +01:00
|
|
|
self.frame_queue = frame_queue
|
2020-11-03 15:15:58 +01:00
|
|
|
self.frame_shape = self.config.frame_shape_yuv
|
|
|
|
self.frame_size = self.frame_shape[0] * self.frame_shape[1]
|
2020-11-29 23:19:59 +01:00
|
|
|
self.stop_event = stop_event
|
2020-10-25 16:05:21 +01:00
|
|
|
|
|
|
|
def run(self):
|
2020-11-29 22:55:53 +01:00
|
|
|
self.start_ffmpeg_detect()
|
|
|
|
|
|
|
|
for c in self.config.ffmpeg_cmds:
|
2021-02-17 14:23:32 +01:00
|
|
|
if "detect" in c["roles"]:
|
2020-11-29 22:55:53 +01:00
|
|
|
continue
|
2021-02-17 14:23:32 +01:00
|
|
|
logpipe = LogPipe(
|
|
|
|
f"ffmpeg.{self.camera_name}.{'_'.join(sorted(c['roles']))}",
|
|
|
|
logging.ERROR,
|
|
|
|
)
|
|
|
|
self.ffmpeg_other_processes.append(
|
|
|
|
{
|
|
|
|
"cmd": c["cmd"],
|
|
|
|
"logpipe": logpipe,
|
|
|
|
"process": start_or_restart_ffmpeg(c["cmd"], self.logger, logpipe),
|
|
|
|
}
|
|
|
|
)
|
|
|
|
|
2020-10-25 16:05:21 +01:00
|
|
|
time.sleep(10)
|
2021-05-21 17:39:14 +02:00
|
|
|
while not self.stop_event.wait(10):
|
2020-10-25 16:05:21 +01:00
|
|
|
now = datetime.datetime.now().timestamp()
|
|
|
|
|
|
|
|
if not self.capture_thread.is_alive():
|
2021-01-30 14:50:17 +01:00
|
|
|
self.logpipe.dump()
|
2020-11-29 22:55:53 +01:00
|
|
|
self.start_ffmpeg_detect()
|
|
|
|
elif now - self.capture_thread.current_frame.value > 20:
|
2021-02-17 14:23:32 +01:00
|
|
|
self.logger.info(
|
|
|
|
f"No frames received from {self.camera_name} in 20 seconds. Exiting ffmpeg..."
|
|
|
|
)
|
2020-11-29 22:55:53 +01:00
|
|
|
self.ffmpeg_detect_process.terminate()
|
2020-10-25 16:05:21 +01:00
|
|
|
try:
|
2020-12-04 13:59:03 +01:00
|
|
|
self.logger.info("Waiting for ffmpeg to exit gracefully...")
|
2020-11-29 22:55:53 +01:00
|
|
|
self.ffmpeg_detect_process.communicate(timeout=30)
|
2020-10-25 16:05:21 +01:00
|
|
|
except sp.TimeoutExpired:
|
2020-12-04 13:59:03 +01:00
|
|
|
self.logger.info("FFmpeg didnt exit. Force killing...")
|
2020-11-29 22:55:53 +01:00
|
|
|
self.ffmpeg_detect_process.kill()
|
|
|
|
self.ffmpeg_detect_process.communicate()
|
2021-02-17 14:23:32 +01:00
|
|
|
|
2020-11-29 22:55:53 +01:00
|
|
|
for p in self.ffmpeg_other_processes:
|
2021-02-17 14:23:32 +01:00
|
|
|
poll = p["process"].poll()
|
2021-06-25 18:37:21 +02:00
|
|
|
if poll is None:
|
2020-11-29 22:55:53 +01:00
|
|
|
continue
|
2021-02-17 14:23:32 +01:00
|
|
|
p["logpipe"].dump()
|
|
|
|
p["process"] = start_or_restart_ffmpeg(
|
|
|
|
p["cmd"], self.logger, p["logpipe"], ffmpeg_process=p["process"]
|
|
|
|
)
|
|
|
|
|
2021-05-21 17:39:14 +02:00
|
|
|
stop_ffmpeg(self.ffmpeg_detect_process, self.logger)
|
|
|
|
for p in self.ffmpeg_other_processes:
|
|
|
|
stop_ffmpeg(p["process"], self.logger)
|
|
|
|
p["logpipe"].close()
|
|
|
|
self.logpipe.close()
|
2021-02-17 14:23:32 +01:00
|
|
|
|
2020-11-29 22:55:53 +01:00
|
|
|
def start_ffmpeg_detect(self):
|
2021-02-17 14:23:32 +01:00
|
|
|
ffmpeg_cmd = [
|
|
|
|
c["cmd"] for c in self.config.ffmpeg_cmds if "detect" in c["roles"]
|
|
|
|
][0]
|
|
|
|
self.ffmpeg_detect_process = start_or_restart_ffmpeg(
|
|
|
|
ffmpeg_cmd, self.logger, self.logpipe, self.frame_size
|
|
|
|
)
|
2020-11-29 22:55:53 +01:00
|
|
|
self.ffmpeg_pid.value = self.ffmpeg_detect_process.pid
|
2021-02-17 14:23:32 +01:00
|
|
|
self.capture_thread = CameraCapture(
|
|
|
|
self.camera_name,
|
|
|
|
self.ffmpeg_detect_process,
|
|
|
|
self.frame_shape,
|
|
|
|
self.frame_queue,
|
|
|
|
self.camera_fps,
|
|
|
|
)
|
2020-11-01 17:55:11 +01:00
|
|
|
self.capture_thread.start()
|
2020-10-25 16:05:21 +01:00
|
|
|
|
2021-02-17 14:23:32 +01:00
|
|
|
|
2020-03-14 21:32:51 +01:00
|
|
|
class CameraCapture(threading.Thread):
|
2020-11-04 13:28:07 +01:00
|
|
|
def __init__(self, camera_name, ffmpeg_process, frame_shape, frame_queue, fps):
|
2020-03-14 21:32:51 +01:00
|
|
|
threading.Thread.__init__(self)
|
2020-11-04 13:28:07 +01:00
|
|
|
self.name = f"capture:{camera_name}"
|
|
|
|
self.camera_name = camera_name
|
2020-03-14 21:32:51 +01:00
|
|
|
self.frame_shape = frame_shape
|
|
|
|
self.frame_queue = frame_queue
|
|
|
|
self.fps = fps
|
2020-04-19 17:07:27 +02:00
|
|
|
self.skipped_fps = EventsPerSecond()
|
2020-09-22 04:02:00 +02:00
|
|
|
self.frame_manager = SharedMemoryFrameManager()
|
2020-03-14 21:32:51 +01:00
|
|
|
self.ffmpeg_process = ffmpeg_process
|
2021-02-17 14:23:32 +01:00
|
|
|
self.current_frame = mp.Value("d", 0.0)
|
2020-04-19 17:07:27 +02:00
|
|
|
self.last_frame = 0
|
2020-03-14 21:32:51 +01:00
|
|
|
|
|
|
|
def run(self):
|
2020-04-19 17:07:27 +02:00
|
|
|
self.skipped_fps.start()
|
2021-02-17 14:23:32 +01:00
|
|
|
capture_frames(
|
|
|
|
self.ffmpeg_process,
|
|
|
|
self.camera_name,
|
|
|
|
self.frame_shape,
|
|
|
|
self.frame_manager,
|
|
|
|
self.frame_queue,
|
|
|
|
self.fps,
|
|
|
|
self.skipped_fps,
|
|
|
|
self.current_frame,
|
|
|
|
)
|
|
|
|
|
2020-03-14 21:32:51 +01:00
|
|
|
|
2020-11-03 15:15:58 +01:00
|
|
|
def capture_camera(name, config: CameraConfig, process_info):
|
2020-11-29 23:19:59 +01:00
|
|
|
stop_event = mp.Event()
|
2021-02-17 14:23:32 +01:00
|
|
|
|
2020-11-29 23:19:59 +01:00
|
|
|
def receiveSignal(signalNumber, frame):
|
|
|
|
stop_event.set()
|
2021-02-17 14:23:32 +01:00
|
|
|
|
2020-11-29 23:19:59 +01:00
|
|
|
signal.signal(signal.SIGTERM, receiveSignal)
|
|
|
|
signal.signal(signal.SIGINT, receiveSignal)
|
|
|
|
|
2021-02-17 14:23:32 +01:00
|
|
|
frame_queue = process_info["frame_queue"]
|
|
|
|
camera_watchdog = CameraWatchdog(
|
|
|
|
name,
|
|
|
|
config,
|
|
|
|
frame_queue,
|
|
|
|
process_info["camera_fps"],
|
|
|
|
process_info["ffmpeg_pid"],
|
|
|
|
stop_event,
|
|
|
|
)
|
2020-10-25 16:05:21 +01:00
|
|
|
camera_watchdog.start()
|
|
|
|
camera_watchdog.join()
|
|
|
|
|
2021-02-17 14:23:32 +01:00
|
|
|
|
|
|
|
def track_camera(
|
|
|
|
name,
|
|
|
|
config: CameraConfig,
|
|
|
|
model_shape,
|
2021-07-08 05:57:19 +02:00
|
|
|
labelmap,
|
2021-02-17 14:23:32 +01:00
|
|
|
detection_queue,
|
|
|
|
result_connection,
|
|
|
|
detected_objects_queue,
|
|
|
|
process_info,
|
|
|
|
):
|
2020-11-29 23:19:59 +01:00
|
|
|
stop_event = mp.Event()
|
2021-02-17 14:23:32 +01:00
|
|
|
|
2020-11-29 23:19:59 +01:00
|
|
|
def receiveSignal(signalNumber, frame):
|
|
|
|
stop_event.set()
|
2021-02-17 14:23:32 +01:00
|
|
|
|
2020-11-29 23:19:59 +01:00
|
|
|
signal.signal(signal.SIGTERM, receiveSignal)
|
|
|
|
signal.signal(signal.SIGINT, receiveSignal)
|
|
|
|
|
2020-11-04 13:28:07 +01:00
|
|
|
threading.current_thread().name = f"process:{name}"
|
2021-01-03 20:41:02 +01:00
|
|
|
setproctitle(f"frigate.process:{name}")
|
2020-03-10 03:12:19 +01:00
|
|
|
listen()
|
2020-02-16 04:07:54 +01:00
|
|
|
|
2021-02-17 14:23:32 +01:00
|
|
|
frame_queue = process_info["frame_queue"]
|
|
|
|
detection_enabled = process_info["detection_enabled"]
|
2020-10-25 16:05:21 +01:00
|
|
|
|
2020-11-03 15:15:58 +01:00
|
|
|
frame_shape = config.frame_shape
|
|
|
|
objects_to_track = config.objects.track
|
|
|
|
object_filters = config.objects.filters
|
2020-02-16 04:07:54 +01:00
|
|
|
|
2021-01-15 14:52:28 +01:00
|
|
|
motion_detector = MotionDetector(frame_shape, config.motion)
|
2021-02-17 14:23:32 +01:00
|
|
|
object_detector = RemoteObjectDetector(
|
2021-07-08 05:57:19 +02:00
|
|
|
name, labelmap, detection_queue, result_connection, model_shape
|
2021-02-17 14:23:32 +01:00
|
|
|
)
|
2020-02-16 04:07:54 +01:00
|
|
|
|
2020-12-19 06:00:13 +01:00
|
|
|
object_tracker = ObjectTracker(config.detect)
|
2020-03-14 21:32:51 +01:00
|
|
|
|
2020-09-22 04:02:00 +02:00
|
|
|
frame_manager = SharedMemoryFrameManager()
|
2020-08-22 14:05:20 +02:00
|
|
|
|
2021-02-17 14:23:32 +01:00
|
|
|
process_frames(
|
|
|
|
name,
|
|
|
|
frame_queue,
|
|
|
|
frame_shape,
|
|
|
|
model_shape,
|
|
|
|
frame_manager,
|
|
|
|
motion_detector,
|
|
|
|
object_detector,
|
|
|
|
object_tracker,
|
|
|
|
detected_objects_queue,
|
|
|
|
process_info,
|
|
|
|
objects_to_track,
|
|
|
|
object_filters,
|
|
|
|
detection_enabled,
|
|
|
|
stop_event,
|
|
|
|
)
|
2020-08-22 14:05:20 +02:00
|
|
|
|
2020-11-04 04:26:39 +01:00
|
|
|
logger.info(f"{name}: exiting subprocess")
|
2020-08-22 14:05:20 +02:00
|
|
|
|
2021-02-17 14:23:32 +01:00
|
|
|
|
2020-08-22 14:05:20 +02:00
|
|
|
def reduce_boxes(boxes):
|
|
|
|
if len(boxes) == 0:
|
|
|
|
return []
|
2021-02-17 14:23:32 +01:00
|
|
|
reduced_boxes = cv2.groupRectangles(
|
|
|
|
[list(b) for b in itertools.chain(boxes, boxes)], 1, 0.2
|
|
|
|
)[0]
|
2020-08-22 14:05:20 +02:00
|
|
|
return [tuple(b) for b in reduced_boxes]
|
|
|
|
|
2021-02-17 14:23:32 +01:00
|
|
|
|
2021-02-03 12:58:33 +01:00
|
|
|
# modified from https://stackoverflow.com/a/40795835
|
|
|
|
def intersects_any(box_a, boxes):
|
|
|
|
for box in boxes:
|
2021-02-17 14:23:32 +01:00
|
|
|
if (
|
|
|
|
box_a[2] < box[0]
|
|
|
|
or box_a[0] > box[2]
|
|
|
|
or box_a[1] > box[3]
|
|
|
|
or box_a[3] < box[1]
|
|
|
|
):
|
2021-02-03 12:58:33 +01:00
|
|
|
continue
|
|
|
|
return True
|
|
|
|
|
2021-02-17 14:23:32 +01:00
|
|
|
|
|
|
|
def detect(
|
|
|
|
object_detector, frame, model_shape, region, objects_to_track, object_filters
|
|
|
|
):
|
2020-12-09 14:18:53 +01:00
|
|
|
tensor_input = create_tensor_input(frame, model_shape, region)
|
2020-08-22 14:05:20 +02:00
|
|
|
|
|
|
|
detections = []
|
|
|
|
region_detections = object_detector.detect(tensor_input)
|
|
|
|
for d in region_detections:
|
|
|
|
box = d[2]
|
2021-02-17 14:23:32 +01:00
|
|
|
size = region[2] - region[0]
|
2020-08-22 14:05:20 +02:00
|
|
|
x_min = int((box[1] * size) + region[0])
|
|
|
|
y_min = int((box[0] * size) + region[1])
|
|
|
|
x_max = int((box[3] * size) + region[0])
|
|
|
|
y_max = int((box[2] * size) + region[1])
|
2021-02-17 14:23:32 +01:00
|
|
|
det = (
|
|
|
|
d[0],
|
2020-08-22 14:05:20 +02:00
|
|
|
d[1],
|
|
|
|
(x_min, y_min, x_max, y_max),
|
2021-02-17 14:23:32 +01:00
|
|
|
(x_max - x_min) * (y_max - y_min),
|
|
|
|
region,
|
|
|
|
)
|
2020-08-22 14:05:20 +02:00
|
|
|
# apply object filters
|
2021-01-15 14:52:28 +01:00
|
|
|
if filtered(det, objects_to_track, object_filters):
|
2020-08-22 14:05:20 +02:00
|
|
|
continue
|
|
|
|
detections.append(det)
|
|
|
|
return detections
|
|
|
|
|
2021-02-17 14:23:32 +01:00
|
|
|
|
|
|
|
def process_frames(
|
|
|
|
camera_name: str,
|
|
|
|
frame_queue: mp.Queue,
|
|
|
|
frame_shape,
|
|
|
|
model_shape,
|
|
|
|
frame_manager: FrameManager,
|
|
|
|
motion_detector: MotionDetector,
|
|
|
|
object_detector: RemoteObjectDetector,
|
|
|
|
object_tracker: ObjectTracker,
|
|
|
|
detected_objects_queue: mp.Queue,
|
|
|
|
process_info: Dict,
|
|
|
|
objects_to_track: List[str],
|
|
|
|
object_filters,
|
|
|
|
detection_enabled: mp.Value,
|
|
|
|
stop_event,
|
|
|
|
exit_on_empty: bool = False,
|
|
|
|
):
|
|
|
|
|
|
|
|
fps = process_info["process_fps"]
|
|
|
|
detection_fps = process_info["detection_fps"]
|
|
|
|
current_frame_time = process_info["detection_frame"]
|
2020-10-25 16:05:21 +01:00
|
|
|
|
2020-02-16 04:07:54 +01:00
|
|
|
fps_tracker = EventsPerSecond()
|
|
|
|
fps_tracker.start()
|
2020-08-22 14:05:20 +02:00
|
|
|
|
2021-05-21 17:39:14 +02:00
|
|
|
while not stop_event.is_set():
|
2020-11-01 17:37:17 +01:00
|
|
|
if exit_on_empty and frame_queue.empty():
|
2020-11-04 04:26:39 +01:00
|
|
|
logger.info(f"Exiting track_objects...")
|
2020-11-01 17:37:17 +01:00
|
|
|
break
|
2020-03-14 21:32:51 +01:00
|
|
|
|
2020-08-22 14:05:20 +02:00
|
|
|
try:
|
|
|
|
frame_time = frame_queue.get(True, 10)
|
|
|
|
except queue.Empty:
|
2020-03-14 21:32:51 +01:00
|
|
|
continue
|
2020-04-19 17:07:27 +02:00
|
|
|
|
2020-08-22 14:05:20 +02:00
|
|
|
current_frame_time.value = frame_time
|
|
|
|
|
2021-02-17 14:23:32 +01:00
|
|
|
frame = frame_manager.get(
|
|
|
|
f"{camera_name}{frame_time}", (frame_shape[0] * 3 // 2, frame_shape[1])
|
|
|
|
)
|
2020-09-14 14:40:26 +02:00
|
|
|
|
|
|
|
if frame is None:
|
2020-11-04 04:26:39 +01:00
|
|
|
logger.info(f"{camera_name}: frame {frame_time} is not in memory store.")
|
2020-09-14 14:40:26 +02:00
|
|
|
continue
|
2020-08-22 14:05:20 +02:00
|
|
|
|
2021-01-16 03:52:59 +01:00
|
|
|
if not detection_enabled.value:
|
|
|
|
fps.value = fps_tracker.eps()
|
|
|
|
object_tracker.match_and_update(frame_time, [])
|
2021-02-17 14:23:32 +01:00
|
|
|
detected_objects_queue.put(
|
|
|
|
(camera_name, frame_time, object_tracker.tracked_objects, [], [])
|
|
|
|
)
|
2021-01-16 03:52:59 +01:00
|
|
|
detection_fps.value = object_detector.fps.eps()
|
|
|
|
frame_manager.close(f"{camera_name}{frame_time}")
|
|
|
|
continue
|
|
|
|
|
2020-02-16 04:07:54 +01:00
|
|
|
# look for motion
|
|
|
|
motion_boxes = motion_detector.detect(frame)
|
|
|
|
|
2021-02-03 12:58:33 +01:00
|
|
|
# only get the tracked object boxes that intersect with motion
|
2021-02-17 14:23:32 +01:00
|
|
|
tracked_object_boxes = [
|
|
|
|
obj["box"]
|
|
|
|
for obj in object_tracker.tracked_objects.values()
|
|
|
|
if intersects_any(obj["box"], motion_boxes)
|
|
|
|
]
|
2020-08-22 14:05:20 +02:00
|
|
|
|
|
|
|
# combine motion boxes with known locations of existing objects
|
|
|
|
combined_boxes = reduce_boxes(motion_boxes + tracked_object_boxes)
|
|
|
|
|
|
|
|
# compute regions
|
2021-02-17 14:23:32 +01:00
|
|
|
regions = [
|
|
|
|
calculate_region(frame_shape, a[0], a[1], a[2], a[3], 1.2)
|
|
|
|
for a in combined_boxes
|
|
|
|
]
|
2020-08-22 14:05:20 +02:00
|
|
|
|
|
|
|
# combine overlapping regions
|
|
|
|
combined_regions = reduce_boxes(regions)
|
2020-02-16 04:07:54 +01:00
|
|
|
|
2020-08-22 14:05:20 +02:00
|
|
|
# re-compute regions
|
2021-02-17 14:23:32 +01:00
|
|
|
regions = [
|
|
|
|
calculate_region(frame_shape, a[0], a[1], a[2], a[3], 1.0)
|
|
|
|
for a in combined_regions
|
|
|
|
]
|
2020-10-10 17:07:14 +02:00
|
|
|
|
2020-02-16 04:07:54 +01:00
|
|
|
# resize regions and detect
|
|
|
|
detections = []
|
2020-08-22 14:05:20 +02:00
|
|
|
for region in regions:
|
2021-02-17 14:23:32 +01:00
|
|
|
detections.extend(
|
|
|
|
detect(
|
|
|
|
object_detector,
|
|
|
|
frame,
|
|
|
|
model_shape,
|
|
|
|
region,
|
|
|
|
objects_to_track,
|
|
|
|
object_filters,
|
|
|
|
)
|
|
|
|
)
|
|
|
|
|
2020-02-16 04:07:54 +01:00
|
|
|
#########
|
2020-08-22 14:05:20 +02:00
|
|
|
# merge objects, check for clipped objects and look again up to 4 times
|
2020-02-16 04:07:54 +01:00
|
|
|
#########
|
|
|
|
refining = True
|
|
|
|
refine_count = 0
|
|
|
|
while refining and refine_count < 4:
|
|
|
|
refining = False
|
|
|
|
|
|
|
|
# group by name
|
|
|
|
detected_object_groups = defaultdict(lambda: [])
|
|
|
|
for detection in detections:
|
|
|
|
detected_object_groups[detection[0]].append(detection)
|
|
|
|
|
|
|
|
selected_objects = []
|
|
|
|
for group in detected_object_groups.values():
|
|
|
|
|
|
|
|
# apply non-maxima suppression to suppress weak, overlapping bounding boxes
|
2021-02-17 14:23:32 +01:00
|
|
|
boxes = [
|
|
|
|
(o[2][0], o[2][1], o[2][2] - o[2][0], o[2][3] - o[2][1])
|
|
|
|
for o in group
|
|
|
|
]
|
2020-02-16 04:07:54 +01:00
|
|
|
confidences = [o[1] for o in group]
|
|
|
|
idxs = cv2.dnn.NMSBoxes(boxes, confidences, 0.5, 0.4)
|
|
|
|
|
|
|
|
for index in idxs:
|
|
|
|
obj = group[index[0]]
|
2020-03-14 21:32:51 +01:00
|
|
|
if clipped(obj, frame_shape):
|
2020-02-16 04:07:54 +01:00
|
|
|
box = obj[2]
|
|
|
|
# calculate a new region that will hopefully get the entire object
|
2021-02-17 14:23:32 +01:00
|
|
|
region = calculate_region(
|
|
|
|
frame_shape, box[0], box[1], box[2], box[3]
|
|
|
|
)
|
2020-12-19 15:22:31 +01:00
|
|
|
|
|
|
|
regions.append(region)
|
2021-02-17 14:23:32 +01:00
|
|
|
|
|
|
|
selected_objects.extend(
|
|
|
|
detect(
|
|
|
|
object_detector,
|
|
|
|
frame,
|
|
|
|
model_shape,
|
|
|
|
region,
|
|
|
|
objects_to_track,
|
|
|
|
object_filters,
|
|
|
|
)
|
|
|
|
)
|
2020-02-16 04:07:54 +01:00
|
|
|
|
|
|
|
refining = True
|
|
|
|
else:
|
2021-02-17 14:23:32 +01:00
|
|
|
selected_objects.append(obj)
|
2020-02-16 04:07:54 +01:00
|
|
|
# set the detections list to only include top, complete objects
|
|
|
|
# and new detections
|
|
|
|
detections = selected_objects
|
|
|
|
|
|
|
|
if refining:
|
|
|
|
refine_count += 1
|
2020-08-22 14:05:20 +02:00
|
|
|
|
2021-02-09 14:24:14 +01:00
|
|
|
# Limit to the detections overlapping with motion areas
|
|
|
|
# to avoid picking up stationary background objects
|
2021-02-17 14:23:32 +01:00
|
|
|
detections_with_motion = [
|
|
|
|
d for d in detections if intersects_any(d[2], motion_boxes)
|
|
|
|
]
|
2021-02-09 14:24:14 +01:00
|
|
|
|
2020-02-16 04:07:54 +01:00
|
|
|
# now that we have refined our detections, we need to track objects
|
2021-02-09 14:24:14 +01:00
|
|
|
object_tracker.match_and_update(frame_time, detections_with_motion)
|
2020-02-16 04:07:54 +01:00
|
|
|
|
2020-10-24 18:36:04 +02:00
|
|
|
# add to the queue if not full
|
2021-02-17 14:23:32 +01:00
|
|
|
if detected_objects_queue.full():
|
2021-01-16 03:52:59 +01:00
|
|
|
frame_manager.delete(f"{camera_name}{frame_time}")
|
|
|
|
continue
|
2020-10-24 18:36:04 +02:00
|
|
|
else:
|
2021-01-16 03:52:59 +01:00
|
|
|
fps_tracker.update()
|
|
|
|
fps.value = fps_tracker.eps()
|
2021-02-17 14:23:32 +01:00
|
|
|
detected_objects_queue.put(
|
|
|
|
(
|
|
|
|
camera_name,
|
|
|
|
frame_time,
|
|
|
|
object_tracker.tracked_objects,
|
|
|
|
motion_boxes,
|
|
|
|
regions,
|
|
|
|
)
|
|
|
|
)
|
2021-01-16 03:52:59 +01:00
|
|
|
detection_fps.value = object_detector.fps.eps()
|
|
|
|
frame_manager.close(f"{camera_name}{frame_time}")
|