blakeblackshear.frigate/detect_objects.py

481 lines
20 KiB
Python
Raw Normal View History

2020-10-07 23:44:34 +02:00
import faulthandler; faulthandler.enable()
import os
2020-08-02 15:46:36 +02:00
import signal
import sys
import traceback
import signal
2019-01-26 15:02:59 +01:00
import cv2
import time
import datetime
2019-03-25 12:24:36 +01:00
import queue
import yaml
import json
2020-02-16 16:19:08 +01:00
import threading
2020-02-16 04:07:54 +01:00
import multiprocessing as mp
import subprocess as sp
2019-01-26 15:02:59 +01:00
import numpy as np
2020-02-18 13:25:24 +01:00
import logging
from flask import Flask, Response, make_response, jsonify, request
2019-02-10 19:00:52 +01:00
import paho.mqtt.client as mqtt
2019-01-26 15:02:59 +01:00
from frigate.video import track_camera, get_ffmpeg_input, get_frame_shape, CameraCapture, start_or_restart_ffmpeg
2020-02-16 04:07:54 +01:00
from frigate.object_processing import TrackedObjectProcessor
2020-07-09 13:57:16 +02:00
from frigate.events import EventProcessor
from frigate.util import EventsPerSecond
2020-02-16 04:07:54 +01:00
from frigate.edgetpu import EdgeTPUProcess
2019-01-26 15:02:59 +01:00
FRIGATE_VARS = {k: v for k, v in os.environ.items() if k.startswith('FRIGATE_')}
CONFIG_FILE = os.environ.get('CONFIG_FILE', '/config/config.yml')
if CONFIG_FILE.endswith(".yml"):
with open(CONFIG_FILE) as f:
CONFIG = yaml.safe_load(f)
elif CONFIG_FILE.endswith(".json"):
with open(CONFIG_FILE) as f:
CONFIG = json.load(f)
2019-01-26 15:02:59 +01:00
2020-10-18 19:05:49 +02:00
CACHE_DIR = CONFIG.get('save_clips', {}).get('cache_dir', '/cache')
2020-10-19 13:49:37 +02:00
CLIPS_DIR = CONFIG.get('save_clips', {}).get('clips_dir', '/clips')
2020-10-18 19:05:49 +02:00
2020-10-19 13:49:37 +02:00
if not os.path.exists(CACHE_DIR) and not os.path.islink(CACHE_DIR):
2020-10-18 19:05:49 +02:00
os.makedirs(CACHE_DIR)
2020-10-19 13:49:37 +02:00
if not os.path.exists(CLIPS_DIR) and not os.path.islink(CLIPS_DIR):
2020-10-18 19:05:49 +02:00
os.makedirs(CLIPS_DIR)
MQTT_HOST = CONFIG['mqtt']['host']
MQTT_PORT = CONFIG.get('mqtt', {}).get('port', 1883)
2019-03-30 02:49:27 +01:00
MQTT_TOPIC_PREFIX = CONFIG.get('mqtt', {}).get('topic_prefix', 'frigate')
MQTT_USER = CONFIG.get('mqtt', {}).get('user')
MQTT_PASS = CONFIG.get('mqtt', {}).get('password')
if not MQTT_PASS is None:
MQTT_PASS = MQTT_PASS.format(**FRIGATE_VARS)
MQTT_CLIENT_ID = CONFIG.get('mqtt', {}).get('client_id', 'frigate')
# Set the default FFmpeg config
FFMPEG_CONFIG = CONFIG.get('ffmpeg', {})
FFMPEG_DEFAULT_CONFIG = {
'global_args': FFMPEG_CONFIG.get('global_args',
['-hide_banner','-loglevel','panic']),
'hwaccel_args': FFMPEG_CONFIG.get('hwaccel_args',
[]),
'input_args': FFMPEG_CONFIG.get('input_args',
['-avoid_negative_ts', 'make_zero',
'-fflags', 'nobuffer',
'-flags', 'low_delay',
'-strict', 'experimental',
'-fflags', '+genpts+discardcorrupt',
'-rtsp_transport', 'tcp',
'-stimeout', '5000000',
'-use_wallclock_as_timestamps', '1']),
'output_args': FFMPEG_CONFIG.get('output_args',
2020-02-16 04:07:54 +01:00
['-f', 'rawvideo',
2020-10-10 17:07:14 +02:00
'-pix_fmt', 'yuv420p'])
}
GLOBAL_OBJECT_CONFIG = CONFIG.get('objects', {})
WEB_PORT = CONFIG.get('web_port', 5000)
2020-10-11 20:07:00 +02:00
DETECTORS = CONFIG.get('detectors', {'coral': {'type': 'edgetpu', 'device': 'usb'}})
2019-02-24 15:41:03 +01:00
2020-02-16 16:19:08 +01:00
class CameraWatchdog(threading.Thread):
def __init__(self, camera_processes, config, detectors, detection_queue, tracked_objects_queue, stop_event):
2020-02-16 16:19:08 +01:00
threading.Thread.__init__(self)
self.camera_processes = camera_processes
self.config = config
self.detectors = detectors
self.detection_queue = detection_queue
2020-02-16 16:19:08 +01:00
self.tracked_objects_queue = tracked_objects_queue
2020-08-02 15:46:36 +02:00
self.stop_event = stop_event
2020-02-16 16:19:08 +01:00
def run(self):
time.sleep(10)
while True:
# wait a bit before checking
time.sleep(10)
2020-08-02 15:46:36 +02:00
if self.stop_event.is_set():
print(f"Exiting watchdog...")
break
now = datetime.datetime.now().timestamp()
2020-02-16 16:19:08 +01:00
# check the detection processes
for detector in self.detectors.values():
detection_start = detector.detection_start.value
if (detection_start > 0.0 and
now - detection_start > 10):
print("Detection appears to be stuck. Restarting detection process")
detector.start_or_restart()
elif not detector.detect_process.is_alive():
print("Detection appears to have stopped. Restarting detection process")
detector.start_or_restart()
# check the camera processes
2020-02-16 16:19:08 +01:00
for name, camera_process in self.camera_processes.items():
process = camera_process['process']
if not process.is_alive():
print(f"Track process for {name} is not alive. Starting again...")
camera_process['process_fps'].value = 0.0
2020-02-23 22:53:00 +01:00
camera_process['detection_fps'].value = 0.0
camera_process['read_start'].value = 0.0
2020-09-09 13:24:46 +02:00
process = mp.Process(target=track_camera, args=(name, self.config[name], camera_process['frame_queue'],
camera_process['frame_shape'], self.detection_queue, self.tracked_objects_queue,
camera_process['process_fps'], camera_process['detection_fps'],
camera_process['read_start'], self.stop_event))
2020-02-16 16:19:08 +01:00
process.daemon = True
camera_process['process'] = process
process.start()
print(f"Track process started for {name}: {process.pid}")
if not camera_process['capture_thread'].is_alive():
frame_shape = camera_process['frame_shape']
frame_size = frame_shape[0] * frame_shape[1] * frame_shape[2]
ffmpeg_process = start_or_restart_ffmpeg(camera_process['ffmpeg_cmd'], frame_size)
camera_capture = CameraCapture(name, ffmpeg_process, frame_shape, camera_process['frame_queue'],
camera_process['take_frame'], camera_process['camera_fps'], self.stop_event)
camera_capture.start()
camera_process['ffmpeg_process'] = ffmpeg_process
camera_process['capture_thread'] = camera_capture
2020-09-07 19:17:42 +02:00
elif now - camera_process['capture_thread'].current_frame.value > 5:
print(f"No frames received from {name} in 5 seconds. Exiting ffmpeg...")
ffmpeg_process = camera_process['ffmpeg_process']
ffmpeg_process.terminate()
try:
print("Waiting for ffmpeg to exit gracefully...")
ffmpeg_process.communicate(timeout=30)
except sp.TimeoutExpired:
print("FFmpeg didnt exit. Force killing...")
ffmpeg_process.kill()
ffmpeg_process.communicate()
2019-01-26 15:02:59 +01:00
def main():
2020-08-02 15:46:36 +02:00
stop_event = threading.Event()
2019-02-26 03:27:02 +01:00
# connect to mqtt and setup last will
2019-03-27 12:17:00 +01:00
def on_connect(client, userdata, flags, rc):
2019-02-28 13:30:34 +01:00
print("On connect called")
2019-05-14 14:34:14 +02:00
if rc != 0:
if rc == 3:
print ("MQTT Server unavailable")
elif rc == 4:
print ("MQTT Bad username or password")
elif rc == 5:
print ("MQTT Not authorized")
else:
print ("Unable to connect to MQTT: Connection refused. Error code: " + str(rc))
2019-02-28 13:30:34 +01:00
# publish a message to signal that the service is running
client.publish(MQTT_TOPIC_PREFIX+'/available', 'online', retain=True)
client = mqtt.Client(client_id=MQTT_CLIENT_ID)
2019-02-28 13:30:34 +01:00
client.on_connect = on_connect
client.will_set(MQTT_TOPIC_PREFIX+'/available', payload='offline', qos=1, retain=True)
if not MQTT_USER is None:
client.username_pw_set(MQTT_USER, password=MQTT_PASS)
client.connect(MQTT_HOST, MQTT_PORT, 60)
client.loop_start()
2019-03-30 02:49:27 +01:00
2020-02-16 04:07:54 +01:00
##
# Setup config defaults for cameras
##
for name, config in CONFIG['cameras'].items():
config['snapshots'] = {
2020-07-25 14:44:07 +02:00
'show_timestamp': config.get('snapshots', {}).get('show_timestamp', True),
'draw_zones': config.get('snapshots', {}).get('draw_zones', False),
'draw_bounding_boxes': config.get('snapshots', {}).get('draw_bounding_boxes', True)
2020-02-16 04:07:54 +01:00
}
2020-09-15 04:03:18 +02:00
config['zones'] = config.get('zones', {})
2020-02-16 04:07:54 +01:00
# Queue for cameras to push tracked objects to
2020-08-02 15:46:36 +02:00
tracked_objects_queue = mp.Queue()
2020-07-09 13:57:16 +02:00
# Queue for clip processing
event_queue = mp.Queue()
2020-10-11 16:40:20 +02:00
# create the detection pipes and shms
out_events = {}
2020-10-11 16:40:20 +02:00
camera_shms = []
for name in CONFIG['cameras'].keys():
out_events[name] = mp.Event()
2020-10-11 16:40:20 +02:00
shm_in = mp.shared_memory.SharedMemory(name=name, create=True, size=300*300*3)
shm_out = mp.shared_memory.SharedMemory(name=f"out-{name}", create=True, size=20*6*4)
camera_shms.append(shm_in)
camera_shms.append(shm_out)
detection_queue = mp.Queue()
detectors = {}
for name, detector in DETECTORS.items():
if detector['type'] == 'cpu':
detectors[name] = EdgeTPUProcess(detection_queue, out_events=out_events, tf_device='cpu')
if detector['type'] == 'edgetpu':
detectors[name] = EdgeTPUProcess(detection_queue, out_events=out_events, tf_device=detector['device'])
2019-02-26 03:27:02 +01:00
# create the camera processes
2020-02-16 16:19:08 +01:00
camera_processes = {}
2020-02-16 04:07:54 +01:00
for name, config in CONFIG['cameras'].items():
# Merge the ffmpeg config with the global config
ffmpeg = config.get('ffmpeg', {})
ffmpeg_input = get_ffmpeg_input(ffmpeg['input'])
ffmpeg_global_args = ffmpeg.get('global_args', FFMPEG_DEFAULT_CONFIG['global_args'])
ffmpeg_hwaccel_args = ffmpeg.get('hwaccel_args', FFMPEG_DEFAULT_CONFIG['hwaccel_args'])
ffmpeg_input_args = ffmpeg.get('input_args', FFMPEG_DEFAULT_CONFIG['input_args'])
ffmpeg_output_args = ffmpeg.get('output_args', FFMPEG_DEFAULT_CONFIG['output_args'])
2020-09-13 14:46:38 +02:00
if not config.get('fps') is None:
ffmpeg_output_args = ["-r", str(config.get('fps'))] + ffmpeg_output_args
2020-07-26 01:59:04 +02:00
if config.get('save_clips', {}).get('enabled', False):
2020-07-09 13:57:16 +02:00
ffmpeg_output_args = [
"-f",
"segment",
"-segment_time",
"10",
"-segment_format",
"mp4",
"-reset_timestamps",
"1",
"-strftime",
"1",
"-c",
"copy",
"-an",
"-map",
"0",
2020-10-18 19:05:49 +02:00
f"{os.path.join(CACHE_DIR, name)}-%Y%m%d%H%M%S.mp4"
2020-07-09 13:57:16 +02:00
] + ffmpeg_output_args
ffmpeg_cmd = (['ffmpeg'] +
ffmpeg_global_args +
ffmpeg_hwaccel_args +
ffmpeg_input_args +
['-i', ffmpeg_input] +
ffmpeg_output_args +
['pipe:'])
if 'width' in config and 'height' in config:
frame_shape = (config['height'], config['width'], 3)
else:
frame_shape = get_frame_shape(ffmpeg_input)
config['frame_shape'] = frame_shape
frame_size = frame_shape[0] * frame_shape[1] * frame_shape[2]
take_frame = config.get('take_frame', 1)
detection_frame = mp.Value('d', 0.0)
ffmpeg_process = start_or_restart_ffmpeg(ffmpeg_cmd, frame_size)
frame_queue = mp.Queue(maxsize=2)
camera_fps = EventsPerSecond()
camera_fps.start()
camera_capture = CameraCapture(name, ffmpeg_process, frame_shape, frame_queue, take_frame, camera_fps, stop_event)
camera_capture.start()
2020-02-16 16:19:08 +01:00
camera_processes[name] = {
'camera_fps': camera_fps,
'take_frame': take_frame,
'process_fps': mp.Value('d', 0.0),
'detection_fps': mp.Value('d', 0.0),
'detection_frame': detection_frame,
'read_start': mp.Value('d', 0.0),
'ffmpeg_process': ffmpeg_process,
'ffmpeg_cmd': ffmpeg_cmd,
'frame_queue': frame_queue,
'frame_shape': frame_shape,
'capture_thread': camera_capture
2020-02-16 04:07:54 +01:00
}
2020-09-07 19:17:42 +02:00
# merge global object config into camera object config
camera_objects_config = config.get('objects', {})
# get objects to track for camera
objects_to_track = camera_objects_config.get('track', GLOBAL_OBJECT_CONFIG.get('track', ['person']))
# get object filters
object_filters = camera_objects_config.get('filters', GLOBAL_OBJECT_CONFIG.get('filters', {}))
2020-09-07 19:17:42 +02:00
config['objects'] = {
'track': objects_to_track,
'filters': object_filters
}
camera_process = mp.Process(target=track_camera, args=(name, config, frame_queue, frame_shape,
detection_queue, out_events[name], tracked_objects_queue, camera_processes[name]['process_fps'],
camera_processes[name]['detection_fps'],
camera_processes[name]['read_start'], camera_processes[name]['detection_frame'], stop_event))
2020-02-16 04:07:54 +01:00
camera_process.daemon = True
2020-02-16 16:19:08 +01:00
camera_processes[name]['process'] = camera_process
2020-02-16 04:07:54 +01:00
# start the camera_processes
2020-02-16 16:19:08 +01:00
for name, camera_process in camera_processes.items():
camera_process['process'].start()
print(f"Camera_process started for {name}: {camera_process['process'].pid}")
2020-07-09 13:57:16 +02:00
2020-10-18 19:05:49 +02:00
event_processor = EventProcessor(CONFIG, camera_processes, CACHE_DIR, CLIPS_DIR, event_queue, stop_event)
2020-07-09 13:57:16 +02:00
event_processor.start()
2020-07-25 14:44:07 +02:00
2020-09-15 04:03:18 +02:00
object_processor = TrackedObjectProcessor(CONFIG['cameras'], client, MQTT_TOPIC_PREFIX, tracked_objects_queue, event_queue, stop_event)
2020-02-16 04:07:54 +01:00
object_processor.start()
camera_watchdog = CameraWatchdog(camera_processes, CONFIG['cameras'], detectors, detection_queue, tracked_objects_queue, stop_event)
camera_watchdog.start()
2019-02-09 15:51:11 +01:00
2020-08-02 15:46:36 +02:00
def receiveSignal(signalNumber, frame):
print('Received:', signalNumber)
stop_event.set()
event_processor.join()
object_processor.join()
camera_watchdog.join()
2020-10-11 16:48:40 +02:00
for camera_name, camera_process in camera_processes.items():
2020-08-02 15:46:36 +02:00
camera_process['capture_thread'].join()
2020-10-11 16:48:40 +02:00
# cleanup the frame queue
while not camera_process['frame_queue'].empty():
frame_time = camera_process['frame_queue'].get()
shm = mp.shared_memory.SharedMemory(name=f"{camera_name}{frame_time}")
shm.close()
shm.unlink()
2020-10-13 15:01:06 +02:00
for detector in detectors.values():
detector.stop()
2020-10-11 16:40:20 +02:00
for shm in camera_shms:
shm.close()
shm.unlink()
2020-08-02 15:46:36 +02:00
sys.exit()
signal.signal(signal.SIGTERM, receiveSignal)
signal.signal(signal.SIGINT, receiveSignal)
2019-02-26 03:27:02 +01:00
# create a flask app that encodes frames a mjpeg on demand
2019-03-30 03:02:40 +01:00
app = Flask(__name__)
2020-02-18 13:25:24 +01:00
log = logging.getLogger('werkzeug')
log.setLevel(logging.ERROR)
2019-03-30 03:02:40 +01:00
@app.route('/')
def ishealthy():
# return a healh
return "Frigate is running. Alive and healthy!"
@app.route('/debug/stack')
def processor_stack():
frame = sys._current_frames().get(object_processor.ident, None)
if frame:
return "<br>".join(traceback.format_stack(frame)), 200
else:
return "no frame found", 200
@app.route('/debug/print_stack')
def print_stack():
pid = int(request.args.get('pid', 0))
if pid == 0:
return "missing pid", 200
else:
os.kill(pid, signal.SIGUSR1)
return "check logs", 200
@app.route('/debug/stats')
def stats():
stats = {}
total_detection_fps = 0
2020-02-16 16:19:08 +01:00
for name, camera_stats in camera_processes.items():
total_detection_fps += camera_stats['detection_fps'].value
capture_thread = camera_stats['capture_thread']
2020-02-16 04:07:54 +01:00
stats[name] = {
'camera_fps': round(capture_thread.fps.eps(), 2),
'process_fps': round(camera_stats['process_fps'].value, 2),
'skipped_fps': round(capture_thread.skipped_fps.eps(), 2),
'detection_fps': round(camera_stats['detection_fps'].value, 2),
'read_start': camera_stats['read_start'].value,
'pid': camera_stats['process'].pid,
'ffmpeg_pid': camera_stats['ffmpeg_process'].pid,
'frame_info': {
2020-09-07 19:17:42 +02:00
'read': capture_thread.current_frame.value,
'detect': camera_stats['detection_frame'].value,
'process': object_processor.camera_data[name]['current_frame_time']
}
2020-02-16 04:07:54 +01:00
}
stats['detectors'] = {}
for name, detector in detectors.items():
stats['detectors'][name] = {
'inference_speed': round(detector.avg_inference_speed.value*1000, 2),
'detection_start': detector.detection_start.value,
'pid': detector.detect_process.pid
}
stats['detection_fps'] = round(total_detection_fps, 2)
return jsonify(stats)
@app.route('/<camera_name>/<label>/best.jpg')
def best(camera_name, label):
2020-02-16 04:07:54 +01:00
if camera_name in CONFIG['cameras']:
best_object = object_processor.get_best(camera_name, label)
2020-10-14 03:57:57 +02:00
best_frame = best_object.get('frame')
if best_frame is None:
best_frame = np.zeros((720,1280,3), np.uint8)
else:
best_frame = cv2.cvtColor(best_frame, cv2.COLOR_YUV2BGR_I420)
crop = bool(request.args.get('crop', 0, type=int))
if crop:
region = best_object.get('region', [0,0,300,300])
best_frame = best_frame[region[1]:region[3], region[0]:region[2]]
2020-08-08 14:20:14 +02:00
height = int(request.args.get('h', str(best_frame.shape[0])))
width = int(height*best_frame.shape[1]/best_frame.shape[0])
best_frame = cv2.resize(best_frame, dsize=(width, height), interpolation=cv2.INTER_AREA)
ret, jpg = cv2.imencode('.jpg', best_frame)
response = make_response(jpg.tobytes())
response.headers['Content-Type'] = 'image/jpg'
return response
else:
return "Camera named {} not found".format(camera_name), 404
2019-03-30 03:02:40 +01:00
@app.route('/<camera_name>')
def mjpeg_feed(camera_name):
fps = int(request.args.get('fps', '3'))
height = int(request.args.get('h', '360'))
2020-02-16 04:07:54 +01:00
if camera_name in CONFIG['cameras']:
# return a multipart response
return Response(imagestream(camera_name, fps, height),
mimetype='multipart/x-mixed-replace; boundary=frame')
else:
return "Camera named {} not found".format(camera_name), 404
2020-07-25 14:54:13 +02:00
@app.route('/<camera_name>/latest.jpg')
def latest_frame(camera_name):
if camera_name in CONFIG['cameras']:
# max out at specified FPS
frame = object_processor.get_current_frame(camera_name)
if frame is None:
2020-08-02 02:20:16 +02:00
frame = np.zeros((720,1280,3), np.uint8)
2020-07-25 14:54:13 +02:00
2020-08-02 02:20:16 +02:00
height = int(request.args.get('h', str(frame.shape[0])))
2020-07-25 14:54:13 +02:00
width = int(height*frame.shape[1]/frame.shape[0])
2020-07-25 14:54:13 +02:00
frame = cv2.resize(frame, dsize=(width, height), interpolation=cv2.INTER_AREA)
ret, jpg = cv2.imencode('.jpg', frame)
response = make_response(jpg.tobytes())
response.headers['Content-Type'] = 'image/jpg'
return response
else:
return "Camera named {} not found".format(camera_name), 404
def imagestream(camera_name, fps, height):
2019-03-30 03:02:40 +01:00
while True:
# max out at specified FPS
time.sleep(1/fps)
2020-10-11 19:16:05 +02:00
frame = object_processor.get_current_frame(camera_name, draw=True)
if frame is None:
frame = np.zeros((height,int(height*16/9),3), np.uint8)
2020-10-11 04:28:12 +02:00
width = int(height*frame.shape[1]/frame.shape[0])
2020-04-24 14:48:19 +02:00
frame = cv2.resize(frame, dsize=(width, height), interpolation=cv2.INTER_LINEAR)
ret, jpg = cv2.imencode('.jpg', frame)
2019-03-30 03:02:40 +01:00
yield (b'--frame\r\n'
b'Content-Type: image/jpeg\r\n\r\n' + jpg.tobytes() + b'\r\n\r\n')
2019-03-30 03:02:40 +01:00
app.run(host='0.0.0.0', port=WEB_PORT, debug=False)
object_processor.join()
2019-01-26 15:02:59 +01:00
if __name__ == '__main__':
2019-05-14 14:34:14 +02:00
main()