mirror of
				https://github.com/blakeblackshear/frigate.git
				synced 2025-10-27 10:52:11 +01:00 
			
		
		
		
	merged shared memory objects and regions and set color of bounding box based on motion
This commit is contained in:
		
							parent
							
								
									30c223711d
								
							
						
					
					
						commit
						f3376a83d6
					
				@ -157,7 +157,17 @@ def main():
 | 
				
			|||||||
            'size': int(region_parts[0]),
 | 
					            'size': int(region_parts[0]),
 | 
				
			||||||
            'x_offset': int(region_parts[1]),
 | 
					            'x_offset': int(region_parts[1]),
 | 
				
			||||||
            'y_offset': int(region_parts[2]),
 | 
					            'y_offset': int(region_parts[2]),
 | 
				
			||||||
            'min_object_size': int(region_parts[3])
 | 
					            'min_object_size': int(region_parts[3]),
 | 
				
			||||||
 | 
					            # shared value for signaling to the capture process that we are ready for the next frame
 | 
				
			||||||
 | 
					            # (1 for ready 0 for not ready)
 | 
				
			||||||
 | 
					            'ready_for_frame': mp.Value('i', 1),
 | 
				
			||||||
 | 
					            # shared value for motion detection signal (1 for motion 0 for no motion)
 | 
				
			||||||
 | 
					            'motion_detected': mp.Value('i', 0),
 | 
				
			||||||
 | 
					            # create shared array for storing 10 detected objects
 | 
				
			||||||
 | 
					            # note: this must be a double even though the value you are storing
 | 
				
			||||||
 | 
					            #       is a float. otherwise it stops updating the value in shared
 | 
				
			||||||
 | 
					            #       memory. probably something to do with the size of the memory block
 | 
				
			||||||
 | 
					            'output_array': mp.Array(ctypes.c_double, 6*10)
 | 
				
			||||||
        })
 | 
					        })
 | 
				
			||||||
    # capture a single frame and check the frame shape so the correct array
 | 
					    # capture a single frame and check the frame shape so the correct array
 | 
				
			||||||
    # size can be allocated in memory
 | 
					    # size can be allocated in memory
 | 
				
			||||||
@ -170,21 +180,6 @@ def main():
 | 
				
			|||||||
        exit(1)
 | 
					        exit(1)
 | 
				
			||||||
    video.release()
 | 
					    video.release()
 | 
				
			||||||
        
 | 
					        
 | 
				
			||||||
    shared_memory_objects = []
 | 
					 | 
				
			||||||
    for region in regions:
 | 
					 | 
				
			||||||
        shared_memory_objects.append({
 | 
					 | 
				
			||||||
            # shared value for signaling to the capture process that we are ready for the next frame
 | 
					 | 
				
			||||||
            # (1 for ready 0 for not ready)
 | 
					 | 
				
			||||||
            'ready_for_frame': mp.Value('i', 1),
 | 
					 | 
				
			||||||
            # shared value for motion detection signal (1 for motion 0 for no motion)
 | 
					 | 
				
			||||||
            'motion_detected': mp.Value('i', 0),
 | 
					 | 
				
			||||||
            # create shared array for storing 10 detected objects
 | 
					 | 
				
			||||||
            # note: this must be a double even though the value you are storing
 | 
					 | 
				
			||||||
            #       is a float. otherwise it stops updating the value in shared
 | 
					 | 
				
			||||||
            #       memory. probably something to do with the size of the memory block
 | 
					 | 
				
			||||||
            'output_array': mp.Array(ctypes.c_double, 6*10)
 | 
					 | 
				
			||||||
        })
 | 
					 | 
				
			||||||
        
 | 
					 | 
				
			||||||
    # compute the flattened array length from the array shape
 | 
					    # compute the flattened array length from the array shape
 | 
				
			||||||
    flat_array_length = frame_shape[0] * frame_shape[1] * frame_shape[2]
 | 
					    flat_array_length = frame_shape[0] * frame_shape[1] * frame_shape[2]
 | 
				
			||||||
    # create shared array for storing the full frame image data
 | 
					    # create shared array for storing the full frame image data
 | 
				
			||||||
@ -194,15 +189,16 @@ def main():
 | 
				
			|||||||
    # shape current frame so it can be treated as an image
 | 
					    # shape current frame so it can be treated as an image
 | 
				
			||||||
    frame_arr = tonumpyarray(shared_arr).reshape(frame_shape)
 | 
					    frame_arr = tonumpyarray(shared_arr).reshape(frame_shape)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
    capture_process = mp.Process(target=fetch_frames, args=(shared_arr, shared_frame_time, [obj['ready_for_frame'] for obj in shared_memory_objects], frame_shape))
 | 
					    capture_process = mp.Process(target=fetch_frames, args=(shared_arr, 
 | 
				
			||||||
 | 
					        shared_frame_time, [region['ready_for_frame'] for region in regions], frame_shape))
 | 
				
			||||||
    capture_process.daemon = True
 | 
					    capture_process.daemon = True
 | 
				
			||||||
 | 
					
 | 
				
			||||||
    detection_processes = []
 | 
					    detection_processes = []
 | 
				
			||||||
    for index, region in enumerate(regions):
 | 
					    for index, region in enumerate(regions):
 | 
				
			||||||
        detection_process = mp.Process(target=process_frames, args=(shared_arr, 
 | 
					        detection_process = mp.Process(target=process_frames, args=(shared_arr, 
 | 
				
			||||||
            shared_memory_objects[index]['output_array'],
 | 
					            region['output_array'],
 | 
				
			||||||
            shared_frame_time,
 | 
					            shared_frame_time,
 | 
				
			||||||
            shared_memory_objects[index]['motion_detected'],
 | 
					            region['motion_detected'],
 | 
				
			||||||
            frame_shape, 
 | 
					            frame_shape, 
 | 
				
			||||||
            region['size'], region['x_offset'], region['y_offset']))
 | 
					            region['size'], region['x_offset'], region['y_offset']))
 | 
				
			||||||
        detection_process.daemon = True
 | 
					        detection_process.daemon = True
 | 
				
			||||||
@ -212,20 +208,20 @@ def main():
 | 
				
			|||||||
    for index, region in enumerate(regions):
 | 
					    for index, region in enumerate(regions):
 | 
				
			||||||
        motion_process = mp.Process(target=detect_motion, args=(shared_arr,
 | 
					        motion_process = mp.Process(target=detect_motion, args=(shared_arr,
 | 
				
			||||||
            shared_frame_time,
 | 
					            shared_frame_time,
 | 
				
			||||||
            shared_memory_objects[index]['ready_for_frame'],
 | 
					            region['ready_for_frame'],
 | 
				
			||||||
            shared_memory_objects[index]['motion_detected'],
 | 
					            region['motion_detected'],
 | 
				
			||||||
            frame_shape, 
 | 
					            frame_shape, 
 | 
				
			||||||
            region['size'], region['x_offset'], region['y_offset'],
 | 
					            region['size'], region['x_offset'], region['y_offset'],
 | 
				
			||||||
            region['min_object_size']))
 | 
					            region['min_object_size']))
 | 
				
			||||||
        motion_process.daemon = True
 | 
					        motion_process.daemon = True
 | 
				
			||||||
        motion_processes.append(motion_process)
 | 
					        motion_processes.append(motion_process)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
    object_parser = ObjectParser([obj['output_array'] for obj in shared_memory_objects])
 | 
					    object_parser = ObjectParser([region['output_array'] for region in regions])
 | 
				
			||||||
    object_parser.start()
 | 
					    object_parser.start()
 | 
				
			||||||
 | 
					
 | 
				
			||||||
    mqtt_publisher = MqttPublisher(MQTT_HOST, MQTT_MOTION_TOPIC, MQTT_OBJECT_TOPIC, 
 | 
					    mqtt_publisher = MqttPublisher(MQTT_HOST, MQTT_MOTION_TOPIC, MQTT_OBJECT_TOPIC, 
 | 
				
			||||||
        MQTT_OBJECT_CLASSES.split(','), 
 | 
					        MQTT_OBJECT_CLASSES.split(','), 
 | 
				
			||||||
        [obj['motion_detected'] for obj in shared_memory_objects])
 | 
					        [region['motion_detected'] for region in regions])
 | 
				
			||||||
    mqtt_publisher.start()
 | 
					    mqtt_publisher.start()
 | 
				
			||||||
 | 
					
 | 
				
			||||||
    capture_process.start()
 | 
					    capture_process.start()
 | 
				
			||||||
@ -268,15 +264,15 @@ def main():
 | 
				
			|||||||
                    use_normalized_coordinates=False)
 | 
					                    use_normalized_coordinates=False)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
            for region in regions:
 | 
					            for region in regions:
 | 
				
			||||||
 | 
					                color = (255,255,255)
 | 
				
			||||||
 | 
					                if region['motion_detected'].value == 1:
 | 
				
			||||||
 | 
					                    color = (0,255,0)
 | 
				
			||||||
                cv2.rectangle(frame, (region['x_offset'], region['y_offset']), 
 | 
					                cv2.rectangle(frame, (region['x_offset'], region['y_offset']), 
 | 
				
			||||||
                    (region['x_offset']+region['size'], region['y_offset']+region['size']), 
 | 
					                    (region['x_offset']+region['size'], region['y_offset']+region['size']), 
 | 
				
			||||||
                    (255,255,255), 2)
 | 
					                    color, 2)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
            motion_status = 'No Motion'
 | 
					            cv2.putText(frame, datetime.datetime.now().strftime("%H:%M:%S"), (1125, 20),
 | 
				
			||||||
            if any(obj['motion_detected'].value == 1 for obj in shared_memory_objects):
 | 
							        cv2.FONT_HERSHEY_SIMPLEX, 0.8, (0, 0, 255), 2)
 | 
				
			||||||
                motion_status = 'Motion'
 | 
					 | 
				
			||||||
            cv2.putText(frame, motion_status, (10, 20),
 | 
					 | 
				
			||||||
		        cv2.FONT_HERSHEY_SIMPLEX, 0.5, (0, 0, 255), 2)
 | 
					 | 
				
			||||||
            # convert back to BGR
 | 
					            # convert back to BGR
 | 
				
			||||||
            frame = cv2.cvtColor(frame, cv2.COLOR_RGB2BGR)
 | 
					            frame = cv2.cvtColor(frame, cv2.COLOR_RGB2BGR)
 | 
				
			||||||
            # encode the image into a jpg
 | 
					            # encode the image into a jpg
 | 
				
			||||||
 | 
				
			|||||||
		Loading…
	
		Reference in New Issue
	
	Block a user