mirror of
https://github.com/blakeblackshear/frigate.git
synced 2024-11-30 19:09:13 +01:00
176af55e8c
* add event_id param to api * exclude query from filtertype * update review pane link for similarity search * update filter group for similarity param and fix switch bug * unneeded prop * update query and input for similarity search param * use undefined instead of empty string for query with similarity search
1047 lines
31 KiB
Python
1047 lines
31 KiB
Python
"""Event apis."""
|
|
|
|
import base64
|
|
import io
|
|
import logging
|
|
import os
|
|
from datetime import datetime
|
|
from functools import reduce
|
|
from pathlib import Path
|
|
from urllib.parse import unquote
|
|
|
|
import cv2
|
|
import numpy as np
|
|
from flask import (
|
|
Blueprint,
|
|
current_app,
|
|
jsonify,
|
|
make_response,
|
|
request,
|
|
)
|
|
from peewee import JOIN, DoesNotExist, fn, operator
|
|
from PIL import Image
|
|
from playhouse.shortcuts import model_to_dict
|
|
|
|
from frigate.const import (
|
|
CLIPS_DIR,
|
|
)
|
|
from frigate.embeddings import EmbeddingsContext
|
|
from frigate.embeddings.embeddings import get_metadata
|
|
from frigate.models import Event, ReviewSegment, Timeline
|
|
from frigate.object_processing import TrackedObject
|
|
from frigate.util.builtin import get_tz_modifiers
|
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
EventBp = Blueprint("events", __name__)
|
|
|
|
DEFAULT_TIME_RANGE = "00:00,24:00"
|
|
|
|
|
|
@EventBp.route("/events")
|
|
def events():
|
|
camera = request.args.get("camera", "all")
|
|
cameras = request.args.get("cameras", "all")
|
|
|
|
# handle old camera arg
|
|
if cameras == "all" and camera != "all":
|
|
cameras = camera
|
|
|
|
label = unquote(request.args.get("label", "all"))
|
|
labels = request.args.get("labels", "all")
|
|
|
|
# handle old label arg
|
|
if labels == "all" and label != "all":
|
|
labels = label
|
|
|
|
sub_label = request.args.get("sub_label", "all")
|
|
sub_labels = request.args.get("sub_labels", "all")
|
|
|
|
# handle old sub_label arg
|
|
if sub_labels == "all" and sub_label != "all":
|
|
sub_labels = sub_label
|
|
|
|
zone = request.args.get("zone", "all")
|
|
zones = request.args.get("zones", "all")
|
|
|
|
# handle old label arg
|
|
if zones == "all" and zone != "all":
|
|
zones = zone
|
|
|
|
limit = request.args.get("limit", 100)
|
|
after = request.args.get("after", type=float)
|
|
before = request.args.get("before", type=float)
|
|
time_range = request.args.get("time_range", DEFAULT_TIME_RANGE)
|
|
has_clip = request.args.get("has_clip", type=int)
|
|
has_snapshot = request.args.get("has_snapshot", type=int)
|
|
in_progress = request.args.get("in_progress", type=int)
|
|
include_thumbnails = request.args.get("include_thumbnails", default=1, type=int)
|
|
favorites = request.args.get("favorites", type=int)
|
|
min_score = request.args.get("min_score", type=float)
|
|
max_score = request.args.get("max_score", type=float)
|
|
is_submitted = request.args.get("is_submitted", type=int)
|
|
min_length = request.args.get("min_length", type=float)
|
|
max_length = request.args.get("max_length", type=float)
|
|
|
|
sort = request.args.get("sort", type=str)
|
|
|
|
clauses = []
|
|
|
|
selected_columns = [
|
|
Event.id,
|
|
Event.camera,
|
|
Event.label,
|
|
Event.zones,
|
|
Event.start_time,
|
|
Event.end_time,
|
|
Event.has_clip,
|
|
Event.has_snapshot,
|
|
Event.plus_id,
|
|
Event.retain_indefinitely,
|
|
Event.sub_label,
|
|
Event.top_score,
|
|
Event.false_positive,
|
|
Event.box,
|
|
Event.data,
|
|
]
|
|
|
|
if camera != "all":
|
|
clauses.append((Event.camera == camera))
|
|
|
|
if cameras != "all":
|
|
camera_list = cameras.split(",")
|
|
clauses.append((Event.camera << camera_list))
|
|
|
|
if labels != "all":
|
|
label_list = labels.split(",")
|
|
clauses.append((Event.label << label_list))
|
|
|
|
if sub_labels != "all":
|
|
# use matching so joined sub labels are included
|
|
# for example a sub label 'bob' would get events
|
|
# with sub labels 'bob' and 'bob, john'
|
|
sub_label_clauses = []
|
|
filtered_sub_labels = sub_labels.split(",")
|
|
|
|
if "None" in filtered_sub_labels:
|
|
filtered_sub_labels.remove("None")
|
|
sub_label_clauses.append((Event.sub_label.is_null()))
|
|
|
|
for label in filtered_sub_labels:
|
|
sub_label_clauses.append(
|
|
(Event.sub_label.cast("text") == label)
|
|
) # include exact matches
|
|
|
|
# include this label when part of a list
|
|
sub_label_clauses.append((Event.sub_label.cast("text") % f"*{label},*"))
|
|
sub_label_clauses.append((Event.sub_label.cast("text") % f"*, {label}*"))
|
|
|
|
sub_label_clause = reduce(operator.or_, sub_label_clauses)
|
|
clauses.append((sub_label_clause))
|
|
|
|
if zones != "all":
|
|
# use matching so events with multiple zones
|
|
# still match on a search where any zone matches
|
|
zone_clauses = []
|
|
filtered_zones = zones.split(",")
|
|
|
|
if "None" in filtered_zones:
|
|
filtered_zones.remove("None")
|
|
zone_clauses.append((Event.zones.length() == 0))
|
|
|
|
for zone in filtered_zones:
|
|
zone_clauses.append((Event.zones.cast("text") % f'*"{zone}"*'))
|
|
|
|
zone_clause = reduce(operator.or_, zone_clauses)
|
|
clauses.append((zone_clause))
|
|
|
|
if after:
|
|
clauses.append((Event.start_time > after))
|
|
|
|
if before:
|
|
clauses.append((Event.start_time < before))
|
|
|
|
if time_range != DEFAULT_TIME_RANGE:
|
|
# get timezone arg to ensure browser times are used
|
|
tz_name = request.args.get("timezone", default="utc", type=str)
|
|
hour_modifier, minute_modifier, _ = get_tz_modifiers(tz_name)
|
|
|
|
times = time_range.split(",")
|
|
time_after = times[0]
|
|
time_before = times[1]
|
|
|
|
start_hour_fun = fn.strftime(
|
|
"%H:%M",
|
|
fn.datetime(Event.start_time, "unixepoch", hour_modifier, minute_modifier),
|
|
)
|
|
|
|
# cases where user wants events overnight, ex: from 20:00 to 06:00
|
|
# should use or operator
|
|
if time_after > time_before:
|
|
clauses.append(
|
|
(
|
|
reduce(
|
|
operator.or_,
|
|
[(start_hour_fun > time_after), (start_hour_fun < time_before)],
|
|
)
|
|
)
|
|
)
|
|
# all other cases should be and operator
|
|
else:
|
|
clauses.append((start_hour_fun > time_after))
|
|
clauses.append((start_hour_fun < time_before))
|
|
|
|
if has_clip is not None:
|
|
clauses.append((Event.has_clip == has_clip))
|
|
|
|
if has_snapshot is not None:
|
|
clauses.append((Event.has_snapshot == has_snapshot))
|
|
|
|
if in_progress is not None:
|
|
clauses.append((Event.end_time.is_null(in_progress)))
|
|
|
|
if include_thumbnails:
|
|
selected_columns.append(Event.thumbnail)
|
|
|
|
if favorites:
|
|
clauses.append((Event.retain_indefinitely == favorites))
|
|
|
|
if max_score is not None:
|
|
clauses.append((Event.data["score"] <= max_score))
|
|
|
|
if min_score is not None:
|
|
clauses.append((Event.data["score"] >= min_score))
|
|
|
|
if min_length is not None:
|
|
clauses.append(((Event.end_time - Event.start_time) >= min_length))
|
|
|
|
if max_length is not None:
|
|
clauses.append(((Event.end_time - Event.start_time) <= max_length))
|
|
|
|
if is_submitted is not None:
|
|
if is_submitted == 0:
|
|
clauses.append((Event.plus_id.is_null()))
|
|
elif is_submitted > 0:
|
|
clauses.append((Event.plus_id != ""))
|
|
|
|
if len(clauses) == 0:
|
|
clauses.append((True))
|
|
|
|
if sort:
|
|
if sort == "score_asc":
|
|
order_by = Event.data["score"].asc()
|
|
elif sort == "score_desc":
|
|
order_by = Event.data["score"].desc()
|
|
elif sort == "date_asc":
|
|
order_by = Event.start_time.asc()
|
|
elif sort == "date_desc":
|
|
order_by = Event.start_time.desc()
|
|
else:
|
|
order_by = Event.start_time.desc()
|
|
|
|
events = (
|
|
Event.select(*selected_columns)
|
|
.where(reduce(operator.and_, clauses))
|
|
.order_by(order_by)
|
|
.limit(limit)
|
|
.dicts()
|
|
.iterator()
|
|
)
|
|
|
|
return jsonify(list(events))
|
|
|
|
|
|
@EventBp.route("/events/explore")
|
|
def events_explore():
|
|
limit = request.args.get("limit", 10, type=int)
|
|
|
|
subquery = Event.select(
|
|
Event.id,
|
|
Event.camera,
|
|
Event.label,
|
|
Event.zones,
|
|
Event.start_time,
|
|
Event.end_time,
|
|
Event.has_clip,
|
|
Event.has_snapshot,
|
|
Event.plus_id,
|
|
Event.retain_indefinitely,
|
|
Event.sub_label,
|
|
Event.top_score,
|
|
Event.false_positive,
|
|
Event.box,
|
|
Event.data,
|
|
fn.rank()
|
|
.over(partition_by=[Event.label], order_by=[Event.start_time.desc()])
|
|
.alias("rank"),
|
|
fn.COUNT(Event.id).over(partition_by=[Event.label]).alias("event_count"),
|
|
).alias("subquery")
|
|
|
|
query = (
|
|
Event.select(
|
|
subquery.c.id,
|
|
subquery.c.camera,
|
|
subquery.c.label,
|
|
subquery.c.zones,
|
|
subquery.c.start_time,
|
|
subquery.c.end_time,
|
|
subquery.c.has_clip,
|
|
subquery.c.has_snapshot,
|
|
subquery.c.plus_id,
|
|
subquery.c.retain_indefinitely,
|
|
subquery.c.sub_label,
|
|
subquery.c.top_score,
|
|
subquery.c.false_positive,
|
|
subquery.c.box,
|
|
subquery.c.data,
|
|
subquery.c.event_count,
|
|
)
|
|
.from_(subquery)
|
|
.where(subquery.c.rank <= limit)
|
|
.order_by(subquery.c.event_count.desc(), subquery.c.start_time.desc())
|
|
.dicts()
|
|
)
|
|
|
|
events = list(query.iterator())
|
|
|
|
processed_events = [
|
|
{k: v for k, v in event.items() if k != "data"}
|
|
| {
|
|
"data": {
|
|
k: v
|
|
for k, v in event["data"].items()
|
|
if k in ["type", "score", "top_score", "description"]
|
|
}
|
|
}
|
|
for event in events
|
|
]
|
|
|
|
return jsonify(processed_events)
|
|
|
|
|
|
@EventBp.route("/event_ids")
|
|
def event_ids():
|
|
idString = request.args.get("ids")
|
|
ids = idString.split(",")
|
|
|
|
if not ids:
|
|
return make_response(
|
|
jsonify({"success": False, "message": "Valid list of ids must be sent"}),
|
|
400,
|
|
)
|
|
|
|
try:
|
|
events = Event.select().where(Event.id << ids).dicts().iterator()
|
|
return jsonify(list(events))
|
|
except Exception:
|
|
return make_response(
|
|
jsonify({"success": False, "message": "Events not found"}), 400
|
|
)
|
|
|
|
|
|
@EventBp.route("/events/search")
|
|
def events_search():
|
|
query = request.args.get("query", type=str)
|
|
search_type = request.args.get("search_type", "thumbnail,description", type=str)
|
|
include_thumbnails = request.args.get("include_thumbnails", default=1, type=int)
|
|
limit = request.args.get("limit", 50, type=int)
|
|
|
|
# Filters
|
|
cameras = request.args.get("cameras", "all", type=str)
|
|
labels = request.args.get("labels", "all", type=str)
|
|
zones = request.args.get("zones", "all", type=str)
|
|
after = request.args.get("after", type=float)
|
|
before = request.args.get("before", type=float)
|
|
|
|
# for similarity search
|
|
event_id = request.args.get("event_id", type=str)
|
|
|
|
if not query and not event_id:
|
|
return make_response(
|
|
jsonify(
|
|
{
|
|
"success": False,
|
|
"message": "A search query must be supplied",
|
|
}
|
|
),
|
|
400,
|
|
)
|
|
|
|
if not current_app.frigate_config.semantic_search.enabled:
|
|
return make_response(
|
|
jsonify(
|
|
{
|
|
"success": False,
|
|
"message": "Semantic search is not enabled",
|
|
}
|
|
),
|
|
400,
|
|
)
|
|
|
|
context: EmbeddingsContext = current_app.embeddings
|
|
|
|
selected_columns = [
|
|
Event.id,
|
|
Event.camera,
|
|
Event.label,
|
|
Event.sub_label,
|
|
Event.zones,
|
|
Event.start_time,
|
|
Event.end_time,
|
|
Event.has_clip,
|
|
Event.has_snapshot,
|
|
Event.data,
|
|
Event.plus_id,
|
|
ReviewSegment.thumb_path,
|
|
]
|
|
|
|
if include_thumbnails:
|
|
selected_columns.append(Event.thumbnail)
|
|
|
|
# Build the where clause for the embeddings query
|
|
embeddings_filters = []
|
|
|
|
if cameras != "all":
|
|
camera_list = cameras.split(",")
|
|
embeddings_filters.append({"camera": {"$in": camera_list}})
|
|
|
|
if labels != "all":
|
|
label_list = labels.split(",")
|
|
embeddings_filters.append({"label": {"$in": label_list}})
|
|
|
|
if zones != "all":
|
|
filtered_zones = zones.split(",")
|
|
zone_filters = [{f"zones_{zone}": {"$eq": True}} for zone in filtered_zones]
|
|
if len(zone_filters) > 1:
|
|
embeddings_filters.append({"$or": zone_filters})
|
|
else:
|
|
embeddings_filters.append(zone_filters[0])
|
|
|
|
if after:
|
|
embeddings_filters.append({"start_time": {"$gt": after}})
|
|
|
|
if before:
|
|
embeddings_filters.append({"start_time": {"$lt": before}})
|
|
|
|
where = None
|
|
if len(embeddings_filters) > 1:
|
|
where = {"$and": embeddings_filters}
|
|
elif len(embeddings_filters) == 1:
|
|
where = embeddings_filters[0]
|
|
|
|
thumb_ids = {}
|
|
desc_ids = {}
|
|
|
|
if search_type == "similarity":
|
|
# Grab the ids of events that match the thumbnail image embeddings
|
|
try:
|
|
search_event: Event = Event.get(Event.id == event_id)
|
|
except DoesNotExist:
|
|
return make_response(
|
|
jsonify(
|
|
{
|
|
"success": False,
|
|
"message": "Event not found",
|
|
}
|
|
),
|
|
404,
|
|
)
|
|
thumbnail = base64.b64decode(search_event.thumbnail)
|
|
img = np.array(Image.open(io.BytesIO(thumbnail)).convert("RGB"))
|
|
thumb_result = context.embeddings.thumbnail.query(
|
|
query_images=[img],
|
|
n_results=limit,
|
|
where=where,
|
|
)
|
|
thumb_ids = dict(
|
|
zip(
|
|
thumb_result["ids"][0],
|
|
context.thumb_stats.normalize(thumb_result["distances"][0]),
|
|
)
|
|
)
|
|
else:
|
|
search_types = search_type.split(",")
|
|
|
|
if "thumbnail" in search_types:
|
|
thumb_result = context.embeddings.thumbnail.query(
|
|
query_texts=[query],
|
|
n_results=limit,
|
|
where=where,
|
|
)
|
|
# Do a rudimentary normalization of the difference in distances returned by CLIP and MiniLM.
|
|
thumb_ids = dict(
|
|
zip(
|
|
thumb_result["ids"][0],
|
|
context.thumb_stats.normalize(thumb_result["distances"][0]),
|
|
)
|
|
)
|
|
|
|
if "description" in search_types:
|
|
desc_result = context.embeddings.description.query(
|
|
query_texts=[query],
|
|
n_results=limit,
|
|
where=where,
|
|
)
|
|
desc_ids = dict(
|
|
zip(
|
|
desc_result["ids"][0],
|
|
context.desc_stats.normalize(desc_result["distances"][0]),
|
|
)
|
|
)
|
|
|
|
results = {}
|
|
for event_id in thumb_ids.keys() | desc_ids:
|
|
min_distance = min(
|
|
i
|
|
for i in (thumb_ids.get(event_id), desc_ids.get(event_id))
|
|
if i is not None
|
|
)
|
|
results[event_id] = {
|
|
"distance": min_distance,
|
|
"source": "thumbnail"
|
|
if min_distance == thumb_ids.get(event_id)
|
|
else "description",
|
|
}
|
|
|
|
if not results:
|
|
return jsonify([])
|
|
|
|
# Get the event data
|
|
events = (
|
|
Event.select(*selected_columns)
|
|
.join(
|
|
ReviewSegment,
|
|
JOIN.LEFT_OUTER,
|
|
on=(fn.json_extract(ReviewSegment.data, "$.detections").contains(Event.id)),
|
|
)
|
|
.where(Event.id << list(results.keys()))
|
|
.dicts()
|
|
.iterator()
|
|
)
|
|
events = list(events)
|
|
|
|
events = [
|
|
{k: v for k, v in event.items() if k != "data"}
|
|
| {
|
|
"data": {
|
|
k: v
|
|
for k, v in event["data"].items()
|
|
if k in ["type", "score", "top_score", "description"]
|
|
}
|
|
}
|
|
| {
|
|
"search_distance": results[event["id"]]["distance"],
|
|
"search_source": results[event["id"]]["source"],
|
|
}
|
|
for event in events
|
|
]
|
|
events = sorted(events, key=lambda x: x["search_distance"])[:limit]
|
|
|
|
return jsonify(events)
|
|
|
|
|
|
@EventBp.route("/events/summary")
|
|
def events_summary():
|
|
tz_name = request.args.get("timezone", default="utc", type=str)
|
|
hour_modifier, minute_modifier, seconds_offset = get_tz_modifiers(tz_name)
|
|
has_clip = request.args.get("has_clip", type=int)
|
|
has_snapshot = request.args.get("has_snapshot", type=int)
|
|
|
|
clauses = []
|
|
|
|
if has_clip is not None:
|
|
clauses.append((Event.has_clip == has_clip))
|
|
|
|
if has_snapshot is not None:
|
|
clauses.append((Event.has_snapshot == has_snapshot))
|
|
|
|
if len(clauses) == 0:
|
|
clauses.append((True))
|
|
|
|
groups = (
|
|
Event.select(
|
|
Event.camera,
|
|
Event.label,
|
|
Event.sub_label,
|
|
fn.strftime(
|
|
"%Y-%m-%d",
|
|
fn.datetime(
|
|
Event.start_time, "unixepoch", hour_modifier, minute_modifier
|
|
),
|
|
).alias("day"),
|
|
Event.zones,
|
|
fn.COUNT(Event.id).alias("count"),
|
|
)
|
|
.where(reduce(operator.and_, clauses))
|
|
.group_by(
|
|
Event.camera,
|
|
Event.label,
|
|
Event.sub_label,
|
|
(Event.start_time + seconds_offset).cast("int") / (3600 * 24),
|
|
Event.zones,
|
|
)
|
|
)
|
|
|
|
return jsonify([e for e in groups.dicts()])
|
|
|
|
|
|
@EventBp.route("/events/<id>", methods=("GET",))
|
|
def event(id):
|
|
try:
|
|
return model_to_dict(Event.get(Event.id == id))
|
|
except DoesNotExist:
|
|
return "Event not found", 404
|
|
|
|
|
|
@EventBp.route("/events/<id>/retain", methods=("POST",))
|
|
def set_retain(id):
|
|
try:
|
|
event = Event.get(Event.id == id)
|
|
except DoesNotExist:
|
|
return make_response(
|
|
jsonify({"success": False, "message": "Event " + id + " not found"}), 404
|
|
)
|
|
|
|
event.retain_indefinitely = True
|
|
event.save()
|
|
|
|
return make_response(
|
|
jsonify({"success": True, "message": "Event " + id + " retained"}), 200
|
|
)
|
|
|
|
|
|
@EventBp.route("/events/<id>/plus", methods=("POST",))
|
|
def send_to_plus(id):
|
|
if not current_app.plus_api.is_active():
|
|
message = "PLUS_API_KEY environment variable is not set"
|
|
logger.error(message)
|
|
return make_response(
|
|
jsonify(
|
|
{
|
|
"success": False,
|
|
"message": message,
|
|
}
|
|
),
|
|
400,
|
|
)
|
|
|
|
include_annotation = (
|
|
request.json.get("include_annotation") if request.is_json else None
|
|
)
|
|
|
|
try:
|
|
event = Event.get(Event.id == id)
|
|
except DoesNotExist:
|
|
message = f"Event {id} not found"
|
|
logger.error(message)
|
|
return make_response(jsonify({"success": False, "message": message}), 404)
|
|
|
|
# events from before the conversion to relative dimensions cant include annotations
|
|
if event.data.get("box") is None:
|
|
include_annotation = None
|
|
|
|
if event.end_time is None:
|
|
logger.error(f"Unable to load clean png for in-progress event: {event.id}")
|
|
return make_response(
|
|
jsonify(
|
|
{
|
|
"success": False,
|
|
"message": "Unable to load clean png for in-progress event",
|
|
}
|
|
),
|
|
400,
|
|
)
|
|
|
|
if event.plus_id:
|
|
message = "Already submitted to plus"
|
|
logger.error(message)
|
|
return make_response(jsonify({"success": False, "message": message}), 400)
|
|
|
|
# load clean.png
|
|
try:
|
|
filename = f"{event.camera}-{event.id}-clean.png"
|
|
image = cv2.imread(os.path.join(CLIPS_DIR, filename))
|
|
except Exception:
|
|
logger.error(f"Unable to load clean png for event: {event.id}")
|
|
return make_response(
|
|
jsonify(
|
|
{"success": False, "message": "Unable to load clean png for event"}
|
|
),
|
|
400,
|
|
)
|
|
|
|
if image is None or image.size == 0:
|
|
logger.error(f"Unable to load clean png for event: {event.id}")
|
|
return make_response(
|
|
jsonify(
|
|
{"success": False, "message": "Unable to load clean png for event"}
|
|
),
|
|
400,
|
|
)
|
|
|
|
try:
|
|
plus_id = current_app.plus_api.upload_image(image, event.camera)
|
|
except Exception as ex:
|
|
logger.exception(ex)
|
|
return make_response(
|
|
jsonify({"success": False, "message": "Error uploading image"}),
|
|
400,
|
|
)
|
|
|
|
# store image id in the database
|
|
event.plus_id = plus_id
|
|
event.save()
|
|
|
|
if include_annotation is not None:
|
|
box = event.data["box"]
|
|
|
|
try:
|
|
current_app.plus_api.add_annotation(
|
|
event.plus_id,
|
|
box,
|
|
event.label,
|
|
)
|
|
except ValueError:
|
|
message = "Error uploading annotation, unsupported label provided."
|
|
logger.error(message)
|
|
return make_response(
|
|
jsonify({"success": False, "message": message}),
|
|
400,
|
|
)
|
|
except Exception as ex:
|
|
logger.exception(ex)
|
|
return make_response(
|
|
jsonify({"success": False, "message": "Error uploading annotation"}),
|
|
400,
|
|
)
|
|
|
|
return make_response(jsonify({"success": True, "plus_id": plus_id}), 200)
|
|
|
|
|
|
@EventBp.route("/events/<id>/false_positive", methods=("PUT",))
|
|
def false_positive(id):
|
|
if not current_app.plus_api.is_active():
|
|
message = "PLUS_API_KEY environment variable is not set"
|
|
logger.error(message)
|
|
return make_response(
|
|
jsonify(
|
|
{
|
|
"success": False,
|
|
"message": message,
|
|
}
|
|
),
|
|
400,
|
|
)
|
|
|
|
try:
|
|
event = Event.get(Event.id == id)
|
|
except DoesNotExist:
|
|
message = f"Event {id} not found"
|
|
logger.error(message)
|
|
return make_response(jsonify({"success": False, "message": message}), 404)
|
|
|
|
# events from before the conversion to relative dimensions cant include annotations
|
|
if event.data.get("box") is None:
|
|
message = "Events prior to 0.13 cannot be submitted as false positives"
|
|
logger.error(message)
|
|
return make_response(jsonify({"success": False, "message": message}), 400)
|
|
|
|
if event.false_positive:
|
|
message = "False positive already submitted to Frigate+"
|
|
logger.error(message)
|
|
return make_response(jsonify({"success": False, "message": message}), 400)
|
|
|
|
if not event.plus_id:
|
|
plus_response = send_to_plus(id)
|
|
if plus_response.status_code != 200:
|
|
return plus_response
|
|
# need to refetch the event now that it has a plus_id
|
|
event = Event.get(Event.id == id)
|
|
|
|
region = event.data["region"]
|
|
box = event.data["box"]
|
|
|
|
# provide top score if score is unavailable
|
|
score = (
|
|
(event.data["top_score"] if event.data["top_score"] else event.top_score)
|
|
if event.data["score"] is None
|
|
else event.data["score"]
|
|
)
|
|
|
|
try:
|
|
current_app.plus_api.add_false_positive(
|
|
event.plus_id,
|
|
region,
|
|
box,
|
|
score,
|
|
event.label,
|
|
event.model_hash,
|
|
event.model_type,
|
|
event.detector_type,
|
|
)
|
|
except ValueError:
|
|
message = "Error uploading false positive, unsupported label provided."
|
|
logger.error(message)
|
|
return make_response(
|
|
jsonify({"success": False, "message": message}),
|
|
400,
|
|
)
|
|
except Exception as ex:
|
|
logger.exception(ex)
|
|
return make_response(
|
|
jsonify({"success": False, "message": "Error uploading false positive"}),
|
|
400,
|
|
)
|
|
|
|
event.false_positive = True
|
|
event.save()
|
|
|
|
return make_response(jsonify({"success": True, "plus_id": event.plus_id}), 200)
|
|
|
|
|
|
@EventBp.route("/events/<id>/retain", methods=("DELETE",))
|
|
def delete_retain(id):
|
|
try:
|
|
event = Event.get(Event.id == id)
|
|
except DoesNotExist:
|
|
return make_response(
|
|
jsonify({"success": False, "message": "Event " + id + " not found"}), 404
|
|
)
|
|
|
|
event.retain_indefinitely = False
|
|
event.save()
|
|
|
|
return make_response(
|
|
jsonify({"success": True, "message": "Event " + id + " un-retained"}), 200
|
|
)
|
|
|
|
|
|
@EventBp.route("/events/<id>/sub_label", methods=("POST",))
|
|
def set_sub_label(id):
|
|
try:
|
|
event: Event = Event.get(Event.id == id)
|
|
except DoesNotExist:
|
|
return make_response(
|
|
jsonify({"success": False, "message": "Event " + id + " not found"}), 404
|
|
)
|
|
|
|
json: dict[str, any] = request.get_json(silent=True) or {}
|
|
new_sub_label = json.get("subLabel")
|
|
new_score = json.get("subLabelScore")
|
|
|
|
if new_sub_label is None:
|
|
return make_response(
|
|
jsonify(
|
|
{
|
|
"success": False,
|
|
"message": "A sub label must be supplied",
|
|
}
|
|
),
|
|
400,
|
|
)
|
|
|
|
if new_sub_label and len(new_sub_label) > 100:
|
|
return make_response(
|
|
jsonify(
|
|
{
|
|
"success": False,
|
|
"message": new_sub_label
|
|
+ " exceeds the 100 character limit for sub_label",
|
|
}
|
|
),
|
|
400,
|
|
)
|
|
|
|
if new_score is not None and (new_score > 1.0 or new_score < 0):
|
|
return make_response(
|
|
jsonify(
|
|
{
|
|
"success": False,
|
|
"message": new_score
|
|
+ " does not fit within the expected bounds 0 <= score <= 1.0",
|
|
}
|
|
),
|
|
400,
|
|
)
|
|
|
|
if not event.end_time:
|
|
# update tracked object
|
|
tracked_obj: TrackedObject = (
|
|
current_app.detected_frames_processor.camera_states[
|
|
event.camera
|
|
].tracked_objects.get(event.id)
|
|
)
|
|
|
|
if tracked_obj:
|
|
tracked_obj.obj_data["sub_label"] = (new_sub_label, new_score)
|
|
|
|
# update timeline items
|
|
Timeline.update(
|
|
data=Timeline.data.update({"sub_label": (new_sub_label, new_score)})
|
|
).where(Timeline.source_id == id).execute()
|
|
|
|
event.sub_label = new_sub_label
|
|
|
|
if new_score:
|
|
data = event.data
|
|
data["sub_label_score"] = new_score
|
|
event.data = data
|
|
|
|
event.save()
|
|
return make_response(
|
|
jsonify(
|
|
{
|
|
"success": True,
|
|
"message": "Event " + id + " sub label set to " + new_sub_label,
|
|
}
|
|
),
|
|
200,
|
|
)
|
|
|
|
|
|
@EventBp.route("/events/<id>/description", methods=("POST",))
|
|
def set_description(id):
|
|
try:
|
|
event: Event = Event.get(Event.id == id)
|
|
except DoesNotExist:
|
|
return make_response(
|
|
jsonify({"success": False, "message": "Event " + id + " not found"}), 404
|
|
)
|
|
|
|
json: dict[str, any] = request.get_json(silent=True) or {}
|
|
new_description = json.get("description")
|
|
|
|
if new_description is None or len(new_description) == 0:
|
|
return make_response(
|
|
jsonify(
|
|
{
|
|
"success": False,
|
|
"message": "description cannot be empty",
|
|
}
|
|
),
|
|
400,
|
|
)
|
|
|
|
event.data["description"] = new_description
|
|
event.save()
|
|
|
|
# If semantic search is enabled, update the index
|
|
if current_app.frigate_config.semantic_search.enabled:
|
|
context: EmbeddingsContext = current_app.embeddings
|
|
context.embeddings.description.upsert(
|
|
documents=[new_description],
|
|
metadatas=[get_metadata(event)],
|
|
ids=[id],
|
|
)
|
|
|
|
return make_response(
|
|
jsonify(
|
|
{
|
|
"success": True,
|
|
"message": "Event " + id + " description set to " + new_description,
|
|
}
|
|
),
|
|
200,
|
|
)
|
|
|
|
|
|
@EventBp.route("/events/<id>", methods=("DELETE",))
|
|
def delete_event(id):
|
|
try:
|
|
event = Event.get(Event.id == id)
|
|
except DoesNotExist:
|
|
return make_response(
|
|
jsonify({"success": False, "message": "Event " + id + " not found"}), 404
|
|
)
|
|
|
|
media_name = f"{event.camera}-{event.id}"
|
|
if event.has_snapshot:
|
|
media = Path(f"{os.path.join(CLIPS_DIR, media_name)}.jpg")
|
|
media.unlink(missing_ok=True)
|
|
media = Path(f"{os.path.join(CLIPS_DIR, media_name)}-clean.png")
|
|
media.unlink(missing_ok=True)
|
|
if event.has_clip:
|
|
media = Path(f"{os.path.join(CLIPS_DIR, media_name)}.mp4")
|
|
media.unlink(missing_ok=True)
|
|
|
|
event.delete_instance()
|
|
Timeline.delete().where(Timeline.source_id == id).execute()
|
|
# If semantic search is enabled, update the index
|
|
if current_app.frigate_config.semantic_search.enabled:
|
|
context: EmbeddingsContext = current_app.embeddings
|
|
context.embeddings.thumbnail.delete(ids=[id])
|
|
context.embeddings.description.delete(ids=[id])
|
|
return make_response(
|
|
jsonify({"success": True, "message": "Event " + id + " deleted"}), 200
|
|
)
|
|
|
|
|
|
@EventBp.route("/events/<camera_name>/<label>/create", methods=["POST"])
|
|
def create_event(camera_name, label):
|
|
if not camera_name or not current_app.frigate_config.cameras.get(camera_name):
|
|
return make_response(
|
|
jsonify(
|
|
{"success": False, "message": f"{camera_name} is not a valid camera."}
|
|
),
|
|
404,
|
|
)
|
|
|
|
if not label:
|
|
return make_response(
|
|
jsonify({"success": False, "message": f"{label} must be set."}), 404
|
|
)
|
|
|
|
json: dict[str, any] = request.get_json(silent=True) or {}
|
|
|
|
try:
|
|
frame = current_app.detected_frames_processor.get_current_frame(camera_name)
|
|
|
|
event_id = current_app.external_processor.create_manual_event(
|
|
camera_name,
|
|
label,
|
|
json.get("source_type", "api"),
|
|
json.get("sub_label", None),
|
|
json.get("score", 0),
|
|
json.get("duration", 30),
|
|
json.get("include_recording", True),
|
|
json.get("draw", {}),
|
|
frame,
|
|
)
|
|
except Exception as e:
|
|
logger.error(e)
|
|
return make_response(
|
|
jsonify({"success": False, "message": "An unknown error occurred"}),
|
|
500,
|
|
)
|
|
|
|
return make_response(
|
|
jsonify(
|
|
{
|
|
"success": True,
|
|
"message": "Successfully created event.",
|
|
"event_id": event_id,
|
|
}
|
|
),
|
|
200,
|
|
)
|
|
|
|
|
|
@EventBp.route("/events/<event_id>/end", methods=["PUT"])
|
|
def end_event(event_id):
|
|
json: dict[str, any] = request.get_json(silent=True) or {}
|
|
|
|
try:
|
|
end_time = json.get("end_time", datetime.now().timestamp())
|
|
current_app.external_processor.finish_manual_event(event_id, end_time)
|
|
except Exception:
|
|
return make_response(
|
|
jsonify(
|
|
{"success": False, "message": f"{event_id} must be set and valid."}
|
|
),
|
|
404,
|
|
)
|
|
|
|
return make_response(
|
|
jsonify({"success": True, "message": "Event successfully ended."}), 200
|
|
)
|