123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380 |
- import datetime
- import itertools
- import logging
- import os
- import random
- import shutil
- import string
- import subprocess as sp
- import threading
- from collections import defaultdict
- from pathlib import Path
- import psutil
- from peewee import JOIN, DoesNotExist
- from frigate.config import FrigateConfig
- from frigate.const import CACHE_DIR, RECORD_DIR
- from frigate.models import Event, Recordings
- logger = logging.getLogger(__name__)
- SECONDS_IN_DAY = 60 * 60 * 24
- def remove_empty_directories(directory):
- # list all directories recursively and sort them by path,
- # longest first
- paths = sorted(
- [x[0] for x in os.walk(RECORD_DIR)],
- key=lambda p: len(str(p)),
- reverse=True,
- )
- for path in paths:
- # don't delete the parent
- if path == RECORD_DIR:
- continue
- if len(os.listdir(path)) == 0:
- os.rmdir(path)
- class RecordingMaintainer(threading.Thread):
- def __init__(self, config: FrigateConfig, stop_event):
- threading.Thread.__init__(self)
- self.name = "recording_maint"
- self.config = config
- self.stop_event = stop_event
- def move_files(self):
- cache_files = [
- d
- for d in os.listdir(CACHE_DIR)
- if os.path.isfile(os.path.join(CACHE_DIR, d))
- and d.endswith(".mp4")
- and not d.startswith("clip_")
- ]
- files_in_use = []
- for process in psutil.process_iter():
- try:
- if process.name() != "ffmpeg":
- continue
- flist = process.open_files()
- if flist:
- for nt in flist:
- if nt.path.startswith(CACHE_DIR):
- files_in_use.append(nt.path.split("/")[-1])
- except:
- continue
- # group recordings by camera
- grouped_recordings = defaultdict(list)
- for f in cache_files:
- # Skip files currently in use
- if f in files_in_use:
- continue
- cache_path = os.path.join(CACHE_DIR, f)
- basename = os.path.splitext(f)[0]
- camera, date = basename.rsplit("-", maxsplit=1)
- start_time = datetime.datetime.strptime(date, "%Y%m%d%H%M%S")
- grouped_recordings[camera].append(
- {
- "cache_path": cache_path,
- "start_time": start_time,
- }
- )
- for camera, recordings in grouped_recordings.items():
- # get all events with the end time after the start of the oldest cache file
- # or with end_time None
- events: Event = (
- Event.select()
- .where(
- Event.camera == camera,
- (Event.end_time == None)
- | (Event.end_time >= recordings[0]["start_time"]),
- Event.has_clip,
- )
- .order_by(Event.start_time)
- )
- for r in recordings:
- cache_path = r["cache_path"]
- start_time = r["start_time"]
- # Just delete files if recordings are turned off
- if (
- not camera in self.config.cameras
- or not self.config.cameras[camera].record.enabled
- ):
- Path(cache_path).unlink(missing_ok=True)
- continue
- ffprobe_cmd = [
- "ffprobe",
- "-v",
- "error",
- "-show_entries",
- "format=duration",
- "-of",
- "default=noprint_wrappers=1:nokey=1",
- f"{cache_path}",
- ]
- p = sp.run(ffprobe_cmd, capture_output=True)
- if p.returncode == 0:
- duration = float(p.stdout.decode().strip())
- end_time = start_time + datetime.timedelta(seconds=duration)
- else:
- logger.warning(f"Discarding a corrupt recording segment: {f}")
- Path(cache_path).unlink(missing_ok=True)
- continue
- # if cached file's start_time is earlier than the retain_days for the camera
- if start_time <= (
- (
- datetime.datetime.now()
- - datetime.timedelta(
- days=self.config.cameras[camera].record.retain_days
- )
- )
- ):
- # if the cached segment overlaps with the events:
- overlaps = False
- for event in events:
- # if the event starts in the future, stop checking events
- # and let this recording segment expire
- if event.start_time > end_time.timestamp():
- overlaps = False
- break
- # if the event is in progress or ends after the recording starts, keep it
- # and stop looking at events
- if event.end_time is None or event.end_time >= start_time:
- overlaps = True
- break
- if overlaps:
- # move from cache to recordings immediately
- self.store_segment(
- camera,
- start_time,
- end_time,
- duration,
- cache_path,
- )
- # else retain_days includes this segment
- else:
- self.store_segment(
- camera, start_time, end_time, duration, cache_path
- )
- if len(recordings) > 2:
- # delete all cached files past the most recent 2
- to_remove = sorted(recordings, key=lambda i: i["start_time"])[:-2]
- for f in to_remove:
- Path(cache_path).unlink(missing_ok=True)
- def store_segment(self, camera, start_time, end_time, duration, cache_path):
- directory = os.path.join(RECORD_DIR, start_time.strftime("%Y-%m/%d/%H"), camera)
- if not os.path.exists(directory):
- os.makedirs(directory)
- file_name = f"{start_time.strftime('%M.%S.mp4')}"
- file_path = os.path.join(directory, file_name)
- # copy then delete is required when recordings are stored on some network drives
- shutil.copyfile(cache_path, file_path)
- os.remove(cache_path)
- rand_id = "".join(random.choices(string.ascii_lowercase + string.digits, k=6))
- Recordings.create(
- id=f"{start_time.timestamp()}-{rand_id}",
- camera=camera,
- path=file_path,
- start_time=start_time.timestamp(),
- end_time=end_time.timestamp(),
- duration=duration,
- )
- def run(self):
- # Check for new files every 5 seconds
- wait_time = 5
- while not self.stop_event.wait(wait_time):
- run_start = datetime.datetime.now().timestamp()
- self.move_files()
- wait_time = max(0, 5 - (datetime.datetime.now().timestamp() - run_start))
- logger.info(f"Exiting recording maintenance...")
- class RecordingCleanup(threading.Thread):
- def __init__(self, config: FrigateConfig, stop_event):
- threading.Thread.__init__(self)
- self.name = "recording_cleanup"
- self.config = config
- self.stop_event = stop_event
- def clean_tmp_clips(self):
- # delete any clips more than 5 minutes old
- for p in Path("/tmp/cache").rglob("clip_*.mp4"):
- logger.debug(f"Checking tmp clip {p}.")
- if p.stat().st_mtime < (datetime.datetime.now().timestamp() - 60 * 1):
- logger.debug("Deleting tmp clip.")
- p.unlink(missing_ok=True)
- def expire_recordings(self):
- logger.debug("Start expire recordings (new).")
- logger.debug("Start deleted cameras.")
- # Handle deleted cameras
- expire_days = self.config.record.retain_days
- expire_before = (
- datetime.datetime.now() - datetime.timedelta(days=expire_days)
- ).timestamp()
- no_camera_recordings: Recordings = Recordings.select().where(
- Recordings.camera.not_in(list(self.config.cameras.keys())),
- Recordings.end_time < expire_before,
- )
- deleted_recordings = set()
- for recording in no_camera_recordings:
- Path(recording.path).unlink(missing_ok=True)
- deleted_recordings.add(recording.id)
- logger.debug(f"Expiring {len(deleted_recordings)} recordings")
- Recordings.delete().where(Recordings.id << deleted_recordings).execute()
- logger.debug("End deleted cameras.")
- logger.debug("Start all cameras.")
- for camera, config in self.config.cameras.items():
- logger.debug(f"Start camera: {camera}.")
- # When deleting recordings without events, we have to keep at LEAST the configured max clip duration
- min_end = (
- datetime.datetime.now()
- - datetime.timedelta(seconds=config.record.events.max_seconds)
- ).timestamp()
- expire_days = config.record.retain_days
- expire_before = (
- datetime.datetime.now() - datetime.timedelta(days=expire_days)
- ).timestamp()
- expire_date = min(min_end, expire_before)
- # Get recordings to check for expiration
- recordings: Recordings = (
- Recordings.select()
- .where(
- Recordings.camera == camera,
- Recordings.end_time < expire_date,
- )
- .order_by(Recordings.start_time)
- )
- # Get all the events to check against
- events: Event = (
- Event.select()
- .where(
- Event.camera == camera,
- # need to ensure segments for all events starting
- # before the expire date are included
- Event.start_time < expire_date,
- Event.has_clip,
- )
- .order_by(Event.start_time)
- .objects()
- )
- # loop over recordings and see if they overlap with any non-expired events
- event_start = 0
- deleted_recordings = set()
- for recording in recordings.objects().iterator():
- keep = False
- # Now look for a reason to keep this recording segment
- for idx in range(event_start, len(events)):
- event = events[idx]
- # if the event starts in the future, stop checking events
- # and let this recording segment expire
- if event.start_time > recording.end_time:
- keep = False
- break
- # if the event is in progress or ends after the recording starts, keep it
- # and stop looking at events
- if event.end_time is None or event.end_time >= recording.start_time:
- keep = True
- break
- # if the event ends before this recording segment starts, skip
- # this event and check the next event for an overlap.
- # since the events and recordings are sorted, we can skip events
- # that end before the previous recording segment started on future segments
- if event.end_time < recording.start_time:
- event_start = idx
- # Delete recordings outside of the retention window
- if not keep:
- Path(recording.path).unlink(missing_ok=True)
- deleted_recordings.add(recording.id)
- logger.debug(f"Expiring {len(deleted_recordings)} recordings")
- Recordings.delete().where(Recordings.id << deleted_recordings).execute()
- logger.debug(f"End camera: {camera}.")
- logger.debug("End all cameras.")
- logger.debug("End expire recordings (new).")
- def expire_files(self):
- logger.debug("Start expire files (legacy).")
- default_expire = (
- datetime.datetime.now().timestamp()
- - SECONDS_IN_DAY * self.config.record.retain_days
- )
- delete_before = {}
- for name, camera in self.config.cameras.items():
- delete_before[name] = (
- datetime.datetime.now().timestamp()
- - SECONDS_IN_DAY * camera.record.retain_days
- )
- # find all the recordings older than the oldest recording in the db
- try:
- oldest_recording = Recordings.select().order_by(Recordings.start_time).get()
- p = Path(oldest_recording.path)
- oldest_timestamp = p.stat().st_mtime - 1
- except DoesNotExist:
- oldest_timestamp = datetime.datetime.now().timestamp()
- logger.debug(f"Oldest recording in the db: {oldest_timestamp}")
- process = sp.run(
- ["find", RECORD_DIR, "-type", "f", "!", "-newermt", f"@{oldest_timestamp}"],
- capture_output=True,
- text=True,
- )
- files_to_check = process.stdout.splitlines()
- for f in files_to_check:
- p = Path(f)
- if p.stat().st_mtime < delete_before.get(p.parent.name, default_expire):
- p.unlink(missing_ok=True)
- logger.debug("End expire files (legacy).")
- def run(self):
- # Expire recordings every minute, clean directories every hour.
- for counter in itertools.cycle(range(60)):
- if self.stop_event.wait(60):
- logger.info(f"Exiting recording cleanup...")
- break
- self.expire_recordings()
- self.clean_tmp_clips()
- if counter == 0:
- self.expire_files()
- remove_empty_directories(RECORD_DIR)
|