I was thinking of setting up a home surveilance system using Frigate, and integrating it with Home Assistant. I’d probably have somewhere on the order of 10-15 1080p 30fps cameras. I’m not sure what components I should get for the server, as I am unsure of the actual processing requirements.
EDIT 1: For some extra information, I did find that Frigate has a recommended hardware page.
The space requirements get super intense with many cameras like that unless you compress the video. And if you compress it then with so many you need the camera to do it on the fly otherwise your server needs to be really beefy to handle real time encoding of 10+ incoming video feeds. Also if the cameras don’t encode then the data flow would congest your network something fierce. Those requirements might push the cameras from cheap to not so cheap though (still far from expensive though imo)
The biggest issue as I see it with so many cameras would be how to find interesting stuff in all that data. If it’s only surveillance then sure you can just retain like a week of feeds and make a vacation mode where you store enough to cover the whole vacation. But if you want to look at what your dogs do etc then trying to track them across 10+ cameras is going to be tricky without some software help, unsure if there is anything open source for that.
The space requirements get super intense with many cameras like that unless you compress the video.
I think Frigate uses h264 if I remember correctly. Also I’m not planning on storing and archiving the recorded data. I most likely would only save a day or a couple days. You do raise a good point about vacations, though - I should probably have enough storage for possible vacations.
Also if the cameras don’t encode then the data flow would congest your network something fierce.
The newtork that the camera feeds would be flowing through would essentially be isolated from the rest of the network. I intend to hook the cameras up to a dedicated network switch, which would then be connected to the camera server.
The biggest issue as I see it with so many cameras would be how to find interesting stuff in all that data.
What’s nitce about Frigate, is that it uses OpenCV, and TensorFlow to analyze the video streams for moving objects.
More Information can be found on Frigate’s website.
That’s quite a few cameras. I would do an audit on how many you will actually need first, because you will likely find you could get by with 5-10.
In terms of what you’ll need - any Intel chip that supports QuickSync will likely do for the main
ffmpeg
processing of the image, but you will definitely want a Google Coral TPU. If you do end up needing 10-15 cameras, you may end up needing the M2 with dual TPU version of the Coral. You will also want some form of reliable storage for your clips (NAS local or NFS), as well as the ability to back up those clips/shots to the cloud somewhere.I’m personally running 4 cameras (3x1080 @ 15fps, 1x4k @ 25fps) through my ~7 year old Synology DS418play NAS using Surveillance Station as the first ingestion point, then restreaming from there to Frigate. Now that Surveillance Station can accept external events via webhook, I may look to swap the direction, and ingest into Frigate first, then restream out to Surveillance Station for long-term storage.
“Why not directly use Frigate?” I hear you ask. Mostly because Frigate is pretty static. It’s all set up via YAML with no config UI currently, whereas I can tweak stuff on Surveillance Station quite easily.
That’s quite a few cameras. I would do an audit on how many you will actually need first, because you will likely find you could get by with 5-10.
That’s a fair point. I haven’t actually methodically gone through to see exactly how many I would need just yet. The numbers that I chose were somewhat just ballpark off the top of my head.
You will also want some form of reliable storage for your clips
I am planning to give the camera server dedicated storage for the data. If I’m really feeling like splurging on it, I may look into getting WD Purple drives, or the like.
as well as the ability to back up those clips/shots to the cloud somewhere.
I’m not sure that I would need this very much. I’m mostly interested in a sort of ephemeral surveilance system; I only really need to store, at most, a few days, and then rewrite over it all.
I’m personally running 4 cameras (3x1080 @ 15fps, 1x4k @ 25fps) through my ~7 year old Synology DS418play NAS
Would you say that 15FPS is a good framerate for surveilance? Or could one get away with even less to lessen the resource requirements?
whereas I can tweak stuff on Surveillance Station quite easily.
What tweaking do you generally need to do for the camera server?
Just reporting back that I did the work last night to change the ingestion order for my cameras. I’m now using the
go2rtc
component offrigate
as the first ingestion point. That component is serving a restream to both Frigate and my NAS’ NVR. It’s working much better now, with less frame delay, and less CPU usage on the NAS.I’m not sure that I would need this very much. I’m mostly interested in a sort of ephemeral surveilance system; I only really need to store, at most, a few days, and then rewrite over it all.
This is exactly what I do. I simply cloud backup any event/object clips but only retain last 5 days. The cloud is if law enforcement needs it, or in the event of hardware failure/catastrophic house damage.
What tweaking do you generally need to do for the camera server?
Recording schedules change based on time of day/when we’re in/out of the house. This is all handled as automations through Home Assistant, but is set up through Surveillance Station NVR.
Would you say that 15FPS is a good framerate for surveilance? Or could one get away with even less to lessen the resource requirements?
If doing CPU-based motion analysis, you could use a lower quality stream (if available from the cameras to avoid transcoding load) for motion detection, then use that to trigger recording on a higher quality stream.
deleted by creator
This is exactly what I do. Record at 15FPS, event detection at 5FPS (at quarter resolution too).
you could use a lower quality stream (…) for motion detection, then use that to trigger recording on a higher quality stream.
Brilliant idea! Thank you for the suggestion!
If doing CPU-based motion analysis
Whyd do you specifically mention CPU-based motion analysis? Does this idea not work with the Google Coral TPU, for example?
I’m using ZoneMinder on my end with more rudimentary motion detection, hence CPU detection. (My current hardware is pre-IOMMU on the mobo, so no pci passthrough for me…)
That said, if you have hardware that can handle X (via CPU, GPU, TPU, etc), then you gotta decide how you want to spend that. Whether resources are spent more in analysis fps or evaluating higher detail frames is up to what you need.
For Frigate especially, you are going to want to use multiple Coral TPUs to handle inferencing. I really can’t speak to the CPU requirements but I know a lot of people like using the Intel NUCs with mic tier processors.