Am 08.03.2022 um 19:09 schrieb Steven Kan:
After 7.5 years of waiting, my banana plant is finally flowering! I want to do
a time-lapse capture of the flowering and fruiting process. Due to its
location, the easiest way for me to get a camera out there is to use a little
WyzeCam v3 with the RTSP firmware and the Wyze lamp socket. Unfortunately the
WyzeCam doesn’t (yet) have a externally accessible JPG snapshot feature, so I
have a cron job set up to:
./ffmpeg -rtsp_transport tcp -i rtsp://anonymous:password@$IPAddress/live
-frames:v 1 $outfile
every hour. The results are OK, but not fantastic:
https://www.kan.org/pictures/BananaTimeLapseFirstImage.jpg
<https://www.kan.org/pictures/BananaTimeLapseFirstImage.jpg>
Is there a way to tell ffmpeg to collect N frames of video and output one
single averaged image to improve the SNR? Even if there’s some wind, the flower
stalk shouldn’t be moving much.
I tried:
./ffmpeg -rtsp_transport tcp -i rtsp://anonymous:[email protected]/live
-frames:v 10 ~/BananaLapse/MultiFrame%03d.jpg
and that results in N JPGs. I suppose I could have a second ffmpeg command that
averages those 10 JPGs, but can this all be done in one pass? Thanks!
You can use the "tmix" filter before you extract the images from the video.
Michael
_______________________________________________
ffmpeg-user mailing list
[email protected]
https://ffmpeg.org/mailman/listinfo/ffmpeg-user
To unsubscribe, visit link above, or email
[email protected] with subject "unsubscribe".