On 15/07/2016 10:20, Carl Eugen Hoyos wrote:

You should teach FFmpeg to read the frames from the camera.
Any ideas how?  Should I consider writing my own container
format library for FFmpeg?
I meant instead of implementing the API in you own piece of
software you could use it in FFmpeg (or reverse engineer it).
Thanks for the advice. I managed to find a workable solution. I have tested piping on the command line before, e.g. "myfeedprogram | ffmpeg -i - [lots of parameters]" and this worked. The problem was that I needed up to three simultaneous streams: live video, video recording and snapshots. This couldn't work because I couldn't have three instances of the same software trying to access the one camera.

The solution was simple, create a pipe, fork the process, connect output end of pipe to stdin of the child process, replace child process with ffmpeg (exec()), then feed raw frames into pipe. This can be done as many times as necessary. So I have one pogram capturing frames from the camera, but then "piping" it to 3 separate instances of ffmpeg.

Thanks again.
Charl
_______________________________________________
ffmpeg-user mailing list
[email protected]
http://ffmpeg.org/mailman/listinfo/ffmpeg-user

To unsubscribe, visit link above, or email
[email protected] with subject "unsubscribe".

Reply via email to