Isnt merging is same as muxing two elementry stream means audio n video in ur cases On Jun 17, 2013 12:32 PM, "jona" <[email protected]> wrote:
> Hello! > > I'm looking to merge various audio and video sources into one. The sources > need to start at different time offsets. My guess is to use avfilter lib. > Would that do the trick? Is that the right direction? > > If avfilter lib is the correct direction I would greatly appreciate if > someone could provide some pseudo code on how I could be done? I'm having a > hard time finding some documentation or examples about using filters. > > I have looked into the filtering_video.c example but I'm not 100% sure how > to go about setting it up and all. > > How do I feed the various audio and video frames to the different input > filter buffers? I'm not sure how the offsets are set or handled or if the > avfilter lib handles that side of things. > > Thanks for any help or direction about this. > - Jona > > > > -- > View this message in context: > http://libav-users.943685.n4.nabble.com/Merging-various-audio-and-video-sources-into-one-using-avfilter-lib-tp4657898.html > Sent from the libav-users mailing list archive at Nabble.com. > _______________________________________________ > Libav-user mailing list > [email protected] > http://ffmpeg.org/mailman/listinfo/libav-user >
_______________________________________________ Libav-user mailing list [email protected] http://ffmpeg.org/mailman/listinfo/libav-user
