>> ffmpeg’s implementation of RTSP isn’t very good. Instead, you could use our >> RTSP client implementation, and just use ffmpeg for the H.264 video >> decoding. (This is what most people do.) > > I can't unfortunately. Another side (https://shinobi.video) uses > ffmpeg internally...
Feel free to file bug reports with that project. > BTW, is everything good with this SDP generated by live555: Almost, but not quite. There’s no "a=fmtp:” line that contains "sprop-parameter-sets”. In other words, the server has not found any H.264 SPS and PPS NAL units at all! > Looks like ffmpeg doesn't like 'npt=0-'? Feel free to file a bug report with “ffmpeg”. >>> Is it absolutely impossible to >>> discard all NALs for the given RTP stream on the server side until >>> SPS/PPS arrive? >> >> If your server is using our “H264VideoFileServerMediaSubsession” class, then >> it’s already doing this. However, the problem is that - for many H.264 >> video sources - the PPS and SPS NAL units appear only at the very start of >> the stream, and never thereafter. That’s why - if you set >> “reuseFirstSource” to True - the second and subsequent receivers will not >> get the SPS/PPS NAL units in the media stream. But, as I’ve already noted, >> this is not something that receivers should be relying on anyway. > > No, it's based on H264LiveServerMediaSubsession I don’t know what “H264LiveServerMediaSubsession” is. We have no class with that name,. Ross Finlayson Live Networks, Inc. http://www.live555.com/ _______________________________________________ live-devel mailing list live-devel@lists.live555.com http://lists.live555.com/mailman/listinfo/live-devel