@RR, i would like to inform you that the issue regarding failed stream for 1st request is solved. Varnish was removing content-length header for 1st request . Enabling Esi processing has resolved this issue.
set beresp.do_esi = true; http://stackoverflow.com/questions/23643233/how-do-i-disable-transfer-encoding-chunked-encoding-in-varnish thanks !! On Sat, Sep 27, 2014 at 10:41 AM, shahzaib shahzaib <[email protected]> wrote: > >>In general it shouldn’t since the ‘?start=’ is handled by nginx and not > varnish, but I’m not exactly sure how the mp4 module of nginx handles a > proxied request. > You have to test it. > > Sure, i'll test it. > > sub vcl_fetch { > return (pass); > } > > You're right about return(pass), coalescing doesn't work with pass. > > >>In worst case scenario imho only the first request (before landing on > the proxy_store server) will “fail” eg play from the beginning instead of > the time set. > Well, i am facing more worse scenario that first request always fail to > stream and player(HTML5) keeps on loading. > > I'm already checking if there's some config issue with varnish or this is > the default behaviour(Which i don't think it is). > > Thanks @RR > > Shahzaib > > > On Fri, Sep 26, 2014 at 2:36 AM, Reinis Rozitis <[email protected]> wrote: > >> It will also prevent users seeking the video because the arguments after >>> "?" will remove whenever user will try to seek the video stream, isn't it ? >>> >> >> In general it shouldn’t since the ‘?start=’ is handled by nginx and not >> varnish, but I’m not exactly sure how the mp4 module of nginx handles a >> proxied request. >> You have to test it. >> >> In worst case scenario imho only the first request (before landing on the >> proxy_store server) will “fail” eg play from the beginning instead of the >> time set. >> >> >> >> Well, only proxy_store is able to fulfill my requirements that is the >>> reason i'll have to stick with it. >>> >> >> Well you can try to use varnish as the streamer, just need some >> (web)player supporting byte-range requests for the seeking ( >> http://flash.flowplayer.org/plugins/streaming/pseudostreaming.html ). >> >> >> I am bit confused about the varnish. Actually, i don't need any kind of >>> caching within the varnish as nginx already doing it via proxy_store. I >>> just need varnish to merge the subsequent requests into 1 and forward it to >>> nginx and i think varnish is doing it pretty well. Nevertheless, i am >>> confused if malloc caching will have any odd effect on the stream behavior ? >>> >> >> >> You can try to pass the request without caching: >> >> sub vcl_fetch { >> return (pass); >> } >> >> (maybe even do it in the vcl_recv stage but again I'm not exactly sure if >> in that case the request coalescing works). >> >> >> >> rr >> _______________________________________________ >> nginx mailing list >> [email protected] >> http://mailman.nginx.org/mailman/listinfo/nginx >> > >
_______________________________________________ nginx mailing list [email protected] http://mailman.nginx.org/mailman/listinfo/nginx
