Also, removing arguments after "?" also disabled the pseudo streaming. So i think i can't apply this method !!
On Mon, Sep 29, 2014 at 6:05 PM, shahzaib shahzaib <[email protected]> wrote: > @RR, i would like to inform you that the issue regarding failed stream for > 1st request is solved. Varnish was removing content-length header for 1st > request . Enabling Esi processing has resolved this issue. > > set beresp.do_esi = true; > > > http://stackoverflow.com/questions/23643233/how-do-i-disable-transfer-encoding-chunked-encoding-in-varnish > > thanks !! > > On Sat, Sep 27, 2014 at 10:41 AM, shahzaib shahzaib <[email protected] > > wrote: > >> >>In general it shouldn’t since the ‘?start=’ is handled by nginx and not >> varnish, but I’m not exactly sure how the mp4 module of nginx handles a >> proxied request. >> You have to test it. >> >> Sure, i'll test it. >> >> sub vcl_fetch { >> return (pass); >> } >> >> You're right about return(pass), coalescing doesn't work with pass. >> >> >>In worst case scenario imho only the first request (before landing on >> the proxy_store server) will “fail” eg play from the beginning instead of >> the time set. >> Well, i am facing more worse scenario that first request always fail to >> stream and player(HTML5) keeps on loading. >> >> I'm already checking if there's some config issue with varnish or this is >> the default behaviour(Which i don't think it is). >> >> Thanks @RR >> >> Shahzaib >> >> >> On Fri, Sep 26, 2014 at 2:36 AM, Reinis Rozitis <[email protected]> wrote: >> >>> It will also prevent users seeking the video because the arguments after >>>> "?" will remove whenever user will try to seek the video stream, isn't it ? >>>> >>> >>> In general it shouldn’t since the ‘?start=’ is handled by nginx and not >>> varnish, but I’m not exactly sure how the mp4 module of nginx handles a >>> proxied request. >>> You have to test it. >>> >>> In worst case scenario imho only the first request (before landing on >>> the proxy_store server) will “fail” eg play from the beginning instead of >>> the time set. >>> >>> >>> >>> Well, only proxy_store is able to fulfill my requirements that is the >>>> reason i'll have to stick with it. >>>> >>> >>> Well you can try to use varnish as the streamer, just need some >>> (web)player supporting byte-range requests for the seeking ( >>> http://flash.flowplayer.org/plugins/streaming/pseudostreaming.html ). >>> >>> >>> I am bit confused about the varnish. Actually, i don't need any kind of >>>> caching within the varnish as nginx already doing it via proxy_store. I >>>> just need varnish to merge the subsequent requests into 1 and forward it to >>>> nginx and i think varnish is doing it pretty well. Nevertheless, i am >>>> confused if malloc caching will have any odd effect on the stream behavior >>>> ? >>>> >>> >>> >>> You can try to pass the request without caching: >>> >>> sub vcl_fetch { >>> return (pass); >>> } >>> >>> (maybe even do it in the vcl_recv stage but again I'm not exactly sure >>> if in that case the request coalescing works). >>> >>> >>> >>> rr >>> _______________________________________________ >>> nginx mailing list >>> [email protected] >>> http://mailman.nginx.org/mailman/listinfo/nginx >>> >> >> >
_______________________________________________ nginx mailing list [email protected] http://mailman.nginx.org/mailman/listinfo/nginx
