> > > This patch is seriously broken.  Request a very large file (100MB or greater) 
>and watch what
> happens
> > > to memory usage.
> > >
> > > The problem is this loop. We basically read the entire content of the file into 
>memory before
> > > sending it out on the network. Haven't given much though on the best way to fix 
>this.
> > >
> > > >   +                APR_BRIGADE_FOREACH(bucket, b) {
> > > >   +                    const char *str;
> > > >   +                    apr_size_t n;
> > > >   +
> > > >   +                    rv = apr_bucket_read(bucket, &str, &n, APR_BLOCK_READ);
> > > >   +                    apr_brigade_write(ctx->b, NULL, NULL, str, n);
> > > >   +                }
> >
> > I don't see how that could happen.  We only enter that section of the
> > core_output_filter if we are saving some data off to the side for
> > keepalive requests.  In fact, we specifically do not enter this loop if we
> > are serving a file from disk.
>
> Attach a debugger and watch what happens.  I am seeing the following buckets...
>
> 1 heap bucket containing the headers
> 1 file bucket with the file descriptor
> 1 eos bucket
>
> The following code is hit and we enter the conditional because the last bucket was 
>an eos and the
> connection is keep-alive.
>
> if ((!fd && !more &&
>              (nbytes < AP_MIN_BYTES_TO_WRITE) && !APR_BUCKET_IS_FLUSH(e))
>             || (APR_BUCKET_IS_EOS(e) && c->keepalive)) {
>
> I think the logic in the conditional is just wrong.

I agree completely.  I think I can fix this in a few minutes.  Watch for a
patch.

Ryan

_______________________________________________________________________________
Ryan Bloom                              [EMAIL PROTECTED]
406 29th St.
San Francisco, CA 94131
-------------------------------------------------------------------------------

Reply via email to