On Jun 12, 2009, at 8:59 AM, Adam Kocoloski wrote:

Hi Damien, I'm not sure I follow. My worry was that, if I built a replicator which only queried _changes to get the list of updates, I'd have to be prepared to process a very large response. I thought one smart way to process this response was to throttle the download at the TCP level by putting the socket into passive mode.

You will have a very large response, but you can stream it, processing one line at a time, then you discard the line and process the next. As long as the writer is using a blocking socket and the reader is only reading as much data as necessary to process a line, you never need to store much of the data in memory on either side. But it seems the HTTP client is buffering the data as it comes in, perhaps unintentionally.

With TCP, the sending side will only send so much data before getting an ACK, acknowledgment that packets sent were actually received. When an ACK isn't received, the sender stops sending, and the TCP calls will block at the sender (or return an error if the socket is in non- blocking mode), until it gets a response or socket timeout.

So if you have a non-buffering reader and a blocking sender, then you can stream the data and only relatively small amounts of data are buffered at any time. The problem is the reader in the HTTP client isn't waiting for the data to be demanded at all, instead as soon as data comes in, it sends it to a receiving erlang process. Erlang processes never block to receive messages, so there is no limit to the amount of data buffered. So if the Erlang process can't process the data fast enough, it starts getting buffered in it's mailbox, consuming unlimited memory.

Assuming I understand the problem correctly, the way to fix it is to have the HTTP client not read the data until it's demanded by the consuming process. Then we are only using the default TCP buffers, not the Erlang message queues as a buffer, and the total amount of memory used at anytime is small.

-Damien



I agree that the HTTP client seems to be at fault, because the option that it exposes to switch to passive mode seems to be a no- op. What exactly did you mean by "streams the data while not buffering the data"? Best,

Adam

On Jun 12, 2009, at 8:03 AM, Damien Katz wrote:

I don't think this is TCPs fault, it's the HTTP client. We need a HTTP client that streams data while not buffering the data (low level TCP already buffers some), instead of sending all the data that comes in to the waiting process, essentially buffering everything.

-Damien


On Jun 11, 2009, at 4:14 PM, Adam Kocoloski wrote:

I had some time to work on a replicator that queries _changes instead of _all_docs_by_seq today. The first question that came to my mind was how to put a spigot on the firehose. If I call _changes without a "since" qs parameter on a 10M document DB I'm going to get 10M chunks of output back.

I thought I might be able to control the flow at the TCP socket level using the inets HTTP client's {stream,{self,once}} option. I still think this would be an elegant option if I can get it to work, but my early tests show that all the chunks still show up immediately in the calling process regardless of whether I stream to self or {self,once}.

All for now, Adam



Reply via email to