Yes it takes about 2-3min for the timeout to appear the query itself should finish in that time. The files are not that big for debugging. I have, but I couldn't find anything relevant or helpful in my situation so far.
On Wed, 20 Sep 2017 at 20:41 Kunal Khatua <[email protected]> wrote: > Do you know in how much time does this timeout occur? There might be some > tuning needed to increase a timeout. Also, I think this (S3 specifically) > has been seen before... So you might find a solution within the mailing > list archives. Did you try looking there? > > > > From: Alan Höng > Sent: Wednesday, September 20, 8:46 AM > Subject: User client timeout with results > 2M rows > To: [email protected] > > > Hello, > > I'm getting errors when trying to fetch results from drill with queries > that evaluate to bigger tables. Surprisingly it works like a charm if the > returned table has less than 2M rows. It also seems like the query is > executed and finishes successfully.... > > I'm querying parquet files with GZIP compression on S3. I'm running drill > in distributed mode with zookeeper. I use version 1.9 from the container > available on dockerhub "harisekhon/apache-drill:1.9". I'm using the pydrill > package which uses the rest api to submit queries and gather results. > > I get the following error message from the client: > > TransportError(500, '{\n "errorMessage" : "CONNECTION ERROR: Connection / > 172.19.0.3:52382<http://172.19.0.3:52382> <--> ef53daab0ef8/ > 172.19.0.6:31010<http://172.19.0.6:31010> (user client) closed > unexpectedly. Drillbit down?\\n\\n\\n[Error Id: > 6a19835b-2325-431e-9bad-dde8f1d3c192 ]"\n}' > > I would appreciate any help with this. > > Best > Alan Höng > > >
