I'm don't know a lot about checking the limits, but I'm hive (on hdfs). Im
able to perform a count on a hive table, so I don't think the issue is
there. The hive client and drillbit sate on different nodes though. I did a
ulimit -a and got:

core file size          (blocks, -c) 0
data seg size           (kbytes, -d) unlimited
scheduling priority             (-e) 0
file size               (blocks, -f) unlimited
pending signals                 (-i) 31403
max locked memory       (kbytes, -l) 64
max memory size         (kbytes, -m) unlimited
open files                      (-n) 1024
pipe size            (512 bytes, -p) 8
POSIX message queues     (bytes, -q) 819200
real-time priority              (-r) 0
stack size              (kbytes, -s) 10240
cpu time               (seconds, -t) unlimited
max user processes              (-u) 1024
virtual memory          (kbytes, -v) unlimited
file locks                      (-x) unlimited


If you have any steps you'd like me to take to help debug, I'd be happy to
attempt them.



On Monday, January 11, 2016, Hanifi GUNES <[email protected]> wrote:

> Interesting. Can you check your procfs to confirm Drill opens excessive
> number of file descriptors after making sure that query is complete through
> Web UI? If so, we likely have some leak. Also, can you confirm soft & hard
> limits for open file descriptors are not set too low on your platform?
>
> -Hanifi
>
> 2016-01-11 13:13 GMT-08:00 Ian Maloney <[email protected]
> <javascript:;>>:
>
> > Hanifi,
> >
> > Thanks for the suggestions, I'm not using any concurrency at the moment.
> In
> > the web UI under profiles I see "No running queries", the error is still
> > happening, even after restarting the bit.
> >
> >
> >
> > On Monday, January 11, 2016, Hanifi GUNES <[email protected]
> <javascript:;>> wrote:
> >
> > > -* Any ideas on how to **keep the (suspected) resource leak from
> > > happening?*
> > >
> > > The answer to this depends on your workload as well. You have mentioned
> > you
> > > are running a lot of queries so Drill might ordinarily open too many
> > > descriptors to serve the high demand. Of course, this assumes that you
> do
> > > not limit the level of concurrency. If so, why don't you try enabling
> > > queues to bound concurrent execution and run your workload again? [1]
> > >
> > > You can always open up web UI to see if your queries are completed or
> > > hanging around. If you see too many queries pending and stuck for a
> long
> > > time, it is a good indicator of the problem I described above [2]
> > >
> > >
> > > -Hanifi
> > >
> > > 1: https://drill.apache.org/docs/enabling-query-queuing/
> > > 2: https://drill.apache.org/docs/query-profiles/
> > >
> > > 2016-01-11 10:35 GMT-08:00 Ian Maloney <[email protected]
> <javascript:;>
> > > <javascript:;>>:
> > >
> > > > Hi Abdel,
> > > >
> > > > I just noticed I'm still on v. 1.2, so maybe upgrading will help. I
> > could
> > > > open a Jira, if need be.
> > > >
> > > > As far as restarting and reproducing, I restarted and after running
> my
> > > app
> > > > with the jdbc logic for a while, I get an IOException: Error
> accessing
> > /
> > > > I paused the app and restarted the specific bit from the jdbc
> > connection
> > > > below, which gave me the "Too many open files" exception again.
> > > >
> > > > Now, restarting that bit won't fix the error.
> > > >
> > > >
> > > > On Monday, January 11, 2016, Abdel Hakim Deneche <
> > [email protected] <javascript:;>
> > > <javascript:;>>
> > > > wrote:
> > > >
> > > > > Hi Ian,
> > > > >
> > > > > Can you open up a JIRA for this ? is it easy to reproduce ?
> > > > >
> > > > > Thanks
> > > > >
> > > > > On Mon, Jan 11, 2016 at 8:59 AM, Ian Maloney <
> > > > [email protected] <javascript:;> <javascript:;>
> > > > > <javascript:;>>
> > > > > wrote:
> > > > >
> > > > > > Hi,
> > > > > >
> > > > > > I've been running a lot of queries via jdbc/drill. I have four
> > > > drillbits,
> > > > > > but I could not get the zk jdbc URL to work so I used:
> > > > > > jdbc:drill:drillbit=a-bits-hostname
> > > > > >
> > > > > > Now I get a SocketException for too many open files, even when
> > > > accessing
> > > > > > via cli. I imagine I could restart the bits, but for something
> > > intended
> > > > > for
> > > > > > production, that doesn't seem like a viable solution. Any ideas
> on
> > > how
> > > > to
> > > > > > keep the (suspected) resource leak from happening?
> > > > > >
> > > > > > I'm closing ResultSet, Statement, and Connection, after each
> query.
> > > > > >
> > > > >
> > > > >
> > > > >
> > > > > --
> > > > >
> > > > > Abdelhakim Deneche
> > > > >
> > > > > Software Engineer
> > > > >
> > > > >   <http://www.mapr.com/>
> > > > >
> > > > >
> > > > > Now Available - Free Hadoop On-Demand Training
> > > > > <
> > > > >
> > > >
> > >
> >
> http://www.mapr.com/training?utm_source=Email&utm_medium=Signature&utm_campaign=Free%20available
> > > > > >
> > > > >
> > > >
> > >
> >
>

Reply via email to