That sounds weird (maybe a reporting issue?), please send a profile
that shows the mem_limit set and the error message referring to the
estimated memory instead - thanks!

On 19 June 2018 at 04:59, Fawze Abujaber <fawz...@gmail.com> wrote:
> Hi Jezy,
>
> Thanks for your quick response.
>
> I’m using impala 2.10, I think since I’m using max memory memory per poo
> with the default memory limit, and here the estimate have to take place in
> order to estimate the max concurrent queries, and if the estimate beyond the
> max memory then the query will not submitted.
>
> In my case I set both values and getting errors of high memory needed for
> the query, it’s not occurring very often.
>
> On Mon, 18 Jun 2018 at 16:41 Jeszy <jes...@gmail.com> wrote:
>>
>> Hey Fawze,
>>
>> Default Query Memory Limit applies here, yes.
>> If you submit a query to a pool with that setting, you should see
>> something like this in the profile:
>> Query Options (set by configuration): MEM_LIMIT=X
>>
>> (YMMV based on version - what version are you running on?)
>> If MEM_LIMIT is present in that line, Impala will (should) disregard
>> estimates.
>>
>> Thanks!
>>
>> On 17 June 2018 at 21:04, Fawze Abujaber <fawz...@gmail.com> wrote:
>> > Hi Jeszy,
>> >
>> > Thanks for your response, Indeed this is what i was thinking about but,
>> > I
>> > have  Default Query Memory Limit and Max memory set per pool which i
>> > think
>> > should be enough to cover this, shouldnot it?  or i should pass the
>> > mem_limit in the default query options?
>> >
>> > On Sun, Jun 17, 2018 at 8:36 PM, Jeszy <jes...@gmail.com> wrote:
>> >>
>> >> Hello Fawze,
>> >>
>> >> Disabling this, per se, is not an option, but an equally simple
>> >> workaround
>> >> is using MEM_LIMIT.
>> >>
>> >> The estimated stats are often very far from actual memory usage and
>> >> shouldn't be relied on - a best practice is to set MEM_LIMIT as a query
>> >> option (preferably have a default value set for each pool). Having that
>> >> set
>> >> will cause Impala to ignore the estimates and rely on this limit for
>> >> admission control purposes. This works decently for well-understood
>> >> workloads (ie. where the memory consumption is known to fit within
>> >> certain
>> >> limits). For ad-hoc workloads, if the query can't be executed within
>> >> the
>> >> default limit of the pool, you can override the limit on a per-query
>> >> basis
>> >> (just issue 'set MEM_LIMIT=...' before running the query).
>> >>
>> >> HTH
>> >>
>> >> On 16 June 2018 at 13:31, Fawze Abujaber <fawz...@gmail.com> wrote:
>> >>>
>> >>> Hi Community ,
>> >>>
>> >>> In the last impala versions, impala is estimating the memory required
>> >>> for
>> >>> the query, and in case the estimated memory required beyonds the
>> >>> configured
>> >>> memory or  the configured memory per pool, impala is not submitting
>> >>> this
>> >>> query, taking the fact that many times and specially running query on
>> >>> tables
>> >>> without stats there is a huge difference between the estimated and the
>> >>> actual memory used, the estimated can be 31 GB per node and the actual
>> >>> use
>> >>> is 1 or 2 GB, that’s mean to submit a query I need at least 1.5 T
>> >>> memory
>> >>> configured which I see it too much.
>> >>>
>> >>> I’m curios to know if there is an option making this to configuration
>> >>> (
>> >>> Un submitting query if we he estimated memory required beyond the
>> >>> configured
>> >>> memory) as an optional choice.
>> >>>
>> >>> Such issue can block using impala dynamic resource pools.
>> >>> --
>> >>> Take Care
>> >>> Fawze Abujaber
>> >>
>> >>
>> >
>> >
>> >
>> > --
>> > Take Care
>> > Fawze Abujaber
>
> --
> Take Care
> Fawze Abujaber

Reply via email to