[jira] [Commented] (IMPALA-7505) impalad webserver hang in getting a lock of QueryExecStatus

2018-08-29 Thread Chen Wang (JIRA)


[ 
https://issues.apache.org/jira/browse/IMPALA-7505?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16597045#comment-16597045
 ] 

Chen Wang commented on IMPALA-7505:
---

[~jeszyb] Thanks. we reproduced IMPALA-1972 on our environment, and plan to 
apply the patch to our codebase. Thank you very much.

> impalad webserver hang in getting a lock of QueryExecStatus
> ---
>
> Key: IMPALA-7505
> URL: https://issues.apache.org/jira/browse/IMPALA-7505
> Project: IMPALA
>  Issue Type: Bug
>  Components: Backend
>Affects Versions: Impala 2.8.0
>Reporter: Chen Wang
>Priority: Major
> Attachments: gdb.out
>
>
> Impalad’s webserver would hang sometimes.
> The following is one of the cases: the webserver threads stuck in getting a 
> lock of QueryExecStatus,  but I can't find where the lock is acquired in the 
> stack. The web requests are sent from the agent of CDH, which is to check the 
> activity of impalad.
> Full gdb log is in the attachment.
> {code}
> Thread 116 (Thread 0x7f288f5e1700 (LWP 31062)):
> #0  0x00378780e334 in __lll_lock_wait () from /lib64/libpthread.so.0
> #1  0x0037878095d8 in _L_lock_854 () from /lib64/libpthread.so.0
> #2  0x0037878094a7 in pthread_mutex_lock () from /lib64/libpthread.so.0
> #3  0x008d6eb8 in pthread_mutex_lock (this=0xcab4f50) at 
> /data/impala/toolchain/boost-1.57.0/include/boost/thread/pthread/mutex.hpp:62
> #4  boost::mutex::lock (this=0xcab4f50) at 
> /data/impala/toolchain/boost-1.57.0/include/boost/thread/pthread/mutex.hpp:116
> #5  0x00b7903c in lock_guard (this=0xa7b5800, query_id=) at 
> /data/impala/toolchain/boost-1.57.0/include/boost/thread/lock_guard.hpp:38
> #6  impala::ImpalaServer::GetRuntimeProfileStr (this=0xa7b5800, query_id=) at 
> /data/impala/be/src/service/impala-server.cc:573
> #7  0x00ba6a8c in 
> impala::ImpalaHttpHandler::QueryProfileEncodedHandler (this=0x3f56be0, args=) 
> at /data/impala/be/src/service/impala-http-handler.cc:219
> #8  0x00cafe75 in operator() (this=) at 
> /data/impala/toolchain/boost-1.57.0/include/boost/function/function_template.hpp:767
> #9  impala::Webserver::RenderUrlWithTemplate (this=) at 
> /data/impala/be/src/util/webserver.cc:443
> #10 0x00cb1295 in impala::Webserver::BeginRequestCallback (this=) at 
> /data/impala/be/src/util/webserver.cc:414
> #11 0x00cc4850 in handle_request ()
> #12 0x00cc6fcd in process_new_connection ()
> #13 0x00cc765d in worker_thread ()
> #14 0x003787807aa1 in start_thread () from /lib64/libpthread.so.0
> #15 0x0037874e8bcd in clone () from /lib64/libc.so.6
> {code}
> The hang situation appears on impala 2.8.0, but I found that code of 
> be/service part hasn’t changed much from 2.8.0 to 2.11.0. so the problem may 
> still exists.
> Hope you experts can give me some guidance of finding the root cause, or 
> workaround plans to deal with these hang situation.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Assigned] (IMPALA-7503) SHOW GRANT USER not showing all privileges

2018-08-29 Thread Adam Holley (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-7503?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Adam Holley reassigned IMPALA-7503:
---

Assignee: Adam Holley

> SHOW GRANT USER not showing all privileges
> --
>
> Key: IMPALA-7503
> URL: https://issues.apache.org/jira/browse/IMPALA-7503
> Project: IMPALA
>  Issue Type: Sub-task
>  Components: Frontend
>Affects Versions: Impala 3.1.0
>Reporter: Adam Holley
>Assignee: Adam Holley
>Priority: Major
>
> SHOW GRANT USER will only show the privileges explicitly granted to the user. 
>  It should show all the privileges that are granted to any groups/roles the 
> user belongs to.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Resolved] (IMPALA-7502) ALTER TABLE RENAME should require ALL on the source table

2018-08-29 Thread Fredy Wijaya (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-7502?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Fredy Wijaya resolved IMPALA-7502.
--
   Resolution: Fixed
Fix Version/s: Impala 3.1.0

> ALTER TABLE RENAME should require ALL on the source table
> -
>
> Key: IMPALA-7502
> URL: https://issues.apache.org/jira/browse/IMPALA-7502
> Project: IMPALA
>  Issue Type: Bug
>  Components: Frontend
>Affects Versions: Impala 3.0
>Reporter: Fredy Wijaya
>Assignee: Fredy Wijaya
>Priority: Major
>  Labels: security
> Fix For: Impala 3.1.0
>
>
> For ALTER TABLE RENAME, currently Impala requires ALTER on the source table 
> and CRATE on the destination database. This may pose a security risk as 
> described in: https://issues.apache.org/jira/browse/SENTRY-2264. The solution 
> is to require ALL at the source table and CREATE at the destination database.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Updated] (IMPALA-7503) SHOW GRANT USER not showing all privileges

2018-08-29 Thread Fredy Wijaya (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-7503?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Fredy Wijaya updated IMPALA-7503:
-
Issue Type: Sub-task  (was: Bug)
Parent: IMPALA-7075

> SHOW GRANT USER not showing all privileges
> --
>
> Key: IMPALA-7503
> URL: https://issues.apache.org/jira/browse/IMPALA-7503
> Project: IMPALA
>  Issue Type: Sub-task
>  Components: Frontend
>Affects Versions: Impala 3.1.0
>Reporter: Adam Holley
>Priority: Major
>
> SHOW GRANT USER will only show the privileges explicitly granted to the user. 
>  It should show all the privileges that are granted to any groups/roles the 
> user belongs to.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Created] (IMPALA-7509) Create table after drop can lead to table not found exception

2018-08-29 Thread Rituparna Agrawal (JIRA)
Rituparna Agrawal created IMPALA-7509:
-

 Summary: Create table after drop can lead to table not found 
exception
 Key: IMPALA-7509
 URL: https://issues.apache.org/jira/browse/IMPALA-7509
 Project: IMPALA
  Issue Type: Sub-task
Reporter: Rituparna Agrawal
 Attachments: failure_snippet.txt

There are two impalads. One running with old mode and one with fetch from 
catalogd mode.

Now create a table and drop it in the impalad running in old mode. Following 
this create the same table from the new mode coodinator. It sometimes throws 
table not found exception.

Test Create and Drop of a table in V1
  Running  1  iterations
Going for :  CREATE TABLE PARTS ( PART_ID DOUBLE, CREATE_TIME DOUBLE, 
LAST_ACCESS_TIME DOUBLE, PART_NAME STRING, SD_ID DOUBLE, TBL_ID DOUBLE) STORED 
AS PARQUETFILE;
Going for :  DROP TABLE PARTS;
Test Create and Drop of a table in V2
  Running  1  iterations
Going for :  CREATE TABLE PARTS ( PART_ID DOUBLE, CREATE_TIME DOUBLE, 
LAST_ACCESS_TIME DOUBLE, PART_NAME STRING, SD_ID DOUBLE, TBL_ID DOUBLE) STORED 
AS PARQUETFILE;
Traceback (most recent call last):
  File "testing.py", line 21, in execute_query
cursor.execute(query)
  File 
"/Users/parna/workspace/virtual_envs/metadata/lib/python2.7/site-packages/impala/hiveserver2.py",
 line 302, in execute
configuration=configuration)
  File 
"/Users/parna/workspace/virtual_envs/metadata/lib/python2.7/site-packages/impala/hiveserver2.py",
 line 343, in execute_async
self._execute_async(op)
  File 
"/Users/parna/workspace/virtual_envs/metadata/lib/python2.7/site-packages/impala/hiveserver2.py",
 line 362, in _execute_async
operation_fn()
  File 
"/Users/parna/workspace/virtual_envs/metadata/lib/python2.7/site-packages/impala/hiveserver2.py",
 line 340, in op
async=True)
  File 
"/Users/parna/workspace/virtual_envs/metadata/lib/python2.7/site-packages/impala/hiveserver2.py",
 line 1027, in execute
return self._operation('ExecuteStatement', req)
  File 
"/Users/parna/workspace/virtual_envs/metadata/lib/python2.7/site-packages/impala/hiveserver2.py",
 line 957, in _operation
resp = self._rpc(kind, request)
  File 
"/Users/parna/workspace/virtual_envs/metadata/lib/python2.7/site-packages/impala/hiveserver2.py",
 line 925, in _rpc
err_if_rpc_not_ok(response)
  File 
"/Users/parna/workspace/virtual_envs/metadata/lib/python2.7/site-packages/impala/hiveserver2.py",
 line 704, in err_if_rpc_not_ok
raise HiveServer2Error(resp.status.errorMessage)
HiveServer2Error: LocalCatalogException: Could not load table parnatest.parts 
from metastore
CAUSED BY: TException: 
TGetPartialCatalogObjectResponse(status:TStatus(status_code:GENERAL, 
error_msgs:[CatalogException: Table not found: parts]))

Going for :  DROP TABLE PARTS;
Traceback (most recent call last):
  File "testing.py", line 21, in execute_query
cursor.execute(query)
  File 
"/Users/parna/workspace/virtual_envs/metadata/lib/python2.7/site-packages/impala/hiveserver2.py",
 line 302, in execute
configuration=configuration)
  File 
"/Users/parna/workspace/virtual_envs/metadata/lib/python2.7/site-packages/impala/hiveserver2.py",
 line 343, in execute_async
self._execute_async(op)
  File 
"/Users/parna/workspace/virtual_envs/metadata/lib/python2.7/site-packages/impala/hiveserver2.py",
 line 362, in _execute_async
operation_fn()
  File 
"/Users/parna/workspace/virtual_envs/metadata/lib/python2.7/site-packages/impala/hiveserver2.py",
 line 340, in op
async=True)
  File 
"/Users/parna/workspace/virtual_envs/metadata/lib/python2.7/site-packages/impala/hiveserver2.py",
 line 1027, in execute
return self._operation('ExecuteStatement', req)
  File 
"/Users/parna/workspace/virtual_envs/metadata/lib/python2.7/site-packages/impala/hiveserver2.py",
 line 957, in _operation
resp = self._rpc(kind, request)
  File 
"/Users/parna/workspace/virtual_envs/metadata/lib/python2.7/site-packages/impala/hiveserver2.py",
 line 925, in _rpc
err_if_rpc_not_ok(response)
  File 
"/Users/parna/workspace/virtual_envs/metadata/lib/python2.7/site-packages/impala/hiveserver2.py",
 line 704, in err_if_rpc_not_ok
raise HiveServer2Error(resp.status.errorMessage)
HiveServer2Error: LocalCatalogException: Could not load table parnatest.parts 
from metastore
CAUSED BY: TException: 
TGetPartialCatalogObjectResponse(status:TStatus(status_code:GENERAL, 
error_msgs:[CatalogException: Table not found: parts]))


Corresponding Impalad log will attached as well. 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-7498) impalad should wait for catalogd during start up

2018-08-29 Thread Philip Zeyliger (JIRA)


[ 
https://issues.apache.org/jira/browse/IMPALA-7498?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16596754#comment-16596754
 ] 

Philip Zeyliger commented on IMPALA-7498:
-

There's a similar issue with the catalog, which fails (in an unpleasant, 
core-dump-y way) if it can't talk to the Hive metastore. To be consistent, we 
should wait a certain amount of time in all cases. (No obligation to fix this 
together; just mentioning it if you end up in the same code.)

> impalad should wait for catalogd during start up
> 
>
> Key: IMPALA-7498
> URL: https://issues.apache.org/jira/browse/IMPALA-7498
> Project: IMPALA
>  Issue Type: Sub-task
>Reporter: Todd Lipcon
>Priority: Major
>
> If you start all daemons simultaneously, impalad with --use_local_catalog 
> enabled will retry three times in a tight loop trying to fetch the DB names, 
> and then exit. Instead it should loop for some amount of time waiting for the 
> catalog to be ready in the same way that the existing implementation does.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Resolved] (IMPALA-7453) Intern HdfsStorageDescriptors

2018-08-29 Thread Todd Lipcon (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-7453?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Todd Lipcon resolved IMPALA-7453.
-
   Resolution: Fixed
Fix Version/s: Impala 3.1.0

> Intern HdfsStorageDescriptors
> -
>
> Key: IMPALA-7453
> URL: https://issues.apache.org/jira/browse/IMPALA-7453
> Project: IMPALA
>  Issue Type: Improvement
>  Components: Catalog
>Reporter: Todd Lipcon
>Priority: Major
> Fix For: Impala 3.1.0
>
>
> Every partition currently has an HdfsStorageDescriptor attached. In most 
> cases, the number of unique storage descriptors in a warehouse is pretty low 
> (most partitions use the same escaping, file formats, etc). For example, in 
> the functional test data load, we only have 24 unique SDs across ~10k 
> partitions. Each object takes 32 bytes (with compressed oops) or 40 
> (without). So, we can get some small memory/object-count savings by interning 
> these.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Work started] (IMPALA-7446) Queries can spill earlier than necessary because of accumulation of free buffers and clean pages

2018-08-29 Thread Tim Armstrong (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-7446?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Work on IMPALA-7446 started by Tim Armstrong.
-
> Queries can spill earlier than necessary because of accumulation of free 
> buffers and clean pages
> 
>
> Key: IMPALA-7446
> URL: https://issues.apache.org/jira/browse/IMPALA-7446
> Project: IMPALA
>  Issue Type: Bug
>  Components: Backend
>Affects Versions: Impala 2.10.0, Impala 2.11.0, Impala 3.0, Impala 2.12.0
>Reporter: Tim Armstrong
>Assignee: Tim Armstrong
>Priority: Critical
>  Labels: resource-management
>
> See IMPALA-7442 for an example where the query started to spill even when 
> memory could have been made available by freeing buffers or evicting clean 
> pages. Usually this would just result in spilling earlier than necessary, but 
> in the case of IMPALA-7442 it lead to a query failure.
> My original intent was that BufferPool::ReleaseMemory() should be called in 
> situations like this, but that was not done.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Updated] (IMPALA-7508) Add Impala GDB commands to find fragment instances and queries in a core file

2018-08-29 Thread Zoram Thanga (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-7508?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zoram Thanga updated IMPALA-7508:
-
Labels: debugging supportability  (was: supportability)

> Add Impala GDB commands to find fragment instances and queries in a core file
> -
>
> Key: IMPALA-7508
> URL: https://issues.apache.org/jira/browse/IMPALA-7508
> Project: IMPALA
>  Issue Type: Improvement
>  Components: Infrastructure
>Affects Versions: Impala 2.12.0, Impala 3.1.0
>Reporter: Zoram Thanga
>Assignee: Zoram Thanga
>Priority: Minor
>  Labels: debugging, supportability
>
> Introduce a Python GDB module, and a couple of commands that helps to find 
> queries and fragment instances that are executing in an impalad at the time 
> the daemon crashes.
> One hopes that folks will keep enhancing the module by adding new and useful 
> GDB commands to aid debugging impalad core dumps.
> The initial patch has these:
> {noformat}
> (gdb) source impala-gdb.py
> (gdb) find-query-ids 
> f74c863dff66a34d:1d983cc3
> 364525e12495932b:73f5dd02
> bc4a3eec25481981:edda04b8
> (gdb) find-fragment-instances
> Fragment Instance IdThread IDs
> 364525e12495932b:73f5dd0200a2   [69]
> 364525e12495932b:73f5dd020171   [196, 136]
> bc4a3eec25481981:edda04b801a8   [252, 237, 206]
> f74c863dff66a34d:1d983cc3009b   [200, 14, 13, 12, 6, 5, 3, 2]
> f74c863dff66a34d:1d983cc3013a   [4]
> {noformat}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-7488) TestShellCommandLine::test_cancellation hangs occasionally

2018-08-29 Thread Thomas Tauber-Marshall (JIRA)


[ 
https://issues.apache.org/jira/browse/IMPALA-7488?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16596641#comment-16596641
 ] 

Thomas Tauber-Marshall commented on IMPALA-7488:


So I can repro this, but the behavior that I'm seeing is strange - it appears 
the python process is hanging on the 'close' rpc, but I can see on the Impala 
side that the call to close() returns. Almost seems like a bug in thrift, 
except that we're not doing anything unusual here so I don't know why we would 
be hitting it in just one test. Continuing to investigate.

> TestShellCommandLine::test_cancellation hangs occasionally
> --
>
> Key: IMPALA-7488
> URL: https://issues.apache.org/jira/browse/IMPALA-7488
> Project: IMPALA
>  Issue Type: Bug
>  Components: Infrastructure
>Affects Versions: Impala 3.1.0
>Reporter: Tim Armstrong
>Assignee: Thomas Tauber-Marshall
>Priority: Major
>  Labels: broken-build
> Attachments: psauxf.txt
>
>
> We've seen a couple of hung builds with no queries running on the cluster. I 
> got "ps auxf" output and it looks like an impala-shell process is hanging 
> around.
> I'm guessing the IMPALA-7407 fix somehow relates to this.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Updated] (IMPALA-7508) Add Impala GDB commands to find fragment instances and queries in a core file

2018-08-29 Thread Zoram Thanga (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-7508?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zoram Thanga updated IMPALA-7508:
-
Labels: supportability  (was: )

> Add Impala GDB commands to find fragment instances and queries in a core file
> -
>
> Key: IMPALA-7508
> URL: https://issues.apache.org/jira/browse/IMPALA-7508
> Project: IMPALA
>  Issue Type: Improvement
>  Components: Infrastructure
>Affects Versions: Impala 2.12.0, Impala 3.1.0
>Reporter: Zoram Thanga
>Assignee: Zoram Thanga
>Priority: Minor
>  Labels: supportability
>
> Introduce a Python GDB module, and a couple of commands that helps to find 
> queries and fragment instances that are executing in an impalad at the time 
> the daemon crashes.
> One hopes that folks will keep enhancing the module by adding new and useful 
> GDB commands to aid debugging impalad core dumps.
> The initial patch has these:
> {noformat}
> (gdb) source impala-gdb.py
> (gdb) find-query-ids 
> f74c863dff66a34d:1d983cc3
> 364525e12495932b:73f5dd02
> bc4a3eec25481981:edda04b8
> (gdb) find-fragment-instances
> Fragment Instance IdThread IDs
> 364525e12495932b:73f5dd0200a2   [69]
> 364525e12495932b:73f5dd020171   [196, 136]
> bc4a3eec25481981:edda04b801a8   [252, 237, 206]
> f74c863dff66a34d:1d983cc3009b   [200, 14, 13, 12, 6, 5, 3, 2]
> f74c863dff66a34d:1d983cc3013a   [4]
> {noformat}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Closed] (IMPALA-6065) Document IMPALA-5743

2018-08-29 Thread Alex Rodoni (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-6065?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alex Rodoni closed IMPALA-6065.
---
   Resolution: Fixed
Fix Version/s: Impala 2.10.0

The flags are already documented in impala_ssl.html.

> Document IMPALA-5743
> 
>
> Key: IMPALA-6065
> URL: https://issues.apache.org/jira/browse/IMPALA-6065
> Project: IMPALA
>  Issue Type: Bug
>  Components: Docs
>Affects Versions: Impala 2.10.0
>Reporter: Tim Armstrong
>Assignee: Alex Rodoni
>Priority: Major
> Fix For: Impala 2.10.0
>
>
> It looks like --ssl_minimum_version went into Impala 2.10 but didn't have any 
> accompanying docs. I believe a common use case is to disallow TLS1.0 and 1.1 
> with --ssl_minimum_version=tlsv1.2



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Created] (IMPALA-7508) Add Impala GDB commands to find fragment instances and queries in a core file

2018-08-29 Thread Zoram Thanga (JIRA)
Zoram Thanga created IMPALA-7508:


 Summary: Add Impala GDB commands to find fragment instances and 
queries in a core file
 Key: IMPALA-7508
 URL: https://issues.apache.org/jira/browse/IMPALA-7508
 Project: IMPALA
  Issue Type: Improvement
  Components: Infrastructure
Affects Versions: Impala 2.12.0, Impala 3.1.0
Reporter: Zoram Thanga
Assignee: Zoram Thanga


Introduce a Python GDB module, and a couple of commands that helps to find 
queries and fragment instances that are executing in an impalad at the time the 
daemon crashes.

One hopes that folks will keep enhancing the module by adding new and useful 
GDB commands to aid debugging impalad core dumps.

The initial patch has these:

{noformat}
(gdb) source impala-gdb.py
(gdb) find-query-ids 
f74c863dff66a34d:1d983cc3
364525e12495932b:73f5dd02
bc4a3eec25481981:edda04b8

(gdb) find-fragment-instances
Fragment Instance IdThread IDs

364525e12495932b:73f5dd0200a2   [69]
364525e12495932b:73f5dd020171   [196, 136]
bc4a3eec25481981:edda04b801a8   [252, 237, 206]
f74c863dff66a34d:1d983cc3009b   [200, 14, 13, 12, 6, 5, 3, 2]
f74c863dff66a34d:1d983cc3013a   [4]

{noformat}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Work started] (IMPALA-7508) Add Impala GDB commands to find fragment instances and queries in a core file

2018-08-29 Thread Zoram Thanga (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-7508?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Work on IMPALA-7508 started by Zoram Thanga.

> Add Impala GDB commands to find fragment instances and queries in a core file
> -
>
> Key: IMPALA-7508
> URL: https://issues.apache.org/jira/browse/IMPALA-7508
> Project: IMPALA
>  Issue Type: Improvement
>  Components: Infrastructure
>Affects Versions: Impala 2.12.0, Impala 3.1.0
>Reporter: Zoram Thanga
>Assignee: Zoram Thanga
>Priority: Minor
>
> Introduce a Python GDB module, and a couple of commands that helps to find 
> queries and fragment instances that are executing in an impalad at the time 
> the daemon crashes.
> One hopes that folks will keep enhancing the module by adding new and useful 
> GDB commands to aid debugging impalad core dumps.
> The initial patch has these:
> {noformat}
> (gdb) source impala-gdb.py
> (gdb) find-query-ids 
> f74c863dff66a34d:1d983cc3
> 364525e12495932b:73f5dd02
> bc4a3eec25481981:edda04b8
> (gdb) find-fragment-instances
> Fragment Instance IdThread IDs
> 364525e12495932b:73f5dd0200a2   [69]
> 364525e12495932b:73f5dd020171   [196, 136]
> bc4a3eec25481981:edda04b801a8   [252, 237, 206]
> f74c863dff66a34d:1d983cc3009b   [200, 14, 13, 12, 6, 5, 3, 2]
> f74c863dff66a34d:1d983cc3013a   [4]
> {noformat}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Resolved] (IMPALA-7286) Command "unset" does not work for shell options

2018-08-29 Thread Csaba Ringhofer (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-7286?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Csaba Ringhofer resolved IMPALA-7286.
-
Resolution: Fixed

> Command "unset" does not work for shell options
> ---
>
> Key: IMPALA-7286
> URL: https://issues.apache.org/jira/browse/IMPALA-7286
> Project: IMPALA
>  Issue Type: Bug
>  Components: Clients
>Reporter: Csaba Ringhofer
>Assignee: Nghia Le
>Priority: Minor
>  Labels: shell
>
> "unset LIVE_PROGRESS;" does not unset LIVE_PROGRESS and leads to the 
> following error message:
> No option called LIVE_PROGRESS is set
> I think that unset should work for shell options too for the sake of 
> consistency.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-7496) Schedule query taking in account the mem available on the impalad nodes

2018-08-29 Thread Adriano (JIRA)


[ 
https://issues.apache.org/jira/browse/IMPALA-7496?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16596341#comment-16596341
 ] 

Adriano commented on IMPALA-7496:
-

The Impala version used during the issue haven't -IMPALA-1575- The specific 
situation appeared once many users connected directly from their workstation to 
the specific impalad (this caused more use of memory, as you guessed).

The query options beside the lack of guarantee are very appreciated.

Thank to include the jira in the backlog.

> Schedule query taking in account the mem available on the impalad nodes
> ---
>
> Key: IMPALA-7496
> URL: https://issues.apache.org/jira/browse/IMPALA-7496
> Project: IMPALA
>  Issue Type: New Feature
>  Components: Backend
>Reporter: Adriano
>Priority: Major
>  Labels: admission-control, scheduler
>
> Environment description: cluster scale (50/100/150 nodes and terabyte of ram 
> available) - Admission Control enabled.
> Issue description:
> Despite the coordinator chosen (with data and statistics unchanged) a query 
> will be planned always in the same way based on the metainfo that the 
> coordinator have.
> The query will be scheduled always on the same nodes if the memory 
> requirements for the admission are satisfied:
> https://github.com/cloudera/Impala/blob/cdh5-2.7.0_5.9.1/be/src/scheduling/admission-controller.cc#L307-L333
> Equal queries are planned/scheduled always in the same way (to hit always the 
> same nodes). 
> This often lead to queue the queries that are hitting the same nodes are 
> queued (not admitted) as on those nodes there's no more memory available 
> within its process limit despite the pool have lot of free memory and the 
> overall cluster load is low.
> When the plan is finished and the query can be evaluated to be admitted often 
> happen that the admission is denied because one of the node have not enough 
> memory to run the query operation (and the query is moved in the pool queue) 
> despite the cluster have 50/100/150 nodes and terabyte of ram available.
> Why the scheduler does not take in consideration the memory available on the 
> nodes involved in the query before to buid the schedule, (maybe preferring a 
> remote read/operation on a free memory node instead to include in the plan 
> always the same nodes that will end to be:
> 1- overloaded
> 2- the query will be not immediately admitted, risking to be timedout in the 
> pool queue
> Since 2.7 the REPLICA_PREFERENCE can possibly help, but it's not good enough 
> as it does not prevent the scheduler to choose busy nodes (with the same 
> potential effect: query queued for lack of resource on specific node despite 
> there are terabytes of free memory).
> Feature Request:
> It would be good if Impala had an option to execute queries (even with worse 
> performance) excluding the nodes overloaded and including different nodes in 
> order to get the query immediately admitted and executed.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-7505) impalad webserver hang in getting a lock of QueryExecStatus

2018-08-29 Thread Balazs Jeszenszky (JIRA)


[ 
https://issues.apache.org/jira/browse/IMPALA-7505?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16596092#comment-16596092
 ] 

Balazs Jeszenszky commented on IMPALA-7505:
---

[~wangchen1ren] FYI, IMPALA-1972 and IMPALA-3882 track some of the issues 
around this.

> impalad webserver hang in getting a lock of QueryExecStatus
> ---
>
> Key: IMPALA-7505
> URL: https://issues.apache.org/jira/browse/IMPALA-7505
> Project: IMPALA
>  Issue Type: Bug
>  Components: Backend
>Affects Versions: Impala 2.8.0
>Reporter: Chen Wang
>Priority: Major
> Attachments: gdb.out
>
>
> Impalad’s webserver would hang sometimes.
> The following is one of the cases: the webserver threads stuck in getting a 
> lock of QueryExecStatus,  but I can't find where the lock is acquired in the 
> stack. The web requests are sent from the agent of CDH, which is to check the 
> activity of impalad.
> Full gdb log is in the attachment.
> {code}
> Thread 116 (Thread 0x7f288f5e1700 (LWP 31062)):
> #0  0x00378780e334 in __lll_lock_wait () from /lib64/libpthread.so.0
> #1  0x0037878095d8 in _L_lock_854 () from /lib64/libpthread.so.0
> #2  0x0037878094a7 in pthread_mutex_lock () from /lib64/libpthread.so.0
> #3  0x008d6eb8 in pthread_mutex_lock (this=0xcab4f50) at 
> /data/impala/toolchain/boost-1.57.0/include/boost/thread/pthread/mutex.hpp:62
> #4  boost::mutex::lock (this=0xcab4f50) at 
> /data/impala/toolchain/boost-1.57.0/include/boost/thread/pthread/mutex.hpp:116
> #5  0x00b7903c in lock_guard (this=0xa7b5800, query_id=) at 
> /data/impala/toolchain/boost-1.57.0/include/boost/thread/lock_guard.hpp:38
> #6  impala::ImpalaServer::GetRuntimeProfileStr (this=0xa7b5800, query_id=) at 
> /data/impala/be/src/service/impala-server.cc:573
> #7  0x00ba6a8c in 
> impala::ImpalaHttpHandler::QueryProfileEncodedHandler (this=0x3f56be0, args=) 
> at /data/impala/be/src/service/impala-http-handler.cc:219
> #8  0x00cafe75 in operator() (this=) at 
> /data/impala/toolchain/boost-1.57.0/include/boost/function/function_template.hpp:767
> #9  impala::Webserver::RenderUrlWithTemplate (this=) at 
> /data/impala/be/src/util/webserver.cc:443
> #10 0x00cb1295 in impala::Webserver::BeginRequestCallback (this=) at 
> /data/impala/be/src/util/webserver.cc:414
> #11 0x00cc4850 in handle_request ()
> #12 0x00cc6fcd in process_new_connection ()
> #13 0x00cc765d in worker_thread ()
> #14 0x003787807aa1 in start_thread () from /lib64/libpthread.so.0
> #15 0x0037874e8bcd in clone () from /lib64/libc.so.6
> {code}
> The hang situation appears on impala 2.8.0, but I found that code of 
> be/service part hasn’t changed much from 2.8.0 to 2.11.0. so the problem may 
> still exists.
> Hope you experts can give me some guidance of finding the root cause, or 
> workaround plans to deal with these hang situation.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Updated] (IMPALA-7505) impalad webserver hang in getting a lock of QueryExecStatus

2018-08-29 Thread Chen Wang (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-7505?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chen Wang updated IMPALA-7505:
--
Affects Version/s: (was: Impala 2.7.1)
  Description: 
Impalad’s webserver would hang sometimes.

The following is one of the cases: the webserver threads stuck in getting a 
lock of QueryExecStatus,  but I can't find where the lock is acquired in the 
stack. The web requests are sent from the agent of CDH, which is to check the 
activity of impalad.
Full gdb log is in the attachment.

{code}
Thread 116 (Thread 0x7f288f5e1700 (LWP 31062)):
#0  0x00378780e334 in __lll_lock_wait () from /lib64/libpthread.so.0
#1  0x0037878095d8 in _L_lock_854 () from /lib64/libpthread.so.0
#2  0x0037878094a7 in pthread_mutex_lock () from /lib64/libpthread.so.0
#3  0x008d6eb8 in pthread_mutex_lock (this=0xcab4f50) at 
/data/impala/toolchain/boost-1.57.0/include/boost/thread/pthread/mutex.hpp:62
#4  boost::mutex::lock (this=0xcab4f50) at 
/data/impala/toolchain/boost-1.57.0/include/boost/thread/pthread/mutex.hpp:116
#5  0x00b7903c in lock_guard (this=0xa7b5800, query_id=) at 
/data/impala/toolchain/boost-1.57.0/include/boost/thread/lock_guard.hpp:38
#6  impala::ImpalaServer::GetRuntimeProfileStr (this=0xa7b5800, query_id=) at 
/data/impala/be/src/service/impala-server.cc:573
#7  0x00ba6a8c in impala::ImpalaHttpHandler::QueryProfileEncodedHandler 
(this=0x3f56be0, args=) at 
/data/impala/be/src/service/impala-http-handler.cc:219
#8  0x00cafe75 in operator() (this=) at 
/data/impala/toolchain/boost-1.57.0/include/boost/function/function_template.hpp:767
#9  impala::Webserver::RenderUrlWithTemplate (this=) at 
/data/impala/be/src/util/webserver.cc:443
#10 0x00cb1295 in impala::Webserver::BeginRequestCallback (this=) at 
/data/impala/be/src/util/webserver.cc:414
#11 0x00cc4850 in handle_request ()
#12 0x00cc6fcd in process_new_connection ()
#13 0x00cc765d in worker_thread ()
#14 0x003787807aa1 in start_thread () from /lib64/libpthread.so.0
#15 0x0037874e8bcd in clone () from /lib64/libc.so.6
{code}

The hang situation appears on impala 2.8.0, but I found that code of be/service 
part hasn’t changed much from 2.8.0 to 2.11.0. so the problem may still exists.
Hope you experts can give me some guidance of finding the root cause, or 
workaround plans to deal with these hang situation.

  was:
Impalad’s webserver would hang sometimes.

The following is one of the cases: the webserver threads stuck in getting a 
lock of QueryExecStatus,  but I can't find where the lock is acquired in the 
stack. The web requests are sent from the agent of CDH, which is to check the 
activity of impalad.
Full gdb log is in the attachment.

{code}
Thread 116 (Thread 0x7f288f5e1700 (LWP 31062)):
#0  0x00378780e334 in __lll_lock_wait () from /lib64/libpthread.so.0
#1  0x0037878095d8 in _L_lock_854 () from /lib64/libpthread.so.0
#2  0x0037878094a7 in pthread_mutex_lock () from /lib64/libpthread.so.0
#3  0x008d6eb8 in pthread_mutex_lock (this=0xcab4f50) at 
/data/impala/toolchain/boost-1.57.0/include/boost/thread/pthread/mutex.hpp:62
#4  boost::mutex::lock (this=0xcab4f50) at 
/data/impala/toolchain/boost-1.57.0/include/boost/thread/pthread/mutex.hpp:116
#5  0x00b7903c in lock_guard (this=0xa7b5800, query_id=) at 
/data/impala/toolchain/boost-1.57.0/include/boost/thread/lock_guard.hpp:38
#6  impala::ImpalaServer::GetRuntimeProfileStr (this=0xa7b5800, query_id=) at 
/data/impala/be/src/service/impala-server.cc:573
#7  0x00ba6a8c in impala::ImpalaHttpHandler::QueryProfileEncodedHandler 
(this=0x3f56be0, args=) at 
/data/impala/be/src/service/impala-http-handler.cc:219
#8  0x00cafe75 in operator() (this=) at 
/data/impala/toolchain/boost-1.57.0/include/boost/function/function_template.hpp:767
#9  impala::Webserver::RenderUrlWithTemplate (this=) at 
/data/impala/be/src/util/webserver.cc:443
#10 0x00cb1295 in impala::Webserver::BeginRequestCallback (this=) at 
/data/impala/be/src/util/webserver.cc:414
#11 0x00cc4850 in handle_request ()
#12 0x00cc6fcd in process_new_connection ()
#13 0x00cc765d in worker_thread ()
#14 0x003787807aa1 in start_thread () from /lib64/libpthread.so.0
#15 0x0037874e8bcd in clone () from /lib64/libc.so.6
{code}

The hang situation appears on impala 2.7.1 & 2.8.0, but I found that code of 
be/service part hasn’t changed much from 2.7.1 to 2.11.0. so the problem may 
still exists.
Hope you experts can give me some guidance of finding the root cause, or 
workaround plans to deal with these hang situation.


> impalad webserver hang in getting a lock of QueryExecStatus
> ---
>
> Key: IMPALA-7505
> URL: https://issues.apache.org/jira/browse/IMPALA-7505
> Project: IMPALA
>

[jira] [Updated] (IMPALA-7506) Support global INVALIDATE METADATA on fetch-on-demand impalad

2018-08-29 Thread Todd Lipcon (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-7506?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Todd Lipcon updated IMPALA-7506:

Description: There is some complexity with how this is implemented in the 
original code: it depends on maintaining the minimum version of any object in 
the impalad's local cache. We can't determine that in an on-demand impalad, so 
INVALIDATE METADATA is not supported currently on "fetch-on-demand".  (was: 
There is some complexity with how this is implemented in the original code: it 
depends on maintaining the minimum version of any object in the impalad's local 
cache. We can't determine that in an on-demand impalad, so SYNC_DDL + 
INVALIDATE METADATA is not supported currently on "fetch-on-demand".)
Summary: Support global INVALIDATE METADATA on fetch-on-demand impalad  
(was: Support INVALIDATE METADATA with SYNC_DDL on fetch-on-demand impalad)

> Support global INVALIDATE METADATA on fetch-on-demand impalad
> -
>
> Key: IMPALA-7506
> URL: https://issues.apache.org/jira/browse/IMPALA-7506
> Project: IMPALA
>  Issue Type: Sub-task
>Reporter: Todd Lipcon
>Priority: Major
>
> There is some complexity with how this is implemented in the original code: 
> it depends on maintaining the minimum version of any object in the impalad's 
> local cache. We can't determine that in an on-demand impalad, so INVALIDATE 
> METADATA is not supported currently on "fetch-on-demand".



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Created] (IMPALA-7507) Clean up user-facing error messages in LocalCatalog mode

2018-08-29 Thread Todd Lipcon (JIRA)
Todd Lipcon created IMPALA-7507:
---

 Summary: Clean up user-facing error messages in LocalCatalog mode
 Key: IMPALA-7507
 URL: https://issues.apache.org/jira/browse/IMPALA-7507
 Project: IMPALA
  Issue Type: Sub-task
Reporter: Todd Lipcon


Currently even normal error messages for things like missing databases are 
quite ugly when running with LocalCatalog:

{code}
ERROR: LocalCatalogException: Could not load table names for database 
'test_minimal_topic_updates_b246004e' from HMS
CAUSED BY: TException: 
TGetPartialCatalogObjectResponse(status:TStatus(status_code:GENERAL, 
error_msgs:[CatalogException: Database not found: 
test_minimal_topic_updates_b246004e]))
{code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Created] (IMPALA-7506) Support INVALIDATE METADATA with SYNC_DDL on fetch-on-demand impalad

2018-08-29 Thread Todd Lipcon (JIRA)
Todd Lipcon created IMPALA-7506:
---

 Summary: Support INVALIDATE METADATA with SYNC_DDL on 
fetch-on-demand impalad
 Key: IMPALA-7506
 URL: https://issues.apache.org/jira/browse/IMPALA-7506
 Project: IMPALA
  Issue Type: Sub-task
Reporter: Todd Lipcon


There is some complexity with how this is implemented in the original code: it 
depends on maintaining the minimum version of any object in the impalad's local 
cache. We can't determine that in an on-demand impalad, so SYNC_DDL + 
INVALIDATE METADATA is not supported currently on "fetch-on-demand".



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org