[jira] [Resolved] (IMPALA-8073) SentryProxy.testAddCatalog() failed in private build because of socket error

2019-01-22 Thread Fredy Wijaya (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8073?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Fredy Wijaya resolved IMPALA-8073.
--
Resolution: Fixed

> SentryProxy.testAddCatalog() failed in private build because of socket error
> 
>
> Key: IMPALA-8073
> URL: https://issues.apache.org/jira/browse/IMPALA-8073
> Project: IMPALA
>  Issue Type: Bug
>  Components: Catalog
>Reporter: Tim Armstrong
>Assignee: Fredy Wijaya
>Priority: Critical
>  Labels: flaky
> Attachments: 
> FeSupport.impala-ec2-centos74-m5-4xlarge-ondemand-0eae.vpc.cloudera.com.jenkins.log.INFO.20190110-184543.10852
>
>
> {noformat}
> org.apache.impala.util.SentryProxyTest.testAddCatalog
> Failing for the past 1 build (Since Failed#4229 )
> Took 3 min 40 sec.
> add description
> Error Message
> Error initializing Catalog. Catalog may be empty.
> Stacktrace
> java.lang.IllegalStateException: Error initializing Catalog. Catalog may be 
> empty.
>   at 
> org.apache.hadoop.hive.metastore.MetaStoreUtils.logAndThrowMetaException(MetaStoreUtils.java:1444)
>   at 
> org.apache.hadoop.hive.metastore.HiveMetaStoreClient.getAllDatabases(HiveMetaStoreClient.java:1351)
>   at sun.reflect.GeneratedMethodAccessor33.invoke(Unknown Source)
>   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>   at java.lang.reflect.Method.invoke(Method.java:498)
>   at 
> org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.invoke(RetryingMetaStoreClient.java:150)
>   at com.sun.proxy.$Proxy16.getAllDatabases(Unknown Source)
>   at 
> org.apache.impala.catalog.CatalogServiceCatalog.reset(CatalogServiceCatalog.java:1181)
>   at 
> org.apache.impala.testutil.CatalogServiceTestCatalog.createWithAuth(CatalogServiceTestCatalog.java:59)
>   at 
> org.apache.impala.util.SentryProxyTest.withAllPrincipalTypes(SentryProxyTest.java:603)
>   at 
> org.apache.impala.util.SentryProxyTest.testAddCatalog(SentryProxyTest.java:140)
> {noformat}
> The error in the log (attached) appears to be a connection to the HMS error.
> {noformat}
> 0110 18:48:08.139935 10853 HiveMetaStoreClient.java:461] Trying to connect to 
> metastore with URI thrift://localhost:9083
> I0110 18:48:08.140183 10853 HiveMetaStoreClient.java:535] Opened a connection 
> to metastore, current connections: 797
> W0110 18:48:28.143384 10853 HiveMetaStoreClient.java:560] set_ugi() not 
> successful, Likely cause: new client talking to old server. Continuing 
> without it.
> Java exception follows:
> org.apache.thrift.transport.TTransportException: java.net.SocketException: 
> Connection reset
> at 
> org.apache.thrift.transport.TIOStreamTransport.read(TIOStreamTransport.java:129)
> at org.apache.thrift.transport.TTransport.readAll(TTransport.java:86)
> at 
> org.apache.thrift.protocol.TBinaryProtocol.readAll(TBinaryProtocol.java:429)
> at 
> org.apache.thrift.protocol.TBinaryProtocol.readI32(TBinaryProtocol.java:318)
> at 
> org.apache.thrift.protocol.TBinaryProtocol.readMessageBegin(TBinaryProtocol.java:219)
> at 
> org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:77)
> at 
> org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.recv_set_ugi(ThriftHiveMetastore.java:4129)
> at 
> org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.set_ugi(ThriftHiveMetastore.java:4115)
> at 
> org.apache.hadoop.hive.metastore.HiveMetaStoreClient.open(HiveMetaStoreClient.java:552)
> at 
> org.apache.hadoop.hive.metastore.HiveMetaStoreClient.(HiveMetaStoreClient.java:299)
> at sun.reflect.GeneratedConstructorAccessor3.newInstance(Unknown 
> Source)
> at 
> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
> at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
> at 
> org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1750)
> at 
> org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.(RetryingMetaStoreClient.java:80)
> at 
> org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:130)
> at 
> org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:101)
> at 
> org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:94)
> at 
> org.apache.impala.catalog.MetaStoreClientPool$MetaStoreClient.(MetaStoreClientPool.java:93)
> at 
> org.apache.impala.catalog.MetaStoreClientPool$MetaStoreClient.(MetaStoreClientPool.java:72)
> at 
> 

[jira] [Commented] (IMPALA-7928) Investigate consistent placement of remote scan ranges

2019-01-22 Thread Alex Rodoni (JIRA)


[ 
https://issues.apache.org/jira/browse/IMPALA-7928?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16749342#comment-16749342
 ] 

Alex Rodoni commented on IMPALA-7928:
-

[~joemcdonnell] Doc impact?

> Investigate consistent placement of remote scan ranges
> --
>
> Key: IMPALA-7928
> URL: https://issues.apache.org/jira/browse/IMPALA-7928
> Project: IMPALA
>  Issue Type: Bug
>  Components: Backend
>Affects Versions: Impala 3.2.0
>Reporter: Joe McDonnell
>Assignee: Joe McDonnell
>Priority: Critical
>
> With the file handle cache, it is useful for repeated scans of the same file 
> to go to the same node, as that node will already have a file handle cached.
> When scheduling remote ranges, the scheduler introduces randomness that can 
> spread reads across all of the nodes. Repeated executions of queries on the 
> same set of files will not schedule the remote reads on the same nodes. This 
> causes a large amount of duplication across file handle caches on different 
> nodes. This reduces the efficiency of the cache significantly.
> It may be useful for the scheduler to introduce some determinism in 
> scheduling remote reads to take advantage of the file handle cache. This is a 
> variation on the well-known tradeoff between skew and locality.
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-7265) Cache remote file handles

2019-01-22 Thread Alex Rodoni (JIRA)


[ 
https://issues.apache.org/jira/browse/IMPALA-7265?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16749340#comment-16749340
 ] 

Alex Rodoni commented on IMPALA-7265:
-

[~joemcdonnell] Is there a doc impact for this feature?

> Cache remote file handles
> -
>
> Key: IMPALA-7265
> URL: https://issues.apache.org/jira/browse/IMPALA-7265
> Project: IMPALA
>  Issue Type: Improvement
>  Components: Backend
>Affects Versions: Impala 3.1.0
>Reporter: Joe McDonnell
>Assignee: Joe McDonnell
>Priority: Critical
>
> The file handle cache currently does not allow caching remote file handles. 
> This means that clusters that have a lot of remote reads can suffer from 
> overloading the NameNode. Impala should be able to cache remote file handles.
> There are some open questions about remote file handles and whether they 
> behave differently from local file handles. In particular:
>  # Is there any resource constraint on the number of remote file handles 
> open? (e.g. do they maintain a network connection?)
>  # Are there any semantic differences in how remote file handles behave when 
> files are deleted, overwritten, or appended?
>  # Are there any extra failure cases for remote file handles? (i.e. if a 
> machine goes down or a remote file handle is left open for an extended period 
> of time)
> The form of caching will depend on the answers, but at the very least, it 
> should be possible to cache a remote file handle at the level of a query so 
> that a Parquet file with multiple columns can share file handles.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-6169) Implement DATE type

2019-01-22 Thread Alex Rodoni (JIRA)


[ 
https://issues.apache.org/jira/browse/IMPALA-6169?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16749337#comment-16749337
 ] 

Alex Rodoni commented on IMPALA-6169:
-

[~attilaj] [~csringhofer] [~zi] What part of this epic will be in 3.2?

> Implement DATE type
> ---
>
> Key: IMPALA-6169
> URL: https://issues.apache.org/jira/browse/IMPALA-6169
> Project: IMPALA
>  Issue Type: New Feature
>  Components: Backend
>Reporter: Quanlong Huang
>Assignee: Attila Jeges
>Priority: Major
>
> In Hive, the Date type describes a particular year/month/day, in the form 
> -­MM-­DD.
> Hive has supported Date type in Parquet two years ago in Hive-1.2.0. (See 
> https://issues.apache.org/jira/browse/HIVE-8119 and 
> https://cwiki.apache.org/confluence/display/Hive/Parquet#Parquet-VersionsandLimitations.)
> We should add support for Date type too.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Updated] (IMPALA-8098) [Docs] Document incompatible changes to :shutdown command

2019-01-22 Thread Alex Rodoni (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8098?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alex Rodoni updated IMPALA-8098:

Target Version: Impala 3.2.0

> [Docs] Document incompatible changes to :shutdown command
> -
>
> Key: IMPALA-8098
> URL: https://issues.apache.org/jira/browse/IMPALA-8098
> Project: IMPALA
>  Issue Type: Task
>Reporter: Andrew Sherman
>Assignee: Alex Rodoni
>Priority: Major
>  Labels: future_release_doc, in_32
>
> The :shutdown command is used to shutdown a remote server. The common case is 
> that a user specifies the impalad to shutdown by specifying a host e.g. 
> :shutdown('host100'). If a user has more than one impalad on a remote host 
> then the form :shutdown(':') can be used to specify the port by 
> which the imapald can be contacted. Prior to IMPALA-7985 this port was the 
> backend port, e.g. :shutdown('host100:22000'). With IMPALA-7985 the port to 
> use is the KRPC port , e.g. :shutdown('host100:27000').



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-3323) impala-shell --ldap_password_cmd has no config file equivalent

2019-01-22 Thread Alex Rodoni (JIRA)


[ 
https://issues.apache.org/jira/browse/IMPALA-3323?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16749331#comment-16749331
 ] 

Alex Rodoni commented on IMPALA-3323:
-

[~fredyw] Is this a user-facing feature / fix with a doc impact?

> impala-shell --ldap_password_cmd has no config file equivalent
> --
>
> Key: IMPALA-3323
> URL: https://issues.apache.org/jira/browse/IMPALA-3323
> Project: IMPALA
>  Issue Type: Bug
>  Components: Clients
>Affects Versions: Impala 2.5.0
>Reporter: John Russell
>Assignee: Fredy Wijaya
>Priority: Minor
>  Labels: newbie
> Fix For: Impala 3.2.0
>
>
> Putting ldap_password_cmd in the .impalarc file results in a parse error when 
> impala-shell starts up. Judging from the impala-shell source, it looks like a 
> dest= parameter is missing so there's no config file equivalent for the 
> --ldap_password_cmd command-line option. E.g. here's one with a config file 
> equivalent and one without:
> {code}
> parser.add_option("--auth_creds_ok_in_clear", dest="creds_ok_in_clear",
> action="store_true", help="If set, LDAP authentication " +
> "may be used with an insecure connection to Impala. " +
> "WARNING: Authentication credentials will therefore be 
> sent " +
> "unencrypted, and may be vulnerable to attack.")
>   parser.add_option("--ldap_password_cmd",
> help="Shell command to run to retrieve the LDAP password")
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Resolved] (IMPALA-3323) impala-shell --ldap_password_cmd has no config file equivalent

2019-01-22 Thread Fredy Wijaya (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-3323?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Fredy Wijaya resolved IMPALA-3323.
--
   Resolution: Fixed
Fix Version/s: Impala 3.2.0

> impala-shell --ldap_password_cmd has no config file equivalent
> --
>
> Key: IMPALA-3323
> URL: https://issues.apache.org/jira/browse/IMPALA-3323
> Project: IMPALA
>  Issue Type: Bug
>  Components: Clients
>Affects Versions: Impala 2.5.0
>Reporter: John Russell
>Assignee: Fredy Wijaya
>Priority: Minor
>  Labels: newbie
> Fix For: Impala 3.2.0
>
>
> Putting ldap_password_cmd in the .impalarc file results in a parse error when 
> impala-shell starts up. Judging from the impala-shell source, it looks like a 
> dest= parameter is missing so there's no config file equivalent for the 
> --ldap_password_cmd command-line option. E.g. here's one with a config file 
> equivalent and one without:
> {code}
> parser.add_option("--auth_creds_ok_in_clear", dest="creds_ok_in_clear",
> action="store_true", help="If set, LDAP authentication " +
> "may be used with an insecure connection to Impala. " +
> "WARNING: Authentication credentials will therefore be 
> sent " +
> "unencrypted, and may be vulnerable to attack.")
>   parser.add_option("--ldap_password_cmd",
> help="Shell command to run to retrieve the LDAP password")
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Updated] (IMPALA-8098) [Docs] Document incompatible changes to :shutdown command

2019-01-22 Thread Alex Rodoni (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8098?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alex Rodoni updated IMPALA-8098:

Labels: future_release_doc in_32  (was: future_release_doc)

> [Docs] Document incompatible changes to :shutdown command
> -
>
> Key: IMPALA-8098
> URL: https://issues.apache.org/jira/browse/IMPALA-8098
> Project: IMPALA
>  Issue Type: Task
>Reporter: Andrew Sherman
>Assignee: Alex Rodoni
>Priority: Major
>  Labels: future_release_doc, in_32
>
> The :shutdown command is used to shutdown a remote server. The common case is 
> that a user specifies the impalad to shutdown by specifying a host e.g. 
> :shutdown('host100'). If a user has more than one impalad on a remote host 
> then the form :shutdown(':') can be used to specify the port by 
> which the imapald can be contacted. Prior to IMPALA-7985 this port was the 
> backend port, e.g. :shutdown('host100:22000'). With IMPALA-7985 the port to 
> use is the KRPC port , e.g. :shutdown('host100:27000').



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Created] (IMPALA-8098) [Docs] Document incompatible changes to :shutdown command

2019-01-22 Thread Andrew Sherman (JIRA)
Andrew Sherman created IMPALA-8098:
--

 Summary: [Docs] Document incompatible changes to :shutdown command
 Key: IMPALA-8098
 URL: https://issues.apache.org/jira/browse/IMPALA-8098
 Project: IMPALA
  Issue Type: Task
Reporter: Andrew Sherman
Assignee: Alex Rodoni


The :shutdown command is used to shutdown a remote server. The common case is 
that a user specifies the impalad to shutdown by specifying a host e.g. 
:shutdown('host100'). If a user has more than one impalad on a remote host then 
the form :shutdown(':') can be used to specify the port by which 
the imapald can be contacted. Prior to IMPALA-7985 this port was the backend 
port, e.g. :shutdown('host100:22000'). With IMPALA-7985 the port to use is the 
KRPC port , e.g. :shutdown('host100:27000').



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-8027) KRPC datastream timing out on both the receiver and sender side even in a minicluster

2019-01-22 Thread Michael Ho (JIRA)


[ 
https://issues.apache.org/jira/browse/IMPALA-8027?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16749308#comment-16749308
 ] 

Michael Ho commented on IMPALA-8027:


The profile didn't suggest much either. The last report time for the 
coordinator fragment was 05:16:55, which was when the profile was initially 
created. So, the coordinator itself didn't receive any update from the 
coordinator fragment itself for two minutes. This suggests the Prepare() didn't 
complete within the time frame.

{noformat}
Instance 194f5b70907ac97c:84a116d6 
(host=impala-ec2-centos74-m5-4xlarge-ondemand-07b3.vpc.cloudera.com:22000)
Last report received time: 2018-12-28 05:16:52.214
{noformat}

FWIW, other fragment instances running on other Impalad peers were able to send 
report so it seems the status reporting handling was working on coordinator 
({{impala-ec2-centos74-m5-4xlarge-ondemand-07b3.vpc.cloudera.com:22000}}):
{noformat}
Instance 194f5b70907ac97c:84a116d60003 
(host=impala-ec2-centos74-m5-4xlarge-ondemand-07b3.vpc.cloudera.com:22001)
Last report received time: 2018-12-28 05:18:56.204
Hdfs split stats (:<# splits>/): -1:4/152.76 KB
{noformat}

The fragment instances running on 
{{impala-ec2-centos74-m5-4xlarge-ondemand-07b3.vpc.cloudera.com:22000}} 
eventually completed.
{noformat}
I1228 05:16:52.217275 26685 query-state.cc:568] Executing instance. 
instance_id=194f5b70907ac97c:84a116d6 fragment_idx=0 
per_fragment_instance_idx=0 coord_state_idx=0 #in-flight=7
I1228 05:20:08.981983 26685 krpc-data-stream-mgr.cc:294] DeregisterRecvr(): 
fragment_instance_id=194f5b70907ac97c:84a116d6, node=2
I1228 05:20:08.982043 26685 query-state.cc:576] Instance completed. 
instance_id=194f5b70907ac97c:84a116d6 #in-flight=5 status=CANCELLED: 
Cancelled
{noformat}

{noformat}
I1228 05:16:52.217344 26686 query-state.cc:568] Executing instance. 
instance_id=194f5b70907ac97c:84a116d60001 fragment_idx=1 
per_fragment_instance_idx=0 coord_state_idx=0 #in-flight=8
I1228 05:20:08.992434 26686 llvm-codegen.cc:1104] For query 
194f5b70907ac97c:84a116d6 the following functions were not finalized 
and have been removed from the module:
I1228 05:20:09.049794 26686 query-state.cc:576] Instance completed. 
instance_id=194f5b70907ac97c:84a116d60001 #in-flight=5 status=CANCELLED: 
Cancelled
I1228 05:20:09.056677 26686 query-exec-mgr.cc:182] ReleaseQueryState(): deleted 
query_id=194f5b70907ac97c:84a116d6
{noformat}

With the data available so far, one could only conclude that the test failed 
due to some flaky scheduling issue which slowed down some query fragments 
running on 
{{impala-ec2-centos74-m5-4xlarge-ondemand-07b3.vpc.cloudera.com:22000}}. Given 
this hasn't happened again since the initial reporting, I don't think there is 
much we need to do at this point.

Longer term, fixing IMPALA-6618 will remove this class of flaky timeout issues.

> KRPC datastream timing out on both the receiver and sender side even in a 
> minicluster
> -
>
> Key: IMPALA-8027
> URL: https://issues.apache.org/jira/browse/IMPALA-8027
> Project: IMPALA
>  Issue Type: Bug
>  Components: Distributed Exec
>Affects Versions: Impala 3.2.0
>Reporter: Bikramjeet Vig
>Assignee: Michael Ho
>Priority: Critical
>  Labels: broken-build
>
> krpc datastreams seem to time out at the same time at both sender and 
> receiver causing two running queries to fail. This happened while running 
> core tests on s3.
> Logs from coordinator:
> {noformat}
> I1228 05:18:56.202587 13396 krpc-data-stream-mgr.cc:353] Sender 127.0.0.1 
> timed out waiting for receiver fragment instance: 
> 8f46b2518734bef1:6ef2d404, dest node: 2
> I1228 05:18:56.203061 13396 rpcz_store.cc:265] Call 
> impala.DataStreamService.TransmitData from 127.0.0.1:53118 (request call id 
> 11274) took 120782ms. Request Metrics: {}
> I1228 05:18:56.203114 13396 krpc-data-stream-mgr.cc:353] Sender 127.0.0.1 
> timed out waiting for receiver fragment instance: 
> 194f5b70907ac97c:84a116d6, dest node: 2
> I1228 05:18:56.203136 13396 rpcz_store.cc:265] Call 
> impala.DataStreamService.TransmitData from 127.0.0.1:53110 (request call id 
> 8637) took 123811ms. Request Metrics: {}
> I1228 05:18:56.203155 13396 krpc-data-stream-mgr.cc:353] Sender 127.0.0.1 
> timed out waiting for receiver fragment instance: 
> 194f5b70907ac97c:84a116d6, dest node: 2
> I1228 05:18:56.203167 13396 rpcz_store.cc:265] Call 
> impala.DataStreamService.TransmitData from 127.0.0.1:53118 (request call id 
> 11273) took 123776ms. Request Metrics: {}
> I1228 05:18:56.203181 13396 krpc-data-stream-mgr.cc:408] Reduced stream ID 
> cache from 413 items, to 410, eviction took: 1ms
> I1228 05:18:56.204746 13377 

[jira] [Work started] (IMPALA-8091) minicluster kudu failure: Cannot initialize clock: failed to wait for clock sync

2019-01-22 Thread Michael Brown (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8091?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Work on IMPALA-8091 started by Michael Brown.
-
> minicluster kudu failure: Cannot initialize clock: failed to wait for clock 
> sync
> 
>
> Key: IMPALA-8091
> URL: https://issues.apache.org/jira/browse/IMPALA-8091
> Project: IMPALA
>  Issue Type: Bug
>  Components: Infrastructure
>Affects Versions: Impala 3.2.0
>Reporter: David Knupp
>Assignee: Michael Brown
>Priority: Blocker
>  Labels: broken-build
>
> *Console log*:
> {noformat}
> 19:10:04 19:10:04 Error executing impala SQL: 
> /data/jenkins/workspace/impala-asf-master-exhaustive-release/repos/Impala/logs/data_loading/sql/functional/create-functional-query-exhaustive-impala-generated-kudu-none-none.sql
>  See: 
> /data/jenkins/workspace/impala-asf-master-exhaustive-release/repos/Impala/logs/data_loading/sql/functional/create-functional-query-exhaustive-impala-generated-kudu-none-none.sql.log
> 19:10:04 Background task Loading functional-query data (pid 4831) failed.
> 19:10:04 Background task Loading TPC-H data (pid 4832) failed.
> 19:10:04 ERROR in 
> /data/jenkins/workspace/impala-asf-master-exhaustive-release/repos/Impala/testdata/bin/create-load-data.sh
>  at line 85: CM_HOST=${2-}
> [...]
> 19:10:05 2019-01-16 19:10:05,461 - archive_core_dumps - INFO - Found core 
> files: ['./core.1547693849.30799.kudu-master', 
> './core.1547693849.30783.kudu-tserver', 
> './core.1547693849.30767.kudu-tserver', 
> './core.1547693849.30808.kudu-tserver']
> 19:10:05 2019-01-16 19:10:05,649 - archive_core_dumps - INFO - [New LWP 30799]
> 19:10:05 [New LWP 30834]
> 19:10:05 [New LWP 30835]
> 19:10:05 [New LWP 30838]
> 19:10:05 [New LWP 30837]
> 19:10:05 [New LWP 30836]
> 19:10:05 Core was generated by 
> `/data/jenkins/workspace/impala-asf-master-exhaustive-release/Impala-Toolchain/c'.
> 19:10:05 Program terminated with signal SIGABRT, Aborted.
> 19:10:05 #0  0x7f9a2cc611f7 in ?? ()
> 19:10:05 
> 19:10:05 2019-01-16 19:10:05,650 - archive_core_dumps - INFO - Found binary 
> path through GDB: 
> /data/jenkins/workspace/impala-asf-master-exhaustive-release/Impala-Toolchain/c
> 19:10:05 2019-01-16 19:10:05,893 - archive_core_dumps - WARNING - Failed to 
> determine binary because multiple candidate binaries were found and none of 
> their paths contained 'latest' to disambiguate:
> 19:10:05 Core:./core.1547693849.30799.kudu-master
> 19:10:05 
> Binaries:['./testdata/cluster/node_templates/common/etc/init.d/kudu-master', 
> './testdata/cluster/cdh6/node-1/etc/init.d/kudu-master']
> 19:10:05 
> 19:10:05 2019-01-16 19:10:05,917 - archive_core_dumps - INFO - [New LWP 30783]
> 19:10:05 [New LWP 30810]
> 19:10:05 [New LWP 30812]
> 19:10:05 [New LWP 30811]
> 19:10:05 [New LWP 30824]
> 19:10:05 [New LWP 30820]
> 19:10:05 Core was generated by 
> `/data/jenkins/workspace/impala-asf-master-exhaustive-release/Impala-Toolchain/c'.
> 19:10:05 Program terminated with signal SIGABRT, Aborted.
> 19:10:05 #0  0x7f0b81fb11f7 in ?? ()
> {noformat}
> *Backtraces*:
> {noformat}
> CORE: ./core.1547693849.30799.kudu-master
> BINARY: ./be/build/latest/service/impalad
> Core was generated by 
> `/data/jenkins/workspace/impala-asf-master-exhaustive-release/Impala-Toolchain/c'.
> Program terminated with signal SIGABRT, Aborted.
> #0  0x7f9a2cc611f7 in ?? ()
> #0  0x7f9a2cc611f7 in ?? ()
> #1  0x7f9a2cc628e8 in ?? ()
> #2  0x0020 in ?? ()
> #3  0x in ?? ()
> CORE: ./core.1547693849.30783.kudu-tserver
> BINARY: ./be/build/latest/service/impalad
> Core was generated by 
> `/data/jenkins/workspace/impala-asf-master-exhaustive-release/Impala-Toolchain/c'.
> Program terminated with signal SIGABRT, Aborted.
> #0  0x7f0b81fb11f7 in ?? ()
> #0  0x7f0b81fb11f7 in ?? ()
> #1  0x7f0b81fb28e8 in ?? ()
> #2  0x0020 in ?? ()
> #3  0x in ?? ()
> CORE: ./core.1547693849.30767.kudu-tserver
> BINARY: ./be/build/latest/service/impalad
> Core was generated by 
> `/data/jenkins/workspace/impala-asf-master-exhaustive-release/Impala-Toolchain/c'.
> Program terminated with signal SIGABRT, Aborted.
> #0  0x7f60b1e2f1f7 in ?? ()
> #0  0x7f60b1e2f1f7 in ?? ()
> #1  0x7f60b1e308e8 in ?? ()
> #2  0x0020 in ?? ()
> #3  0x in ?? ()
> CORE: ./core.1547693849.30808.kudu-tserver
> BINARY: ./be/build/latest/service/impalad
> Core was generated by 
> `/data/jenkins/workspace/impala-asf-master-exhaustive-release/Impala-Toolchain/c'.
> Program terminated with signal SIGABRT, Aborted.
> #0  0x7fa3cb5591f7 in ?? ()
> #0  0x7fa3cb5591f7 in ?? ()
> #1  0x7fa3cb55a8e8 in ?? ()
> #2  0x0020 in ?? ()
> #3  0x in ?? 

[jira] [Assigned] (IMPALA-8091) minicluster kudu failure: Cannot initialize clock: failed to wait for clock sync

2019-01-22 Thread Michael Brown (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8091?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Michael Brown reassigned IMPALA-8091:
-

Assignee: Michael Brown  (was: Thomas Tauber-Marshall)

> minicluster kudu failure: Cannot initialize clock: failed to wait for clock 
> sync
> 
>
> Key: IMPALA-8091
> URL: https://issues.apache.org/jira/browse/IMPALA-8091
> Project: IMPALA
>  Issue Type: Bug
>  Components: Infrastructure
>Affects Versions: Impala 3.2.0
>Reporter: David Knupp
>Assignee: Michael Brown
>Priority: Blocker
>  Labels: broken-build
>
> *Console log*:
> {noformat}
> 19:10:04 19:10:04 Error executing impala SQL: 
> /data/jenkins/workspace/impala-asf-master-exhaustive-release/repos/Impala/logs/data_loading/sql/functional/create-functional-query-exhaustive-impala-generated-kudu-none-none.sql
>  See: 
> /data/jenkins/workspace/impala-asf-master-exhaustive-release/repos/Impala/logs/data_loading/sql/functional/create-functional-query-exhaustive-impala-generated-kudu-none-none.sql.log
> 19:10:04 Background task Loading functional-query data (pid 4831) failed.
> 19:10:04 Background task Loading TPC-H data (pid 4832) failed.
> 19:10:04 ERROR in 
> /data/jenkins/workspace/impala-asf-master-exhaustive-release/repos/Impala/testdata/bin/create-load-data.sh
>  at line 85: CM_HOST=${2-}
> [...]
> 19:10:05 2019-01-16 19:10:05,461 - archive_core_dumps - INFO - Found core 
> files: ['./core.1547693849.30799.kudu-master', 
> './core.1547693849.30783.kudu-tserver', 
> './core.1547693849.30767.kudu-tserver', 
> './core.1547693849.30808.kudu-tserver']
> 19:10:05 2019-01-16 19:10:05,649 - archive_core_dumps - INFO - [New LWP 30799]
> 19:10:05 [New LWP 30834]
> 19:10:05 [New LWP 30835]
> 19:10:05 [New LWP 30838]
> 19:10:05 [New LWP 30837]
> 19:10:05 [New LWP 30836]
> 19:10:05 Core was generated by 
> `/data/jenkins/workspace/impala-asf-master-exhaustive-release/Impala-Toolchain/c'.
> 19:10:05 Program terminated with signal SIGABRT, Aborted.
> 19:10:05 #0  0x7f9a2cc611f7 in ?? ()
> 19:10:05 
> 19:10:05 2019-01-16 19:10:05,650 - archive_core_dumps - INFO - Found binary 
> path through GDB: 
> /data/jenkins/workspace/impala-asf-master-exhaustive-release/Impala-Toolchain/c
> 19:10:05 2019-01-16 19:10:05,893 - archive_core_dumps - WARNING - Failed to 
> determine binary because multiple candidate binaries were found and none of 
> their paths contained 'latest' to disambiguate:
> 19:10:05 Core:./core.1547693849.30799.kudu-master
> 19:10:05 
> Binaries:['./testdata/cluster/node_templates/common/etc/init.d/kudu-master', 
> './testdata/cluster/cdh6/node-1/etc/init.d/kudu-master']
> 19:10:05 
> 19:10:05 2019-01-16 19:10:05,917 - archive_core_dumps - INFO - [New LWP 30783]
> 19:10:05 [New LWP 30810]
> 19:10:05 [New LWP 30812]
> 19:10:05 [New LWP 30811]
> 19:10:05 [New LWP 30824]
> 19:10:05 [New LWP 30820]
> 19:10:05 Core was generated by 
> `/data/jenkins/workspace/impala-asf-master-exhaustive-release/Impala-Toolchain/c'.
> 19:10:05 Program terminated with signal SIGABRT, Aborted.
> 19:10:05 #0  0x7f0b81fb11f7 in ?? ()
> {noformat}
> *Backtraces*:
> {noformat}
> CORE: ./core.1547693849.30799.kudu-master
> BINARY: ./be/build/latest/service/impalad
> Core was generated by 
> `/data/jenkins/workspace/impala-asf-master-exhaustive-release/Impala-Toolchain/c'.
> Program terminated with signal SIGABRT, Aborted.
> #0  0x7f9a2cc611f7 in ?? ()
> #0  0x7f9a2cc611f7 in ?? ()
> #1  0x7f9a2cc628e8 in ?? ()
> #2  0x0020 in ?? ()
> #3  0x in ?? ()
> CORE: ./core.1547693849.30783.kudu-tserver
> BINARY: ./be/build/latest/service/impalad
> Core was generated by 
> `/data/jenkins/workspace/impala-asf-master-exhaustive-release/Impala-Toolchain/c'.
> Program terminated with signal SIGABRT, Aborted.
> #0  0x7f0b81fb11f7 in ?? ()
> #0  0x7f0b81fb11f7 in ?? ()
> #1  0x7f0b81fb28e8 in ?? ()
> #2  0x0020 in ?? ()
> #3  0x in ?? ()
> CORE: ./core.1547693849.30767.kudu-tserver
> BINARY: ./be/build/latest/service/impalad
> Core was generated by 
> `/data/jenkins/workspace/impala-asf-master-exhaustive-release/Impala-Toolchain/c'.
> Program terminated with signal SIGABRT, Aborted.
> #0  0x7f60b1e2f1f7 in ?? ()
> #0  0x7f60b1e2f1f7 in ?? ()
> #1  0x7f60b1e308e8 in ?? ()
> #2  0x0020 in ?? ()
> #3  0x in ?? ()
> CORE: ./core.1547693849.30808.kudu-tserver
> BINARY: ./be/build/latest/service/impalad
> Core was generated by 
> `/data/jenkins/workspace/impala-asf-master-exhaustive-release/Impala-Toolchain/c'.
> Program terminated with signal SIGABRT, Aborted.
> #0  0x7fa3cb5591f7 in ?? ()
> #0  0x7fa3cb5591f7 in ?? ()
> #1  0x7fa3cb55a8e8 in ?? ()
> #2  

[jira] [Work started] (IMPALA-7928) Investigate consistent placement of remote scan ranges

2019-01-22 Thread Joe McDonnell (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-7928?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Work on IMPALA-7928 started by Joe McDonnell.
-
> Investigate consistent placement of remote scan ranges
> --
>
> Key: IMPALA-7928
> URL: https://issues.apache.org/jira/browse/IMPALA-7928
> Project: IMPALA
>  Issue Type: Bug
>  Components: Backend
>Affects Versions: Impala 3.2.0
>Reporter: Joe McDonnell
>Assignee: Joe McDonnell
>Priority: Critical
>
> With the file handle cache, it is useful for repeated scans of the same file 
> to go to the same node, as that node will already have a file handle cached.
> When scheduling remote ranges, the scheduler introduces randomness that can 
> spread reads across all of the nodes. Repeated executions of queries on the 
> same set of files will not schedule the remote reads on the same nodes. This 
> causes a large amount of duplication across file handle caches on different 
> nodes. This reduces the efficiency of the cache significantly.
> It may be useful for the scheduler to introduce some determinism in 
> scheduling remote reads to take advantage of the file handle cache. This is a 
> variation on the well-known tradeoff between skew and locality.
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Created] (IMPALA-8097) Experimental flag for running all queries with mt_dop

2019-01-22 Thread Tim Armstrong (JIRA)
Tim Armstrong created IMPALA-8097:
-

 Summary: Experimental flag for running all queries with mt_dop
 Key: IMPALA-8097
 URL: https://issues.apache.org/jira/browse/IMPALA-8097
 Project: IMPALA
  Issue Type: Sub-task
Reporter: Tim Armstrong
Assignee: Tim Armstrong


It's possible to execute queries with joins and inserts with mt_dop with some 
small modifications to the Impala code (without the separate join build work). 
This isn't production-ready because of the other subtasks of IMPALA-3902 but it 
would be useful to have an experimental flag for people to play around with to 
test the functionality.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-8091) minicluster kudu failure: Cannot initialize clock: failed to wait for clock sync

2019-01-22 Thread Michael Brown (JIRA)


[ 
https://issues.apache.org/jira/browse/IMPALA-8091?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16749232#comment-16749232
 ] 

Michael Brown commented on IMPALA-8091:
---

Relevant code in {{testdata/cluster/admin}}:
{noformat}
  # line 333
  if [[ "${SERVICE}" == "kudu" ]]; then
NTP_CANARY=pool.ntp.org
if ! ping -c 1 -w 5 "${NTP_CANARY}" >/dev/null 2>/dev/null; then
  echo "WARNING: cannot reach ${NTP_CANARY}; ntp sync recommended for 
Kudu"
elif ntp-wait --help >/dev/null 2>/dev/null && ! ntp-wait -v; then
  echo "ntp-wait failed; cannot start kudu"
  return 1
fi
  fi
{noformat}

We silently move on if ntp-wait isn't present. We should warn. We don't try to 
install ntp-wait, which I get, because it's rude to do that to somebody's 
system. We should install ntp-perl for CentOS 7 in bootstrap_system.sh.

> minicluster kudu failure: Cannot initialize clock: failed to wait for clock 
> sync
> 
>
> Key: IMPALA-8091
> URL: https://issues.apache.org/jira/browse/IMPALA-8091
> Project: IMPALA
>  Issue Type: Bug
>  Components: Infrastructure
>Affects Versions: Impala 3.2.0
>Reporter: David Knupp
>Assignee: Thomas Tauber-Marshall
>Priority: Blocker
>  Labels: broken-build
>
> *Console log*:
> {noformat}
> 19:10:04 19:10:04 Error executing impala SQL: 
> /data/jenkins/workspace/impala-asf-master-exhaustive-release/repos/Impala/logs/data_loading/sql/functional/create-functional-query-exhaustive-impala-generated-kudu-none-none.sql
>  See: 
> /data/jenkins/workspace/impala-asf-master-exhaustive-release/repos/Impala/logs/data_loading/sql/functional/create-functional-query-exhaustive-impala-generated-kudu-none-none.sql.log
> 19:10:04 Background task Loading functional-query data (pid 4831) failed.
> 19:10:04 Background task Loading TPC-H data (pid 4832) failed.
> 19:10:04 ERROR in 
> /data/jenkins/workspace/impala-asf-master-exhaustive-release/repos/Impala/testdata/bin/create-load-data.sh
>  at line 85: CM_HOST=${2-}
> [...]
> 19:10:05 2019-01-16 19:10:05,461 - archive_core_dumps - INFO - Found core 
> files: ['./core.1547693849.30799.kudu-master', 
> './core.1547693849.30783.kudu-tserver', 
> './core.1547693849.30767.kudu-tserver', 
> './core.1547693849.30808.kudu-tserver']
> 19:10:05 2019-01-16 19:10:05,649 - archive_core_dumps - INFO - [New LWP 30799]
> 19:10:05 [New LWP 30834]
> 19:10:05 [New LWP 30835]
> 19:10:05 [New LWP 30838]
> 19:10:05 [New LWP 30837]
> 19:10:05 [New LWP 30836]
> 19:10:05 Core was generated by 
> `/data/jenkins/workspace/impala-asf-master-exhaustive-release/Impala-Toolchain/c'.
> 19:10:05 Program terminated with signal SIGABRT, Aborted.
> 19:10:05 #0  0x7f9a2cc611f7 in ?? ()
> 19:10:05 
> 19:10:05 2019-01-16 19:10:05,650 - archive_core_dumps - INFO - Found binary 
> path through GDB: 
> /data/jenkins/workspace/impala-asf-master-exhaustive-release/Impala-Toolchain/c
> 19:10:05 2019-01-16 19:10:05,893 - archive_core_dumps - WARNING - Failed to 
> determine binary because multiple candidate binaries were found and none of 
> their paths contained 'latest' to disambiguate:
> 19:10:05 Core:./core.1547693849.30799.kudu-master
> 19:10:05 
> Binaries:['./testdata/cluster/node_templates/common/etc/init.d/kudu-master', 
> './testdata/cluster/cdh6/node-1/etc/init.d/kudu-master']
> 19:10:05 
> 19:10:05 2019-01-16 19:10:05,917 - archive_core_dumps - INFO - [New LWP 30783]
> 19:10:05 [New LWP 30810]
> 19:10:05 [New LWP 30812]
> 19:10:05 [New LWP 30811]
> 19:10:05 [New LWP 30824]
> 19:10:05 [New LWP 30820]
> 19:10:05 Core was generated by 
> `/data/jenkins/workspace/impala-asf-master-exhaustive-release/Impala-Toolchain/c'.
> 19:10:05 Program terminated with signal SIGABRT, Aborted.
> 19:10:05 #0  0x7f0b81fb11f7 in ?? ()
> {noformat}
> *Backtraces*:
> {noformat}
> CORE: ./core.1547693849.30799.kudu-master
> BINARY: ./be/build/latest/service/impalad
> Core was generated by 
> `/data/jenkins/workspace/impala-asf-master-exhaustive-release/Impala-Toolchain/c'.
> Program terminated with signal SIGABRT, Aborted.
> #0  0x7f9a2cc611f7 in ?? ()
> #0  0x7f9a2cc611f7 in ?? ()
> #1  0x7f9a2cc628e8 in ?? ()
> #2  0x0020 in ?? ()
> #3  0x in ?? ()
> CORE: ./core.1547693849.30783.kudu-tserver
> BINARY: ./be/build/latest/service/impalad
> Core was generated by 
> `/data/jenkins/workspace/impala-asf-master-exhaustive-release/Impala-Toolchain/c'.
> Program terminated with signal SIGABRT, Aborted.
> #0  0x7f0b81fb11f7 in ?? ()
> #0  0x7f0b81fb11f7 in ?? ()
> #1  0x7f0b81fb28e8 in ?? ()
> #2  0x0020 in ?? ()
> #3  0x in ?? ()
> CORE: ./core.1547693849.30767.kudu-tserver
> BINARY: ./be/build/latest/service/impalad
> Core 

[jira] [Commented] (IMPALA-8091) minicluster kudu failure: Cannot initialize clock: failed to wait for clock sync

2019-01-22 Thread Michael Brown (JIRA)


[ 
https://issues.apache.org/jira/browse/IMPALA-8091?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16749190#comment-16749190
 ] 

Michael Brown commented on IMPALA-8091:
---

On Ubuntu 16, the {{ntp}} package provides {{ntp-wait}}. On CentOS 7, it's 
provided by {{ntp-perl}}. This is happening in a downstream environment where 
{{ntp-perl}} is not installed. On the same downstream environment, {{chrony}} 
seems to be installed but isn't running a daemon (ok). I'm looking to see if 
there is an upstream improvement that can be made here.

> minicluster kudu failure: Cannot initialize clock: failed to wait for clock 
> sync
> 
>
> Key: IMPALA-8091
> URL: https://issues.apache.org/jira/browse/IMPALA-8091
> Project: IMPALA
>  Issue Type: Bug
>  Components: Infrastructure
>Affects Versions: Impala 3.2.0
>Reporter: David Knupp
>Assignee: Thomas Tauber-Marshall
>Priority: Blocker
>  Labels: broken-build
>
> *Console log*:
> {noformat}
> 19:10:04 19:10:04 Error executing impala SQL: 
> /data/jenkins/workspace/impala-asf-master-exhaustive-release/repos/Impala/logs/data_loading/sql/functional/create-functional-query-exhaustive-impala-generated-kudu-none-none.sql
>  See: 
> /data/jenkins/workspace/impala-asf-master-exhaustive-release/repos/Impala/logs/data_loading/sql/functional/create-functional-query-exhaustive-impala-generated-kudu-none-none.sql.log
> 19:10:04 Background task Loading functional-query data (pid 4831) failed.
> 19:10:04 Background task Loading TPC-H data (pid 4832) failed.
> 19:10:04 ERROR in 
> /data/jenkins/workspace/impala-asf-master-exhaustive-release/repos/Impala/testdata/bin/create-load-data.sh
>  at line 85: CM_HOST=${2-}
> [...]
> 19:10:05 2019-01-16 19:10:05,461 - archive_core_dumps - INFO - Found core 
> files: ['./core.1547693849.30799.kudu-master', 
> './core.1547693849.30783.kudu-tserver', 
> './core.1547693849.30767.kudu-tserver', 
> './core.1547693849.30808.kudu-tserver']
> 19:10:05 2019-01-16 19:10:05,649 - archive_core_dumps - INFO - [New LWP 30799]
> 19:10:05 [New LWP 30834]
> 19:10:05 [New LWP 30835]
> 19:10:05 [New LWP 30838]
> 19:10:05 [New LWP 30837]
> 19:10:05 [New LWP 30836]
> 19:10:05 Core was generated by 
> `/data/jenkins/workspace/impala-asf-master-exhaustive-release/Impala-Toolchain/c'.
> 19:10:05 Program terminated with signal SIGABRT, Aborted.
> 19:10:05 #0  0x7f9a2cc611f7 in ?? ()
> 19:10:05 
> 19:10:05 2019-01-16 19:10:05,650 - archive_core_dumps - INFO - Found binary 
> path through GDB: 
> /data/jenkins/workspace/impala-asf-master-exhaustive-release/Impala-Toolchain/c
> 19:10:05 2019-01-16 19:10:05,893 - archive_core_dumps - WARNING - Failed to 
> determine binary because multiple candidate binaries were found and none of 
> their paths contained 'latest' to disambiguate:
> 19:10:05 Core:./core.1547693849.30799.kudu-master
> 19:10:05 
> Binaries:['./testdata/cluster/node_templates/common/etc/init.d/kudu-master', 
> './testdata/cluster/cdh6/node-1/etc/init.d/kudu-master']
> 19:10:05 
> 19:10:05 2019-01-16 19:10:05,917 - archive_core_dumps - INFO - [New LWP 30783]
> 19:10:05 [New LWP 30810]
> 19:10:05 [New LWP 30812]
> 19:10:05 [New LWP 30811]
> 19:10:05 [New LWP 30824]
> 19:10:05 [New LWP 30820]
> 19:10:05 Core was generated by 
> `/data/jenkins/workspace/impala-asf-master-exhaustive-release/Impala-Toolchain/c'.
> 19:10:05 Program terminated with signal SIGABRT, Aborted.
> 19:10:05 #0  0x7f0b81fb11f7 in ?? ()
> {noformat}
> *Backtraces*:
> {noformat}
> CORE: ./core.1547693849.30799.kudu-master
> BINARY: ./be/build/latest/service/impalad
> Core was generated by 
> `/data/jenkins/workspace/impala-asf-master-exhaustive-release/Impala-Toolchain/c'.
> Program terminated with signal SIGABRT, Aborted.
> #0  0x7f9a2cc611f7 in ?? ()
> #0  0x7f9a2cc611f7 in ?? ()
> #1  0x7f9a2cc628e8 in ?? ()
> #2  0x0020 in ?? ()
> #3  0x in ?? ()
> CORE: ./core.1547693849.30783.kudu-tserver
> BINARY: ./be/build/latest/service/impalad
> Core was generated by 
> `/data/jenkins/workspace/impala-asf-master-exhaustive-release/Impala-Toolchain/c'.
> Program terminated with signal SIGABRT, Aborted.
> #0  0x7f0b81fb11f7 in ?? ()
> #0  0x7f0b81fb11f7 in ?? ()
> #1  0x7f0b81fb28e8 in ?? ()
> #2  0x0020 in ?? ()
> #3  0x in ?? ()
> CORE: ./core.1547693849.30767.kudu-tserver
> BINARY: ./be/build/latest/service/impalad
> Core was generated by 
> `/data/jenkins/workspace/impala-asf-master-exhaustive-release/Impala-Toolchain/c'.
> Program terminated with signal SIGABRT, Aborted.
> #0  0x7f60b1e2f1f7 in ?? ()
> #0  0x7f60b1e2f1f7 in ?? ()
> #1  0x7f60b1e308e8 in ?? ()
> #2  0x0020 in ?? ()
> #3  0x in ?? ()
> CORE: 

[jira] [Work started] (IMPALA-7934) Switch to using Java 8's Base64 impl for incremental stats encoding

2019-01-22 Thread Fredy Wijaya (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-7934?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Work on IMPALA-7934 started by Fredy Wijaya.

> Switch to using Java 8's Base64 impl for incremental stats encoding
> ---
>
> Key: IMPALA-7934
> URL: https://issues.apache.org/jira/browse/IMPALA-7934
> Project: IMPALA
>  Issue Type: Bug
>  Components: Catalog
>Affects Versions: Impala 3.1.0
>Reporter: bharath v
>Assignee: Fredy Wijaya
>Priority: Major
>  Labels: ramp-up
> Attachments: base64.png
>
>
> Incremental stats are compressed and Base64 encoded before they are chunked 
> and written to the HMS' partition parameters map. When they are read back, we 
> need to Base64 decode and decompress. 
> For certain incremental stats heavy tables, we noticed that a significant 
> amount of time is spent in these base64 classes (see the attached image for 
> the stack. Unfortunately, I don't have the text version of it).
> Java 8 comes with its own Base64 implementation and that has shown much 
> better perf results [1] compared to apache codec's impl. So consider 
> switching to Java 8's base64 impl.
>  [1] http://java-performance.info/base64-encoding-and-decoding-performance/
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-8091) minicluster kudu failure: Cannot initialize clock: failed to wait for clock sync

2019-01-22 Thread Michael Brown (JIRA)


[ 
https://issues.apache.org/jira/browse/IMPALA-8091?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16749176#comment-16749176
 ] 

Michael Brown commented on IMPALA-8091:
---

This affected numerous runs over a quiet weekend without much commit churn.

> minicluster kudu failure: Cannot initialize clock: failed to wait for clock 
> sync
> 
>
> Key: IMPALA-8091
> URL: https://issues.apache.org/jira/browse/IMPALA-8091
> Project: IMPALA
>  Issue Type: Bug
>  Components: Infrastructure
>Affects Versions: Impala 3.2.0
>Reporter: David Knupp
>Assignee: Thomas Tauber-Marshall
>Priority: Blocker
>  Labels: broken-build
>
> *Console log*:
> {noformat}
> 19:10:04 19:10:04 Error executing impala SQL: 
> /data/jenkins/workspace/impala-asf-master-exhaustive-release/repos/Impala/logs/data_loading/sql/functional/create-functional-query-exhaustive-impala-generated-kudu-none-none.sql
>  See: 
> /data/jenkins/workspace/impala-asf-master-exhaustive-release/repos/Impala/logs/data_loading/sql/functional/create-functional-query-exhaustive-impala-generated-kudu-none-none.sql.log
> 19:10:04 Background task Loading functional-query data (pid 4831) failed.
> 19:10:04 Background task Loading TPC-H data (pid 4832) failed.
> 19:10:04 ERROR in 
> /data/jenkins/workspace/impala-asf-master-exhaustive-release/repos/Impala/testdata/bin/create-load-data.sh
>  at line 85: CM_HOST=${2-}
> [...]
> 19:10:05 2019-01-16 19:10:05,461 - archive_core_dumps - INFO - Found core 
> files: ['./core.1547693849.30799.kudu-master', 
> './core.1547693849.30783.kudu-tserver', 
> './core.1547693849.30767.kudu-tserver', 
> './core.1547693849.30808.kudu-tserver']
> 19:10:05 2019-01-16 19:10:05,649 - archive_core_dumps - INFO - [New LWP 30799]
> 19:10:05 [New LWP 30834]
> 19:10:05 [New LWP 30835]
> 19:10:05 [New LWP 30838]
> 19:10:05 [New LWP 30837]
> 19:10:05 [New LWP 30836]
> 19:10:05 Core was generated by 
> `/data/jenkins/workspace/impala-asf-master-exhaustive-release/Impala-Toolchain/c'.
> 19:10:05 Program terminated with signal SIGABRT, Aborted.
> 19:10:05 #0  0x7f9a2cc611f7 in ?? ()
> 19:10:05 
> 19:10:05 2019-01-16 19:10:05,650 - archive_core_dumps - INFO - Found binary 
> path through GDB: 
> /data/jenkins/workspace/impala-asf-master-exhaustive-release/Impala-Toolchain/c
> 19:10:05 2019-01-16 19:10:05,893 - archive_core_dumps - WARNING - Failed to 
> determine binary because multiple candidate binaries were found and none of 
> their paths contained 'latest' to disambiguate:
> 19:10:05 Core:./core.1547693849.30799.kudu-master
> 19:10:05 
> Binaries:['./testdata/cluster/node_templates/common/etc/init.d/kudu-master', 
> './testdata/cluster/cdh6/node-1/etc/init.d/kudu-master']
> 19:10:05 
> 19:10:05 2019-01-16 19:10:05,917 - archive_core_dumps - INFO - [New LWP 30783]
> 19:10:05 [New LWP 30810]
> 19:10:05 [New LWP 30812]
> 19:10:05 [New LWP 30811]
> 19:10:05 [New LWP 30824]
> 19:10:05 [New LWP 30820]
> 19:10:05 Core was generated by 
> `/data/jenkins/workspace/impala-asf-master-exhaustive-release/Impala-Toolchain/c'.
> 19:10:05 Program terminated with signal SIGABRT, Aborted.
> 19:10:05 #0  0x7f0b81fb11f7 in ?? ()
> {noformat}
> *Backtraces*:
> {noformat}
> CORE: ./core.1547693849.30799.kudu-master
> BINARY: ./be/build/latest/service/impalad
> Core was generated by 
> `/data/jenkins/workspace/impala-asf-master-exhaustive-release/Impala-Toolchain/c'.
> Program terminated with signal SIGABRT, Aborted.
> #0  0x7f9a2cc611f7 in ?? ()
> #0  0x7f9a2cc611f7 in ?? ()
> #1  0x7f9a2cc628e8 in ?? ()
> #2  0x0020 in ?? ()
> #3  0x in ?? ()
> CORE: ./core.1547693849.30783.kudu-tserver
> BINARY: ./be/build/latest/service/impalad
> Core was generated by 
> `/data/jenkins/workspace/impala-asf-master-exhaustive-release/Impala-Toolchain/c'.
> Program terminated with signal SIGABRT, Aborted.
> #0  0x7f0b81fb11f7 in ?? ()
> #0  0x7f0b81fb11f7 in ?? ()
> #1  0x7f0b81fb28e8 in ?? ()
> #2  0x0020 in ?? ()
> #3  0x in ?? ()
> CORE: ./core.1547693849.30767.kudu-tserver
> BINARY: ./be/build/latest/service/impalad
> Core was generated by 
> `/data/jenkins/workspace/impala-asf-master-exhaustive-release/Impala-Toolchain/c'.
> Program terminated with signal SIGABRT, Aborted.
> #0  0x7f60b1e2f1f7 in ?? ()
> #0  0x7f60b1e2f1f7 in ?? ()
> #1  0x7f60b1e308e8 in ?? ()
> #2  0x0020 in ?? ()
> #3  0x in ?? ()
> CORE: ./core.1547693849.30808.kudu-tserver
> BINARY: ./be/build/latest/service/impalad
> Core was generated by 
> `/data/jenkins/workspace/impala-asf-master-exhaustive-release/Impala-Toolchain/c'.
> Program terminated with signal SIGABRT, Aborted.
> #0  0x7fa3cb5591f7 in ?? ()
> #0  0x7fa3cb5591f7 in 

[jira] [Updated] (IMPALA-8091) minicluster kudu failure: Cannot initialize clock: failed to wait for clock sync

2019-01-22 Thread Michael Brown (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8091?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Michael Brown updated IMPALA-8091:
--
Priority: Blocker  (was: Critical)

> minicluster kudu failure: Cannot initialize clock: failed to wait for clock 
> sync
> 
>
> Key: IMPALA-8091
> URL: https://issues.apache.org/jira/browse/IMPALA-8091
> Project: IMPALA
>  Issue Type: Bug
>  Components: Infrastructure
>Affects Versions: Impala 3.2.0
>Reporter: David Knupp
>Assignee: Thomas Tauber-Marshall
>Priority: Blocker
>  Labels: broken-build
>
> *Console log*:
> {noformat}
> 19:10:04 19:10:04 Error executing impala SQL: 
> /data/jenkins/workspace/impala-asf-master-exhaustive-release/repos/Impala/logs/data_loading/sql/functional/create-functional-query-exhaustive-impala-generated-kudu-none-none.sql
>  See: 
> /data/jenkins/workspace/impala-asf-master-exhaustive-release/repos/Impala/logs/data_loading/sql/functional/create-functional-query-exhaustive-impala-generated-kudu-none-none.sql.log
> 19:10:04 Background task Loading functional-query data (pid 4831) failed.
> 19:10:04 Background task Loading TPC-H data (pid 4832) failed.
> 19:10:04 ERROR in 
> /data/jenkins/workspace/impala-asf-master-exhaustive-release/repos/Impala/testdata/bin/create-load-data.sh
>  at line 85: CM_HOST=${2-}
> [...]
> 19:10:05 2019-01-16 19:10:05,461 - archive_core_dumps - INFO - Found core 
> files: ['./core.1547693849.30799.kudu-master', 
> './core.1547693849.30783.kudu-tserver', 
> './core.1547693849.30767.kudu-tserver', 
> './core.1547693849.30808.kudu-tserver']
> 19:10:05 2019-01-16 19:10:05,649 - archive_core_dumps - INFO - [New LWP 30799]
> 19:10:05 [New LWP 30834]
> 19:10:05 [New LWP 30835]
> 19:10:05 [New LWP 30838]
> 19:10:05 [New LWP 30837]
> 19:10:05 [New LWP 30836]
> 19:10:05 Core was generated by 
> `/data/jenkins/workspace/impala-asf-master-exhaustive-release/Impala-Toolchain/c'.
> 19:10:05 Program terminated with signal SIGABRT, Aborted.
> 19:10:05 #0  0x7f9a2cc611f7 in ?? ()
> 19:10:05 
> 19:10:05 2019-01-16 19:10:05,650 - archive_core_dumps - INFO - Found binary 
> path through GDB: 
> /data/jenkins/workspace/impala-asf-master-exhaustive-release/Impala-Toolchain/c
> 19:10:05 2019-01-16 19:10:05,893 - archive_core_dumps - WARNING - Failed to 
> determine binary because multiple candidate binaries were found and none of 
> their paths contained 'latest' to disambiguate:
> 19:10:05 Core:./core.1547693849.30799.kudu-master
> 19:10:05 
> Binaries:['./testdata/cluster/node_templates/common/etc/init.d/kudu-master', 
> './testdata/cluster/cdh6/node-1/etc/init.d/kudu-master']
> 19:10:05 
> 19:10:05 2019-01-16 19:10:05,917 - archive_core_dumps - INFO - [New LWP 30783]
> 19:10:05 [New LWP 30810]
> 19:10:05 [New LWP 30812]
> 19:10:05 [New LWP 30811]
> 19:10:05 [New LWP 30824]
> 19:10:05 [New LWP 30820]
> 19:10:05 Core was generated by 
> `/data/jenkins/workspace/impala-asf-master-exhaustive-release/Impala-Toolchain/c'.
> 19:10:05 Program terminated with signal SIGABRT, Aborted.
> 19:10:05 #0  0x7f0b81fb11f7 in ?? ()
> {noformat}
> *Backtraces*:
> {noformat}
> CORE: ./core.1547693849.30799.kudu-master
> BINARY: ./be/build/latest/service/impalad
> Core was generated by 
> `/data/jenkins/workspace/impala-asf-master-exhaustive-release/Impala-Toolchain/c'.
> Program terminated with signal SIGABRT, Aborted.
> #0  0x7f9a2cc611f7 in ?? ()
> #0  0x7f9a2cc611f7 in ?? ()
> #1  0x7f9a2cc628e8 in ?? ()
> #2  0x0020 in ?? ()
> #3  0x in ?? ()
> CORE: ./core.1547693849.30783.kudu-tserver
> BINARY: ./be/build/latest/service/impalad
> Core was generated by 
> `/data/jenkins/workspace/impala-asf-master-exhaustive-release/Impala-Toolchain/c'.
> Program terminated with signal SIGABRT, Aborted.
> #0  0x7f0b81fb11f7 in ?? ()
> #0  0x7f0b81fb11f7 in ?? ()
> #1  0x7f0b81fb28e8 in ?? ()
> #2  0x0020 in ?? ()
> #3  0x in ?? ()
> CORE: ./core.1547693849.30767.kudu-tserver
> BINARY: ./be/build/latest/service/impalad
> Core was generated by 
> `/data/jenkins/workspace/impala-asf-master-exhaustive-release/Impala-Toolchain/c'.
> Program terminated with signal SIGABRT, Aborted.
> #0  0x7f60b1e2f1f7 in ?? ()
> #0  0x7f60b1e2f1f7 in ?? ()
> #1  0x7f60b1e308e8 in ?? ()
> #2  0x0020 in ?? ()
> #3  0x in ?? ()
> CORE: ./core.1547693849.30808.kudu-tserver
> BINARY: ./be/build/latest/service/impalad
> Core was generated by 
> `/data/jenkins/workspace/impala-asf-master-exhaustive-release/Impala-Toolchain/c'.
> Program terminated with signal SIGABRT, Aborted.
> #0  0x7fa3cb5591f7 in ?? ()
> #0  0x7fa3cb5591f7 in ?? ()
> #1  0x7fa3cb55a8e8 in ?? ()
> #2  0x0020 in ?? ()
> 

[jira] [Assigned] (IMPALA-8096) Limit on #rows returned from query

2019-01-22 Thread Pooja Nilangekar (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8096?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Pooja Nilangekar reassigned IMPALA-8096:


Assignee: Pooja Nilangekar

> Limit on #rows returned from query
> --
>
> Key: IMPALA-8096
> URL: https://issues.apache.org/jira/browse/IMPALA-8096
> Project: IMPALA
>  Issue Type: Sub-task
>  Components: Backend
>Reporter: Tim Armstrong
>Assignee: Pooja Nilangekar
>Priority: Major
>  Labels: resource-management
>
> Sometimes users accidentally run queries that return a large number of rows, 
> e.g.
> {code}
> SELECT * FROM table
> {code}
> When they really only need to look at a subset of the rows. It would be 
> useful to have a guardrail to fail queries the return more rows than a 
> particular limit. Maybe it would make sense to integrate with IMPALA-4268 so 
> that the query is failed when the buffer fills up, but it may also be useful 
> to have an easier-to-understand option based on #rows.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-8091) minicluster kudu failure: Cannot initialize clock: failed to wait for clock sync

2019-01-22 Thread Michael Brown (JIRA)


[ 
https://issues.apache.org/jira/browse/IMPALA-8091?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16749157#comment-16749157
 ] 

Michael Brown commented on IMPALA-8091:
---

This is a lot of diagnostic info...
{noformat}
I0116 18:56:30.146215 30799 master_main.cc:80] Initializing master server...
I0116 18:56:30.146333 30799 system_ntp.cc:122] Waiting up to 
--ntp_initial_sync_wait_secs=60 seconds for the clock to synchronize
W0116 18:56:30.147728 30799 system_ntp.cc:132] Could not find ntp-wait; trying 
chrony waitsync instead: Not found: Unable to find binary: ntp-wait
E0116 18:57:29.300457 30799 system_ntp.cc:157] Dumping NTP diagnostics
E0116 18:57:29.302417 30799 system_ntp.cc:102] /sbin/ntptime
--
stdout:
ntp_gettime() returns code 5 (ERROR)
  time dfea6d99.4d59a2ec  Wed, Jan 16 2019 18:57:29.302, (.302149739),
  maximum error 1600 us, estimated error 1600 us, TAI offset 0
ntp_adjtime() returns code 5 (ERROR)
  modes 0x0 (),
  offset 0.000 us, frequency 129.100 ppm, interval 1 s,
  maximum error 1600 us, estimated error 1600 us,
  status 0x2041 (PLL,UNSYNC,NANO),
  time constant 6, precision 0.001 us, tolerance 500 ppm,

E0116 18:57:29.330603 30799 system_ntp.cc:102] /sbin/ntpq -n -c timeout 1000 -c 
readvar -c sysinfo -c lpeers -c opeers -c version
--
stdout:
associd=0 status=c618 leap_alarm, sync_ntp, 1 event, no_sys_peer,
version="ntpd 4.2.6p5@1.2349-o Wed Apr 12 21:24:06 UTC 2017 (1)",
processor="x86_64", system="Linux/3.10.0-693.5.2.el7.x86_64", leap=11,
stratum=2, precision=-24, rootdelay=65.438, rootdisp=43.467,
refid=96.126.122.39,
reftime=dfea6d72.e0d7d2b9  Wed, Jan 16 2019 18:56:50.878,
clock=dfea6d99.4f1265e8  Wed, Jan 16 2019 18:57:29.308, peer=64381, tc=6,
mintc=3, offset=0.000, frequency=129.100, sys_jitter=11.220,
clk_jitter=12.567, clk_wander=0.221
 remote   refid  st t when poll reach   delay   offset  jitter
==
+66.228.48.38127.67.113.922 u   39   647   70.236  -35.592  28.619
+208.75.89.4 63.145.169.3 2 u   37   647   64.246  -20.776  29.110
*96.126.122.39   .GPS.1 u   38   647   63.424  -31.810  28.640
+69.89.207.199   212.215.1.1572 u   36   647   37.729  -37.207  28.643
 remote   local  st t when poll reach   delay   offsetdisp
==
+66.228.48.38172.28.204.722 u   39   647   70.236  -35.592   0.913
+208.75.89.4 172.28.204.722 u   37   647   64.246  -20.776   0.921
*96.126.122.39   172.28.204.721 u   38   647   63.424  -31.810   0.906
+69.89.207.199   172.28.204.722 u   36   647   37.729  -37.207   0.914
ntpq 4.2.6p5@1.2349-o Wed Apr 12 21:24:08 UTC 2017 (1)

stderr:
***Command `sysinfo' unknown

E0116 18:57:29.333056 30799 system_ntp.cc:102] /sbin/ntpdc -n -c timeout 1000 
-c peers -c sysinfo -c sysstats -c version
--
stdout:
 remote   local  st poll reach  delay   offsetdisp
===
*96.126.122.39   172.28.204.721   647 0.06342 -0.031810 0.03009
=208.75.89.4 172.28.204.722   647 0.06424 -0.020776 0.03033
=69.89.207.199   172.28.204.722   647 0.03772 -0.037207 0.03023
=66.228.48.38172.28.204.722   647 0.07022 -0.035592 0.03021
system peer:  96.126.122.39
system peer mode: client
leap indicator:   11
stratum:  2
precision:-24
root distance:0.06543 s
root dispersion:  0.04346 s
reference ID: [96.126.122.39]
reference time:   dfea6d72.e0d7d2b9  Wed, Jan 16 2019 18:56:50.878
system flags: auth ntp kernel stats
jitter:   0.011215 s
stability:0.000 ppm
broadcastdelay:   0.00 s
authdelay:0.00 s
time since restart: 1459
time since reset:   1459
packets received:   180
packets processed:  124
current version:124
previous version:   0
declined:   0
access denied:  0
bad length or format:   0
bad authentication: 0
rate exceeded:  0
ntpdc 4.2.6p5@1.2349-o Wed Apr 12 21:24:08 UTC 2017 (1)

E0116 18:57:29.336165 30799 system_ntp.cc:102] /usr/bin/chronyc -n tracking
--
stdout:
506 Cannot talk to daemon

E0116 18:57:29.339103 30799 system_ntp.cc:102] /usr/bin/chronyc -n sources
--
stdout:
506 Cannot talk to daemon

F0116 18:57:29.339143 30799 master_main.cc:81] Check failed: _s.ok() Bad 
status: Runtime error: Cannot initialize clock: failed to wait for clock sync 
using command '/usr/bin/chronyc waitsync 60 0 0 1': /usr/bin/chronyc: process 
exited 

[jira] [Updated] (IMPALA-8091) minicluster kudu failure: Cannot initialize clock: failed to wait for clock sync

2019-01-22 Thread Michael Brown (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8091?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Michael Brown updated IMPALA-8091:
--
Component/s: (was: Backend)
 Infrastructure

> minicluster kudu failure: Cannot initialize clock: failed to wait for clock 
> sync
> 
>
> Key: IMPALA-8091
> URL: https://issues.apache.org/jira/browse/IMPALA-8091
> Project: IMPALA
>  Issue Type: Bug
>  Components: Infrastructure
>Affects Versions: Impala 3.2.0
>Reporter: David Knupp
>Assignee: Thomas Tauber-Marshall
>Priority: Critical
>  Labels: broken-build
>
> *Console log*:
> {noformat}
> 19:10:04 19:10:04 Error executing impala SQL: 
> /data/jenkins/workspace/impala-asf-master-exhaustive-release/repos/Impala/logs/data_loading/sql/functional/create-functional-query-exhaustive-impala-generated-kudu-none-none.sql
>  See: 
> /data/jenkins/workspace/impala-asf-master-exhaustive-release/repos/Impala/logs/data_loading/sql/functional/create-functional-query-exhaustive-impala-generated-kudu-none-none.sql.log
> 19:10:04 Background task Loading functional-query data (pid 4831) failed.
> 19:10:04 Background task Loading TPC-H data (pid 4832) failed.
> 19:10:04 ERROR in 
> /data/jenkins/workspace/impala-asf-master-exhaustive-release/repos/Impala/testdata/bin/create-load-data.sh
>  at line 85: CM_HOST=${2-}
> [...]
> 19:10:05 2019-01-16 19:10:05,461 - archive_core_dumps - INFO - Found core 
> files: ['./core.1547693849.30799.kudu-master', 
> './core.1547693849.30783.kudu-tserver', 
> './core.1547693849.30767.kudu-tserver', 
> './core.1547693849.30808.kudu-tserver']
> 19:10:05 2019-01-16 19:10:05,649 - archive_core_dumps - INFO - [New LWP 30799]
> 19:10:05 [New LWP 30834]
> 19:10:05 [New LWP 30835]
> 19:10:05 [New LWP 30838]
> 19:10:05 [New LWP 30837]
> 19:10:05 [New LWP 30836]
> 19:10:05 Core was generated by 
> `/data/jenkins/workspace/impala-asf-master-exhaustive-release/Impala-Toolchain/c'.
> 19:10:05 Program terminated with signal SIGABRT, Aborted.
> 19:10:05 #0  0x7f9a2cc611f7 in ?? ()
> 19:10:05 
> 19:10:05 2019-01-16 19:10:05,650 - archive_core_dumps - INFO - Found binary 
> path through GDB: 
> /data/jenkins/workspace/impala-asf-master-exhaustive-release/Impala-Toolchain/c
> 19:10:05 2019-01-16 19:10:05,893 - archive_core_dumps - WARNING - Failed to 
> determine binary because multiple candidate binaries were found and none of 
> their paths contained 'latest' to disambiguate:
> 19:10:05 Core:./core.1547693849.30799.kudu-master
> 19:10:05 
> Binaries:['./testdata/cluster/node_templates/common/etc/init.d/kudu-master', 
> './testdata/cluster/cdh6/node-1/etc/init.d/kudu-master']
> 19:10:05 
> 19:10:05 2019-01-16 19:10:05,917 - archive_core_dumps - INFO - [New LWP 30783]
> 19:10:05 [New LWP 30810]
> 19:10:05 [New LWP 30812]
> 19:10:05 [New LWP 30811]
> 19:10:05 [New LWP 30824]
> 19:10:05 [New LWP 30820]
> 19:10:05 Core was generated by 
> `/data/jenkins/workspace/impala-asf-master-exhaustive-release/Impala-Toolchain/c'.
> 19:10:05 Program terminated with signal SIGABRT, Aborted.
> 19:10:05 #0  0x7f0b81fb11f7 in ?? ()
> {noformat}
> *Backtraces*:
> {noformat}
> CORE: ./core.1547693849.30799.kudu-master
> BINARY: ./be/build/latest/service/impalad
> Core was generated by 
> `/data/jenkins/workspace/impala-asf-master-exhaustive-release/Impala-Toolchain/c'.
> Program terminated with signal SIGABRT, Aborted.
> #0  0x7f9a2cc611f7 in ?? ()
> #0  0x7f9a2cc611f7 in ?? ()
> #1  0x7f9a2cc628e8 in ?? ()
> #2  0x0020 in ?? ()
> #3  0x in ?? ()
> CORE: ./core.1547693849.30783.kudu-tserver
> BINARY: ./be/build/latest/service/impalad
> Core was generated by 
> `/data/jenkins/workspace/impala-asf-master-exhaustive-release/Impala-Toolchain/c'.
> Program terminated with signal SIGABRT, Aborted.
> #0  0x7f0b81fb11f7 in ?? ()
> #0  0x7f0b81fb11f7 in ?? ()
> #1  0x7f0b81fb28e8 in ?? ()
> #2  0x0020 in ?? ()
> #3  0x in ?? ()
> CORE: ./core.1547693849.30767.kudu-tserver
> BINARY: ./be/build/latest/service/impalad
> Core was generated by 
> `/data/jenkins/workspace/impala-asf-master-exhaustive-release/Impala-Toolchain/c'.
> Program terminated with signal SIGABRT, Aborted.
> #0  0x7f60b1e2f1f7 in ?? ()
> #0  0x7f60b1e2f1f7 in ?? ()
> #1  0x7f60b1e308e8 in ?? ()
> #2  0x0020 in ?? ()
> #3  0x in ?? ()
> CORE: ./core.1547693849.30808.kudu-tserver
> BINARY: ./be/build/latest/service/impalad
> Core was generated by 
> `/data/jenkins/workspace/impala-asf-master-exhaustive-release/Impala-Toolchain/c'.
> Program terminated with signal SIGABRT, Aborted.
> #0  0x7fa3cb5591f7 in ?? ()
> #0  0x7fa3cb5591f7 in ?? ()
> #1  0x7fa3cb55a8e8 in ?? ()
> #2  

[jira] [Updated] (IMPALA-8091) minicluster kudu failure: Cannot initialize clock: failed to wait for clock sync

2019-01-22 Thread Michael Brown (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8091?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Michael Brown updated IMPALA-8091:
--
Summary: minicluster kudu failure: Cannot initialize clock: failed to wait 
for clock sync  (was: Kudu SIGABRT'ed during dataload on Impala release build)

> minicluster kudu failure: Cannot initialize clock: failed to wait for clock 
> sync
> 
>
> Key: IMPALA-8091
> URL: https://issues.apache.org/jira/browse/IMPALA-8091
> Project: IMPALA
>  Issue Type: Bug
>  Components: Backend
>Affects Versions: Impala 3.2.0
>Reporter: David Knupp
>Assignee: Thomas Tauber-Marshall
>Priority: Critical
>  Labels: broken-build
>
> *Console log*:
> {noformat}
> 19:10:04 19:10:04 Error executing impala SQL: 
> /data/jenkins/workspace/impala-asf-master-exhaustive-release/repos/Impala/logs/data_loading/sql/functional/create-functional-query-exhaustive-impala-generated-kudu-none-none.sql
>  See: 
> /data/jenkins/workspace/impala-asf-master-exhaustive-release/repos/Impala/logs/data_loading/sql/functional/create-functional-query-exhaustive-impala-generated-kudu-none-none.sql.log
> 19:10:04 Background task Loading functional-query data (pid 4831) failed.
> 19:10:04 Background task Loading TPC-H data (pid 4832) failed.
> 19:10:04 ERROR in 
> /data/jenkins/workspace/impala-asf-master-exhaustive-release/repos/Impala/testdata/bin/create-load-data.sh
>  at line 85: CM_HOST=${2-}
> [...]
> 19:10:05 2019-01-16 19:10:05,461 - archive_core_dumps - INFO - Found core 
> files: ['./core.1547693849.30799.kudu-master', 
> './core.1547693849.30783.kudu-tserver', 
> './core.1547693849.30767.kudu-tserver', 
> './core.1547693849.30808.kudu-tserver']
> 19:10:05 2019-01-16 19:10:05,649 - archive_core_dumps - INFO - [New LWP 30799]
> 19:10:05 [New LWP 30834]
> 19:10:05 [New LWP 30835]
> 19:10:05 [New LWP 30838]
> 19:10:05 [New LWP 30837]
> 19:10:05 [New LWP 30836]
> 19:10:05 Core was generated by 
> `/data/jenkins/workspace/impala-asf-master-exhaustive-release/Impala-Toolchain/c'.
> 19:10:05 Program terminated with signal SIGABRT, Aborted.
> 19:10:05 #0  0x7f9a2cc611f7 in ?? ()
> 19:10:05 
> 19:10:05 2019-01-16 19:10:05,650 - archive_core_dumps - INFO - Found binary 
> path through GDB: 
> /data/jenkins/workspace/impala-asf-master-exhaustive-release/Impala-Toolchain/c
> 19:10:05 2019-01-16 19:10:05,893 - archive_core_dumps - WARNING - Failed to 
> determine binary because multiple candidate binaries were found and none of 
> their paths contained 'latest' to disambiguate:
> 19:10:05 Core:./core.1547693849.30799.kudu-master
> 19:10:05 
> Binaries:['./testdata/cluster/node_templates/common/etc/init.d/kudu-master', 
> './testdata/cluster/cdh6/node-1/etc/init.d/kudu-master']
> 19:10:05 
> 19:10:05 2019-01-16 19:10:05,917 - archive_core_dumps - INFO - [New LWP 30783]
> 19:10:05 [New LWP 30810]
> 19:10:05 [New LWP 30812]
> 19:10:05 [New LWP 30811]
> 19:10:05 [New LWP 30824]
> 19:10:05 [New LWP 30820]
> 19:10:05 Core was generated by 
> `/data/jenkins/workspace/impala-asf-master-exhaustive-release/Impala-Toolchain/c'.
> 19:10:05 Program terminated with signal SIGABRT, Aborted.
> 19:10:05 #0  0x7f0b81fb11f7 in ?? ()
> {noformat}
> *Backtraces*:
> {noformat}
> CORE: ./core.1547693849.30799.kudu-master
> BINARY: ./be/build/latest/service/impalad
> Core was generated by 
> `/data/jenkins/workspace/impala-asf-master-exhaustive-release/Impala-Toolchain/c'.
> Program terminated with signal SIGABRT, Aborted.
> #0  0x7f9a2cc611f7 in ?? ()
> #0  0x7f9a2cc611f7 in ?? ()
> #1  0x7f9a2cc628e8 in ?? ()
> #2  0x0020 in ?? ()
> #3  0x in ?? ()
> CORE: ./core.1547693849.30783.kudu-tserver
> BINARY: ./be/build/latest/service/impalad
> Core was generated by 
> `/data/jenkins/workspace/impala-asf-master-exhaustive-release/Impala-Toolchain/c'.
> Program terminated with signal SIGABRT, Aborted.
> #0  0x7f0b81fb11f7 in ?? ()
> #0  0x7f0b81fb11f7 in ?? ()
> #1  0x7f0b81fb28e8 in ?? ()
> #2  0x0020 in ?? ()
> #3  0x in ?? ()
> CORE: ./core.1547693849.30767.kudu-tserver
> BINARY: ./be/build/latest/service/impalad
> Core was generated by 
> `/data/jenkins/workspace/impala-asf-master-exhaustive-release/Impala-Toolchain/c'.
> Program terminated with signal SIGABRT, Aborted.
> #0  0x7f60b1e2f1f7 in ?? ()
> #0  0x7f60b1e2f1f7 in ?? ()
> #1  0x7f60b1e308e8 in ?? ()
> #2  0x0020 in ?? ()
> #3  0x in ?? ()
> CORE: ./core.1547693849.30808.kudu-tserver
> BINARY: ./be/build/latest/service/impalad
> Core was generated by 
> `/data/jenkins/workspace/impala-asf-master-exhaustive-release/Impala-Toolchain/c'.
> Program terminated with signal SIGABRT, Aborted.
> #0  0x7fa3cb5591f7 

[jira] [Commented] (IMPALA-8091) Kudu SIGABRT'ed during dataload on Impala release build

2019-01-22 Thread Michael Brown (JIRA)


[ 
https://issues.apache.org/jira/browse/IMPALA-8091?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16749143#comment-16749143
 ] 

Michael Brown commented on IMPALA-8091:
---

The failure isn't easy to triage from a console log. There, it looks like:
{noformat}
11:40:19 11:40:19 Error executing impala SQL: 
/data/jenkins/workspace/impala-asf-master-core/repos/Impala/logs/data_loading/sql/functional/create-functional-query-exhaustive-impala-generated-kudu-none-none.sql
 See: 
/data/jenkins/workspace/impala-asf-master-core/repos/Impala/logs/data_loading/sql/functional/create-functional-query-exhaustive-impala-generated-kudu-none-none.sql.log
{noformat}
That log looks like this:
{noformat}
Traceback (most recent call last):
  File 
"/data/jenkins/workspace/impala-asf-master-exhaustive-release/repos/Impala/bin/load-data.py",
 line 207, in exec_impala_query_from_file
result = impala_client.execute(query)
  File 
"/data/jenkins/workspace/impala-asf-master-exhaustive-release/repos/Impala/tests/beeswax/impala_beeswax.py",
 line 182, in execute
handle = self.__execute_query(query_string.strip(), user=user)
  File 
"/data/jenkins/workspace/impala-asf-master-exhaustive-release/repos/Impala/tests/beeswax/impala_beeswax.py",
 line 357, in __execute_query
handle = self.execute_query_async(query_string, user=user)
  File 
"/data/jenkins/workspace/impala-asf-master-exhaustive-release/repos/Impala/tests/beeswax/impala_beeswax.py",
 line 351, in execute_query_async
handle = self.__do_rpc(lambda: self.imp_service.query(query,))
  File 
"/data/jenkins/workspace/impala-asf-master-exhaustive-release/repos/Impala/tests/beeswax/impala_beeswax.py",
 line 509, in __do_rpc
raise ImpalaBeeswaxException(self.__build_error_message(b), b)
ImpalaBeeswaxException: ImpalaBeeswaxException:
 INNER EXCEPTION: 
 MESSAGE: ImpalaRuntimeException: Error creating Kudu table 
'impala::tpch_kudu.lineitem'
CAUSED BY: NonRecoverableException: too many attempts: 
KuduRpc(method=ListTables, tablet=null, attempt=101, 
DeadlineTracker(timeout=18, elapsed=177706), Traces: [0ms] querying master, 
[45ms] Sub rpc
: ConnectToMaster sending RPC to server master-127.0.0.1:7051, [100ms] Sub rpc: 
ConnectToMaster received from server master-127.0.0.1:7051 response Network 
error: java.net.ConnectException: Connection r
efused: /127.0.0.1:7051, [107ms] delaying RPC due to Service unavailable: 
Master config (127.0.0.1:7051) has no leader.
{noformat}
This is the kudu service in the minicluster dropping out from under Impala. 
Note btw that Kudu initially started ok.
{noformat}
11:26:26 Starting kudu (Web UI - http://localhost:8051)
11:26:41 The cluster is running
{noformat}
Finally, there's evidence in the minicluster logs for Kudu that all 4 services 
(1 master, 3 tablet servers) all failed the same dcheck as pasted above.

> Kudu SIGABRT'ed during dataload on Impala release build
> ---
>
> Key: IMPALA-8091
> URL: https://issues.apache.org/jira/browse/IMPALA-8091
> Project: IMPALA
>  Issue Type: Bug
>  Components: Backend
>Affects Versions: Impala 3.2.0
>Reporter: David Knupp
>Assignee: Thomas Tauber-Marshall
>Priority: Critical
>  Labels: broken-build
>
> *Console log*:
> {noformat}
> 19:10:04 19:10:04 Error executing impala SQL: 
> /data/jenkins/workspace/impala-asf-master-exhaustive-release/repos/Impala/logs/data_loading/sql/functional/create-functional-query-exhaustive-impala-generated-kudu-none-none.sql
>  See: 
> /data/jenkins/workspace/impala-asf-master-exhaustive-release/repos/Impala/logs/data_loading/sql/functional/create-functional-query-exhaustive-impala-generated-kudu-none-none.sql.log
> 19:10:04 Background task Loading functional-query data (pid 4831) failed.
> 19:10:04 Background task Loading TPC-H data (pid 4832) failed.
> 19:10:04 ERROR in 
> /data/jenkins/workspace/impala-asf-master-exhaustive-release/repos/Impala/testdata/bin/create-load-data.sh
>  at line 85: CM_HOST=${2-}
> [...]
> 19:10:05 2019-01-16 19:10:05,461 - archive_core_dumps - INFO - Found core 
> files: ['./core.1547693849.30799.kudu-master', 
> './core.1547693849.30783.kudu-tserver', 
> './core.1547693849.30767.kudu-tserver', 
> './core.1547693849.30808.kudu-tserver']
> 19:10:05 2019-01-16 19:10:05,649 - archive_core_dumps - INFO - [New LWP 30799]
> 19:10:05 [New LWP 30834]
> 19:10:05 [New LWP 30835]
> 19:10:05 [New LWP 30838]
> 19:10:05 [New LWP 30837]
> 19:10:05 [New LWP 30836]
> 19:10:05 Core was generated by 
> `/data/jenkins/workspace/impala-asf-master-exhaustive-release/Impala-Toolchain/c'.
> 19:10:05 Program terminated with signal SIGABRT, Aborted.
> 19:10:05 #0  0x7f9a2cc611f7 in ?? ()
> 19:10:05 
> 19:10:05 2019-01-16 19:10:05,650 - archive_core_dumps - INFO - Found binary 
> path through GDB: 
> 

[jira] [Commented] (IMPALA-8091) Kudu SIGABRT'ed during dataload on Impala release build

2019-01-22 Thread Michael Brown (JIRA)


[ 
https://issues.apache.org/jira/browse/IMPALA-8091?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16749130#comment-16749130
 ] 

Michael Brown commented on IMPALA-8091:
---

I've seen this too, and I think this is likely an environmental problem.
{noformat}
F0116 18:57:29.339143 30799 master_main.cc:81] Check failed: _s.ok() Bad 
status: Runtime error: Cannot initialize clock: failed to wait for clock sync 
using command '/usr/bin/chronyc waitsync 60 0 0 1': /usr/bin/chronyc: process 
exited with non-zero status 1
{noformat}

> Kudu SIGABRT'ed during dataload on Impala release build
> ---
>
> Key: IMPALA-8091
> URL: https://issues.apache.org/jira/browse/IMPALA-8091
> Project: IMPALA
>  Issue Type: Bug
>  Components: Backend
>Affects Versions: Impala 3.2.0
>Reporter: David Knupp
>Assignee: Thomas Tauber-Marshall
>Priority: Critical
>  Labels: broken-build
>
> *Console log*:
> {noformat}
> 19:10:04 19:10:04 Error executing impala SQL: 
> /data/jenkins/workspace/impala-asf-master-exhaustive-release/repos/Impala/logs/data_loading/sql/functional/create-functional-query-exhaustive-impala-generated-kudu-none-none.sql
>  See: 
> /data/jenkins/workspace/impala-asf-master-exhaustive-release/repos/Impala/logs/data_loading/sql/functional/create-functional-query-exhaustive-impala-generated-kudu-none-none.sql.log
> 19:10:04 Background task Loading functional-query data (pid 4831) failed.
> 19:10:04 Background task Loading TPC-H data (pid 4832) failed.
> 19:10:04 ERROR in 
> /data/jenkins/workspace/impala-asf-master-exhaustive-release/repos/Impala/testdata/bin/create-load-data.sh
>  at line 85: CM_HOST=${2-}
> [...]
> 19:10:05 2019-01-16 19:10:05,461 - archive_core_dumps - INFO - Found core 
> files: ['./core.1547693849.30799.kudu-master', 
> './core.1547693849.30783.kudu-tserver', 
> './core.1547693849.30767.kudu-tserver', 
> './core.1547693849.30808.kudu-tserver']
> 19:10:05 2019-01-16 19:10:05,649 - archive_core_dumps - INFO - [New LWP 30799]
> 19:10:05 [New LWP 30834]
> 19:10:05 [New LWP 30835]
> 19:10:05 [New LWP 30838]
> 19:10:05 [New LWP 30837]
> 19:10:05 [New LWP 30836]
> 19:10:05 Core was generated by 
> `/data/jenkins/workspace/impala-asf-master-exhaustive-release/Impala-Toolchain/c'.
> 19:10:05 Program terminated with signal SIGABRT, Aborted.
> 19:10:05 #0  0x7f9a2cc611f7 in ?? ()
> 19:10:05 
> 19:10:05 2019-01-16 19:10:05,650 - archive_core_dumps - INFO - Found binary 
> path through GDB: 
> /data/jenkins/workspace/impala-asf-master-exhaustive-release/Impala-Toolchain/c
> 19:10:05 2019-01-16 19:10:05,893 - archive_core_dumps - WARNING - Failed to 
> determine binary because multiple candidate binaries were found and none of 
> their paths contained 'latest' to disambiguate:
> 19:10:05 Core:./core.1547693849.30799.kudu-master
> 19:10:05 
> Binaries:['./testdata/cluster/node_templates/common/etc/init.d/kudu-master', 
> './testdata/cluster/cdh6/node-1/etc/init.d/kudu-master']
> 19:10:05 
> 19:10:05 2019-01-16 19:10:05,917 - archive_core_dumps - INFO - [New LWP 30783]
> 19:10:05 [New LWP 30810]
> 19:10:05 [New LWP 30812]
> 19:10:05 [New LWP 30811]
> 19:10:05 [New LWP 30824]
> 19:10:05 [New LWP 30820]
> 19:10:05 Core was generated by 
> `/data/jenkins/workspace/impala-asf-master-exhaustive-release/Impala-Toolchain/c'.
> 19:10:05 Program terminated with signal SIGABRT, Aborted.
> 19:10:05 #0  0x7f0b81fb11f7 in ?? ()
> {noformat}
> *Backtraces*:
> {noformat}
> CORE: ./core.1547693849.30799.kudu-master
> BINARY: ./be/build/latest/service/impalad
> Core was generated by 
> `/data/jenkins/workspace/impala-asf-master-exhaustive-release/Impala-Toolchain/c'.
> Program terminated with signal SIGABRT, Aborted.
> #0  0x7f9a2cc611f7 in ?? ()
> #0  0x7f9a2cc611f7 in ?? ()
> #1  0x7f9a2cc628e8 in ?? ()
> #2  0x0020 in ?? ()
> #3  0x in ?? ()
> CORE: ./core.1547693849.30783.kudu-tserver
> BINARY: ./be/build/latest/service/impalad
> Core was generated by 
> `/data/jenkins/workspace/impala-asf-master-exhaustive-release/Impala-Toolchain/c'.
> Program terminated with signal SIGABRT, Aborted.
> #0  0x7f0b81fb11f7 in ?? ()
> #0  0x7f0b81fb11f7 in ?? ()
> #1  0x7f0b81fb28e8 in ?? ()
> #2  0x0020 in ?? ()
> #3  0x in ?? ()
> CORE: ./core.1547693849.30767.kudu-tserver
> BINARY: ./be/build/latest/service/impalad
> Core was generated by 
> `/data/jenkins/workspace/impala-asf-master-exhaustive-release/Impala-Toolchain/c'.
> Program terminated with signal SIGABRT, Aborted.
> #0  0x7f60b1e2f1f7 in ?? ()
> #0  0x7f60b1e2f1f7 in ?? ()
> #1  0x7f60b1e308e8 in ?? ()
> #2  0x0020 in ?? ()
> #3  0x in ?? ()
> CORE: ./core.1547693849.30808.kudu-tserver
> BINARY: ./be/build/latest/service/impalad

[jira] [Assigned] (IMPALA-8093) Profiles prefix counters inconsistently

2019-01-22 Thread Yongzhi Chen (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8093?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yongzhi Chen reassigned IMPALA-8093:


Assignee: Yongzhi Chen

> Profiles prefix counters inconsistently
> ---
>
> Key: IMPALA-8093
> URL: https://issues.apache.org/jira/browse/IMPALA-8093
> Project: IMPALA
>  Issue Type: Bug
>  Components: Backend
>Affects Versions: Impala 3.2.0
>Reporter: Lars Volker
>Assignee: Yongzhi Chen
>Priority: Major
>  Labels: newbie, observability, profile, ramp-up
>
> In pretty printed profiles, plain counters are prefixed with \{{ - }} but 
> TimeSeriesCounters are not. We should make them consistent.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Created] (IMPALA-8096) Limit on #rows returned from query

2019-01-22 Thread Tim Armstrong (JIRA)
Tim Armstrong created IMPALA-8096:
-

 Summary: Limit on #rows returned from query
 Key: IMPALA-8096
 URL: https://issues.apache.org/jira/browse/IMPALA-8096
 Project: IMPALA
  Issue Type: Sub-task
  Components: Backend
Reporter: Tim Armstrong


Sometimes users accidentally run queries that return a large number of rows, 
e.g.
{code}
SELECT * FROM table
{code}

When they really only need to look at a subset of the rows. It would be useful 
to have a guardrail to fail queries the return more rows than a particular 
limit. Maybe it would make sense to integrate with IMPALA-4268 so that the 
query is failed when the buffer fills up, but it may also be useful to have an 
easier-to-understand option based on #rows.




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-8027) KRPC datastream timing out on both the receiver and sender side even in a minicluster

2019-01-22 Thread Michael Ho (JIRA)


[ 
https://issues.apache.org/jira/browse/IMPALA-8027?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16749067#comment-16749067
 ] 

Michael Ho commented on IMPALA-8027:


Looks like the coordinator seems to be slow somehow between 05:16:52 and 
05:18:56. Both failing queries had to do with time out waiting for the 
coordinator fragment not becoming ready within 2 minutes.

Log from coordinators:
{noformat}
I1228 05:16:55.374305 26714 query-state.cc:568] Executing instance. 
instance_id=8f46b2518734bef1:6ef2d404 fragment_idx=0 
per_fragment_instance_idx=0 coord_state_idx=0 #in-flight=9
I1228 05:16:55.374305 26710 coordinator.cc:368] started execution on 2 backends 
for query_id=8f46b2518734bef1:6ef2d404
{noformat}

{noformat}
I1228 05:16:52.217275 26685 query-state.cc:568] Executing instance. 
instance_id=194f5b70907ac97c:84a116d6 fragment_idx=0 
per_fragment_instance_idx=0 coord_state_idx=0 #in-flight=7
I1228 05:16:52.217211 26678 coordinator.cc:368] started execution on 3 backends 
for query_id=194f5b70907ac97c:84a116d6
{noformat}

The top output snapshots from that time period didn't suggest much other than 
Impalad wasn't scheduled much in them. Looking into query profiles next.


> KRPC datastream timing out on both the receiver and sender side even in a 
> minicluster
> -
>
> Key: IMPALA-8027
> URL: https://issues.apache.org/jira/browse/IMPALA-8027
> Project: IMPALA
>  Issue Type: Bug
>  Components: Distributed Exec
>Affects Versions: Impala 3.2.0
>Reporter: Bikramjeet Vig
>Assignee: Michael Ho
>Priority: Critical
>  Labels: broken-build
>
> krpc datastreams seem to time out at the same time at both sender and 
> receiver causing two running queries to fail. This happened while running 
> core tests on s3.
> Logs from coordinator:
> {noformat}
> I1228 05:18:56.202587 13396 krpc-data-stream-mgr.cc:353] Sender 127.0.0.1 
> timed out waiting for receiver fragment instance: 
> 8f46b2518734bef1:6ef2d404, dest node: 2
> I1228 05:18:56.203061 13396 rpcz_store.cc:265] Call 
> impala.DataStreamService.TransmitData from 127.0.0.1:53118 (request call id 
> 11274) took 120782ms. Request Metrics: {}
> I1228 05:18:56.203114 13396 krpc-data-stream-mgr.cc:353] Sender 127.0.0.1 
> timed out waiting for receiver fragment instance: 
> 194f5b70907ac97c:84a116d6, dest node: 2
> I1228 05:18:56.203136 13396 rpcz_store.cc:265] Call 
> impala.DataStreamService.TransmitData from 127.0.0.1:53110 (request call id 
> 8637) took 123811ms. Request Metrics: {}
> I1228 05:18:56.203155 13396 krpc-data-stream-mgr.cc:353] Sender 127.0.0.1 
> timed out waiting for receiver fragment instance: 
> 194f5b70907ac97c:84a116d6, dest node: 2
> I1228 05:18:56.203167 13396 rpcz_store.cc:265] Call 
> impala.DataStreamService.TransmitData from 127.0.0.1:53118 (request call id 
> 11273) took 123776ms. Request Metrics: {}
> I1228 05:18:56.203181 13396 krpc-data-stream-mgr.cc:408] Reduced stream ID 
> cache from 413 items, to 410, eviction took: 1ms
> I1228 05:18:56.204746 13377 coordinator.cc:707] Backend completed: 
> host=impala-ec2-centos74-m5-4xlarge-ondemand-07b3.vpc.cloudera.com:22001 
> remaining=2 query_id=8f46b2518734bef1:6ef2d404
> I1228 05:18:56.204756 13377 coordinator-backend-state.cc:262] 
> query_id=8f46b2518734bef1:6ef2d404: first in-progress backend: 
> impala-ec2-centos74-m5-4xlarge-ondemand-07b3.vpc.cloudera.com:22000
> I1228 05:18:56.204769 13377 coordinator.cc:522] ExecState: query 
> id=8f46b2518734bef1:6ef2d404 
> finstance=8f46b2518734bef1:6ef2d4040001 on 
> host=impala-ec2-centos74-m5-4xlarge-ondemand-07b3.vpc.cloudera.com:22001 
> (EXECUTING -> ERROR) status=Sender 127.0.0.1 timed out waiting for receiver 
> fragment instance: 8f46b2518734bef1:6ef2d404, dest node: 2
> {noformat}
> Logs from executor:
> {noformat}
> E1228 05:18:56.203181 26715 krpc-data-stream-sender.cc:343] channel send to 
> 127.0.0.1:27000 failed: 
> (fragment_instance_id=8f46b2518734bef1:6ef2d404): Sender 127.0.0.1 
> timed out waiting for receiver fragment instance: 
> 8f46b2518734bef1:6ef2d404, dest node: 2
> E1228 05:18:56.203256 26682 krpc-data-stream-sender.cc:343] channel send to 
> 127.0.0.1:27000 failed: 
> (fragment_instance_id=194f5b70907ac97c:84a116d6): Sender 127.0.0.1 
> timed out waiting for receiver fragment instance: 
> 194f5b70907ac97c:84a116d6, dest node: 2
> I1228 05:18:56.203451 26715 query-state.cc:576] Instance completed. 
> instance_id=8f46b2518734bef1:6ef2d4040001 #in-flight=3 
> status=DATASTREAM_SENDER_TIMEOUT: Sender 127.0.0.1 timed out waiting for 
> receiver fragment instance: 8f46b2518734bef1:6ef2d404, dest node: 2
> I1228 

[jira] [Updated] (IMPALA-8093) Profiles prefix counters inconsistently

2019-01-22 Thread Lars Volker (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8093?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Volker updated IMPALA-8093:

Labels: newbie observability profile  (was: observability profile)

> Profiles prefix counters inconsistently
> ---
>
> Key: IMPALA-8093
> URL: https://issues.apache.org/jira/browse/IMPALA-8093
> Project: IMPALA
>  Issue Type: Bug
>  Components: Backend
>Affects Versions: Impala 3.2.0
>Reporter: Lars Volker
>Priority: Major
>  Labels: newbie, observability, profile
>
> In pretty printed profiles, plain counters are prefixed with \{{ - }} but 
> TimeSeriesCounters are not. We should make them consistent.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-4249) Clarify lifetime of DiskIoMgr::ScanRange objects

2019-01-22 Thread Tim Armstrong (JIRA)


[ 
https://issues.apache.org/jira/browse/IMPALA-4249?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16749057#comment-16749057
 ] 

Tim Armstrong commented on IMPALA-4249:
---

We could consider using std::shared_ptr here since it does seem to be a genuine 
case of shared ownership.

> Clarify lifetime of DiskIoMgr::ScanRange objects
> 
>
> Key: IMPALA-4249
> URL: https://issues.apache.org/jira/browse/IMPALA-4249
> Project: IMPALA
>  Issue Type: Improvement
>  Components: Backend
>Affects Versions: Impala 2.8.0
>Reporter: Tim Armstrong
>Priority: Minor
>  Labels: resource-management
>
> When working on the DiskIoMgr code I noticed that the lifetime of the 
> ScanRange objects isn't very clear. In particular, it's not clear when it is 
> safe for a client to free the ScanRange object. We should at least document 
> the expectation in the code.
> It seems like in principle that the client should be able to free the scan 
> range after it hits eosr() or cancels the scan range. However, that's not 
> true because a) BufferDescriptors hold internal pointers to the ScanRange and 
> b) the DiskIoMgr touches the ScanRange in the below code after enqueueing the 
> last buffer:
> {code}
>   bool queue_full = scan_range->EnqueueBuffer(buffer);
>   if (eosr) {
> // For cached buffers, we can't close the range until the cached buffer 
> is returned.
> // Close() is called from DiskIoMgr::ReturnBuffer().
> if (!is_cached) scan_range->Close();
>   } else {
> ...
> {code}
> It seems like we (mostly?) avoid issues by sticking the ScanRanges in object 
> pools where they will outlive the reader contexts. This seems suboptimal - we 
> shouldn't really leave lots of small unused objects lying around.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Updated] (IMPALA-8093) Profiles prefix counters inconsistently

2019-01-22 Thread Lars Volker (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8093?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Volker updated IMPALA-8093:

Labels: newbie observability profile ramp-up  (was: newbie observability 
profile)

> Profiles prefix counters inconsistently
> ---
>
> Key: IMPALA-8093
> URL: https://issues.apache.org/jira/browse/IMPALA-8093
> Project: IMPALA
>  Issue Type: Bug
>  Components: Backend
>Affects Versions: Impala 3.2.0
>Reporter: Lars Volker
>Priority: Major
>  Labels: newbie, observability, profile, ramp-up
>
> In pretty printed profiles, plain counters are prefixed with \{{ - }} but 
> TimeSeriesCounters are not. We should make them consistent.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Resolved] (IMPALA-5847) Some query options do not work as expected in .test files

2019-01-22 Thread Thomas Tauber-Marshall (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-5847?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Thomas Tauber-Marshall resolved IMPALA-5847.

   Resolution: Fixed
Fix Version/s: Impala 3.2.0

> Some query options do not work as expected in .test files
> -
>
> Key: IMPALA-5847
> URL: https://issues.apache.org/jira/browse/IMPALA-5847
> Project: IMPALA
>  Issue Type: Improvement
>  Components: Infrastructure
>Reporter: Alexander Behm
>Assignee: Thomas Tauber-Marshall
>Priority: Minor
> Fix For: Impala 3.2.0
>
>
> We often use "set" in .test files to alter query options. Theoretically, a 
> "set" command should change the session-level query options and in most cases 
> a single .test file is executed from the same Impala session. However, for 
> some options using "set" within a query section does not seem to work. For 
> example, "num_nodes" does not work as expected as shown below.
> PyTest:
> {code}
> import pytest
> from tests.common.impala_test_suite import ImpalaTestSuite
> class TestStringQueries(ImpalaTestSuite):
>   @classmethod
>   def get_workload(cls):
> return 'functional-query'
>   def test_set_bug(self, vector):
> self.run_test_case('QueryTest/set_bug', vector)
> {code}
> Corresponding .test file:
> {code}
> 
>  QUERY
> set num_nodes=1;
> select count(*) from functional.alltypes;
> select count(*) from functional.alltypes;
> select count(*) from functional.alltypes;
>  RESULTS
> 7300
>  TYPES
> BIGINT
> 
> {code}
> After running the test above, I validated that the 3 queries were run from 
> the same session, and that the queries run a distributed plan. The 
> "num_nodes" option was definitely not picked up. I am not sure which query 
> options are affected. In several .test files setting other query options does 
> seem to work as expected.
> I suspect that the test framework might keep its own list of default query 
> options which get submitted together with the query, so the session-level 
> options are overridden on a per-request basis. For example, if I change the 
> pytest to remove the "num_nodes" dictionary entry, then the test works as 
> expected.
> PyTest workaround:
> {code}
> import pytest
> from tests.common.impala_test_suite import ImpalaTestSuite
> class TestStringQueries(ImpalaTestSuite):
>   @classmethod
>   def get_workload(cls):
> return 'functional-query'
>   def test_set_bug(self, vector):
> # Workaround SET bug
> vector.get_value('exec_option').pop('num_nodes', None)
> self.run_test_case('QueryTest/set_bug', vector)
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-5847) Some query options do not work as expected in .test files

2019-01-22 Thread ASF subversion and git services (JIRA)


[ 
https://issues.apache.org/jira/browse/IMPALA-5847?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16749012#comment-16749012
 ] 

ASF subversion and git services commented on IMPALA-5847:
-

Commit 6fa35478d5930cc3ec6d88c6b2e3f6bedd3d8ed9 in impala's branch 
refs/heads/master from Thomas Tauber-Marshall
[ https://gitbox.apache.org/repos/asf?p=impala.git;h=6fa3547 ]

IMPALA-5847: Fix incorrect use of SET in .test files

The '.test' files are used to run queries for tests. These files are
run with a vector of default query options. They also sometimes
include SET queries that modify query options. If SET is used on a
query option that is included in the vector, the default value from
the vector will override the value from the SET, leading to tests that
don't actually run with the query options they appear to.

This patch asserts that '.test' files don't use SET for values present
in the default vector. It also fixes various tests that already had
this incorrect behavior.

Testing:
- Passed a full exhaustive run.

Change-Id: I4e4c0f31bf4850642b624acdb1f6cb8837957990
Reviewed-on: http://gerrit.cloudera.org:8080/12220
Reviewed-by: Thomas Marshall 
Tested-by: Impala Public Jenkins 


> Some query options do not work as expected in .test files
> -
>
> Key: IMPALA-5847
> URL: https://issues.apache.org/jira/browse/IMPALA-5847
> Project: IMPALA
>  Issue Type: Improvement
>  Components: Infrastructure
>Reporter: Alexander Behm
>Assignee: Thomas Tauber-Marshall
>Priority: Minor
>
> We often use "set" in .test files to alter query options. Theoretically, a 
> "set" command should change the session-level query options and in most cases 
> a single .test file is executed from the same Impala session. However, for 
> some options using "set" within a query section does not seem to work. For 
> example, "num_nodes" does not work as expected as shown below.
> PyTest:
> {code}
> import pytest
> from tests.common.impala_test_suite import ImpalaTestSuite
> class TestStringQueries(ImpalaTestSuite):
>   @classmethod
>   def get_workload(cls):
> return 'functional-query'
>   def test_set_bug(self, vector):
> self.run_test_case('QueryTest/set_bug', vector)
> {code}
> Corresponding .test file:
> {code}
> 
>  QUERY
> set num_nodes=1;
> select count(*) from functional.alltypes;
> select count(*) from functional.alltypes;
> select count(*) from functional.alltypes;
>  RESULTS
> 7300
>  TYPES
> BIGINT
> 
> {code}
> After running the test above, I validated that the 3 queries were run from 
> the same session, and that the queries run a distributed plan. The 
> "num_nodes" option was definitely not picked up. I am not sure which query 
> options are affected. In several .test files setting other query options does 
> seem to work as expected.
> I suspect that the test framework might keep its own list of default query 
> options which get submitted together with the query, so the session-level 
> options are overridden on a per-request basis. For example, if I change the 
> pytest to remove the "num_nodes" dictionary entry, then the test works as 
> expected.
> PyTest workaround:
> {code}
> import pytest
> from tests.common.impala_test_suite import ImpalaTestSuite
> class TestStringQueries(ImpalaTestSuite):
>   @classmethod
>   def get_workload(cls):
> return 'functional-query'
>   def test_set_bug(self, vector):
> # Workaround SET bug
> vector.get_value('exec_option').pop('num_nodes', None)
> self.run_test_case('QueryTest/set_bug', vector)
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org