[jira] [Updated] (IMPALA-8072) Clean up config files in docker containers
[ https://issues.apache.org/jira/browse/IMPALA-8072?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Tim Armstrong updated IMPALA-8072: -- Description: Currently the docker containers include a bunch of config files copied indiscriminately from the dev environment. Mostly these aren't valid for a production container and it's expected that the real config files will be mounted at /opt/impala/conf. We should instead include a more reasonable set of default configs (e.g. for admission control), plus placeholders for other config files that may need to be overridden with site-specific configs. was: Currently the docker containers include a bunch of config files copied indiscriminately from the dev environment. Mostly these aren't valid for a production container and it's expected that the real config files will be mounted at /opt/impala/conf. We should stop including those config files or at least include a more reasonable set of default configs. > Clean up config files in docker containers > -- > > Key: IMPALA-8072 > URL: https://issues.apache.org/jira/browse/IMPALA-8072 > Project: IMPALA > Issue Type: Sub-task > Components: Infrastructure >Affects Versions: Impala 3.2.0 >Reporter: Tim Armstrong >Assignee: Tim Armstrong >Priority: Major > Labels: docker > > Currently the docker containers include a bunch of config files copied > indiscriminately from the dev environment. Mostly these aren't valid for a > production container and it's expected that the real config files will be > mounted at /opt/impala/conf. > We should instead include a more reasonable set of default configs (e.g. for > admission control), plus placeholders for other config files that may need to > be overridden with site-specific configs. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org For additional commands, e-mail: issues-all-h...@impala.apache.org
[jira] [Assigned] (IMPALA-8114) Build test failure in test_breakpad.py
[ https://issues.apache.org/jira/browse/IMPALA-8114?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Lars Volker reassigned IMPALA-8114: --- Assignee: Lars Volker (was: Tim Armstrong) > Build test failure in test_breakpad.py > -- > > Key: IMPALA-8114 > URL: https://issues.apache.org/jira/browse/IMPALA-8114 > Project: IMPALA > Issue Type: Bug > Components: Infrastructure >Affects Versions: Impala 3.1.0 >Reporter: Paul Rogers >Assignee: Lars Volker >Priority: Blocker > > Recent builds have failed due to a failure in {{test_breakpad.py}}. Assigning > to Tim as the person who most recently touched this file. > Test output: > {noformat} > 09:04:35 ERRORS > > 09:04:35 ___ ERROR at teardown of > TestBreakpadExhaustive.test_minidump_cleanup_thread ___ > 09:04:35 custom_cluster/test_breakpad.py:49: in teardown_method > 09:04:35 self.kill_cluster(SIGKILL) > 09:04:35 custom_cluster/test_breakpad.py:80: in kill_cluster > 09:04:35 self.kill_processes(processes, signal) > 09:04:35 custom_cluster/test_breakpad.py:85: in kill_processes > 09:04:35 process.kill(signal) > 09:04:35 common/impala_cluster.py:330: in kill > 09:04:35 assert 0, "No processes %s found" % self.cmd > 09:04:35 E AssertionError: No processes > ['/data/jenkins/workspace/impala-cdh6.x-exhaustive-release/repos/Impala/be/build/latest/service/impalad', > '-kudu_client_rpc_timeout_ms', '0', '-kudu_master_hosts', 'localhost', > '--mem_limit=12884901888', '-logbufsecs=5', '-v=1', '-max_log_files=0', > '-log_filename=impalad', > '-log_dir=/data/jenkins/workspace/impala-cdh6.x-exhaustive-release/repos/Impala/logs/custom_cluster_tests', > '-beeswax_port=21000', '-hs2_port=21050', '-be_port=22000', > '-krpc_port=27000', '-state_store_subscriber_port=23000', > '-webserver_port=25000', '-max_minidumps=2', '-logbufsecs=1', > '-minidump_path=/tmp/tmpKaSw_w', '--default_query_options='] found > {noformat} > Distilled {{TEST-impala-custom-cluster.xml}} output: > {noformat} > -- 2019-01-23 08:00:43,585 INFO MainThread: Found 3 impalad/1 > statestored/1 catalogd process(es) > … > -- 2019-01-23 08:00:43,667 INFO MainThread: Killing: > /data/jenkins/workspace/impala-cdh6.x-exhaustive-release/repos/Impala/be/build/latest/service/statestored > -logbufsecs=5 -v=1 -max_log_files=0 -log_filename=statestored > -log_dir=/data/jenkins/workspace/impala-cdh6.x-exhaustive-release/repos/Impala/logs/custom_cluster_tests > -max_minidumps=2 -logbufsecs=1 -minidump_path=/tmp/tmpKaSw_w (PID: 16809) > with signal 10 > -- 2019-01-23 08:00:43,692 INFO MainThread: Found 6 impalad/1 > statestored/1 catalogd process(es) > ... > E AssertionError: No processes > [/data/jenkins/workspace/impala-cdh6.x-exhaustive-release/repos/Impala/be/build/latest/service/impalad > {noformat} > Notice that the main thread appaars to be killing statestore, but fails to > kill impalad. Notice that a message appears that says that all impalads are > running in the midst of the code that tries to shut down the cluster. Is this > test multi-threaded? Is there more than one “main thread” Are these main > threads working at cross purposes? What recent change may have caused this? > Also, looks like the script is sending signal 10 (SIGUSR1) while the > statestore (in its log) says it got a SIGTERM (15): > {noformat} > I0123 08:00:44.086009 16868 thrift-client.cc:78] Couldn't open transport for > impala-ec2-centoCaught signal: SIGTERM. Daemon will exit. > {noformat} > Not terribly familiar with this area of the product, so bumping it over to > the BE team. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org For additional commands, e-mail: issues-all-h...@impala.apache.org
[jira] [Created] (IMPALA-8116) Impala Doc: Create Impala Limitations doc
Alex Rodoni created IMPALA-8116: --- Summary: Impala Doc: Create Impala Limitations doc Key: IMPALA-8116 URL: https://issues.apache.org/jira/browse/IMPALA-8116 Project: IMPALA Issue Type: Improvement Components: Docs Affects Versions: Impala 3.1.0 Reporter: Alex Rodoni Assignee: Alex Rodoni Create a separate document that focuses on design limitations more than bugs. It could also include functional limitations like "cannot write nested types", etc. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Created] (IMPALA-8116) Impala Doc: Create Impala Limitations doc
Alex Rodoni created IMPALA-8116: --- Summary: Impala Doc: Create Impala Limitations doc Key: IMPALA-8116 URL: https://issues.apache.org/jira/browse/IMPALA-8116 Project: IMPALA Issue Type: Improvement Components: Docs Affects Versions: Impala 3.1.0 Reporter: Alex Rodoni Assignee: Alex Rodoni Create a separate document that focuses on design limitations more than bugs. It could also include functional limitations like "cannot write nested types", etc. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org For additional commands, e-mail: issues-all-h...@impala.apache.org
[jira] [Resolved] (IMPALA-8090) DiskIoMgrTest.SyncReadTest hits file_ != nullptr DCHECK in LocalFileReader::ReadFromPos()
[ https://issues.apache.org/jira/browse/IMPALA-8090?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Tim Armstrong resolved IMPALA-8090. --- Resolution: Fixed Fix Version/s: Impala 3.2.0 > DiskIoMgrTest.SyncReadTest hits file_ != nullptr DCHECK in > LocalFileReader::ReadFromPos() > - > > Key: IMPALA-8090 > URL: https://issues.apache.org/jira/browse/IMPALA-8090 > Project: IMPALA > Issue Type: Bug > Components: Backend >Affects Versions: Impala 3.2.0 >Reporter: David Knupp >Assignee: Tim Armstrong >Priority: Critical > Fix For: Impala 3.2.0 > > > *Test output*: > {noformat} > 45/99 Test #45: disk-io-mgr-test .***Exception: Other 43.29 > sec > Turning perftools heap leak checking off > [==] Running 25 tests from 1 test case. > [--] Global test environment set-up. > [--] 25 tests from DiskIoMgrTest > [ RUN ] DiskIoMgrTest.SingleWriter > 19/01/16 15:57:09 INFO util.JvmPauseMonitor: Starting JVM pause monitor > [ OK ] DiskIoMgrTest.SingleWriter (3407 ms) > [ RUN ] DiskIoMgrTest.InvalidWrite > [ OK ] DiskIoMgrTest.InvalidWrite (281 ms) > [ RUN ] DiskIoMgrTest.WriteErrors > [ OK ] DiskIoMgrTest.WriteErrors (235 ms) > [ RUN ] DiskIoMgrTest.SingleWriterCancel > [ OK ] DiskIoMgrTest.SingleWriterCancel (1165 ms) > [ RUN ] DiskIoMgrTest.SingleReader > [ OK ] DiskIoMgrTest.SingleReader (5835 ms) > [ RUN ] DiskIoMgrTest.SingleReaderSubRanges > [ OK ] DiskIoMgrTest.SingleReaderSubRanges (16404 ms) > [ RUN ] DiskIoMgrTest.AddScanRangeTest > [ OK ] DiskIoMgrTest.AddScanRangeTest (1210 ms) > [ RUN ] DiskIoMgrTest.SyncReadTest > *** Check failure stack trace: *** > @ 0x4825dcc > @ 0x4827671 > @ 0x48257a6 > @ 0x4828d6d > @ 0x1af39ec > @ 0x1ae90a4 > @ 0x1ac30ea > @ 0x1accad3 > @ 0x1acc660 > @ 0x1acbf3e > @ 0x1acb62d > @ 0x1b03671 > @ 0x1f79988 > @ 0x1f82b60 > @ 0x1f82a84 > @ 0x1f82a47 > @ 0x3751579 > @ 0x3ea4807850 > @ 0x3ea44e894c > Wrote minidump to > /data/jenkins/workspace/<...>/repos/Impala/logs/be_tests/minidumps/disk-io-mgr-test/5bbf76f7-e5d6-4ac9-bdae9d9b-065c32ec.dmp > {noformat} > *Error*: > {noformat} > Operating system: Linux > 0.0.0 Linux 2.6.32-358.14.1.el6.centos.plus.x86_64 #1 SMP > Tue Jul 16 21:33:24 UTC 2013 x86_64 > CPU: amd64 > family 6 model 45 stepping 7 > 8 CPUs > GPU: UNKNOWN > Crash reason: SIGABRT > Crash address: 0x4522fa1 > Process uptime: not available > Thread 205 (crashed) > 0 libc-2.12.so + 0x328e5 > rax = 0x rdx = 0x0006 > rcx = 0x rbx = 0x06adf9c0 > rsi = 0x0563 rdi = 0x2fa1 > rbp = 0x7f8009b8ffe0 rsp = 0x7f8009b8fc78 > r8 = 0x7f8009b8fd00r9 = 0x0563 > r10 = 0x0008 r11 = 0x0202 > r12 = 0x06adfa40 r13 = 0x001f > r14 = 0x06ae7384 r15 = 0x06adf9c0 > rip = 0x003ea44328e5 > Found by: given as instruction pointer in context > 1 libc-2.12.so + 0x340c5 > rbp = 0x7f8009b8ffe0 rsp = 0x7f8009b8fc80 > rip = 0x003ea44340c5 > Found by: stack scanning > 2 disk-io-mgr-test!boost::_bi::bind_t impala::io::DiskQueue, impala::io::DiskIoMgr*>, > boost::_bi::list2, > boost::_bi::value > >::operator()() > [bind_template.hpp : 20 + 0x21] > rbp = 0x7f8009b8ffe0 rsp = 0x7f8009b8fc88 > rip = 0x01acbf3e > Found by: stack scanning > 3 disk-io-mgr-test!google::LogMessage::Flush() + 0x157 > rbx = 0x0007 rbp = 0x06adf980 > rsp = 0x7f8009b8fff0 rip = 0x048257a7 > Found by: call frame info > 4 disk-io-mgr-test!google::LogMessageFatal::~LogMessageFatal() + 0xe > rbx = 0x7f8009b90110 rbp = 0x7f8009b903f0 > rsp = 0x7f8009b90070 r12 = 0x0001 > r13 = 0x06aee8b8 r14 = 0x0c213538 > r15 = 0x0007 rip = 0x04828d6e > Found by: call frame info > 5 disk-io-mgr-test!impala::io::LocalFileReader::ReadFromPos(long, unsigned > char*, long, long*, bool*) [local-file-reader.cc : 67 + 0x10] > rbx = 0x0001 rbp = 0x7f8009b903f0 > rsp = 0x7f8009b90090 r12 = 0x0001 > r13 = 0x06aee8b8 r14 = 0x0c213538 > r15 = 0x0007 rip = 0x01af39ed > Found by: call frame
[jira] [Commented] (IMPALA-8090) DiskIoMgrTest.SyncReadTest hits file_ != nullptr DCHECK in LocalFileReader::ReadFromPos()
[ https://issues.apache.org/jira/browse/IMPALA-8090?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16751732#comment-16751732 ] ASF subversion and git services commented on IMPALA-8090: - Commit aa603d8cb9e870919b60d80828951beee1db6622 in impala's branch refs/heads/master from Tim Armstrong [ https://gitbox.apache.org/repos/asf?p=impala.git;h=aa603d8 ] IMPALA-8090: race when reusing ScanRange in test The issue occurs in test code where two reads are issued with the same ScanRange object back-to-back, e.g. in SyncReadTest. The DiskIoMgr enqueues the last buffer or cancels the scan range *before* closing the 'file_reader_', which means that the client thread thinks the ScanRange is done and can re-enqueue it into the DiskIoMgr. This is not an issue in Impala itself because scan ranges are not reused in this fashion. I evaluated adding a DCHECK but the lifecycle of the ScanRange objects is complicated enough that there wasn't a straightforward invariant to enforce. Testing: Looped the previously-failing test overnight. Change-Id: I3122e5b2efea60ffe82d780930301d5be108876b Reviewed-on: http://gerrit.cloudera.org:8080/12238 Reviewed-by: Joe McDonnell Tested-by: Impala Public Jenkins > DiskIoMgrTest.SyncReadTest hits file_ != nullptr DCHECK in > LocalFileReader::ReadFromPos() > - > > Key: IMPALA-8090 > URL: https://issues.apache.org/jira/browse/IMPALA-8090 > Project: IMPALA > Issue Type: Bug > Components: Backend >Affects Versions: Impala 3.2.0 >Reporter: David Knupp >Assignee: Tim Armstrong >Priority: Critical > Fix For: Impala 3.2.0 > > > *Test output*: > {noformat} > 45/99 Test #45: disk-io-mgr-test .***Exception: Other 43.29 > sec > Turning perftools heap leak checking off > [==] Running 25 tests from 1 test case. > [--] Global test environment set-up. > [--] 25 tests from DiskIoMgrTest > [ RUN ] DiskIoMgrTest.SingleWriter > 19/01/16 15:57:09 INFO util.JvmPauseMonitor: Starting JVM pause monitor > [ OK ] DiskIoMgrTest.SingleWriter (3407 ms) > [ RUN ] DiskIoMgrTest.InvalidWrite > [ OK ] DiskIoMgrTest.InvalidWrite (281 ms) > [ RUN ] DiskIoMgrTest.WriteErrors > [ OK ] DiskIoMgrTest.WriteErrors (235 ms) > [ RUN ] DiskIoMgrTest.SingleWriterCancel > [ OK ] DiskIoMgrTest.SingleWriterCancel (1165 ms) > [ RUN ] DiskIoMgrTest.SingleReader > [ OK ] DiskIoMgrTest.SingleReader (5835 ms) > [ RUN ] DiskIoMgrTest.SingleReaderSubRanges > [ OK ] DiskIoMgrTest.SingleReaderSubRanges (16404 ms) > [ RUN ] DiskIoMgrTest.AddScanRangeTest > [ OK ] DiskIoMgrTest.AddScanRangeTest (1210 ms) > [ RUN ] DiskIoMgrTest.SyncReadTest > *** Check failure stack trace: *** > @ 0x4825dcc > @ 0x4827671 > @ 0x48257a6 > @ 0x4828d6d > @ 0x1af39ec > @ 0x1ae90a4 > @ 0x1ac30ea > @ 0x1accad3 > @ 0x1acc660 > @ 0x1acbf3e > @ 0x1acb62d > @ 0x1b03671 > @ 0x1f79988 > @ 0x1f82b60 > @ 0x1f82a84 > @ 0x1f82a47 > @ 0x3751579 > @ 0x3ea4807850 > @ 0x3ea44e894c > Wrote minidump to > /data/jenkins/workspace/<...>/repos/Impala/logs/be_tests/minidumps/disk-io-mgr-test/5bbf76f7-e5d6-4ac9-bdae9d9b-065c32ec.dmp > {noformat} > *Error*: > {noformat} > Operating system: Linux > 0.0.0 Linux 2.6.32-358.14.1.el6.centos.plus.x86_64 #1 SMP > Tue Jul 16 21:33:24 UTC 2013 x86_64 > CPU: amd64 > family 6 model 45 stepping 7 > 8 CPUs > GPU: UNKNOWN > Crash reason: SIGABRT > Crash address: 0x4522fa1 > Process uptime: not available > Thread 205 (crashed) > 0 libc-2.12.so + 0x328e5 > rax = 0x rdx = 0x0006 > rcx = 0x rbx = 0x06adf9c0 > rsi = 0x0563 rdi = 0x2fa1 > rbp = 0x7f8009b8ffe0 rsp = 0x7f8009b8fc78 > r8 = 0x7f8009b8fd00r9 = 0x0563 > r10 = 0x0008 r11 = 0x0202 > r12 = 0x06adfa40 r13 = 0x001f > r14 = 0x06ae7384 r15 = 0x06adf9c0 > rip = 0x003ea44328e5 > Found by: given as instruction pointer in context > 1 libc-2.12.so + 0x340c5 > rbp = 0x7f8009b8ffe0 rsp = 0x7f8009b8fc80 > rip = 0x003ea44340c5 > Found by: stack scanning > 2 disk-io-mgr-test!boost::_bi::bind_t impala::io::DiskQueue, impala::io::DiskIoMgr*>, > boost::_bi::list2, > boost::_bi::value > >::operator()() > [bind_template.hpp : 20 + 0x21] > rbp = 0x7f8009b8ffe0
[jira] [Resolved] (IMPALA-8107) Support EXEC_TIME_LIMIT_S in resource pool setting
[ https://issues.apache.org/jira/browse/IMPALA-8107?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Quanlong Huang resolved IMPALA-8107. Resolution: Resolved Already supported by IMPALA-2538. Close this JIRA. > Support EXEC_TIME_LIMIT_S in resource pool setting > -- > > Key: IMPALA-8107 > URL: https://issues.apache.org/jira/browse/IMPALA-8107 > Project: IMPALA > Issue Type: New Feature >Reporter: Quanlong Huang >Priority: Major > Labels: admission-control > > Timeout limit should be different for different kinds of queries. For > example, resource pool for adhoc queries may set EXEC_TIME_LIMIT_S to 60s. > Resource pool for building pre-aggregaions or other ETL may need a larger > EXEC_TIME_LIMIT_S like 30 minutes. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Resolved] (IMPALA-8107) Support EXEC_TIME_LIMIT_S in resource pool setting
[ https://issues.apache.org/jira/browse/IMPALA-8107?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Quanlong Huang resolved IMPALA-8107. Resolution: Resolved Already supported by IMPALA-2538. Close this JIRA. > Support EXEC_TIME_LIMIT_S in resource pool setting > -- > > Key: IMPALA-8107 > URL: https://issues.apache.org/jira/browse/IMPALA-8107 > Project: IMPALA > Issue Type: New Feature >Reporter: Quanlong Huang >Priority: Major > Labels: admission-control > > Timeout limit should be different for different kinds of queries. For > example, resource pool for adhoc queries may set EXEC_TIME_LIMIT_S to 60s. > Resource pool for building pre-aggregaions or other ETL may need a larger > EXEC_TIME_LIMIT_S like 30 minutes. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org For additional commands, e-mail: issues-all-h...@impala.apache.org
[jira] [Commented] (IMPALA-8115) some jenkins workers slow to do real work due to dpkg lock conflicts
[ https://issues.apache.org/jira/browse/IMPALA-8115?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16751703#comment-16751703 ] Michael Brown commented on IMPALA-8115: --- I noticed this because I posted https://gerrit.cloudera.org/#/c/12272/ at 2:42 PT and at 3:45 PT there is still no update that the pre-review job completed. https://jenkins.impala.io/job/gerrit-code-review-checks/1879/console shows a slow timeline: {noformat} Triggered by Gerrit: http://gerrit.cloudera.org:8080/12272 Running in Durability level: MAX_SURVIVABILITY [Pipeline] timestamps [Pipeline] { [Pipeline] ansiColor [Pipeline] { [Pipeline] timeout 22:42:09 Timeout set to expire in 10 hr [Pipeline] { [Pipeline] parallel [Pipeline] { (Branch: Tidy) [Pipeline] { (Branch: BuildOnly) [Pipeline] { (Branch: Python26Compatibility) [Pipeline] { (Branch: Rat) [Pipeline] build (Building clang-tidy-ub1604) 22:42:09 Scheduling project: clang-tidy-ub1604 [Pipeline] build (Building ubuntu-16.04-build-only) 22:42:09 Scheduling project: ubuntu-16.04-build-only [Pipeline] build (Building python26-incompatibility-check) 22:42:09 Scheduling project: python26-incompatibility-check [Pipeline] build (Building rat-check-ub1604) 22:42:09 Scheduling project: rat-check-ub1604 22:44:22 Starting building: python26-incompatibility-check #1548 [Pipeline] } 23:35:15 Starting building: ubuntu-16.04-build-only #5027 23:36:34 Starting building: rat-check-ub1604 #5404 [Pipeline] } 23:37:14 Starting building: clang-tidy-ub1604 #4922 {noformat} It's taken nearly an hour to start rat-check-ub1604 #5404 and clang-tidy-ub1604 #4922. My 15 minutes math doesn't account for all of this. I'm not yet sure where else we could accelerate/improve the process. > some jenkins workers slow to do real work due to dpkg lock conflicts > > > Key: IMPALA-8115 > URL: https://issues.apache.org/jira/browse/IMPALA-8115 > Project: IMPALA > Issue Type: Bug > Components: Infrastructure >Reporter: Michael Brown >Priority: Major > > A Jenkins worker for label {{ubuntu-16.04}} took about 15 minutes to start > doing real work. I noticed that it was retrying {{apt-get update}}: > {noformat} > ++ sudo apt-get --yes install openjdk-8-jdk > E: Could not get lock /var/lib/dpkg/lock - open (11: Resource temporarily > unavailable) > E: Unable to lock the administration directory (/var/lib/dpkg/), is another > process using it? > ++ date > Thu Jan 24 23:37:33 UTC 2019 > ++ sudo apt-get update > ++ sleep 10 > ++ sudo apt-get --yes install openjdk-8-jdk > [etc] > {noformat} > I ssh'd into a host and saw that, yes, something else was holding onto the > dpkg log (confirmed with lsof and not pasted here. dpkg process PID 11459 was > the culprit) > {noformat} > root 1750 0.0 0.0 4508 1664 ?Ss 23:21 0:00 /bin/sh > /usr/lib/apt/apt.systemd.daily > root 1804 12.3 0.1 141076 80452 ?S23:22 1:24 \_ > /usr/bin/python3 /usr/bin/unattended-upgrade > root 3263 0.0 0.1 140960 72896 ?S23:23 0:00 \_ > /usr/bin/python3 /usr/bin/unattended-upgrade > root 11459 0.6 0.0 45920 25184 pts/1Ss+ 23:24 0:03 \_ > /usr/bin/dpkg --status-fd 10 --unpack --auto-deconfigure > /var/cache/apt/archives/tzdata_2018i-0ubuntu0.16.04_all.deb > /var/cache/apt/archives/distro-info-data_0.28ubuntu0.9_all.deb > /var/cache/apt/archives/file_1%3a5.25-2ubuntu1.1_amd64.deb > /var/cache/apt/archives/libmagic1_1%3a5.25-2ubuntu1.1_amd64.deb > /var/cache/apt/archives/libisc-export160_1%3a9.10.3.dfsg.P4-8ubuntu1.11_amd64.deb > > /var/cache/apt/archives/libdns-export162_1%3a9.10.3.dfsg.P4-8ubuntu1.11_amd64.deb > /var/cache/apt/archives/isc-dhcp-client_4.3.3-5ubuntu12.9_amd64.deb > /var/cache/apt/archives/isc-dhcp-common_4.3.3-5ubuntu12.9_amd64.deb > /var/cache/apt/archives/libidn11_1.32-3ubuntu1.2_amd64.deb > /var/cache/apt/archives/libpng12-0_1.2.54-1ubuntu1.1_amd64.deb > /var/cache/apt/archives/libtasn1-6_4.7-3ubuntu0.16.04.3_amd64.deb > /var/cache/apt/archives/libapparmor-perl_2.10.95-0ubuntu2.10_amd64.deb > /var/cache/apt/archives/apparmor_2.10.95-0ubuntu2.10_amd64.deb > /var/cache/apt/archives/curl_7.47.0-1ubuntu2.11_amd64.deb > /var/cache/apt/archives/libgssapi-krb5-2_1.13.2+dfsg-5ubuntu2.1_amd64.deb > /var/cache/apt/archives/libkrb5-3_1.13.2+dfsg-5ubuntu2.1_amd64.deb > /var/cache/apt/archives/libkrb5support0_1.13.2+dfsg-5ubuntu2.1_amd64.deb > /var/cache/apt/archives/libk5crypto3_1.13.2+dfsg-5ubuntu2.1_amd64.deb > /var/cache/apt/archives/libcurl3-gnutls_7.47.0-1ubuntu2.11_amd64.deb > /var/cache/apt/archives/apt-transport-https_1.2.29ubuntu0.1_amd64.deb > /var/cache/apt/archives/libicu55_55.1-7ubuntu0.4_amd64.deb > /var/cache/apt/archives/libxml2_2.9.3+dfsg1-1ubuntu0.6_amd64.deb >
[jira] [Commented] (IMPALA-8107) Support EXEC_TIME_LIMIT_S in resource pool setting
[ https://issues.apache.org/jira/browse/IMPALA-8107?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16751711#comment-16751711 ] Quanlong Huang commented on IMPALA-8107: Oh, I really missed the impala.admission-control.pool-default-query-options settings! Thanks for pointing that! > Support EXEC_TIME_LIMIT_S in resource pool setting > -- > > Key: IMPALA-8107 > URL: https://issues.apache.org/jira/browse/IMPALA-8107 > Project: IMPALA > Issue Type: New Feature >Reporter: Quanlong Huang >Priority: Major > Labels: admission-control > > Timeout limit should be different for different kinds of queries. For > example, resource pool for adhoc queries may set EXEC_TIME_LIMIT_S to 60s. > Resource pool for building pre-aggregaions or other ETL may need a larger > EXEC_TIME_LIMIT_S like 30 minutes. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org For additional commands, e-mail: issues-all-h...@impala.apache.org
[jira] [Updated] (IMPALA-8115) some jenkins workers slow to do real work due to dpkg lock conflicts
[ https://issues.apache.org/jira/browse/IMPALA-8115?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Michael Brown updated IMPALA-8115: -- Summary: some jenkins workers slow to do real work due to dpkg lock conflicts (was: some jenkins workers slow to spawn to to dpkg lock conflicts) > some jenkins workers slow to do real work due to dpkg lock conflicts > > > Key: IMPALA-8115 > URL: https://issues.apache.org/jira/browse/IMPALA-8115 > Project: IMPALA > Issue Type: Bug > Components: Infrastructure >Reporter: Michael Brown >Priority: Major > > A Jenkins worker for label {{ubuntu-16.04}} took about 15 minutes to start > doing real work. I noticed that it was retrying {{apt-get update}}: > {noformat} > ++ sudo apt-get --yes install openjdk-8-jdk > E: Could not get lock /var/lib/dpkg/lock - open (11: Resource temporarily > unavailable) > E: Unable to lock the administration directory (/var/lib/dpkg/), is another > process using it? > ++ date > Thu Jan 24 23:37:33 UTC 2019 > ++ sudo apt-get update > ++ sleep 10 > ++ sudo apt-get --yes install openjdk-8-jdk > [etc] > {noformat} > I ssh'd into a host and saw that, yes, something else was holding onto the > dpkg log (confirmed with lsof and not pasted here. dpkg process PID 11459 was > the culprit) > {noformat} > root 1750 0.0 0.0 4508 1664 ?Ss 23:21 0:00 /bin/sh > /usr/lib/apt/apt.systemd.daily > root 1804 12.3 0.1 141076 80452 ?S23:22 1:24 \_ > /usr/bin/python3 /usr/bin/unattended-upgrade > root 3263 0.0 0.1 140960 72896 ?S23:23 0:00 \_ > /usr/bin/python3 /usr/bin/unattended-upgrade > root 11459 0.6 0.0 45920 25184 pts/1Ss+ 23:24 0:03 \_ > /usr/bin/dpkg --status-fd 10 --unpack --auto-deconfigure > /var/cache/apt/archives/tzdata_2018i-0ubuntu0.16.04_all.deb > /var/cache/apt/archives/distro-info-data_0.28ubuntu0.9_all.deb > /var/cache/apt/archives/file_1%3a5.25-2ubuntu1.1_amd64.deb > /var/cache/apt/archives/libmagic1_1%3a5.25-2ubuntu1.1_amd64.deb > /var/cache/apt/archives/libisc-export160_1%3a9.10.3.dfsg.P4-8ubuntu1.11_amd64.deb > > /var/cache/apt/archives/libdns-export162_1%3a9.10.3.dfsg.P4-8ubuntu1.11_amd64.deb > /var/cache/apt/archives/isc-dhcp-client_4.3.3-5ubuntu12.9_amd64.deb > /var/cache/apt/archives/isc-dhcp-common_4.3.3-5ubuntu12.9_amd64.deb > /var/cache/apt/archives/libidn11_1.32-3ubuntu1.2_amd64.deb > /var/cache/apt/archives/libpng12-0_1.2.54-1ubuntu1.1_amd64.deb > /var/cache/apt/archives/libtasn1-6_4.7-3ubuntu0.16.04.3_amd64.deb > /var/cache/apt/archives/libapparmor-perl_2.10.95-0ubuntu2.10_amd64.deb > /var/cache/apt/archives/apparmor_2.10.95-0ubuntu2.10_amd64.deb > /var/cache/apt/archives/curl_7.47.0-1ubuntu2.11_amd64.deb > /var/cache/apt/archives/libgssapi-krb5-2_1.13.2+dfsg-5ubuntu2.1_amd64.deb > /var/cache/apt/archives/libkrb5-3_1.13.2+dfsg-5ubuntu2.1_amd64.deb > /var/cache/apt/archives/libkrb5support0_1.13.2+dfsg-5ubuntu2.1_amd64.deb > /var/cache/apt/archives/libk5crypto3_1.13.2+dfsg-5ubuntu2.1_amd64.deb > /var/cache/apt/archives/libcurl3-gnutls_7.47.0-1ubuntu2.11_amd64.deb > /var/cache/apt/archives/apt-transport-https_1.2.29ubuntu0.1_amd64.deb > /var/cache/apt/archives/libicu55_55.1-7ubuntu0.4_amd64.deb > /var/cache/apt/archives/libxml2_2.9.3+dfsg1-1ubuntu0.6_amd64.deb > /var/cache/apt/archives/bind9-host_1%3a9.10.3.dfsg.P4-8ubuntu1.11_amd64.deb > /var/cache/apt/archives/dnsutils_1%3a9.10.3.dfsg.P4-8ubuntu1.11_amd64.deb > /var/cache/apt/archives/libisc160_1%3a9.10.3.dfsg.P4-8ubuntu1.11_amd64.deb > /var/cache/apt/archives/libdns162_1%3a9.10.3.dfsg.P4-8ubuntu1.11_amd64.deb > /var/cache/apt/archives/libisccc140_1%3a9.10.3.dfsg.P4-8ubuntu1.11_amd64.deb > /var/cache/apt/archives/libisccfg140_1%3a9.10.3.dfsg.P4-8ubuntu1.11_amd64.deb > /var/cache/apt/archives/liblwres141_1%3a9.10.3.dfsg.P4-8ubuntu1.11_amd64.deb > /var/cache/apt/archives/libbind9-140_1%3a9.10.3.dfsg.P4-8ubuntu1.11_amd64.deb > /var/cache/apt/archives/openssl_1.0.2g-1ubuntu4.14_amd64.deb > /var/cache/apt/archives/ca-certificates_20170717~16.04.1_all.deb > /var/cache/apt/archives/libasprintf0v5_0.19.7-2ubuntu3.1_amd64.deb > /var/cache/apt/archives/gettext-base_0.19.7-2ubuntu3.1_amd64.deb > /var/cache/apt/archives/krb5-locales_1.13.2+dfsg-5ubuntu2.1_all.deb > /var/cache/apt/archives/libelf1_0.165-3ubuntu1.1_amd64.deb > /var/cache/apt/archives/libglib2.0-data_2.48.2-0ubuntu4.1_all.deb > /var/cache/apt/archives/libnuma1_2.0.11-1ubuntu1.1_amd64.deb > /var/cache/apt/archives/libpolkit-gobject-1-0_0.105-14.1ubuntu0.4_amd64.deb > /var/cache/apt/archives/libx11-data_2%3a1.6.3-1ubuntu2.1_all.deb > /var/cache/apt/archives/libx11-6_2%3a1.6.3-1ubuntu2.1_amd64.deb > /var/cache/apt/archives/openssh-sftp-server_1%3a7.2p2-4ubuntu2.6_amd64.deb
[jira] [Updated] (IMPALA-8115) some jenkins workers slow to spawn to to dpkg lock conflicts
[ https://issues.apache.org/jira/browse/IMPALA-8115?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Michael Brown updated IMPALA-8115: -- Description: A Jenkins worker for label {{ubuntu-16.04}} took about 15 minutes to start doing real work. I noticed that it was retrying {{apt-get update}}: {noformat} ++ sudo apt-get --yes install openjdk-8-jdk E: Could not get lock /var/lib/dpkg/lock - open (11: Resource temporarily unavailable) E: Unable to lock the administration directory (/var/lib/dpkg/), is another process using it? ++ date Thu Jan 24 23:37:33 UTC 2019 ++ sudo apt-get update ++ sleep 10 ++ sudo apt-get --yes install openjdk-8-jdk [etc] {noformat} I ssh'd into a host and saw that, yes, something else was holding onto the dpkg log (confirmed with lsof and not pasted here. dpkg process PID 11459 was the culprit) {noformat} root 1750 0.0 0.0 4508 1664 ?Ss 23:21 0:00 /bin/sh /usr/lib/apt/apt.systemd.daily root 1804 12.3 0.1 141076 80452 ?S23:22 1:24 \_ /usr/bin/python3 /usr/bin/unattended-upgrade root 3263 0.0 0.1 140960 72896 ?S23:23 0:00 \_ /usr/bin/python3 /usr/bin/unattended-upgrade root 11459 0.6 0.0 45920 25184 pts/1Ss+ 23:24 0:03 \_ /usr/bin/dpkg --status-fd 10 --unpack --auto-deconfigure /var/cache/apt/archives/tzdata_2018i-0ubuntu0.16.04_all.deb /var/cache/apt/archives/distro-info-data_0.28ubuntu0.9_all.deb /var/cache/apt/archives/file_1%3a5.25-2ubuntu1.1_amd64.deb /var/cache/apt/archives/libmagic1_1%3a5.25-2ubuntu1.1_amd64.deb /var/cache/apt/archives/libisc-export160_1%3a9.10.3.dfsg.P4-8ubuntu1.11_amd64.deb /var/cache/apt/archives/libdns-export162_1%3a9.10.3.dfsg.P4-8ubuntu1.11_amd64.deb /var/cache/apt/archives/isc-dhcp-client_4.3.3-5ubuntu12.9_amd64.deb /var/cache/apt/archives/isc-dhcp-common_4.3.3-5ubuntu12.9_amd64.deb /var/cache/apt/archives/libidn11_1.32-3ubuntu1.2_amd64.deb /var/cache/apt/archives/libpng12-0_1.2.54-1ubuntu1.1_amd64.deb /var/cache/apt/archives/libtasn1-6_4.7-3ubuntu0.16.04.3_amd64.deb /var/cache/apt/archives/libapparmor-perl_2.10.95-0ubuntu2.10_amd64.deb /var/cache/apt/archives/apparmor_2.10.95-0ubuntu2.10_amd64.deb /var/cache/apt/archives/curl_7.47.0-1ubuntu2.11_amd64.deb /var/cache/apt/archives/libgssapi-krb5-2_1.13.2+dfsg-5ubuntu2.1_amd64.deb /var/cache/apt/archives/libkrb5-3_1.13.2+dfsg-5ubuntu2.1_amd64.deb /var/cache/apt/archives/libkrb5support0_1.13.2+dfsg-5ubuntu2.1_amd64.deb /var/cache/apt/archives/libk5crypto3_1.13.2+dfsg-5ubuntu2.1_amd64.deb /var/cache/apt/archives/libcurl3-gnutls_7.47.0-1ubuntu2.11_amd64.deb /var/cache/apt/archives/apt-transport-https_1.2.29ubuntu0.1_amd64.deb /var/cache/apt/archives/libicu55_55.1-7ubuntu0.4_amd64.deb /var/cache/apt/archives/libxml2_2.9.3+dfsg1-1ubuntu0.6_amd64.deb /var/cache/apt/archives/bind9-host_1%3a9.10.3.dfsg.P4-8ubuntu1.11_amd64.deb /var/cache/apt/archives/dnsutils_1%3a9.10.3.dfsg.P4-8ubuntu1.11_amd64.deb /var/cache/apt/archives/libisc160_1%3a9.10.3.dfsg.P4-8ubuntu1.11_amd64.deb /var/cache/apt/archives/libdns162_1%3a9.10.3.dfsg.P4-8ubuntu1.11_amd64.deb /var/cache/apt/archives/libisccc140_1%3a9.10.3.dfsg.P4-8ubuntu1.11_amd64.deb /var/cache/apt/archives/libisccfg140_1%3a9.10.3.dfsg.P4-8ubuntu1.11_amd64.deb /var/cache/apt/archives/liblwres141_1%3a9.10.3.dfsg.P4-8ubuntu1.11_amd64.deb /var/cache/apt/archives/libbind9-140_1%3a9.10.3.dfsg.P4-8ubuntu1.11_amd64.deb /var/cache/apt/archives/openssl_1.0.2g-1ubuntu4.14_amd64.deb /var/cache/apt/archives/ca-certificates_20170717~16.04.1_all.deb /var/cache/apt/archives/libasprintf0v5_0.19.7-2ubuntu3.1_amd64.deb /var/cache/apt/archives/gettext-base_0.19.7-2ubuntu3.1_amd64.deb /var/cache/apt/archives/krb5-locales_1.13.2+dfsg-5ubuntu2.1_all.deb /var/cache/apt/archives/libelf1_0.165-3ubuntu1.1_amd64.deb /var/cache/apt/archives/libglib2.0-data_2.48.2-0ubuntu4.1_all.deb /var/cache/apt/archives/libnuma1_2.0.11-1ubuntu1.1_amd64.deb /var/cache/apt/archives/libpolkit-gobject-1-0_0.105-14.1ubuntu0.4_amd64.deb /var/cache/apt/archives/libx11-data_2%3a1.6.3-1ubuntu2.1_all.deb /var/cache/apt/archives/libx11-6_2%3a1.6.3-1ubuntu2.1_amd64.deb /var/cache/apt/archives/openssh-sftp-server_1%3a7.2p2-4ubuntu2.6_amd64.deb /var/cache/apt/archives/openssh-server_1%3a7.2p2-4ubuntu2.6_amd64.deb /var/cache/apt/archives/openssh-client_1%3a7.2p2-4ubuntu2.6_amd64.deb /var/cache/apt/archives/rsync_3.1.1-3ubuntu1.2_amd64.deb /var/cache/apt/archives/tcpdump_4.9.2-0ubuntu0.16.04.1_amd64.deb /var/cache/apt/archives/wget_1.17.1-1ubuntu1.4_amd64.deb /var/cache/apt/archives/python3-problem-report_2.20.1-0ubuntu2.18_all.deb /var/cache/apt/archives/python3-apport_2.20.1-0ubuntu2.18_all.deb /var/cache/apt/archives/apport_2.20.1-0ubuntu2.18_all.deb /var/cache/apt/archives/dns-root-data_2018013001~16.04.1_all.deb /var/cache/apt/archives/dnsmasq-base_2.75-1ubuntu0.16.04.5_amd64.deb
[jira] [Created] (IMPALA-8115) some jenkins workers slow to spawn to to dpkg lock conflicts
Michael Brown created IMPALA-8115: - Summary: some jenkins workers slow to spawn to to dpkg lock conflicts Key: IMPALA-8115 URL: https://issues.apache.org/jira/browse/IMPALA-8115 Project: IMPALA Issue Type: Bug Components: Infrastructure Reporter: Michael Brown A Jenkins worker for label {{ubuntu-16.04}} took about 15 minutes to start doing real work. I noticed that it was retrying {{apt-get update}}: {noformat} ++ sudo apt-get --yes install openjdk-8-jdk E: Could not get lock /var/lib/dpkg/lock - open (11: Resource temporarily unavailable) E: Unable to lock the administration directory (/var/lib/dpkg/), is another process using it? ++ date Thu Jan 24 23:37:33 UTC 2019 ++ sudo apt-get update ++ sleep 10 ++ sudo apt-get --yes install openjdk-8-jdk [etc] {noformat} I ssh'd into a host and saw that, yes, something else was holding onto the dpkg log (confirmed with lsof and not pasted here. dpkg process PID 11459 was the culprit) {noformat} root 1750 0.0 0.0 4508 1664 ?Ss 23:21 0:00 /bin/sh /usr/lib/apt/apt.systemd.daily root 1804 12.3 0.1 141076 80452 ?S23:22 1:24 \_ /usr/bin/python3 /usr/bin/unattended-upgrade root 3263 0.0 0.1 140960 72896 ?S23:23 0:00 \_ /usr/bin/python3 /usr/bin/unattended-upgrade root 11459 0.6 0.0 45920 25184 pts/1Ss+ 23:24 0:03 \_ /usr/bin/dpkg --status-fd 10 --unpack --auto-deconfigure /var/cache/apt/archives/tzdata_2018i-0ubuntu0.16.04_all.deb /var/cache/apt/archives/distro-info-data_0.28ubuntu0.9_all.deb /var/cache/apt/archives/file_1%3a5.25-2ubuntu1.1_amd64.deb /var/cache/apt/archives/libmagic1_1%3a5.25-2ubuntu1.1_amd64.deb /var/cache/apt/archives/libisc-export160_1%3a9.10.3.dfsg.P4-8ubuntu1.11_amd64.deb /var/cache/apt/archives/libdns-export162_1%3a9.10.3.dfsg.P4-8ubuntu1.11_amd64.deb /var/cache/apt/archives/isc-dhcp-client_4.3.3-5ubuntu12.9_amd64.deb /var/cache/apt/archives/isc-dhcp-common_4.3.3-5ubuntu12.9_amd64.deb /var/cache/apt/archives/libidn11_1.32-3ubuntu1.2_amd64.deb /var/cache/apt/archives/libpng12-0_1.2.54-1ubuntu1.1_amd64.deb /var/cache/apt/archives/libtasn1-6_4.7-3ubuntu0.16.04.3_amd64.deb /var/cache/apt/archives/libapparmor-perl_2.10.95-0ubuntu2.10_amd64.deb /var/cache/apt/archives/apparmor_2.10.95-0ubuntu2.10_amd64.deb /var/cache/apt/archives/curl_7.47.0-1ubuntu2.11_amd64.deb /var/cache/apt/archives/libgssapi-krb5-2_1.13.2+dfsg-5ubuntu2.1_amd64.deb /var/cache/apt/archives/libkrb5-3_1.13.2+dfsg-5ubuntu2.1_amd64.deb /var/cache/apt/archives/libkrb5support0_1.13.2+dfsg-5ubuntu2.1_amd64.deb /var/cache/apt/archives/libk5crypto3_1.13.2+dfsg-5ubuntu2.1_amd64.deb /var/cache/apt/archives/libcurl3-gnutls_7.47.0-1ubuntu2.11_amd64.deb /var/cache/apt/archives/apt-transport-https_1.2.29ubuntu0.1_amd64.deb /var/cache/apt/archives/libicu55_55.1-7ubuntu0.4_amd64.deb /var/cache/apt/archives/libxml2_2.9.3+dfsg1-1ubuntu0.6_amd64.deb /var/cache/apt/archives/bind9-host_1%3a9.10.3.dfsg.P4-8ubuntu1.11_amd64.deb /var/cache/apt/archives/dnsutils_1%3a9.10.3.dfsg.P4-8ubuntu1.11_amd64.deb /var/cache/apt/archives/libisc160_1%3a9.10.3.dfsg.P4-8ubuntu1.11_amd64.deb /var/cache/apt/archives/libdns162_1%3a9.10.3.dfsg.P4-8ubuntu1.11_amd64.deb /var/cache/apt/archives/libisccc140_1%3a9.10.3.dfsg.P4-8ubuntu1.11_amd64.deb /var/cache/apt/archives/libisccfg140_1%3a9.10.3.dfsg.P4-8ubuntu1.11_amd64.deb /var/cache/apt/archives/liblwres141_1%3a9.10.3.dfsg.P4-8ubuntu1.11_amd64.deb /var/cache/apt/archives/libbind9-140_1%3a9.10.3.dfsg.P4-8ubuntu1.11_amd64.deb /var/cache/apt/archives/openssl_1.0.2g-1ubuntu4.14_amd64.deb /var/cache/apt/archives/ca-certificates_20170717~16.04.1_all.deb /var/cache/apt/archives/libasprintf0v5_0.19.7-2ubuntu3.1_amd64.deb /var/cache/apt/archives/gettext-base_0.19.7-2ubuntu3.1_amd64.deb /var/cache/apt/archives/krb5-locales_1.13.2+dfsg-5ubuntu2.1_all.deb /var/cache/apt/archives/libelf1_0.165-3ubuntu1.1_amd64.deb /var/cache/apt/archives/libglib2.0-data_2.48.2-0ubuntu4.1_all.deb /var/cache/apt/archives/libnuma1_2.0.11-1ubuntu1.1_amd64.deb /var/cache/apt/archives/libpolkit-gobject-1-0_0.105-14.1ubuntu0.4_amd64.deb /var/cache/apt/archives/libx11-data_2%3a1.6.3-1ubuntu2.1_all.deb /var/cache/apt/archives/libx11-6_2%3a1.6.3-1ubuntu2.1_amd64.deb /var/cache/apt/archives/openssh-sftp-server_1%3a7.2p2-4ubuntu2.6_amd64.deb /var/cache/apt/archives/openssh-server_1%3a7.2p2-4ubuntu2.6_amd64.deb /var/cache/apt/archives/openssh-client_1%3a7.2p2-4ubuntu2.6_amd64.deb /var/cache/apt/archives/rsync_3.1.1-3ubuntu1.2_amd64.deb /var/cache/apt/archives/tcpdump_4.9.2-0ubuntu0.16.04.1_amd64.deb /var/cache/apt/archives/wget_1.17.1-1ubuntu1.4_amd64.deb /var/cache/apt/archives/python3-problem-report_2.20.1-0ubuntu2.18_all.deb /var/cache/apt/archives/python3-apport_2.20.1-0ubuntu2.18_all.deb
[jira] [Created] (IMPALA-8115) some jenkins workers slow to spawn to to dpkg lock conflicts
Michael Brown created IMPALA-8115: - Summary: some jenkins workers slow to spawn to to dpkg lock conflicts Key: IMPALA-8115 URL: https://issues.apache.org/jira/browse/IMPALA-8115 Project: IMPALA Issue Type: Bug Components: Infrastructure Reporter: Michael Brown A Jenkins worker for label {{ubuntu-16.04}} took about 15 minutes to start doing real work. I noticed that it was retrying {{apt-get update}}: {noformat} ++ sudo apt-get --yes install openjdk-8-jdk E: Could not get lock /var/lib/dpkg/lock - open (11: Resource temporarily unavailable) E: Unable to lock the administration directory (/var/lib/dpkg/), is another process using it? ++ date Thu Jan 24 23:37:33 UTC 2019 ++ sudo apt-get update ++ sleep 10 ++ sudo apt-get --yes install openjdk-8-jdk [etc] {noformat} I ssh'd into a host and saw that, yes, something else was holding onto the dpkg log (confirmed with lsof and not pasted here. dpkg process PID 11459 was the culprit) {noformat} root 1750 0.0 0.0 4508 1664 ?Ss 23:21 0:00 /bin/sh /usr/lib/apt/apt.systemd.daily root 1804 12.3 0.1 141076 80452 ?S23:22 1:24 \_ /usr/bin/python3 /usr/bin/unattended-upgrade root 3263 0.0 0.1 140960 72896 ?S23:23 0:00 \_ /usr/bin/python3 /usr/bin/unattended-upgrade root 11459 0.6 0.0 45920 25184 pts/1Ss+ 23:24 0:03 \_ /usr/bin/dpkg --status-fd 10 --unpack --auto-deconfigure /var/cache/apt/archives/tzdata_2018i-0ubuntu0.16.04_all.deb /var/cache/apt/archives/distro-info-data_0.28ubuntu0.9_all.deb /var/cache/apt/archives/file_1%3a5.25-2ubuntu1.1_amd64.deb /var/cache/apt/archives/libmagic1_1%3a5.25-2ubuntu1.1_amd64.deb /var/cache/apt/archives/libisc-export160_1%3a9.10.3.dfsg.P4-8ubuntu1.11_amd64.deb /var/cache/apt/archives/libdns-export162_1%3a9.10.3.dfsg.P4-8ubuntu1.11_amd64.deb /var/cache/apt/archives/isc-dhcp-client_4.3.3-5ubuntu12.9_amd64.deb /var/cache/apt/archives/isc-dhcp-common_4.3.3-5ubuntu12.9_amd64.deb /var/cache/apt/archives/libidn11_1.32-3ubuntu1.2_amd64.deb /var/cache/apt/archives/libpng12-0_1.2.54-1ubuntu1.1_amd64.deb /var/cache/apt/archives/libtasn1-6_4.7-3ubuntu0.16.04.3_amd64.deb /var/cache/apt/archives/libapparmor-perl_2.10.95-0ubuntu2.10_amd64.deb /var/cache/apt/archives/apparmor_2.10.95-0ubuntu2.10_amd64.deb /var/cache/apt/archives/curl_7.47.0-1ubuntu2.11_amd64.deb /var/cache/apt/archives/libgssapi-krb5-2_1.13.2+dfsg-5ubuntu2.1_amd64.deb /var/cache/apt/archives/libkrb5-3_1.13.2+dfsg-5ubuntu2.1_amd64.deb /var/cache/apt/archives/libkrb5support0_1.13.2+dfsg-5ubuntu2.1_amd64.deb /var/cache/apt/archives/libk5crypto3_1.13.2+dfsg-5ubuntu2.1_amd64.deb /var/cache/apt/archives/libcurl3-gnutls_7.47.0-1ubuntu2.11_amd64.deb /var/cache/apt/archives/apt-transport-https_1.2.29ubuntu0.1_amd64.deb /var/cache/apt/archives/libicu55_55.1-7ubuntu0.4_amd64.deb /var/cache/apt/archives/libxml2_2.9.3+dfsg1-1ubuntu0.6_amd64.deb /var/cache/apt/archives/bind9-host_1%3a9.10.3.dfsg.P4-8ubuntu1.11_amd64.deb /var/cache/apt/archives/dnsutils_1%3a9.10.3.dfsg.P4-8ubuntu1.11_amd64.deb /var/cache/apt/archives/libisc160_1%3a9.10.3.dfsg.P4-8ubuntu1.11_amd64.deb /var/cache/apt/archives/libdns162_1%3a9.10.3.dfsg.P4-8ubuntu1.11_amd64.deb /var/cache/apt/archives/libisccc140_1%3a9.10.3.dfsg.P4-8ubuntu1.11_amd64.deb /var/cache/apt/archives/libisccfg140_1%3a9.10.3.dfsg.P4-8ubuntu1.11_amd64.deb /var/cache/apt/archives/liblwres141_1%3a9.10.3.dfsg.P4-8ubuntu1.11_amd64.deb /var/cache/apt/archives/libbind9-140_1%3a9.10.3.dfsg.P4-8ubuntu1.11_amd64.deb /var/cache/apt/archives/openssl_1.0.2g-1ubuntu4.14_amd64.deb /var/cache/apt/archives/ca-certificates_20170717~16.04.1_all.deb /var/cache/apt/archives/libasprintf0v5_0.19.7-2ubuntu3.1_amd64.deb /var/cache/apt/archives/gettext-base_0.19.7-2ubuntu3.1_amd64.deb /var/cache/apt/archives/krb5-locales_1.13.2+dfsg-5ubuntu2.1_all.deb /var/cache/apt/archives/libelf1_0.165-3ubuntu1.1_amd64.deb /var/cache/apt/archives/libglib2.0-data_2.48.2-0ubuntu4.1_all.deb /var/cache/apt/archives/libnuma1_2.0.11-1ubuntu1.1_amd64.deb /var/cache/apt/archives/libpolkit-gobject-1-0_0.105-14.1ubuntu0.4_amd64.deb /var/cache/apt/archives/libx11-data_2%3a1.6.3-1ubuntu2.1_all.deb /var/cache/apt/archives/libx11-6_2%3a1.6.3-1ubuntu2.1_amd64.deb /var/cache/apt/archives/openssh-sftp-server_1%3a7.2p2-4ubuntu2.6_amd64.deb /var/cache/apt/archives/openssh-server_1%3a7.2p2-4ubuntu2.6_amd64.deb /var/cache/apt/archives/openssh-client_1%3a7.2p2-4ubuntu2.6_amd64.deb /var/cache/apt/archives/rsync_3.1.1-3ubuntu1.2_amd64.deb /var/cache/apt/archives/tcpdump_4.9.2-0ubuntu0.16.04.1_amd64.deb /var/cache/apt/archives/wget_1.17.1-1ubuntu1.4_amd64.deb /var/cache/apt/archives/python3-problem-report_2.20.1-0ubuntu2.18_all.deb /var/cache/apt/archives/python3-apport_2.20.1-0ubuntu2.18_all.deb
[jira] [Updated] (IMPALA-6852) Add a section to the query profile covering number of fragments instances and total bytes scanned per host
[ https://issues.apache.org/jira/browse/IMPALA-6852?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bikramjeet Vig updated IMPALA-6852: --- Summary: Add a section to the query profile covering number of fragments instances and total bytes scanned per host (was: Add a section to the query profile covering number of fragments and total bytes scanned per host) > Add a section to the query profile covering number of fragments instances and > total bytes scanned per host > -- > > Key: IMPALA-6852 > URL: https://issues.apache.org/jira/browse/IMPALA-6852 > Project: IMPALA > Issue Type: Improvement > Components: Backend >Affects Versions: Impala 2.13.0 >Reporter: Mostafa Mokhtar >Assignee: Alice Fan >Priority: Major > Labels: ramp-up, supportability > > Root causing performance issues usually comes down to skew in terms of work > done across hosts. Un-even assignment of fragments or bytes scanned per host > is a common. > Making this information readily available in the query profile will help > speedup RCA. > Proposal is to add two tables to the query profile that cover > * Number of fragments per host > * Number of bytes scanned per host -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org For additional commands, e-mail: issues-all-h...@impala.apache.org
[jira] [Assigned] (IMPALA-7957) UNION ALL query returns incorrect results
[ https://issues.apache.org/jira/browse/IMPALA-7957?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Paul Rogers reassigned IMPALA-7957: --- Assignee: Paul Rogers (was: Janaki Lahorani) > UNION ALL query returns incorrect results > - > > Key: IMPALA-7957 > URL: https://issues.apache.org/jira/browse/IMPALA-7957 > Project: IMPALA > Issue Type: Bug > Components: Frontend >Affects Versions: Impala 2.12.0 >Reporter: Luis E Martinez-Poblete >Assignee: Paul Rogers >Priority: Blocker > Labels: correctness > > Synopsis: > = > UNION ALL query returns incorrect results > Problem: > > Customer reported a UNION ALL query returning incorrect results. The UNION > ALL query has 2 legs, but Impala is only returning information from one leg. > Issue can be reproduced in the latest version of Impala. Below is the > reproduction case: > {noformat} > create table mytest_t (c1 timestamp, c2 timestamp, c3 int, c4 int); > insert into mytest_t values (now(), ADDDATE (now(),1), 1,1); > insert into mytest_t values (now(), ADDDATE (now(),1), 2,2); > insert into mytest_t values (now(), ADDDATE (now(),1), 3,3); > SELECT t.c1 > FROM > (SELECT c1, c2 > FROM mytest_t) t > LEFT JOIN > (SELECT c1, c2 > FROM mytest_t > WHERE c2 = c1) t2 ON (t.c2 = t2.c2) > UNION ALL > VALUES (NULL) > {noformat} > The above query produces the following execution plan: > {noformat} > ++ > | Explain String >| > ++ > | Max Per-Host Resource Reservation: Memory=34.02MB Threads=5 >| > | Per-Host Resource Estimates: Memory=2.06GB >| > | WARNING: The following tables are missing relevant table and/or column > statistics. | > | default.mytest_t >| > | >| > | PLAN-ROOT SINK >| > | | >| > | 06:EXCHANGE [UNPARTITIONED] >| > | | >| > | 00:UNION >| > | | constant-operands=1 >| > | | >| > | 04:SELECT >| > | | predicates: default.mytest_t.c1 = default.mytest_t.c2 >| > | | >| > | 03:HASH JOIN [LEFT OUTER JOIN, BROADCAST] >| > | | hash predicates: c2 = c2 >| > | | >| > | |--05:EXCHANGE [BROADCAST] >| > | | | >| > | | 02:SCAN HDFS [default.mytest_t] >| > | | partitions=1/1 files=3 size=192B >| > | | predicates: c2 = c1 >| > | | >| > | 01:SCAN HDFS [default.mytest_t] >| > |partitions=1/1 files=3 size=192B >| > ++ > {noformat} > The issue is in operator 4: > {noformat} > | 04:SELECT | > | | predicates: default.mytest_t.c1 = default.mytest_t.c2 | > {noformat} > It's definitely a bug with predicate placement - that c1 = c2 predicate > shouldn't be evaluated outside the right branch of the LEFT JOIN. > Thanks, > Luis Martinez. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org For additional commands, e-mail: issues-all-h...@impala.apache.org
[jira] [Assigned] (IMPALA-8064) test_min_max_filters is flaky
[ https://issues.apache.org/jira/browse/IMPALA-8064?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Paul Rogers reassigned IMPALA-8064: --- Assignee: Pooja Nilangekar (was: Janaki Lahorani) > test_min_max_filters is flaky > -- > > Key: IMPALA-8064 > URL: https://issues.apache.org/jira/browse/IMPALA-8064 > Project: IMPALA > Issue Type: Bug >Reporter: Pooja Nilangekar >Assignee: Pooja Nilangekar >Priority: Blocker > Labels: broken-build, flaky-test > Attachments: profile.txt > > > The following configuration of the test_min_max_filters: > {code:java} > query_test.test_runtime_filters.TestMinMaxFilters.test_min_max_filters[protocol: > beeswax | exec_option: {'batch_size': 0, 'num_nodes': 0, > 'disable_codegen_rows_threshold': 0, 'disable_codegen': False, > 'abort_on_error': 1, 'debug_action': None, 'exec_single_node_rows_threshold': > 0} | table_format: kudu/none]{code} > It produces a higher aggregation of sum over the proberows than expected: > {code:java} > query_test/test_runtime_filters.py:113: in test_min_max_filters > self.run_test_case('QueryTest/min_max_filters', vector) > common/impala_test_suite.py:518: in run_test_case > update_section=pytest.config.option.update_results) > common/test_result_verifier.py:612: in verify_runtime_profile % > (function, field, expected_value, actual_value, actual)) > E AssertionError: Aggregation of SUM over ProbeRows did not match expected > results. > E EXPECTED VALUE: E 619 > EACTUAL VALUE: E 652 > {code} > This test was introduced in the patch for IMPALA-6533. The failure occurred > during an ASAN build. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org For additional commands, e-mail: issues-all-h...@impala.apache.org
[jira] [Created] (IMPALA-8114) Build test failure in test_breakpad.py
Paul Rogers created IMPALA-8114: --- Summary: Build test failure in test_breakpad.py Key: IMPALA-8114 URL: https://issues.apache.org/jira/browse/IMPALA-8114 Project: IMPALA Issue Type: Bug Components: Infrastructure Affects Versions: Impala 3.1.0 Reporter: Paul Rogers Assignee: Tim Armstrong Recent builds have failed due to a failure in {{test_breakpad.py}}. Assigning to Tim as the person who most recently touched this file. Test output: {noformat} 09:04:35 ERRORS 09:04:35 ___ ERROR at teardown of TestBreakpadExhaustive.test_minidump_cleanup_thread ___ 09:04:35 custom_cluster/test_breakpad.py:49: in teardown_method 09:04:35 self.kill_cluster(SIGKILL) 09:04:35 custom_cluster/test_breakpad.py:80: in kill_cluster 09:04:35 self.kill_processes(processes, signal) 09:04:35 custom_cluster/test_breakpad.py:85: in kill_processes 09:04:35 process.kill(signal) 09:04:35 common/impala_cluster.py:330: in kill 09:04:35 assert 0, "No processes %s found" % self.cmd 09:04:35 E AssertionError: No processes ['/data/jenkins/workspace/impala-cdh6.x-exhaustive-release/repos/Impala/be/build/latest/service/impalad', '-kudu_client_rpc_timeout_ms', '0', '-kudu_master_hosts', 'localhost', '--mem_limit=12884901888', '-logbufsecs=5', '-v=1', '-max_log_files=0', '-log_filename=impalad', '-log_dir=/data/jenkins/workspace/impala-cdh6.x-exhaustive-release/repos/Impala/logs/custom_cluster_tests', '-beeswax_port=21000', '-hs2_port=21050', '-be_port=22000', '-krpc_port=27000', '-state_store_subscriber_port=23000', '-webserver_port=25000', '-max_minidumps=2', '-logbufsecs=1', '-minidump_path=/tmp/tmpKaSw_w', '--default_query_options='] found {noformat} Distilled {{TEST-impala-custom-cluster.xml}} output: {noformat} -- 2019-01-23 08:00:43,585 INFO MainThread: Found 3 impalad/1 statestored/1 catalogd process(es) … -- 2019-01-23 08:00:43,667 INFO MainThread: Killing: /data/jenkins/workspace/impala-cdh6.x-exhaustive-release/repos/Impala/be/build/latest/service/statestored -logbufsecs=5 -v=1 -max_log_files=0 -log_filename=statestored -log_dir=/data/jenkins/workspace/impala-cdh6.x-exhaustive-release/repos/Impala/logs/custom_cluster_tests -max_minidumps=2 -logbufsecs=1 -minidump_path=/tmp/tmpKaSw_w (PID: 16809) with signal 10 -- 2019-01-23 08:00:43,692 INFO MainThread: Found 6 impalad/1 statestored/1 catalogd process(es) ... E AssertionError: No processes [/data/jenkins/workspace/impala-cdh6.x-exhaustive-release/repos/Impala/be/build/latest/service/impalad {noformat} Notice that the main thread appaars to be killing statestore, but fails to kill impalad. Notice that a message appears that says that all impalads are running in the midst of the code that tries to shut down the cluster. Is this test multi-threaded? Is there more than one “main thread” Are these main threads working at cross purposes? What recent change may have caused this? Also, looks like the script is sending signal 10 (SIGUSR1) while the statestore (in its log) says it got a SIGTERM (15): {noformat} I0123 08:00:44.086009 16868 thrift-client.cc:78] Couldn't open transport for impala-ec2-centoCaught signal: SIGTERM. Daemon will exit. {noformat} Not terribly familiar with this area of the product, so bumping it over to the BE team. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org For additional commands, e-mail: issues-all-h...@impala.apache.org
[jira] [Created] (IMPALA-8114) Build test failure in test_breakpad.py
Paul Rogers created IMPALA-8114: --- Summary: Build test failure in test_breakpad.py Key: IMPALA-8114 URL: https://issues.apache.org/jira/browse/IMPALA-8114 Project: IMPALA Issue Type: Bug Components: Infrastructure Affects Versions: Impala 3.1.0 Reporter: Paul Rogers Assignee: Tim Armstrong Recent builds have failed due to a failure in {{test_breakpad.py}}. Assigning to Tim as the person who most recently touched this file. Test output: {noformat} 09:04:35 ERRORS 09:04:35 ___ ERROR at teardown of TestBreakpadExhaustive.test_minidump_cleanup_thread ___ 09:04:35 custom_cluster/test_breakpad.py:49: in teardown_method 09:04:35 self.kill_cluster(SIGKILL) 09:04:35 custom_cluster/test_breakpad.py:80: in kill_cluster 09:04:35 self.kill_processes(processes, signal) 09:04:35 custom_cluster/test_breakpad.py:85: in kill_processes 09:04:35 process.kill(signal) 09:04:35 common/impala_cluster.py:330: in kill 09:04:35 assert 0, "No processes %s found" % self.cmd 09:04:35 E AssertionError: No processes ['/data/jenkins/workspace/impala-cdh6.x-exhaustive-release/repos/Impala/be/build/latest/service/impalad', '-kudu_client_rpc_timeout_ms', '0', '-kudu_master_hosts', 'localhost', '--mem_limit=12884901888', '-logbufsecs=5', '-v=1', '-max_log_files=0', '-log_filename=impalad', '-log_dir=/data/jenkins/workspace/impala-cdh6.x-exhaustive-release/repos/Impala/logs/custom_cluster_tests', '-beeswax_port=21000', '-hs2_port=21050', '-be_port=22000', '-krpc_port=27000', '-state_store_subscriber_port=23000', '-webserver_port=25000', '-max_minidumps=2', '-logbufsecs=1', '-minidump_path=/tmp/tmpKaSw_w', '--default_query_options='] found {noformat} Distilled {{TEST-impala-custom-cluster.xml}} output: {noformat} -- 2019-01-23 08:00:43,585 INFO MainThread: Found 3 impalad/1 statestored/1 catalogd process(es) … -- 2019-01-23 08:00:43,667 INFO MainThread: Killing: /data/jenkins/workspace/impala-cdh6.x-exhaustive-release/repos/Impala/be/build/latest/service/statestored -logbufsecs=5 -v=1 -max_log_files=0 -log_filename=statestored -log_dir=/data/jenkins/workspace/impala-cdh6.x-exhaustive-release/repos/Impala/logs/custom_cluster_tests -max_minidumps=2 -logbufsecs=1 -minidump_path=/tmp/tmpKaSw_w (PID: 16809) with signal 10 -- 2019-01-23 08:00:43,692 INFO MainThread: Found 6 impalad/1 statestored/1 catalogd process(es) ... E AssertionError: No processes [/data/jenkins/workspace/impala-cdh6.x-exhaustive-release/repos/Impala/be/build/latest/service/impalad {noformat} Notice that the main thread appaars to be killing statestore, but fails to kill impalad. Notice that a message appears that says that all impalads are running in the midst of the code that tries to shut down the cluster. Is this test multi-threaded? Is there more than one “main thread” Are these main threads working at cross purposes? What recent change may have caused this? Also, looks like the script is sending signal 10 (SIGUSR1) while the statestore (in its log) says it got a SIGTERM (15): {noformat} I0123 08:00:44.086009 16868 thrift-client.cc:78] Couldn't open transport for impala-ec2-centoCaught signal: SIGTERM. Daemon will exit. {noformat} Not terribly familiar with this area of the product, so bumping it over to the BE team. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Created] (IMPALA-8113) test_aggregation and test_avro_primitive_in_list fail in S3
Michael Brown created IMPALA-8113: - Summary: test_aggregation and test_avro_primitive_in_list fail in S3 Key: IMPALA-8113 URL: https://issues.apache.org/jira/browse/IMPALA-8113 Project: IMPALA Issue Type: Bug Components: Infrastructure Affects Versions: Impala 3.2.0 Reporter: Michael Brown Assignee: Michael Brown Likely more victims of our infra in S3. {noformat} query_test/test_aggregation.py:138: in test_aggregation result = self.execute_query(query, vector.get_value('exec_option')) common/impala_test_suite.py:597: in wrapper return function(*args, **kwargs) common/impala_test_suite.py:628: in execute_query return self.__execute_query(self.client, query, query_options) common/impala_test_suite.py:695: in __execute_query return impalad_client.execute(query, user=user) common/impala_connection.py:174: in execute return self.__beeswax_client.execute(sql_stmt, user=user) beeswax/impala_beeswax.py:182: in execute handle = self.__execute_query(query_string.strip(), user=user) beeswax/impala_beeswax.py:359: in __execute_query self.wait_for_finished(handle) beeswax/impala_beeswax.py:380: in wait_for_finished raise ImpalaBeeswaxException("Query aborted:" + error_log, None) E ImpalaBeeswaxException: ImpalaBeeswaxException: EQuery aborted:Disk I/O error: Error reading from HDFS file: s3a://impala-test-uswest2-1/test-warehouse/alltypesagg_parquet/year=2010/month=1/day=8/5642b2da93dae1ad-494132e5_592013737_data.0.parq E Error(255): Unknown error 255 E Root cause: SdkClientException: Data read has a different length than the expected: dataLength=0; expectedLength=45494; includeSkipped=true; in.getClass()=class com.amazonaws.services.s3.AmazonS3Client$2; markedSupported=false; marked=0; resetSinceLastMarked=false; markCount=0; resetCount=0 {noformat} {noformat} query_test/test_nested_types.py:263: in test_avro_primitive_in_list "AvroPrimitiveInList.parquet", vector) query_test/test_nested_types.py:287: in __test_primitive_in_list result = self.execute_query("select item from %s.col1" % full_name, qopts) common/impala_test_suite.py:597: in wrapper return function(*args, **kwargs) common/impala_test_suite.py:628: in execute_query return self.__execute_query(self.client, query, query_options) common/impala_test_suite.py:695: in __execute_query return impalad_client.execute(query, user=user) common/impala_connection.py:174: in execute return self.__beeswax_client.execute(sql_stmt, user=user) beeswax/impala_beeswax.py:182: in execute handle = self.__execute_query(query_string.strip(), user=user) beeswax/impala_beeswax.py:359: in __execute_query self.wait_for_finished(handle) beeswax/impala_beeswax.py:380: in wait_for_finished raise ImpalaBeeswaxException("Query aborted:" + error_log, None) E ImpalaBeeswaxException: ImpalaBeeswaxException: EQuery aborted:Disk I/O error: Failed to open HDFS file s3a://impala-test-uswest2-1/test-warehouse/test_avro_primitive_in_list_38f182c4.db/AvroPrimitiveInList/AvroPrimitiveInList.parquet E Error(2): No such file or directory E Root cause: FileNotFoundException: No such file or directory: s3a://impala-test-uswest2-1/test-warehouse/test_avro_primitive_in_list_38f182c4.db/AvroPrimitiveInList/AvroPrimitiveInList.parquet {noformat} -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org For additional commands, e-mail: issues-all-h...@impala.apache.org
[jira] [Created] (IMPALA-8113) test_aggregation and test_avro_primitive_in_list fail in S3
Michael Brown created IMPALA-8113: - Summary: test_aggregation and test_avro_primitive_in_list fail in S3 Key: IMPALA-8113 URL: https://issues.apache.org/jira/browse/IMPALA-8113 Project: IMPALA Issue Type: Bug Components: Infrastructure Affects Versions: Impala 3.2.0 Reporter: Michael Brown Assignee: Michael Brown Likely more victims of our infra in S3. {noformat} query_test/test_aggregation.py:138: in test_aggregation result = self.execute_query(query, vector.get_value('exec_option')) common/impala_test_suite.py:597: in wrapper return function(*args, **kwargs) common/impala_test_suite.py:628: in execute_query return self.__execute_query(self.client, query, query_options) common/impala_test_suite.py:695: in __execute_query return impalad_client.execute(query, user=user) common/impala_connection.py:174: in execute return self.__beeswax_client.execute(sql_stmt, user=user) beeswax/impala_beeswax.py:182: in execute handle = self.__execute_query(query_string.strip(), user=user) beeswax/impala_beeswax.py:359: in __execute_query self.wait_for_finished(handle) beeswax/impala_beeswax.py:380: in wait_for_finished raise ImpalaBeeswaxException("Query aborted:" + error_log, None) E ImpalaBeeswaxException: ImpalaBeeswaxException: EQuery aborted:Disk I/O error: Error reading from HDFS file: s3a://impala-test-uswest2-1/test-warehouse/alltypesagg_parquet/year=2010/month=1/day=8/5642b2da93dae1ad-494132e5_592013737_data.0.parq E Error(255): Unknown error 255 E Root cause: SdkClientException: Data read has a different length than the expected: dataLength=0; expectedLength=45494; includeSkipped=true; in.getClass()=class com.amazonaws.services.s3.AmazonS3Client$2; markedSupported=false; marked=0; resetSinceLastMarked=false; markCount=0; resetCount=0 {noformat} {noformat} query_test/test_nested_types.py:263: in test_avro_primitive_in_list "AvroPrimitiveInList.parquet", vector) query_test/test_nested_types.py:287: in __test_primitive_in_list result = self.execute_query("select item from %s.col1" % full_name, qopts) common/impala_test_suite.py:597: in wrapper return function(*args, **kwargs) common/impala_test_suite.py:628: in execute_query return self.__execute_query(self.client, query, query_options) common/impala_test_suite.py:695: in __execute_query return impalad_client.execute(query, user=user) common/impala_connection.py:174: in execute return self.__beeswax_client.execute(sql_stmt, user=user) beeswax/impala_beeswax.py:182: in execute handle = self.__execute_query(query_string.strip(), user=user) beeswax/impala_beeswax.py:359: in __execute_query self.wait_for_finished(handle) beeswax/impala_beeswax.py:380: in wait_for_finished raise ImpalaBeeswaxException("Query aborted:" + error_log, None) E ImpalaBeeswaxException: ImpalaBeeswaxException: EQuery aborted:Disk I/O error: Failed to open HDFS file s3a://impala-test-uswest2-1/test-warehouse/test_avro_primitive_in_list_38f182c4.db/AvroPrimitiveInList/AvroPrimitiveInList.parquet E Error(2): No such file or directory E Root cause: FileNotFoundException: No such file or directory: s3a://impala-test-uswest2-1/test-warehouse/test_avro_primitive_in_list_38f182c4.db/AvroPrimitiveInList/AvroPrimitiveInList.parquet {noformat} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (IMPALA-8112) test_cancel_select with debug action failed with unexpected error
[ https://issues.apache.org/jira/browse/IMPALA-8112?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Michael Brown updated IMPALA-8112: -- Labels: flaky (was: ) > test_cancel_select with debug action failed with unexpected error > - > > Key: IMPALA-8112 > URL: https://issues.apache.org/jira/browse/IMPALA-8112 > Project: IMPALA > Issue Type: Bug > Components: Backend >Affects Versions: Impala 3.2.0 >Reporter: Michael Brown >Assignee: Andrew Sherman >Priority: Critical > Labels: flaky > > Stacktrace > {noformat} > query_test/test_cancellation.py:241: in test_cancel_select > self.execute_cancel_test(vector) > query_test/test_cancellation.py:213: in execute_cancel_test > assert 'Cancelled' in str(thread.fetch_results_error) > E assert 'Cancelled' in "ImpalaBeeswaxException:\n INNER EXCEPTION: 'beeswaxd.ttypes.BeeswaxException'>\n MESSAGE: Unable to open Kudu table: > Network error: recv error from 0.0.0.0:0: Transport endpoint is not connected > (error 107)\n" > E+ where "ImpalaBeeswaxException:\n INNER EXCEPTION: 'beeswaxd.ttypes.BeeswaxException'>\n MESSAGE: Unable to open Kudu table: > Network error: recv error from 0.0.0.0:0: Transport endpoint is not connected > (error 107)\n" = str(ImpalaBeeswaxException()) > E+where ImpalaBeeswaxException() = 140481071658752)>.fetch_results_error > {noformat} > Standard Error > {noformat} > SET > client_identifier=query_test/test_cancellation.py::TestCancellationParallel::()::test_cancel_select[protocol:beeswax|table_format:kudu/none|exec_option:{'batch_size':0;'num_nodes':0;'disable_codegen_rows_threshold':0;'disable_codegen':False;'abort_on_error':1;'debug_action; > -- executing against localhost:21000 > use tpch_kudu; > -- 2019-01-18 17:50:03,100 INFO MainThread: Started query > 4e4b3ab4cc7d:11efc3f5 > SET > client_identifier=query_test/test_cancellation.py::TestCancellationParallel::()::test_cancel_select[protocol:beeswax|table_format:kudu/none|exec_option:{'batch_size':0;'num_nodes':0;'disable_codegen_rows_threshold':0;'disable_codegen':False;'abort_on_error':1;'debug_action; > SET batch_size=0; > SET num_nodes=0; > SET disable_codegen_rows_threshold=0; > SET disable_codegen=False; > SET abort_on_error=1; > SET cpu_limit_s=10; > SET debug_action=0:GETNEXT:WAIT|COORD_CANCEL_QUERY_FINSTANCES_RPC:FAIL; > SET exec_single_node_rows_threshold=0; > SET buffer_pool_limit=0; > -- executing async: localhost:21000 > select l_returnflag from lineitem; > -- 2019-01-18 17:50:03,139 INFO MainThread: Started query > fa4ddb9e62a01240:54c86ad > SET > client_identifier=query_test/test_cancellation.py::TestCancellationParallel::()::test_cancel_select[protocol:beeswax|table_format:kudu/none|exec_option:{'batch_size':0;'num_nodes':0;'disable_codegen_rows_threshold':0;'disable_codegen':False;'abort_on_error':1;'debug_action; > -- connecting to: localhost:21000 > -- fetching results from: object at 0x6235e90> > -- getting state for operation: > > -- canceling operation: object at 0x6235e90> > -- 2019-01-18 17:50:08,196 INFO Thread-4: Starting new HTTP connection > (1): localhost > -- closing query for operation handle: > > {noformat} > [~asherman] please take a look since it looks like you touched code around > this area last. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org For additional commands, e-mail: issues-all-h...@impala.apache.org
[jira] [Issue Comment Deleted] (IMPALA-8112) test_cancel_select with debug action failed with unexpected error
[ https://issues.apache.org/jira/browse/IMPALA-8112?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Michael Brown updated IMPALA-8112: -- Comment: was deleted (was: This has happened once on a downstream CI job https://master-02.jenkins.cloudera.com/job/impala-cdh6.x-core/532/ .) > test_cancel_select with debug action failed with unexpected error > - > > Key: IMPALA-8112 > URL: https://issues.apache.org/jira/browse/IMPALA-8112 > Project: IMPALA > Issue Type: Bug > Components: Backend >Affects Versions: Impala 3.2.0 >Reporter: Michael Brown >Assignee: Andrew Sherman >Priority: Critical > Labels: flaky > > Stacktrace > {noformat} > query_test/test_cancellation.py:241: in test_cancel_select > self.execute_cancel_test(vector) > query_test/test_cancellation.py:213: in execute_cancel_test > assert 'Cancelled' in str(thread.fetch_results_error) > E assert 'Cancelled' in "ImpalaBeeswaxException:\n INNER EXCEPTION: 'beeswaxd.ttypes.BeeswaxException'>\n MESSAGE: Unable to open Kudu table: > Network error: recv error from 0.0.0.0:0: Transport endpoint is not connected > (error 107)\n" > E+ where "ImpalaBeeswaxException:\n INNER EXCEPTION: 'beeswaxd.ttypes.BeeswaxException'>\n MESSAGE: Unable to open Kudu table: > Network error: recv error from 0.0.0.0:0: Transport endpoint is not connected > (error 107)\n" = str(ImpalaBeeswaxException()) > E+where ImpalaBeeswaxException() = 140481071658752)>.fetch_results_error > {noformat} > Standard Error > {noformat} > SET > client_identifier=query_test/test_cancellation.py::TestCancellationParallel::()::test_cancel_select[protocol:beeswax|table_format:kudu/none|exec_option:{'batch_size':0;'num_nodes':0;'disable_codegen_rows_threshold':0;'disable_codegen':False;'abort_on_error':1;'debug_action; > -- executing against localhost:21000 > use tpch_kudu; > -- 2019-01-18 17:50:03,100 INFO MainThread: Started query > 4e4b3ab4cc7d:11efc3f5 > SET > client_identifier=query_test/test_cancellation.py::TestCancellationParallel::()::test_cancel_select[protocol:beeswax|table_format:kudu/none|exec_option:{'batch_size':0;'num_nodes':0;'disable_codegen_rows_threshold':0;'disable_codegen':False;'abort_on_error':1;'debug_action; > SET batch_size=0; > SET num_nodes=0; > SET disable_codegen_rows_threshold=0; > SET disable_codegen=False; > SET abort_on_error=1; > SET cpu_limit_s=10; > SET debug_action=0:GETNEXT:WAIT|COORD_CANCEL_QUERY_FINSTANCES_RPC:FAIL; > SET exec_single_node_rows_threshold=0; > SET buffer_pool_limit=0; > -- executing async: localhost:21000 > select l_returnflag from lineitem; > -- 2019-01-18 17:50:03,139 INFO MainThread: Started query > fa4ddb9e62a01240:54c86ad > SET > client_identifier=query_test/test_cancellation.py::TestCancellationParallel::()::test_cancel_select[protocol:beeswax|table_format:kudu/none|exec_option:{'batch_size':0;'num_nodes':0;'disable_codegen_rows_threshold':0;'disable_codegen':False;'abort_on_error':1;'debug_action; > -- connecting to: localhost:21000 > -- fetching results from: object at 0x6235e90> > -- getting state for operation: > > -- canceling operation: object at 0x6235e90> > -- 2019-01-18 17:50:08,196 INFO Thread-4: Starting new HTTP connection > (1): localhost > -- closing query for operation handle: > > {noformat} > [~asherman] please take a look since it looks like you touched code around > this area last. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org For additional commands, e-mail: issues-all-h...@impala.apache.org
[jira] [Commented] (IMPALA-8112) test_cancel_select with debug action failed with unexpected error
[ https://issues.apache.org/jira/browse/IMPALA-8112?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16751610#comment-16751610 ] Michael Brown commented on IMPALA-8112: --- This has happened once on a downstream CI job https://master-02.jenkins.cloudera.com/job/impala-cdh6.x-core/532/ . > test_cancel_select with debug action failed with unexpected error > - > > Key: IMPALA-8112 > URL: https://issues.apache.org/jira/browse/IMPALA-8112 > Project: IMPALA > Issue Type: Bug > Components: Backend >Affects Versions: Impala 3.2.0 >Reporter: Michael Brown >Assignee: Andrew Sherman >Priority: Critical > > Stacktrace > {noformat} > query_test/test_cancellation.py:241: in test_cancel_select > self.execute_cancel_test(vector) > query_test/test_cancellation.py:213: in execute_cancel_test > assert 'Cancelled' in str(thread.fetch_results_error) > E assert 'Cancelled' in "ImpalaBeeswaxException:\n INNER EXCEPTION: 'beeswaxd.ttypes.BeeswaxException'>\n MESSAGE: Unable to open Kudu table: > Network error: recv error from 0.0.0.0:0: Transport endpoint is not connected > (error 107)\n" > E+ where "ImpalaBeeswaxException:\n INNER EXCEPTION: 'beeswaxd.ttypes.BeeswaxException'>\n MESSAGE: Unable to open Kudu table: > Network error: recv error from 0.0.0.0:0: Transport endpoint is not connected > (error 107)\n" = str(ImpalaBeeswaxException()) > E+where ImpalaBeeswaxException() = 140481071658752)>.fetch_results_error > {noformat} > Standard Error > {noformat} > SET > client_identifier=query_test/test_cancellation.py::TestCancellationParallel::()::test_cancel_select[protocol:beeswax|table_format:kudu/none|exec_option:{'batch_size':0;'num_nodes':0;'disable_codegen_rows_threshold':0;'disable_codegen':False;'abort_on_error':1;'debug_action; > -- executing against localhost:21000 > use tpch_kudu; > -- 2019-01-18 17:50:03,100 INFO MainThread: Started query > 4e4b3ab4cc7d:11efc3f5 > SET > client_identifier=query_test/test_cancellation.py::TestCancellationParallel::()::test_cancel_select[protocol:beeswax|table_format:kudu/none|exec_option:{'batch_size':0;'num_nodes':0;'disable_codegen_rows_threshold':0;'disable_codegen':False;'abort_on_error':1;'debug_action; > SET batch_size=0; > SET num_nodes=0; > SET disable_codegen_rows_threshold=0; > SET disable_codegen=False; > SET abort_on_error=1; > SET cpu_limit_s=10; > SET debug_action=0:GETNEXT:WAIT|COORD_CANCEL_QUERY_FINSTANCES_RPC:FAIL; > SET exec_single_node_rows_threshold=0; > SET buffer_pool_limit=0; > -- executing async: localhost:21000 > select l_returnflag from lineitem; > -- 2019-01-18 17:50:03,139 INFO MainThread: Started query > fa4ddb9e62a01240:54c86ad > SET > client_identifier=query_test/test_cancellation.py::TestCancellationParallel::()::test_cancel_select[protocol:beeswax|table_format:kudu/none|exec_option:{'batch_size':0;'num_nodes':0;'disable_codegen_rows_threshold':0;'disable_codegen':False;'abort_on_error':1;'debug_action; > -- connecting to: localhost:21000 > -- fetching results from: object at 0x6235e90> > -- getting state for operation: > > -- canceling operation: object at 0x6235e90> > -- 2019-01-18 17:50:08,196 INFO Thread-4: Starting new HTTP connection > (1): localhost > -- closing query for operation handle: > > {noformat} > [~asherman] please take a look since it looks like you touched code around > this area last. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org For additional commands, e-mail: issues-all-h...@impala.apache.org
[jira] [Created] (IMPALA-8112) test_cancel_select with debug action failed with unexpected error
Michael Brown created IMPALA-8112: - Summary: test_cancel_select with debug action failed with unexpected error Key: IMPALA-8112 URL: https://issues.apache.org/jira/browse/IMPALA-8112 Project: IMPALA Issue Type: Bug Components: Backend Affects Versions: Impala 3.2.0 Reporter: Michael Brown Assignee: Andrew Sherman Stacktrace {noformat} query_test/test_cancellation.py:241: in test_cancel_select self.execute_cancel_test(vector) query_test/test_cancellation.py:213: in execute_cancel_test assert 'Cancelled' in str(thread.fetch_results_error) E assert 'Cancelled' in "ImpalaBeeswaxException:\n INNER EXCEPTION: \n MESSAGE: Unable to open Kudu table: Network error: recv error from 0.0.0.0:0: Transport endpoint is not connected (error 107)\n" E+ where "ImpalaBeeswaxException:\n INNER EXCEPTION: \n MESSAGE: Unable to open Kudu table: Network error: recv error from 0.0.0.0:0: Transport endpoint is not connected (error 107)\n" = str(ImpalaBeeswaxException()) E+where ImpalaBeeswaxException() = .fetch_results_error {noformat} Standard Error {noformat} SET client_identifier=query_test/test_cancellation.py::TestCancellationParallel::()::test_cancel_select[protocol:beeswax|table_format:kudu/none|exec_option:{'batch_size':0;'num_nodes':0;'disable_codegen_rows_threshold':0;'disable_codegen':False;'abort_on_error':1;'debug_action; -- executing against localhost:21000 use tpch_kudu; -- 2019-01-18 17:50:03,100 INFO MainThread: Started query 4e4b3ab4cc7d:11efc3f5 SET client_identifier=query_test/test_cancellation.py::TestCancellationParallel::()::test_cancel_select[protocol:beeswax|table_format:kudu/none|exec_option:{'batch_size':0;'num_nodes':0;'disable_codegen_rows_threshold':0;'disable_codegen':False;'abort_on_error':1;'debug_action; SET batch_size=0; SET num_nodes=0; SET disable_codegen_rows_threshold=0; SET disable_codegen=False; SET abort_on_error=1; SET cpu_limit_s=10; SET debug_action=0:GETNEXT:WAIT|COORD_CANCEL_QUERY_FINSTANCES_RPC:FAIL; SET exec_single_node_rows_threshold=0; SET buffer_pool_limit=0; -- executing async: localhost:21000 select l_returnflag from lineitem; -- 2019-01-18 17:50:03,139 INFO MainThread: Started query fa4ddb9e62a01240:54c86ad SET client_identifier=query_test/test_cancellation.py::TestCancellationParallel::()::test_cancel_select[protocol:beeswax|table_format:kudu/none|exec_option:{'batch_size':0;'num_nodes':0;'disable_codegen_rows_threshold':0;'disable_codegen':False;'abort_on_error':1;'debug_action; -- connecting to: localhost:21000 -- fetching results from: -- getting state for operation: -- canceling operation: -- 2019-01-18 17:50:08,196 INFO Thread-4: Starting new HTTP connection (1): localhost -- closing query for operation handle: {noformat} [~asherman] please take a look since it looks like you touched code around this area last. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Created] (IMPALA-8112) test_cancel_select with debug action failed with unexpected error
Michael Brown created IMPALA-8112: - Summary: test_cancel_select with debug action failed with unexpected error Key: IMPALA-8112 URL: https://issues.apache.org/jira/browse/IMPALA-8112 Project: IMPALA Issue Type: Bug Components: Backend Affects Versions: Impala 3.2.0 Reporter: Michael Brown Assignee: Andrew Sherman Stacktrace {noformat} query_test/test_cancellation.py:241: in test_cancel_select self.execute_cancel_test(vector) query_test/test_cancellation.py:213: in execute_cancel_test assert 'Cancelled' in str(thread.fetch_results_error) E assert 'Cancelled' in "ImpalaBeeswaxException:\n INNER EXCEPTION: \n MESSAGE: Unable to open Kudu table: Network error: recv error from 0.0.0.0:0: Transport endpoint is not connected (error 107)\n" E+ where "ImpalaBeeswaxException:\n INNER EXCEPTION: \n MESSAGE: Unable to open Kudu table: Network error: recv error from 0.0.0.0:0: Transport endpoint is not connected (error 107)\n" = str(ImpalaBeeswaxException()) E+where ImpalaBeeswaxException() = .fetch_results_error {noformat} Standard Error {noformat} SET client_identifier=query_test/test_cancellation.py::TestCancellationParallel::()::test_cancel_select[protocol:beeswax|table_format:kudu/none|exec_option:{'batch_size':0;'num_nodes':0;'disable_codegen_rows_threshold':0;'disable_codegen':False;'abort_on_error':1;'debug_action; -- executing against localhost:21000 use tpch_kudu; -- 2019-01-18 17:50:03,100 INFO MainThread: Started query 4e4b3ab4cc7d:11efc3f5 SET client_identifier=query_test/test_cancellation.py::TestCancellationParallel::()::test_cancel_select[protocol:beeswax|table_format:kudu/none|exec_option:{'batch_size':0;'num_nodes':0;'disable_codegen_rows_threshold':0;'disable_codegen':False;'abort_on_error':1;'debug_action; SET batch_size=0; SET num_nodes=0; SET disable_codegen_rows_threshold=0; SET disable_codegen=False; SET abort_on_error=1; SET cpu_limit_s=10; SET debug_action=0:GETNEXT:WAIT|COORD_CANCEL_QUERY_FINSTANCES_RPC:FAIL; SET exec_single_node_rows_threshold=0; SET buffer_pool_limit=0; -- executing async: localhost:21000 select l_returnflag from lineitem; -- 2019-01-18 17:50:03,139 INFO MainThread: Started query fa4ddb9e62a01240:54c86ad SET client_identifier=query_test/test_cancellation.py::TestCancellationParallel::()::test_cancel_select[protocol:beeswax|table_format:kudu/none|exec_option:{'batch_size':0;'num_nodes':0;'disable_codegen_rows_threshold':0;'disable_codegen':False;'abort_on_error':1;'debug_action; -- connecting to: localhost:21000 -- fetching results from: -- getting state for operation: -- canceling operation: -- 2019-01-18 17:50:08,196 INFO Thread-4: Starting new HTTP connection (1): localhost -- closing query for operation handle: {noformat} [~asherman] please take a look since it looks like you touched code around this area last. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org For additional commands, e-mail: issues-all-h...@impala.apache.org
[jira] [Commented] (IMPALA-7351) Add memory estimates for plan nodes and sinks with missing estimates
[ https://issues.apache.org/jira/browse/IMPALA-7351?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16751586#comment-16751586 ] Bikramjeet Vig commented on IMPALA-7351: [~tarmstrong] Only PlanRootSink is left and since IMPALA-4268 can change mem requirement drastically, I have deferred closing it till then. Also added that task as a dependency. > Add memory estimates for plan nodes and sinks with missing estimates > > > Key: IMPALA-7351 > URL: https://issues.apache.org/jira/browse/IMPALA-7351 > Project: IMPALA > Issue Type: Sub-task > Components: Frontend >Reporter: Tim Armstrong >Assignee: Bikramjeet Vig >Priority: Major > Labels: admission-control, resource-management > > Many plan nodes and sinks, e.g. KuduScanNode, KuduTableSink, ExchangeNode, > etc are missing memory estimates entirely. > We should add a basic estimate for all these cases based on experiments and > data from real workloads. In some cases 0 may be the right estimate (e.g. for > streaming nodes like SelectNode that just pass through data) but we should > remove TODOs and document the reasoning in those cases. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org For additional commands, e-mail: issues-all-h...@impala.apache.org
[jira] [Commented] (IMPALA-8111) Document workaround for some authentication issues with KRPC
[ https://issues.apache.org/jira/browse/IMPALA-8111?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16751567#comment-16751567 ] Alex Rodoni commented on IMPALA-8111: - [~kwho] I will add the 3 issues to the known issues doc. > Document workaround for some authentication issues with KRPC > > > Key: IMPALA-8111 > URL: https://issues.apache.org/jira/browse/IMPALA-8111 > Project: IMPALA > Issue Type: Task > Components: Docs >Affects Versions: Impala 2.12.0, Impala 3.1.0 >Reporter: Michael Ho >Assignee: Alex Rodoni >Priority: Major > Labels: future_release_doc, in_32 > > There have been complaints from users about not being able to use Impala > after upgrading to Impala version with KRPC enabled due to authentication > issues. Please document them in the known issues or best practice guide. > 1. https://issues.apache.org/jira/browse/IMPALA-7585: > *Symptoms*: When using Impala with LDAP enabled, a user may hit the > following: > {noformat} > Not authorized: Client connection negotiation failed: client connection to > 127.0.0.1:27000: SASL(-1): generic failure: All-whitespace username. > {noformat} > *Root cause*: The following sequence can lead to the user "impala" not being > created in /etc/passwd. > {quote}time 1: no impala in LDAP; things get installed; impala created in > /etc/passwd > time 2: impala added to LDAP > time 3: new machine added > {quote} > *Workaround*: > - Manually edit /etc/passwd to add the impala user > - Upgrade to a version of Impala with the patch IMPALA-7585 > 2. https://issues.apache.org/jira/browse/IMPALA-7298 > *Symptoms*: When running with Kerberos enabled, a user may hit the following > error: > {noformat} > WARNINGS: TransmitData() to X.X.X.X:27000 failed: Not authorized: Client > connection negotiation failed: client connection to X.X.X.X:27000: Server > impala/x.x@vpc.cloudera.com not found in Kerberos database > {noformat} > *Root cause*: > KrpcDataStreamSender passes a resolved IP address when creating a proxy. > Instead, we should pass both the resolved address and the hostname when > creating the proxy so that we won't end up using the IP address as the > hostname in the Kerberos principal. > *Workaround*: > - Set rdns=true in /etc/krb5.conf > - Upgrade to a version of Impala with the fix of IMPALA-7298 > 3. https://issues.apache.org/jira/browse/KUDU-2198 > *Symptoms*: When running with Kerberos enabled, a user may hit the following > error message where is some random string which doesn't match > the primary in the Kerberos principal > {noformat} > WARNINGS: TransmitData() to X.X.X.X:27000 failed: Remote error: Not > authorized: {username='', principal='impala/redacted'} is not > allowed to access DataStreamService > {noformat} > *Root cause*: > Due to system "auth_to_local" mapping, the principal may be mapped to some > local name. > *Workaround*: > - Start Impala with the flag {{--use_system_auth_to_local=false}} -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org For additional commands, e-mail: issues-all-h...@impala.apache.org
[jira] [Commented] (IMPALA-7351) Add memory estimates for plan nodes and sinks with missing estimates
[ https://issues.apache.org/jira/browse/IMPALA-7351?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16751495#comment-16751495 ] Tim Armstrong commented on IMPALA-7351: --- [~bikramjeet.vig] can we close this now? > Add memory estimates for plan nodes and sinks with missing estimates > > > Key: IMPALA-7351 > URL: https://issues.apache.org/jira/browse/IMPALA-7351 > Project: IMPALA > Issue Type: Sub-task > Components: Frontend >Reporter: Tim Armstrong >Assignee: Bikramjeet Vig >Priority: Major > Labels: admission-control, resource-management > > Many plan nodes and sinks, e.g. KuduScanNode, KuduTableSink, ExchangeNode, > etc are missing memory estimates entirely. > We should add a basic estimate for all these cases based on experiments and > data from real workloads. In some cases 0 may be the right estimate (e.g. for > streaming nodes like SelectNode that just pass through data) but we should > remove TODOs and document the reasoning in those cases. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org For additional commands, e-mail: issues-all-h...@impala.apache.org
[jira] [Updated] (IMPALA-8111) Document workaround for some authentication issues with KRPC
[ https://issues.apache.org/jira/browse/IMPALA-8111?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Alex Rodoni updated IMPALA-8111: Labels: future_release_doc in_32 (was: ) > Document workaround for some authentication issues with KRPC > > > Key: IMPALA-8111 > URL: https://issues.apache.org/jira/browse/IMPALA-8111 > Project: IMPALA > Issue Type: Task > Components: Docs >Affects Versions: Impala 2.12.0, Impala 3.1.0 >Reporter: Michael Ho >Assignee: Alex Rodoni >Priority: Major > Labels: future_release_doc, in_32 > > There have been complaints from users about not being able to use Impala > after upgrading to Impala version with KRPC enabled due to authentication > issues. Please document them in the known issues or best practice guide. > 1. https://issues.apache.org/jira/browse/IMPALA-7585: > *Symptoms*: When using Impala with LDAP enabled, a user may hit the > following: > {noformat} > Not authorized: Client connection negotiation failed: client connection to > 127.0.0.1:27000: SASL(-1): generic failure: All-whitespace username. > {noformat} > *Root cause*: The following sequence can lead to the user "impala" not being > created in /etc/passwd. > {quote}time 1: no impala in LDAP; things get installed; impala created in > /etc/passwd > time 2: impala added to LDAP > time 3: new machine added > {quote} > *Workaround*: > - Manually edit /etc/passwd to add the impala user > - Upgrade to a version of Impala with the patch IMPALA-7585 > 2. https://issues.apache.org/jira/browse/IMPALA-7298 > *Symptoms*: When running with Kerberos enabled, a user may hit the following > error: > {noformat} > WARNINGS: TransmitData() to X.X.X.X:27000 failed: Not authorized: Client > connection negotiation failed: client connection to X.X.X.X:27000: Server > impala/x.x@vpc.cloudera.com not found in Kerberos database > {noformat} > *Root cause*: > KrpcDataStreamSender passes a resolved IP address when creating a proxy. > Instead, we should pass both the resolved address and the hostname when > creating the proxy so that we won't end up using the IP address as the > hostname in the Kerberos principal. > *Workaround*: > - Set rdns=true in /etc/krb5.conf > - Upgrade to a version of Impala with the fix of IMPALA-7298 > 3. https://issues.apache.org/jira/browse/KUDU-2198 > *Symptoms*: When running with Kerberos enabled, a user may hit the following > error message where is some random string which doesn't match > the primary in the Kerberos principal > {noformat} > WARNINGS: TransmitData() to X.X.X.X:27000 failed: Remote error: Not > authorized: {username='', principal='impala/redacted'} is not > allowed to access DataStreamService > {noformat} > *Root cause*: > Due to system "auth_to_local" mapping, the principal may be mapped to some > local name. > *Workaround*: > - Start Impala with the flag {{--use_system_auth_to_local=false}} -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org For additional commands, e-mail: issues-all-h...@impala.apache.org
[jira] [Commented] (IMPALA-8109) Impala cannot read the gzip files bigger than 2 GB
[ https://issues.apache.org/jira/browse/IMPALA-8109?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16751462#comment-16751462 ] Tim Armstrong commented on IMPALA-8109: --- bytes_read_ used to be an int, but was switched to an int64_t by IMPALA-7543. I think that probably fixed this - I didn't see another way that the offset could overflow. [~boroknagyz] do you remember if you reproduced a bug like this when making that change or just did it as part of code cleanup? > Impala cannot read the gzip files bigger than 2 GB > -- > > Key: IMPALA-8109 > URL: https://issues.apache.org/jira/browse/IMPALA-8109 > Project: IMPALA > Issue Type: Bug > Components: Backend >Affects Versions: Impala 2.12.0 >Reporter: hakki >Priority: Minor > > When querying a partition containing gzip files, the query fails with the > error below: > WARNINGS: Disk I/O error: Error seeking to -2147483648 in file: > hdfs://HADOOP_CLUSTER/user/hive/AAA/BBB/datehour=20180910/XXX.gz: > Error(255): Unknown error 255 > Root cause: EOFException: Cannot seek to negative offset > hdfs://HADOOP_CLUSTER/user/hive/AAA/BBB/datehour=20180910/XXX.gz file is > a delimited text file and has a size of bigger than 2 GB (approx: 2.4 GB) The > uncompressed size is ~13GB > The impalad version is : 2.12.0-cdh5.15.0 -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org For additional commands, e-mail: issues-all-h...@impala.apache.org
[jira] [Created] (IMPALA-8111) Document workaround for some authentication issues with KRPC
Michael Ho created IMPALA-8111: -- Summary: Document workaround for some authentication issues with KRPC Key: IMPALA-8111 URL: https://issues.apache.org/jira/browse/IMPALA-8111 Project: IMPALA Issue Type: Task Components: Docs Affects Versions: Impala 3.1.0, Impala 2.12.0 Reporter: Michael Ho Assignee: Alex Rodoni There have been complaints from users about not being able to use Impala after upgrading to Impala version with KRPC enabled due to authentication issues. Please document them in the known issues or best practice guide. 1. https://issues.apache.org/jira/browse/IMPALA-7585: *Symptoms*: When using Impala with LDAP enabled, a user may hit the following: {noformat} Not authorized: Client connection negotiation failed: client connection to 127.0.0.1:27000: SASL(-1): generic failure: All-whitespace username. {noformat} *Root cause*: The following sequence can lead to the user "impala" not being created in /etc/passwd. {quote}time 1: no impala in LDAP; things get installed; impala created in /etc/passwd time 2: impala added to LDAP time 3: new machine added {quote} *Workaround*: - Manually edit /etc/passwd to add the impala user - Upgrade to a version of Impala with the patch IMPALA-7585 2. https://issues.apache.org/jira/browse/IMPALA-7298 *Symptoms*: When running with Kerberos enabled, a user may hit the following error: {noformat} WARNINGS: TransmitData() to X.X.X.X:27000 failed: Not authorized: Client connection negotiation failed: client connection to X.X.X.X:27000: Server impala/x.x@vpc.cloudera.com not found in Kerberos database {noformat} *Root cause*: KrpcDataStreamSender passes a resolved IP address when creating a proxy. Instead, we should pass both the resolved address and the hostname when creating the proxy so that we won't end up using the IP address as the hostname in the Kerberos principal. *Workaround*: - Set rdns=true in /etc/krb5.conf - Upgrade to a version of Impala with the fix of IMPALA-7298 3. https://issues.apache.org/jira/browse/KUDU-2198 *Symptoms*: When running with Kerberos enabled, a user may hit the following error message where is some random string which doesn't match the primary in the Kerberos principal {noformat} WARNINGS: TransmitData() to X.X.X.X:27000 failed: Remote error: Not authorized: {username='', principal='impala/redacted'} is not allowed to access DataStreamService {noformat} *Root cause*: Due to system "auth_to_local" mapping, the principal may be mapped to some local name. *Workaround*: - Start Impala with the flag {{--use_system_auth_to_local=false}} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Created] (IMPALA-8111) Document workaround for some authentication issues with KRPC
Michael Ho created IMPALA-8111: -- Summary: Document workaround for some authentication issues with KRPC Key: IMPALA-8111 URL: https://issues.apache.org/jira/browse/IMPALA-8111 Project: IMPALA Issue Type: Task Components: Docs Affects Versions: Impala 3.1.0, Impala 2.12.0 Reporter: Michael Ho Assignee: Alex Rodoni There have been complaints from users about not being able to use Impala after upgrading to Impala version with KRPC enabled due to authentication issues. Please document them in the known issues or best practice guide. 1. https://issues.apache.org/jira/browse/IMPALA-7585: *Symptoms*: When using Impala with LDAP enabled, a user may hit the following: {noformat} Not authorized: Client connection negotiation failed: client connection to 127.0.0.1:27000: SASL(-1): generic failure: All-whitespace username. {noformat} *Root cause*: The following sequence can lead to the user "impala" not being created in /etc/passwd. {quote}time 1: no impala in LDAP; things get installed; impala created in /etc/passwd time 2: impala added to LDAP time 3: new machine added {quote} *Workaround*: - Manually edit /etc/passwd to add the impala user - Upgrade to a version of Impala with the patch IMPALA-7585 2. https://issues.apache.org/jira/browse/IMPALA-7298 *Symptoms*: When running with Kerberos enabled, a user may hit the following error: {noformat} WARNINGS: TransmitData() to X.X.X.X:27000 failed: Not authorized: Client connection negotiation failed: client connection to X.X.X.X:27000: Server impala/x.x@vpc.cloudera.com not found in Kerberos database {noformat} *Root cause*: KrpcDataStreamSender passes a resolved IP address when creating a proxy. Instead, we should pass both the resolved address and the hostname when creating the proxy so that we won't end up using the IP address as the hostname in the Kerberos principal. *Workaround*: - Set rdns=true in /etc/krb5.conf - Upgrade to a version of Impala with the fix of IMPALA-7298 3. https://issues.apache.org/jira/browse/KUDU-2198 *Symptoms*: When running with Kerberos enabled, a user may hit the following error message where is some random string which doesn't match the primary in the Kerberos principal {noformat} WARNINGS: TransmitData() to X.X.X.X:27000 failed: Remote error: Not authorized: {username='', principal='impala/redacted'} is not allowed to access DataStreamService {noformat} *Root cause*: Due to system "auth_to_local" mapping, the principal may be mapped to some local name. *Workaround*: - Start Impala with the flag {{--use_system_auth_to_local=false}} -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org For additional commands, e-mail: issues-all-h...@impala.apache.org
[jira] [Commented] (IMPALA-8107) Support EXEC_TIME_LIMIT_S in resource pool setting
[ https://issues.apache.org/jira/browse/IMPALA-8107?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16751393#comment-16751393 ] Tim Armstrong commented on IMPALA-8107: --- You can specify default query options for any resource pool - https://impala.apache.org/docs/build/html/topics/impala_admission.html That's generally how I'd expect this to be used. > Support EXEC_TIME_LIMIT_S in resource pool setting > -- > > Key: IMPALA-8107 > URL: https://issues.apache.org/jira/browse/IMPALA-8107 > Project: IMPALA > Issue Type: New Feature >Reporter: Quanlong Huang >Priority: Major > Labels: admission-control > > Timeout limit should be different for different kinds of queries. For > example, resource pool for adhoc queries may set EXEC_TIME_LIMIT_S to 60s. > Resource pool for building pre-aggregaions or other ETL may need a larger > EXEC_TIME_LIMIT_S like 30 minutes. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org For additional commands, e-mail: issues-all-h...@impala.apache.org
[jira] [Commented] (IMPALA-7832) Support IF NOT EXISTS in alter table add columns
[ https://issues.apache.org/jira/browse/IMPALA-7832?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16751384#comment-16751384 ] ASF subversion and git services commented on IMPALA-7832: - Commit bfb9ccc8e02be20fb8b57bae4d55e4094ab7ea3f in impala's branch refs/heads/master from Fredy Wijaya [ https://gitbox.apache.org/repos/asf?p=impala.git;h=bfb9ccc ] IMPALA-7832: Support for IF NOT EXISTS in ALTER TABLE ADD COLUMN(S) This patch adds IF NOT EXISTS support in ALTER TABLE ADD COLUMN and ALTER TABLE ADD COLUMNS. If IF NOT EXISTS is specified and a column already exists with this name, no error is thrown. If IF NOT EXISTS is specified for multiple columns and a column already exists, no error is thrown and a new column that does not exist will be added. Syntax: ALTER TABLE tbl ADD COLUMN [IF NOT EXISTS] i int ALTER TABLE tbl ADD [IF NOT EXISTS] COLUMNS (i int, j int) Testing: - Added new FE tests - Ran all FE tests - Updated E2E DDL tests - Ran all E2E DDL tests Change-Id: I60ed22c8a8eefa10e94ad3dedf32fe67c16642d9 Reviewed-on: http://gerrit.cloudera.org:8080/12181 Reviewed-by: Impala Public Jenkins Tested-by: Impala Public Jenkins > Support IF NOT EXISTS in alter table add columns > > > Key: IMPALA-7832 > URL: https://issues.apache.org/jira/browse/IMPALA-7832 > Project: IMPALA > Issue Type: New Feature > Components: Frontend >Affects Versions: Impala 3.1.0 >Reporter: Thomas Tauber-Marshall >Assignee: Fredy Wijaya >Priority: Minor > Labels: ramp-up > Fix For: Impala 3.2.0 > > > alter table add [if not exists] columns ( [, > ...]) > would add the column only if a column of the same name does not already exist > Probably worth checking out what other databases do in different situations, > eg. if the column already exists but with a different type, if "replace" is > used instead of "add", etc. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org For additional commands, e-mail: issues-all-h...@impala.apache.org
[jira] [Commented] (IMPALA-7905) ToSqlUtils does not correctly quote lower-case Hive keywords
[ https://issues.apache.org/jira/browse/IMPALA-7905?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16751382#comment-16751382 ] ASF subversion and git services commented on IMPALA-7905: - Commit 85a8b34645a46038fd217c03e64326b72d9669b5 in impala's branch refs/heads/master from Paul Rogers [ https://gitbox.apache.org/repos/asf?p=impala.git;h=85a8b34 ] IMPALA-7905: Hive keywords not quoted for identifiers Impala often generates SQL for statements using the toSql() call. Generated SQL is often used during testing or when writing the query plan. Impala keywords such as "create", when used as identifiers, must be quoted: SELECT `select`, `from` FROM `order` ... The code in ToSqlUtils.getIdentSql() quotes the identifier if it is an Impala or Hive keyword, or if it does not follow the identifier pattern. The code uses the Hive lexer to detect a keyword. But, the code contained a flaw: the lexer expects a case-insensitive input. We provide a case sensitive input. As a result, "MONTH" is caught as a Hive keyword and quoted, but "month" is not. This patch fixes that flaw. This patch also fixes: IMPALA-8051: Compute stats fails on a column with comment character in name The code uses the Hive lexical analyzer to check names. Since "#" and "--" are comment characters, a name like "foo#" is parsed as "foo" which does not need quotes, hence we don't quote "foo#", which causes issues. Added a special check for "#" and "--" to resolve this issue. Testing: * Refactored getIdentSql() easier testing. * Added a tests to the recently added ToSqlUtilsTest for this case and several others. * Making this change caused the columns `month`, `year`, and `key` to be quoted when before they were not. Updated many tests as a result. * Added a new identSql() function, for use in tests, to match the quoting that Impala uses, and to handle the wildcard, and multi-part names. Used this in ToSqlTest to handle the quoted names. * PlannerTest emits statement SQL to the output file wrapped to 80 columns and sometimes leaves trailing spaces at the end of the line. Some tools remove that trailing space, resulting in trivial file differences. Fixed this to remove trailing spaces in order to simplify file comparisons. * Tweaked the "In pipelines" output to avoid trailing spaces when no pipelines are listed. * Reran all FE tests. Change-Id: I06cc20b052a3a66535a171c36b4b31477c0ba6d0 Reviewed-on: http://gerrit.cloudera.org:8080/12009 Reviewed-by: Impala Public Jenkins Tested-by: Impala Public Jenkins > ToSqlUtils does not correctly quote lower-case Hive keywords > > > Key: IMPALA-7905 > URL: https://issues.apache.org/jira/browse/IMPALA-7905 > Project: IMPALA > Issue Type: Bug > Components: Frontend >Affects Versions: Impala 3.0 >Reporter: Paul Rogers >Assignee: Paul Rogers >Priority: Minor > > Not sure yet how to reproduce this error via the shell, but here is the code > analysis. > The {{ToSqlUtils}} class generates a {{CREATE TABLE}} statement, which uses a > method {{getIdentSql()}} to possibly quote a table or column name. This same > method is used in multiple places in the {{toSql()}} logic for various > statements. > The comment for the method says: > bq. returns an identifier lexable by Impala and Hive, possibly by enclosing > the original identifier in "`" quotes. > To check for a Hive-compatible identifier, the code uses the Hive lexer: > {code:java} > HiveLexer hiveLexer = new HiveLexer(new ANTLRStringStream(ident)); > {code} > A unit test shows that this logic fails to catch lower case keywords: > "select", say, while it does catch upper-case keywords: "SELECT". > Checking the Hive source, it appears we're using the lexer wrong: > {code:java} > HiveLexerX lexer = new HiveLexerX(new ANTLRNoCaseStringStream(command)); > {code} > The fix is simple: upper-case the symbol before using he Hive lexer. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org For additional commands, e-mail: issues-all-h...@impala.apache.org
[jira] [Commented] (IMPALA-7698) Add centos/redhat 6/7 support to bootstrap_system.sh
[ https://issues.apache.org/jira/browse/IMPALA-7698?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16751381#comment-16751381 ] ASF subversion and git services commented on IMPALA-7698: - Commit 81d0bcb3c967bf76b6858b221416e6fcb863b187 in impala's branch refs/heads/master from Philip Zeyliger [ https://gitbox.apache.org/repos/asf?p=impala.git;h=81d0bcb ] Support centos:7 for test-with-docker. As a follow-on to IMPALA-7698, adds various incantations so that centos:7 can build under test-with-docker. The core issue is that the centos:7 image doesn't let you start sshd (necessary for the HBase startup scripts, and probably could be worked around) or postgresql (harder to work around) with systemctl, because systemd isn't "running." To avoid this, we start them manually with /usr/sbin/sshd and pg_ctl. Change-Id: I7577949b6eaaa2239bcf0fadf64e1490c2106b08 Reviewed-on: http://gerrit.cloudera.org:8080/12139 Reviewed-by: Impala Public Jenkins Tested-by: Impala Public Jenkins > Add centos/redhat 6/7 support to bootstrap_system.sh > > > Key: IMPALA-7698 > URL: https://issues.apache.org/jira/browse/IMPALA-7698 > Project: IMPALA > Issue Type: Task > Components: Infrastructure >Reporter: Philip Zeyliger >Assignee: Philip Zeyliger >Priority: Major > > {{bootstrap_system.sh}} currently only works on Ubuntu. Making it work on > CentOS/Redhat would open the door to running automated tests on those > platforms more readily, including using {{test-with-docker}}. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org For additional commands, e-mail: issues-all-h...@impala.apache.org
[jira] [Commented] (IMPALA-8051) Compute stats fails on a column with comment character in name
[ https://issues.apache.org/jira/browse/IMPALA-8051?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16751383#comment-16751383 ] ASF subversion and git services commented on IMPALA-8051: - Commit 85a8b34645a46038fd217c03e64326b72d9669b5 in impala's branch refs/heads/master from Paul Rogers [ https://gitbox.apache.org/repos/asf?p=impala.git;h=85a8b34 ] IMPALA-7905: Hive keywords not quoted for identifiers Impala often generates SQL for statements using the toSql() call. Generated SQL is often used during testing or when writing the query plan. Impala keywords such as "create", when used as identifiers, must be quoted: SELECT `select`, `from` FROM `order` ... The code in ToSqlUtils.getIdentSql() quotes the identifier if it is an Impala or Hive keyword, or if it does not follow the identifier pattern. The code uses the Hive lexer to detect a keyword. But, the code contained a flaw: the lexer expects a case-insensitive input. We provide a case sensitive input. As a result, "MONTH" is caught as a Hive keyword and quoted, but "month" is not. This patch fixes that flaw. This patch also fixes: IMPALA-8051: Compute stats fails on a column with comment character in name The code uses the Hive lexical analyzer to check names. Since "#" and "--" are comment characters, a name like "foo#" is parsed as "foo" which does not need quotes, hence we don't quote "foo#", which causes issues. Added a special check for "#" and "--" to resolve this issue. Testing: * Refactored getIdentSql() easier testing. * Added a tests to the recently added ToSqlUtilsTest for this case and several others. * Making this change caused the columns `month`, `year`, and `key` to be quoted when before they were not. Updated many tests as a result. * Added a new identSql() function, for use in tests, to match the quoting that Impala uses, and to handle the wildcard, and multi-part names. Used this in ToSqlTest to handle the quoted names. * PlannerTest emits statement SQL to the output file wrapped to 80 columns and sometimes leaves trailing spaces at the end of the line. Some tools remove that trailing space, resulting in trivial file differences. Fixed this to remove trailing spaces in order to simplify file comparisons. * Tweaked the "In pipelines" output to avoid trailing spaces when no pipelines are listed. * Reran all FE tests. Change-Id: I06cc20b052a3a66535a171c36b4b31477c0ba6d0 Reviewed-on: http://gerrit.cloudera.org:8080/12009 Reviewed-by: Impala Public Jenkins Tested-by: Impala Public Jenkins > Compute stats fails on a column with comment character in name > -- > > Key: IMPALA-8051 > URL: https://issues.apache.org/jira/browse/IMPALA-8051 > Project: IMPALA > Issue Type: Bug > Components: Frontend >Affects Versions: Impala 3.1.0 >Reporter: Paul Rogers >Assignee: Paul Rogers >Priority: Minor > > Problem - "compute stats" query executed on a table containing a special > character "#" in one of its columns is failing with below error: > WARNINGS: AnalysisException: Syntax error in line 1: > ...length(cola)), NDV(colb#) AS colb#, CAST(-1 as BIG... > ^ > Encountered: Unexpected character > Expected: ADD, ALTER, AND, ARRAY, AS, ASC, BETWEEN, BIGINT, BINARY, > BLOCK_SIZE, BOOLEAN, CACHED, CASCADE, CHANGE, CHAR, COMMENT, COMPRESSION, > CROSS, DATE, DATETIME, DECIMAL, DEFAULT, DESC, DIV, REAL, DROP, ELSE, > ENCODING, END, FLOAT, FOLLOWING, FROM, FULL, GROUP, IGNORE, HAVING, ILIKE, > IN, INNER, INTEGER, IREGEXP, IS, JOIN, LEFT, LIKE, LIMIT, LOCATION, MAP, NOT, > NULL, NULLS, OFFSET, ON, OR, ORDER, PARTITION, PARTITIONED, PRECEDING, > PRIMARY, PURGE, RANGE, RECOVER, REGEXP, RENAME, REPLACE, RESTRICT, RIGHT, > RLIKE, ROW, ROWS, SELECT, SET, SMALLINT, SORT, STORED, STRAIGHT_JOIN, STRING, > STRUCT, TABLESAMPLE, TBLPROPERTIES, THEN, TIMESTAMP, TINYINT, TO, UNCACHED, > UNION, USING, VALUES, VARCHAR, WHEN, WHERE, WITH, COMMA, IDENTIFIER > Steps to reproduce the issue - > # Create a table containing special character in one of it columns from Hive. > For example: > {code:sql} > CREATE TABLE test_special_character (`id#` int); > {code} > # Execute "INVALIDATE METADATA test_special_character" from Impala. > # Execute "COMPUTE STATS test_special_character" from Impala and it'll lead > to above mentioned error. > Impala does not allow to create tables with columns containing special > characters but Hive allows it by using back ticks (``) to escape it. However, > Impala still can load the metadata of table and can read from column > containing special character as well by escaping the special character using > back ticks (``). For example, below query can be executed from Impala - > {code:sql} > select `id#` from test_special_character; >
[jira] [Resolved] (IMPALA-7832) Support IF NOT EXISTS in alter table add columns
[ https://issues.apache.org/jira/browse/IMPALA-7832?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Fredy Wijaya resolved IMPALA-7832. -- Resolution: Fixed Fix Version/s: Impala 3.2.0 > Support IF NOT EXISTS in alter table add columns > > > Key: IMPALA-7832 > URL: https://issues.apache.org/jira/browse/IMPALA-7832 > Project: IMPALA > Issue Type: New Feature > Components: Frontend >Affects Versions: Impala 3.1.0 >Reporter: Thomas Tauber-Marshall >Assignee: Fredy Wijaya >Priority: Minor > Labels: ramp-up > Fix For: Impala 3.2.0 > > > alter table add [if not exists] columns ( [, > ...]) > would add the column only if a column of the same name does not already exist > Probably worth checking out what other databases do in different situations, > eg. if the column already exists but with a different type, if "replace" is > used instead of "add", etc. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Resolved] (IMPALA-7832) Support IF NOT EXISTS in alter table add columns
[ https://issues.apache.org/jira/browse/IMPALA-7832?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Fredy Wijaya resolved IMPALA-7832. -- Resolution: Fixed Fix Version/s: Impala 3.2.0 > Support IF NOT EXISTS in alter table add columns > > > Key: IMPALA-7832 > URL: https://issues.apache.org/jira/browse/IMPALA-7832 > Project: IMPALA > Issue Type: New Feature > Components: Frontend >Affects Versions: Impala 3.1.0 >Reporter: Thomas Tauber-Marshall >Assignee: Fredy Wijaya >Priority: Minor > Labels: ramp-up > Fix For: Impala 3.2.0 > > > alter table add [if not exists] columns ( [, > ...]) > would add the column only if a column of the same name does not already exist > Probably worth checking out what other databases do in different situations, > eg. if the column already exists but with a different type, if "replace" is > used instead of "add", etc. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org For additional commands, e-mail: issues-all-h...@impala.apache.org
[jira] [Created] (IMPALA-8110) Parquet stat filtering does not handle narrowed int types correctly
Csaba Ringhofer created IMPALA-8110: --- Summary: Parquet stat filtering does not handle narrowed int types correctly Key: IMPALA-8110 URL: https://issues.apache.org/jira/browse/IMPALA-8110 Project: IMPALA Issue Type: Improvement Components: Backend Reporter: Csaba Ringhofer Impala can read int32 Parquet columns as tiny/smallint SQL columns. If the value does not fit into the 8/16 bit signed int's range, the value will overflow, e.g writing 128 as int32 and then rereading it as int8 will return -128. This is normal as far as I understand, but min/max stat filtering does not handle this case correctly: create table tnarrow (i int) stored as parquet; insert into tnarrow values (1), (201); alter table tnarrow change column i i tinyint; set PARQUET_READ_STATISTICS=0; select * from tnarrow where i < 0; -> returns 1 row: -56 set PARQUET_READ_STATISTICS=1; -> returns 0 row -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org For additional commands, e-mail: issues-all-h...@impala.apache.org
[jira] [Created] (IMPALA-8110) Parquet stat filtering does not handle narrowed int types correctly
Csaba Ringhofer created IMPALA-8110: --- Summary: Parquet stat filtering does not handle narrowed int types correctly Key: IMPALA-8110 URL: https://issues.apache.org/jira/browse/IMPALA-8110 Project: IMPALA Issue Type: Improvement Components: Backend Reporter: Csaba Ringhofer Impala can read int32 Parquet columns as tiny/smallint SQL columns. If the value does not fit into the 8/16 bit signed int's range, the value will overflow, e.g writing 128 as int32 and then rereading it as int8 will return -128. This is normal as far as I understand, but min/max stat filtering does not handle this case correctly: create table tnarrow (i int) stored as parquet; insert into tnarrow values (1), (201); alter table tnarrow change column i i tinyint; set PARQUET_READ_STATISTICS=0; select * from tnarrow where i < 0; -> returns 1 row: -56 set PARQUET_READ_STATISTICS=1; -> returns 0 row -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (IMPALA-8109) Impala cannot read the gzip files bigger than 2 GB
[ https://issues.apache.org/jira/browse/IMPALA-8109?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] hakki updated IMPALA-8109: -- Description: When querying a partition containing gzip files, the query fails with the error below: WARNINGS: Disk I/O error: Error seeking to -2147483648 in file: hdfs://HADOOP_CLUSTER/user/hive/AAA/BBB/datehour=20180910/XXX.gz: Error(255): Unknown error 255 Root cause: EOFException: Cannot seek to negative offset hdfs://HADOOP_CLUSTER/user/hive/AAA/BBB/datehour=20180910/XXX.gz file is a delimited text file and has a size of bigger than 2 GB (approx: 2.4 GB) The uncompressed size is ~13GB The impalad version is : 2.12.0-cdh5.15.0 was: When querying a partition containing gzip files, the query fails with the error below: WARNINGS: Disk I/O error: Error seeking to -2147483648 in file: hdfs://HADOOP_CLUSTER/user/hive/AAA/BBB/datehour=20180910/XXX.gz: Error(255): Unknown error 255 Root cause: EOFException: Cannot seek to negative offset hdfs://HADOOP_CLUSTER/user/hive/AAA/BBB/datehour=20180910/XXX.gz file has a size of bigger than 2 GB (approx: 2.4 GB) The exact version is : 2.12.0-cdh5.15.0 > Impala cannot read the gzip files bigger than 2 GB > -- > > Key: IMPALA-8109 > URL: https://issues.apache.org/jira/browse/IMPALA-8109 > Project: IMPALA > Issue Type: Bug > Components: Backend >Affects Versions: Impala 2.12.0 >Reporter: hakki >Priority: Minor > > When querying a partition containing gzip files, the query fails with the > error below: > WARNINGS: Disk I/O error: Error seeking to -2147483648 in file: > hdfs://HADOOP_CLUSTER/user/hive/AAA/BBB/datehour=20180910/XXX.gz: > Error(255): Unknown error 255 > Root cause: EOFException: Cannot seek to negative offset > hdfs://HADOOP_CLUSTER/user/hive/AAA/BBB/datehour=20180910/XXX.gz file is > a delimited text file and has a size of bigger than 2 GB (approx: 2.4 GB) The > uncompressed size is ~13GB > The impalad version is : 2.12.0-cdh5.15.0 -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org For additional commands, e-mail: issues-all-h...@impala.apache.org
[jira] [Updated] (IMPALA-8109) Impala cannot read the gzip files bigger than 2 GB
[ https://issues.apache.org/jira/browse/IMPALA-8109?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] hakki updated IMPALA-8109: -- Description: When querying a partition containing gzip files, the query fails with the error below: WARNINGS: Disk I/O error: Error seeking to -2147483648 in file: hdfs://HADOOP_CLUSTER/user/hive/AAA/BBB/datehour=20180910/XXX.gz: Error(255): Unknown error 255 Root cause: EOFException: Cannot seek to negative offset hdfs://HADOOP_CLUSTER/user/hive/AAA/BBB/datehour=20180910/XXX.gz file has a size of bigger than 2 GB (approx: 2.4 GB) The exact version is : 2.12.0-cdh5.15.0 was: When querying a partition containing gzip files, the query fails with the error below: WARNINGS: Disk I/O error: Error seeking to -2147483648 in file: hdfs://HADOOP_CLUSTER/user/hive/AAA/BBB/datehour=20180910/XXX.gz: Error(255): Unknown error 255 Root cause: EOFException: Cannot seek to negative offset hdfs://HADOOP_CLUSTER/user/hive/AAA/BBB/datehour=20180910/XXX.gz file has a size of bigger than 2 GB (approx: 2.4 GB) > Impala cannot read the gzip files bigger than 2 GB > -- > > Key: IMPALA-8109 > URL: https://issues.apache.org/jira/browse/IMPALA-8109 > Project: IMPALA > Issue Type: Bug > Components: Backend >Affects Versions: Impala 2.12.0 >Reporter: hakki >Priority: Minor > > When querying a partition containing gzip files, the query fails with the > error below: > WARNINGS: Disk I/O error: Error seeking to -2147483648 in file: > hdfs://HADOOP_CLUSTER/user/hive/AAA/BBB/datehour=20180910/XXX.gz: > Error(255): Unknown error 255 > Root cause: EOFException: Cannot seek to negative offset > hdfs://HADOOP_CLUSTER/user/hive/AAA/BBB/datehour=20180910/XXX.gz file has > a size of bigger than 2 GB (approx: 2.4 GB) > The exact version is : 2.12.0-cdh5.15.0 -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org For additional commands, e-mail: issues-all-h...@impala.apache.org
[jira] [Created] (IMPALA-8109) Impala cannot read the gzip files bigger than 2 GB
hakki created IMPALA-8109: - Summary: Impala cannot read the gzip files bigger than 2 GB Key: IMPALA-8109 URL: https://issues.apache.org/jira/browse/IMPALA-8109 Project: IMPALA Issue Type: Bug Components: Backend Affects Versions: Impala 2.12.0 Reporter: hakki When querying a partition containing gzip files, the query fails with the error below: WARNINGS: Disk I/O error: Error seeking to -2147483648 in file: hdfs://HADOOP_CLUSTER/user/hive/AAA/BBB/datehour=20180910/XXX.gz: Error(255): Unknown error 255 Root cause: EOFException: Cannot seek to negative offset hdfs://HADOOP_CLUSTER/user/hive/AAA/BBB/datehour=20180910/XXX.gz file has a size of bigger than 2 GB (approx: 2.4 GB) -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Created] (IMPALA-8109) Impala cannot read the gzip files bigger than 2 GB
hakki created IMPALA-8109: - Summary: Impala cannot read the gzip files bigger than 2 GB Key: IMPALA-8109 URL: https://issues.apache.org/jira/browse/IMPALA-8109 Project: IMPALA Issue Type: Bug Components: Backend Affects Versions: Impala 2.12.0 Reporter: hakki When querying a partition containing gzip files, the query fails with the error below: WARNINGS: Disk I/O error: Error seeking to -2147483648 in file: hdfs://HADOOP_CLUSTER/user/hive/AAA/BBB/datehour=20180910/XXX.gz: Error(255): Unknown error 255 Root cause: EOFException: Cannot seek to negative offset hdfs://HADOOP_CLUSTER/user/hive/AAA/BBB/datehour=20180910/XXX.gz file has a size of bigger than 2 GB (approx: 2.4 GB) -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org For additional commands, e-mail: issues-all-h...@impala.apache.org
[jira] [Created] (IMPALA-8108) Impala query returns TIMESTAMP values in different types
Robbie Zhang created IMPALA-8108: Summary: Impala query returns TIMESTAMP values in different types Key: IMPALA-8108 URL: https://issues.apache.org/jira/browse/IMPALA-8108 Project: IMPALA Issue Type: Improvement Components: Backend Reporter: Robbie Zhang When a timestamp has a .000 or .00 or .0 (when fraction value is zeros) the timestamp is displayed with no fraction of second. For example: {code:java} select cast(ts as timestamp) from (values ('2019-01-11 10:40:18' as ts), ('2019-01-11 10:40:19.0'), ('2019-01-11 10:40:19.00'), ('2019-01-11 10:40:19.000'), ('2019-01-11 10:40:19.'), ('2019-01-11 10:40:19.0'), ('2019-01-11 10:40:19.00'), ('2019-01-11 10:40:19.000'), ('2019-01-11 10:40:19.'), ('2019-01-11 10:40:19.0'), ('2019-01-11 10:40:19.1') ) t;{code} The output is: {code:java} +---+ |cast(ts as timestamp)| +---+ |2019-01-11 10:40:18| |2019-01-11 10:40:19| |2019-01-11 10:40:19| |2019-01-11 10:40:19| |2019-01-11 10:40:19| |2019-01-11 10:40:19| |2019-01-11 10:40:19| |2019-01-11 10:40:19| |2019-01-11 10:40:19| |2019-01-11 10:40:19| |2019-01-11 10:40:19.1| +---+ {code} As we can see, values of the same column are returned in two different types. The inconsistency breaks some downstream use cases. The reason is that impala uses function boost::posix_time::to_simple_string(time_duration) to convert timestamp to a string and to_simple_string() remove fractional seconds if they are all zeros. Perhaps we can append ".0" if the length of the string is 8 (HH:MM:SS). For now we can work around it by using function from_timestamp(ts, '-mm-dd hh:mm.ss.s') to unify the output (convert to string), or using function millisecond(ts) to get fractional seconds. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org For additional commands, e-mail: issues-all-h...@impala.apache.org
[jira] [Created] (IMPALA-8108) Impala query returns TIMESTAMP values in different types
Robbie Zhang created IMPALA-8108: Summary: Impala query returns TIMESTAMP values in different types Key: IMPALA-8108 URL: https://issues.apache.org/jira/browse/IMPALA-8108 Project: IMPALA Issue Type: Improvement Components: Backend Reporter: Robbie Zhang When a timestamp has a .000 or .00 or .0 (when fraction value is zeros) the timestamp is displayed with no fraction of second. For example: {code:java} select cast(ts as timestamp) from (values ('2019-01-11 10:40:18' as ts), ('2019-01-11 10:40:19.0'), ('2019-01-11 10:40:19.00'), ('2019-01-11 10:40:19.000'), ('2019-01-11 10:40:19.'), ('2019-01-11 10:40:19.0'), ('2019-01-11 10:40:19.00'), ('2019-01-11 10:40:19.000'), ('2019-01-11 10:40:19.'), ('2019-01-11 10:40:19.0'), ('2019-01-11 10:40:19.1') ) t;{code} The output is: {code:java} +---+ |cast(ts as timestamp)| +---+ |2019-01-11 10:40:18| |2019-01-11 10:40:19| |2019-01-11 10:40:19| |2019-01-11 10:40:19| |2019-01-11 10:40:19| |2019-01-11 10:40:19| |2019-01-11 10:40:19| |2019-01-11 10:40:19| |2019-01-11 10:40:19| |2019-01-11 10:40:19| |2019-01-11 10:40:19.1| +---+ {code} As we can see, values of the same column are returned in two different types. The inconsistency breaks some downstream use cases. The reason is that impala uses function boost::posix_time::to_simple_string(time_duration) to convert timestamp to a string and to_simple_string() remove fractional seconds if they are all zeros. Perhaps we can append ".0" if the length of the string is 8 (HH:MM:SS). For now we can work around it by using function from_timestamp(ts, '-mm-dd hh:mm.ss.s') to unify the output (convert to string), or using function millisecond(ts) to get fractional seconds. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Created] (IMPALA-8107) Support EXEC_TIME_LIMIT_S in resource pool setting
Quanlong Huang created IMPALA-8107: -- Summary: Support EXEC_TIME_LIMIT_S in resource pool setting Key: IMPALA-8107 URL: https://issues.apache.org/jira/browse/IMPALA-8107 Project: IMPALA Issue Type: New Feature Reporter: Quanlong Huang Timeout limit should be different for different kinds of queries. For example, resource pool for adhoc queries may set EXEC_TIME_LIMIT_S to 60s. Resource pool for building pre-aggregaions or other ETL may need a larger EXEC_TIME_LIMIT_S like 30 minutes. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org For additional commands, e-mail: issues-all-h...@impala.apache.org
[jira] [Created] (IMPALA-8107) Support EXEC_TIME_LIMIT_S in resource pool setting
Quanlong Huang created IMPALA-8107: -- Summary: Support EXEC_TIME_LIMIT_S in resource pool setting Key: IMPALA-8107 URL: https://issues.apache.org/jira/browse/IMPALA-8107 Project: IMPALA Issue Type: New Feature Reporter: Quanlong Huang Timeout limit should be different for different kinds of queries. For example, resource pool for adhoc queries may set EXEC_TIME_LIMIT_S to 60s. Resource pool for building pre-aggregaions or other ETL may need a larger EXEC_TIME_LIMIT_S like 30 minutes. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Work started] (IMPALA-4018) Add support for SQL:2016 datetime templates/patterns/masks to CAST(... AS ... FORMAT )
[ https://issues.apache.org/jira/browse/IMPALA-4018?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Work on IMPALA-4018 started by Gabor Kaszab. > Add support for SQL:2016 datetime templates/patterns/masks to CAST(... AS ... > FORMAT ) > > > Key: IMPALA-4018 > URL: https://issues.apache.org/jira/browse/IMPALA-4018 > Project: IMPALA > Issue Type: New Feature > Components: Frontend >Affects Versions: Impala 2.2.4 >Reporter: Greg Rahn >Assignee: Gabor Kaszab >Priority: Critical > Labels: ansi-sql, compatibility, sql-language > > *Summary* > The format masks/templates for currently are implemented using the [Java > SimpleDateFormat > patterns|http://docs.oracle.com/javase/8/docs/api/java/text/SimpleDateFormat.html], > and although this is what Hive has implemented, it is not what most standard > SQL systems implement. For example see > [Vertica|https://my.vertica.com/docs/7.2.x/HTML/Content/Authoring/SQLReferenceManual/Functions/Formatting/TemplatePatternsForDateTimeFormatting.htm], > > [Netezza|http://www.ibm.com/support/knowledgecenter/SSULQD_7.2.1/com.ibm.nz.dbu.doc/r_dbuser_ntz_sql_extns_templ_patterns_date_time_conv.html], > > [Oracle|https://docs.oracle.com/database/121/SQLRF/sql_elements004.htm#SQLRF00212], > and > [PostgreSQL|https://www.postgresql.org/docs/9.5/static/functions-formatting.html#FUNCTIONS-FORMATTING-DATETIME-TABLE]. > > *Examples of incompatibilities* > {noformat} > -- PostgreSQL/Netezza/Vertica/Oracle > select to_timestamp('May 15, 2015 12:00:00', 'mon dd, hh:mi:ss'); > -- Impala > select to_timestamp('May 15, 2015 12:00:00', 'MMM dd, HH:mm:ss'); > -- PostgreSQL/Netezza/Vertica/Oracle > select to_timestamp('2015-02-14 20:19:07','-mm-dd hh24:mi:ss'); > -- Impala > select to_timestamp('2015-02-14 20:19:07','-MM-dd HH:mm:ss'); > -- Vertica/Oracle > select to_timestamp('2015-02-14 20:19:07.123456','-mm-dd hh24:mi:ss.ff'); > -- Impala > select to_timestamp('2015-02-14 20:19:07.123456','-MM-dd > HH:mm:ss.SS'); > {noformat} > *Considerations* > Because this is a change in default behavior for to_timestamp(), if possible, > having a feature flag to revert to the legacy Java SimpleDateFormat patterns > should be strongly considered. This would allow users to chose the behavior > they desire and scope it to a session if need be. > SQL:2016 defines the following datetime templates > {noformat} > ::= > { }... > ::= > > | > ::= > > | > | > | > | > | > | > | > | > | > | > | > | > | > ::= > > | > | > | > | > | > | > | > ::= > | YYY | YY | Y > ::= > | RR > ::= > MM > ::= > DD > ::= > DDD > ::= > HH | HH12 > ::= > HH24 > ::= > MI > ::= > SS > ::= > S > ::= > FF1 | FF2 | FF3 | FF4 | FF5 | FF6 | FF7 | FF8 | FF9 > ::= > A.M. | P.M. > ::= > TZH > ::= > TZM > {noformat} > SQL:2016 also introduced the FORMAT clause for CAST which is the standard way > to do string <> datetime conversions > {noformat} > ::= > CAST >AS > [ FORMAT ] > > ::= > > | > ::= > > | > ::= > > {noformat} > For example: > {noformat} > CAST( AS [FORMAT ]) > CAST( AS [FORMAT ]) > cast(dt as string format 'DD-MM-') > cast('01-05-2017' as date format 'DD-MM-') > {noformat} -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org For additional commands, e-mail: issues-all-h...@impala.apache.org