See <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/542/changes>
Changes:
[pauly] HIVE-1818 Call frequency and duration metrics for HiveMetaStore via jmx
(Sushanth Sowmyan via pauly)
[jvs] HIVE-1970. Modify build to run all tests regardless of subproject
failures.
(Carl Steinbach via jvs)
------------------------------------------
[...truncated 22603 lines...]
[junit] POSTHOOK: query: drop table testhivedrivertable
[junit] POSTHOOK: type: DROPTABLE
[junit] OK
[junit] PREHOOK: query: create table testhivedrivertable (num int)
[junit] PREHOOK: type: CREATETABLE
[junit] POSTHOOK: query: create table testhivedrivertable (num int)
[junit] POSTHOOK: type: CREATETABLE
[junit] POSTHOOK: Output: default@testhivedrivertable
[junit] OK
[junit] PREHOOK: query: drop table testhivedrivertable
[junit] PREHOOK: type: DROPTABLE
[junit] PREHOOK: Input: default@testhivedrivertable
[junit] PREHOOK: Output: default@testhivedrivertable
[junit] POSTHOOK: query: drop table testhivedrivertable
[junit] POSTHOOK: type: DROPTABLE
[junit] POSTHOOK: Input: default@testhivedrivertable
[junit] POSTHOOK: Output: default@testhivedrivertable
[junit] OK
[junit] Hive history
file=<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/service/tmp/hive_job_log_hudson_201102081943_1314495100.txt>
[junit] Hive history
file=<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/service/tmp/hive_job_log_hudson_201102081943_474408860.txt>
[junit] PREHOOK: query: drop table testhivedrivertable
[junit] PREHOOK: type: DROPTABLE
[junit] POSTHOOK: query: drop table testhivedrivertable
[junit] POSTHOOK: type: DROPTABLE
[junit] OK
[junit] PREHOOK: query: create table testhivedrivertable (key int, value
string)
[junit] PREHOOK: type: CREATETABLE
[junit] POSTHOOK: query: create table testhivedrivertable (key int, value
string)
[junit] POSTHOOK: type: CREATETABLE
[junit] POSTHOOK: Output: default@testhivedrivertable
[junit] OK
[junit] PREHOOK: query: load data local inpath
'<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt'>
into table testhivedrivertable
[junit] PREHOOK: type: LOAD
[junit] Copying data from
<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
[junit] Loading data to table testhivedrivertable
[junit] POSTHOOK: query: load data local inpath
'<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt'>
into table testhivedrivertable
[junit] POSTHOOK: type: LOAD
[junit] POSTHOOK: Output: default@testhivedrivertable
[junit] OK
[junit] PREHOOK: query: select key, value from testhivedrivertable
[junit] PREHOOK: type: QUERY
[junit] PREHOOK: Input: default@testhivedrivertable
[junit] PREHOOK: Output:
file:/tmp/hudson/hive_2011-02-08_19-43-39_257_7044498286683889035/-mr-10000
[junit] Total MapReduce jobs = 1
[junit] Launching Job 1 out of 1
[junit] Number of reduce tasks is set to 0 since there's no reduce operator
[junit] Job running in-process (local Hadoop)
[junit] 2011-02-08 19:43:41,823 null map = 100%, reduce = 0%
[junit] Ended Job = job_local_0001
[junit] POSTHOOK: query: select key, value from testhivedrivertable
[junit] POSTHOOK: type: QUERY
[junit] POSTHOOK: Input: default@testhivedrivertable
[junit] POSTHOOK: Output:
file:/tmp/hudson/hive_2011-02-08_19-43-39_257_7044498286683889035/-mr-10000
[junit] OK
[junit] PREHOOK: query: select key, value from testhivedrivertable
[junit] PREHOOK: type: QUERY
[junit] PREHOOK: Input: default@testhivedrivertable
[junit] PREHOOK: Output:
file:/tmp/hudson/hive_2011-02-08_19-43-41_974_4178845283739002672/-mr-10000
[junit] Total MapReduce jobs = 1
[junit] Launching Job 1 out of 1
[junit] Number of reduce tasks is set to 0 since there's no reduce operator
[junit] Job running in-process (local Hadoop)
[junit] 2011-02-08 19:43:44,534 null map = 100%, reduce = 0%
[junit] Ended Job = job_local_0001
[junit] POSTHOOK: query: select key, value from testhivedrivertable
[junit] POSTHOOK: type: QUERY
[junit] POSTHOOK: Input: default@testhivedrivertable
[junit] POSTHOOK: Output:
file:/tmp/hudson/hive_2011-02-08_19-43-41_974_4178845283739002672/-mr-10000
[junit] OK
[junit] PREHOOK: query: select key, value from testhivedrivertable
[junit] PREHOOK: type: QUERY
[junit] PREHOOK: Input: default@testhivedrivertable
[junit] PREHOOK: Output:
file:/tmp/hudson/hive_2011-02-08_19-43-44_715_3499711993732994623/-mr-10000
[junit] Total MapReduce jobs = 1
[junit] Launching Job 1 out of 1
[junit] Number of reduce tasks is set to 0 since there's no reduce operator
[junit] Job running in-process (local Hadoop)
[junit] 2011-02-08 19:43:47,282 null map = 100%, reduce = 0%
[junit] Ended Job = job_local_0001
[junit] POSTHOOK: query: select key, value from testhivedrivertable
[junit] POSTHOOK: type: QUERY
[junit] POSTHOOK: Input: default@testhivedrivertable
[junit] POSTHOOK: Output:
file:/tmp/hudson/hive_2011-02-08_19-43-44_715_3499711993732994623/-mr-10000
[junit] OK
[junit] Hive history
file=<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/service/tmp/hive_job_log_hudson_201102081943_1130149402.txt>
[junit] PREHOOK: query: drop table testhivedrivertable
[junit] PREHOOK: type: DROPTABLE
[junit] PREHOOK: Input: default@testhivedrivertable
[junit] PREHOOK: Output: default@testhivedrivertable
[junit] POSTHOOK: query: drop table testhivedrivertable
[junit] POSTHOOK: type: DROPTABLE
[junit] POSTHOOK: Input: default@testhivedrivertable
[junit] POSTHOOK: Output: default@testhivedrivertable
[junit] OK
[junit] PREHOOK: query: create table testhivedrivertable (key int, value
string)
[junit] PREHOOK: type: CREATETABLE
[junit] POSTHOOK: query: create table testhivedrivertable (key int, value
string)
[junit] POSTHOOK: type: CREATETABLE
[junit] POSTHOOK: Output: default@testhivedrivertable
[junit] OK
[junit] PREHOOK: query: load data local inpath
'<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt'>
into table testhivedrivertable
[junit] PREHOOK: type: LOAD
[junit] Copying data from
<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt>
[junit] Loading data to table testhivedrivertable
[junit] POSTHOOK: query: load data local inpath
'<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt'>
into table testhivedrivertable
[junit] POSTHOOK: type: LOAD
[junit] POSTHOOK: Output: default@testhivedrivertable
[junit] OK
[junit] PREHOOK: query: select key, value from testhivedrivertable where
key > 10
[junit] PREHOOK: type: QUERY
[junit] PREHOOK: Input: default@testhivedrivertable
[junit] PREHOOK: Output:
file:/tmp/hudson/hive_2011-02-08_19-43-47_983_4465070353321866234/-mr-10000
[junit] Total MapReduce jobs = 1
[junit] Launching Job 1 out of 1
[junit] Number of reduce tasks is set to 0 since there's no reduce operator
[junit] Job running in-process (local Hadoop)
[junit] 2011-02-08 19:43:50,658 null map = 100%, reduce = 0%
[junit] Ended Job = job_local_0001
[junit] POSTHOOK: query: select key, value from testhivedrivertable where
key > 10
[junit] POSTHOOK: type: QUERY
[junit] POSTHOOK: Input: default@testhivedrivertable
[junit] POSTHOOK: Output:
file:/tmp/hudson/hive_2011-02-08_19-43-47_983_4465070353321866234/-mr-10000
[junit] OK
[junit] PREHOOK: query: select count(1) as c from testhivedrivertable
[junit] PREHOOK: type: QUERY
[junit] PREHOOK: Input: default@testhivedrivertable
[junit] PREHOOK: Output:
file:/tmp/hudson/hive_2011-02-08_19-43-50_858_3703174997344666152/-mr-10000
[junit] Total MapReduce jobs = 1
[junit] Launching Job 1 out of 1
[junit] Number of reduce tasks determined at compile time: 1
[junit] In order to change the average load for a reducer (in bytes):
[junit] set hive.exec.reducers.bytes.per.reducer=<number>
[junit] In order to limit the maximum number of reducers:
[junit] set hive.exec.reducers.max=<number>
[junit] In order to set a constant number of reducers:
[junit] set mapred.reduce.tasks=<number>
[junit] Job running in-process (local Hadoop)
[junit] 2011-02-08 19:43:53,566 null map = 100%, reduce = 100%
[junit] Ended Job = job_local_0001
[junit] POSTHOOK: query: select count(1) as c from testhivedrivertable
[junit] POSTHOOK: type: QUERY
[junit] POSTHOOK: Input: default@testhivedrivertable
[junit] POSTHOOK: Output:
file:/tmp/hudson/hive_2011-02-08_19-43-50_858_3703174997344666152/-mr-10000
[junit] OK
[junit] ------------- ---------------- ---------------
[junit]
[junit] Testcase: testExecute took 9.148 sec
[junit] Testcase: testNonHiveCommand took 0.77 sec
[junit] Testcase: testMetastore took 0.249 sec
[junit] Testcase: testGetClusterStatus took 0.089 sec
[junit] Testcase: testFetch took 8.661 sec
[junit] Testcase: testDynamicSerde took 6.298 sec
test-conditions:
gen-test:
create-dirs:
compile-ant-tasks:
create-dirs:
init:
compile:
[echo] Compiling: anttasks
[javac]
<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/ant/build.xml>:40:
warning: 'includeantruntime' was not set, defaulting to
build.sysclasspath=last; set to false for repeatable builds
deploy-ant-tasks:
create-dirs:
init:
compile:
[echo] Compiling: anttasks
[javac]
<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/ant/build.xml>:40:
warning: 'includeantruntime' was not set, defaulting to
build.sysclasspath=last; set to false for repeatable builds
jar:
init:
compile:
ivy-init-dirs:
ivy-download:
[get] Getting:
http://repo2.maven.org/maven2/org/apache/ivy/ivy/2.1.0/ivy-2.1.0.jar
[get] To:
<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ivy/lib/ivy-2.1.0.jar>
[get] Not modified - so not downloaded
ivy-probe-antlib:
ivy-init-antlib:
ivy-init:
ivy-retrieve-hadoop-source:
[ivy:retrieve] :: Ivy 2.1.0 - 20090925235825 :: http://ant.apache.org/ivy/ ::
[ivy:retrieve] :: loading settings :: file =
<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/ivy/ivysettings.xml>
[ivy:retrieve] :: resolving dependencies ::
org.apache.hadoop.hive#shims;working@minerva
[ivy:retrieve] confs: [default]
[ivy:retrieve] found hadoop#core;0.20.0 in hadoop-source
[ivy:retrieve] found hadoop#core;0.20.3-CDH3-SNAPSHOT in hadoop-source
[ivy:retrieve] :: resolution report :: resolve 2765ms :: artifacts dl 2ms
---------------------------------------------------------------------
| | modules || artifacts |
| conf | number| search|dwnlded|evicted|| number|dwnlded|
---------------------------------------------------------------------
| default | 2 | 0 | 0 | 0 || 2 | 0 |
---------------------------------------------------------------------
[ivy:retrieve] :: retrieving :: org.apache.hadoop.hive#shims
[ivy:retrieve] confs: [default]
[ivy:retrieve] 0 artifacts copied, 2 already retrieved (0kB/3ms)
install-hadoopcore-internal:
build_shims:
[echo] Compiling shims against hadoop 0.20.0
(<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/hadoopcore/hadoop-0.20.0)>
[javac]
<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/shims/build.xml>:53:
warning: 'includeantruntime' was not set, defaulting to
build.sysclasspath=last; set to false for repeatable builds
ivy-init-dirs:
ivy-download:
[get] Getting:
http://repo2.maven.org/maven2/org/apache/ivy/ivy/2.1.0/ivy-2.1.0.jar
[get] To:
<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ivy/lib/ivy-2.1.0.jar>
[get] Not modified - so not downloaded
ivy-probe-antlib:
ivy-init-antlib:
ivy-init:
ivy-retrieve-hadoop-source:
[ivy:retrieve] :: Ivy 2.1.0 - 20090925235825 :: http://ant.apache.org/ivy/ ::
[ivy:retrieve] :: loading settings :: file =
<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/ivy/ivysettings.xml>
[for]
<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/shims/build.xml>:
The following error occurred while executing this line:
[for]
<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build.xml>:213:
The following error occurred while executing this line:
[for]
<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/shims/build.xml>:65:
The following error occurred while executing this line:
[for]
<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build-common.xml>:189:
java.lang.OutOfMemoryError: PermGen space
BUILD FAILED
<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build.xml>:208:
Keepgoing execution: 4 of 11 iterations failed.
Total time: 170 minutes 13 seconds
Archiving artifacts
Recording test results