See <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/559/>
------------------------------------------ [...truncated 25836 lines...] [junit] POSTHOOK: Output: default@testhivedrivertable [junit] OK [junit] PREHOOK: query: drop table testhivedrivertable [junit] PREHOOK: type: DROPTABLE [junit] PREHOOK: Input: default@testhivedrivertable [junit] PREHOOK: Output: default@testhivedrivertable [junit] POSTHOOK: query: drop table testhivedrivertable [junit] POSTHOOK: type: DROPTABLE [junit] POSTHOOK: Input: default@testhivedrivertable [junit] POSTHOOK: Output: default@testhivedrivertable [junit] OK [junit] Hive history file=<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/service/tmp/hive_job_log_hudson_201102151612_801321682.txt> [junit] Hive history file=<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/service/tmp/hive_job_log_hudson_201102151612_395095848.txt> [junit] PREHOOK: query: drop table testhivedrivertable [junit] PREHOOK: type: DROPTABLE [junit] POSTHOOK: query: drop table testhivedrivertable [junit] POSTHOOK: type: DROPTABLE [junit] OK [junit] PREHOOK: query: create table testhivedrivertable (key int, value string) [junit] PREHOOK: type: CREATETABLE [junit] POSTHOOK: query: create table testhivedrivertable (key int, value string) [junit] POSTHOOK: type: CREATETABLE [junit] POSTHOOK: Output: default@testhivedrivertable [junit] OK [junit] PREHOOK: query: load data local inpath '<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt'> into table testhivedrivertable [junit] PREHOOK: type: LOAD [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt> [junit] Loading data to table testhivedrivertable [junit] POSTHOOK: query: load data local inpath '<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt'> into table testhivedrivertable [junit] POSTHOOK: type: LOAD [junit] POSTHOOK: Output: default@testhivedrivertable [junit] OK [junit] PREHOOK: query: select key, value from testhivedrivertable [junit] PREHOOK: type: QUERY [junit] PREHOOK: Input: default@testhivedrivertable [junit] PREHOOK: Output: file:/tmp/hudson/hive_2011-02-15_16-12-10_689_182555332694834192/-mr-10000 [junit] Total MapReduce jobs = 1 [junit] Launching Job 1 out of 1 [junit] Number of reduce tasks is set to 0 since there's no reduce operator [junit] Job running in-process (local Hadoop) [junit] Hadoop job information for null: number of mappers: 0; number of reducers: 0 [junit] 2011-02-15 16:12:13,254 null map = 100%, reduce = 0% [junit] Ended Job = job_local_0001 [junit] POSTHOOK: query: select key, value from testhivedrivertable [junit] POSTHOOK: type: QUERY [junit] POSTHOOK: Input: default@testhivedrivertable [junit] POSTHOOK: Output: file:/tmp/hudson/hive_2011-02-15_16-12-10_689_182555332694834192/-mr-10000 [junit] OK [junit] PREHOOK: query: select key, value from testhivedrivertable [junit] PREHOOK: type: QUERY [junit] PREHOOK: Input: default@testhivedrivertable [junit] PREHOOK: Output: file:/tmp/hudson/hive_2011-02-15_16-12-13_411_5648131512347137248/-mr-10000 [junit] Total MapReduce jobs = 1 [junit] Launching Job 1 out of 1 [junit] Number of reduce tasks is set to 0 since there's no reduce operator [junit] Job running in-process (local Hadoop) [junit] Hadoop job information for null: number of mappers: 0; number of reducers: 0 [junit] 2011-02-15 16:12:15,956 null map = 100%, reduce = 0% [junit] Ended Job = job_local_0001 [junit] POSTHOOK: query: select key, value from testhivedrivertable [junit] POSTHOOK: type: QUERY [junit] POSTHOOK: Input: default@testhivedrivertable [junit] POSTHOOK: Output: file:/tmp/hudson/hive_2011-02-15_16-12-13_411_5648131512347137248/-mr-10000 [junit] OK [junit] PREHOOK: query: select key, value from testhivedrivertable [junit] PREHOOK: type: QUERY [junit] PREHOOK: Input: default@testhivedrivertable [junit] PREHOOK: Output: file:/tmp/hudson/hive_2011-02-15_16-12-16_122_5975070769629814430/-mr-10000 [junit] Total MapReduce jobs = 1 [junit] Launching Job 1 out of 1 [junit] Number of reduce tasks is set to 0 since there's no reduce operator [junit] Job running in-process (local Hadoop) [junit] Hadoop job information for null: number of mappers: 0; number of reducers: 0 [junit] 2011-02-15 16:12:18,674 null map = 100%, reduce = 0% [junit] Ended Job = job_local_0001 [junit] POSTHOOK: query: select key, value from testhivedrivertable [junit] POSTHOOK: type: QUERY [junit] POSTHOOK: Input: default@testhivedrivertable [junit] POSTHOOK: Output: file:/tmp/hudson/hive_2011-02-15_16-12-16_122_5975070769629814430/-mr-10000 [junit] OK [junit] Hive history file=<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/service/tmp/hive_job_log_hudson_201102151612_1197890842.txt> [junit] PREHOOK: query: drop table testhivedrivertable [junit] PREHOOK: type: DROPTABLE [junit] PREHOOK: Input: default@testhivedrivertable [junit] PREHOOK: Output: default@testhivedrivertable [junit] POSTHOOK: query: drop table testhivedrivertable [junit] POSTHOOK: type: DROPTABLE [junit] POSTHOOK: Input: default@testhivedrivertable [junit] POSTHOOK: Output: default@testhivedrivertable [junit] OK [junit] PREHOOK: query: create table testhivedrivertable (key int, value string) [junit] PREHOOK: type: CREATETABLE [junit] POSTHOOK: query: create table testhivedrivertable (key int, value string) [junit] POSTHOOK: type: CREATETABLE [junit] POSTHOOK: Output: default@testhivedrivertable [junit] OK [junit] PREHOOK: query: load data local inpath '<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt'> into table testhivedrivertable [junit] PREHOOK: type: LOAD [junit] Copying data from <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt> [junit] Loading data to table testhivedrivertable [junit] POSTHOOK: query: load data local inpath '<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/data/files/kv1.txt'> into table testhivedrivertable [junit] POSTHOOK: type: LOAD [junit] POSTHOOK: Output: default@testhivedrivertable [junit] OK [junit] PREHOOK: query: select key, value from testhivedrivertable where key > 10 [junit] PREHOOK: type: QUERY [junit] PREHOOK: Input: default@testhivedrivertable [junit] PREHOOK: Output: file:/tmp/hudson/hive_2011-02-15_16-12-19_430_944035855518782289/-mr-10000 [junit] Total MapReduce jobs = 1 [junit] Launching Job 1 out of 1 [junit] Number of reduce tasks is set to 0 since there's no reduce operator [junit] Job running in-process (local Hadoop) [junit] Hadoop job information for null: number of mappers: 0; number of reducers: 0 [junit] 2011-02-15 16:12:22,070 null map = 100%, reduce = 0% [junit] Ended Job = job_local_0001 [junit] POSTHOOK: query: select key, value from testhivedrivertable where key > 10 [junit] POSTHOOK: type: QUERY [junit] POSTHOOK: Input: default@testhivedrivertable [junit] POSTHOOK: Output: file:/tmp/hudson/hive_2011-02-15_16-12-19_430_944035855518782289/-mr-10000 [junit] OK [junit] PREHOOK: query: select count(1) as c from testhivedrivertable [junit] PREHOOK: type: QUERY [junit] PREHOOK: Input: default@testhivedrivertable [junit] PREHOOK: Output: file:/tmp/hudson/hive_2011-02-15_16-12-22_266_8574901807619109460/-mr-10000 [junit] Total MapReduce jobs = 1 [junit] Launching Job 1 out of 1 [junit] Number of reduce tasks determined at compile time: 1 [junit] In order to change the average load for a reducer (in bytes): [junit] set hive.exec.reducers.bytes.per.reducer=<number> [junit] In order to limit the maximum number of reducers: [junit] set hive.exec.reducers.max=<number> [junit] In order to set a constant number of reducers: [junit] set mapred.reduce.tasks=<number> [junit] Job running in-process (local Hadoop) [junit] Hadoop job information for null: number of mappers: 0; number of reducers: 0 [junit] 2011-02-15 16:12:24,949 null map = 100%, reduce = 100% [junit] Ended Job = job_local_0001 [junit] POSTHOOK: query: select count(1) as c from testhivedrivertable [junit] POSTHOOK: type: QUERY [junit] POSTHOOK: Input: default@testhivedrivertable [junit] POSTHOOK: Output: file:/tmp/hudson/hive_2011-02-15_16-12-22_266_8574901807619109460/-mr-10000 [junit] OK [junit] ------------- ---------------- --------------- [junit] [junit] Testcase: testExecute took 9.233 sec [junit] Testcase: testNonHiveCommand took 0.806 sec [junit] Testcase: testMetastore took 0.249 sec [junit] Testcase: testGetClusterStatus took 0.088 sec [junit] Testcase: testFetch took 8.695 sec [junit] Testcase: testDynamicSerde took 6.284 sec test-conditions: gen-test: create-dirs: compile-ant-tasks: create-dirs: init: compile: [echo] Compiling: anttasks [javac] <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/ant/build.xml>:40: warning: 'includeantruntime' was not set, defaulting to build.sysclasspath=last; set to false for repeatable builds deploy-ant-tasks: create-dirs: init: compile: [echo] Compiling: anttasks [javac] <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/ant/build.xml>:40: warning: 'includeantruntime' was not set, defaulting to build.sysclasspath=last; set to false for repeatable builds jar: init: compile: ivy-init-dirs: ivy-download: [get] Getting: http://repo2.maven.org/maven2/org/apache/ivy/ivy/2.1.0/ivy-2.1.0.jar [get] To: <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ivy/lib/ivy-2.1.0.jar> [get] Not modified - so not downloaded ivy-probe-antlib: ivy-init-antlib: ivy-init: ivy-retrieve-hadoop-source: [ivy:retrieve] :: Ivy 2.1.0 - 20090925235825 :: http://ant.apache.org/ivy/ :: [ivy:retrieve] :: loading settings :: file = <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/ivy/ivysettings.xml> [ivy:retrieve] :: resolving dependencies :: org.apache.hadoop.hive#shims;working@minerva [ivy:retrieve] confs: [default] [ivy:retrieve] found hadoop#core;0.20.0 in hadoop-source [ivy:retrieve] found hadoop#core;0.20.3-CDH3-SNAPSHOT in hadoop-source [ivy:retrieve] :: resolution report :: resolve 2724ms :: artifacts dl 2ms --------------------------------------------------------------------- | | modules || artifacts | | conf | number| search|dwnlded|evicted|| number|dwnlded| --------------------------------------------------------------------- | default | 2 | 0 | 0 | 0 || 2 | 0 | --------------------------------------------------------------------- [ivy:retrieve] :: retrieving :: org.apache.hadoop.hive#shims [ivy:retrieve] confs: [default] [ivy:retrieve] 0 artifacts copied, 2 already retrieved (0kB/3ms) install-hadoopcore-internal: build_shims: [echo] Compiling shims against hadoop 0.20.0 (<https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/hadoopcore/hadoop-0.20.0)> [javac] <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/shims/build.xml>:53: warning: 'includeantruntime' was not set, defaulting to build.sysclasspath=last; set to false for repeatable builds ivy-init-dirs: ivy-download: [get] Getting: http://repo2.maven.org/maven2/org/apache/ivy/ivy/2.1.0/ivy-2.1.0.jar [get] To: <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build/ivy/lib/ivy-2.1.0.jar> [get] Not modified - so not downloaded ivy-probe-antlib: ivy-init-antlib: ivy-init: ivy-retrieve-hadoop-source: [ivy:retrieve] :: Ivy 2.1.0 - 20090925235825 :: http://ant.apache.org/ivy/ :: [ivy:retrieve] :: loading settings :: file = <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/ivy/ivysettings.xml> [ivy:retrieve] :: resolving dependencies :: org.apache.hadoop.hive#shims;working@minerva [ivy:retrieve] confs: [default] [for] <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/shims/build.xml>: The following error occurred while executing this line: [for] <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build.xml>:212: The following error occurred while executing this line: [for] <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/shims/build.xml>:65: The following error occurred while executing this line: [for] <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build-common.xml>:189: java.lang.OutOfMemoryError: PermGen space BUILD FAILED <https://hudson.apache.org/hudson/job/Hive-trunk-h0.20/ws/hive/build.xml>:207: Keepgoing execution: 2 of 11 iterations failed. Total time: 190 minutes 8 seconds Archiving artifacts Recording test results