----------------------------------------------------------- This is an automatically generated e-mail. To reply, visit: https://reviews.apache.org/r/27052/#review57885 -----------------------------------------------------------
Ship it! Ship It! - Mahadev Konar On Oct. 22, 2014, 8:05 p.m., Andrew Onischuk wrote: > > ----------------------------------------------------------- > This is an automatically generated e-mail. To reply, visit: > https://reviews.apache.org/r/27052/ > ----------------------------------------------------------- > > (Updated Oct. 22, 2014, 8:05 p.m.) > > > Review request for Ambari and Dmitro Lisnichenko. > > > Bugs: AMBARI-7910 > https://issues.apache.org/jira/browse/AMBARI-7910 > > > Repository: ambari > > > Description > ------- > > Compression jobs are failing due to missing Lzo package. > > **console.log** > > > > 2014-10-18 > 21:21:34,621|main|INFO|19267|139929561851648|MainThread|RUNNING TEST > "test_Compression[com.hadoop.compression.lzo.LzoCodec-org.apache.hadoop.io.compress.DefaultCodec-NONE-TextFormat]" > at location "tests/mapred/mapred_1/Compression/test_Compression_20.py" at > line number "72" > 2014-10-18 > 21:21:34,622|beaver.machine|INFO|19267|139929561851648|MainThread|RUNNING: > /usr/hdp/current/hadoop-client/bin/hadoop jar > /usr/hdp/current/hadoop-mapreduce-client/hadoop-mapreduce-examples.jar sort > -Dmapreduce.map.output.compress=true > -Dmapreduce.map.output.compress.codec=org.apache.hadoop.io.compress.DefaultCodec > -Dmapreduce.output.fileoutputformat.compress=true > -Dmapreduce.output.fileoutputformat.compression.type=NONE > -Dmapreduce.output.fileoutputformat.compress.codec=com.hadoop.compression.lzo.LzoCodec > -outKey org.apache.hadoop.io.Text -outValue org.apache.hadoop.io.Text > Compression/textinput Compression/textoutput-1413667294.62 > 2014-10-18 > 21:21:36,591|beaver.machine|INFO|19267|139929561851648|MainThread|14/10/18 > 21:21:36 INFO client.RMProxy: Connecting to ResourceManager at > ambari-rerun-su-1.cs1cloud.internal/172.18.146.170:8050 > 2014-10-18 > 21:21:37,982|beaver.machine|INFO|19267|139929561851648|MainThread|Running on > 1 nodes to sort from > hdfs://ambari-rerun-su-1.cs1cloud.internal:8020/user/hrt_qa/Compression/textinput > into > hdfs://ambari-rerun-su-1.cs1cloud.internal:8020/user/hrt_qa/Compression/textoutput-1413667294.62 > with 1 reduces. > 2014-10-18 > 21:21:37,988|beaver.machine|INFO|19267|139929561851648|MainThread|Job > started: Sat Oct 18 21:21:37 UTC 2014 > 2014-10-18 > 21:21:38,026|beaver.machine|INFO|19267|139929561851648|MainThread|14/10/18 > 21:21:38 INFO client.RMProxy: Connecting to ResourceManager at > ambari-rerun-su-1.cs1cloud.internal/172.18.146.170:8050 > 2014-10-18 > 21:21:38,113|beaver.machine|INFO|19267|139929561851648|MainThread|14/10/18 > 21:21:38 INFO hdfs.DFSClient: Created HDFS_DELEGATION_TOKEN token 82 for > hrt_qa on 172.18.146.170:8020 > 2014-10-18 > 21:21:38,138|beaver.machine|INFO|19267|139929561851648|MainThread|14/10/18 > 21:21:38 INFO security.TokenCache: Got dt for > hdfs://ambari-rerun-su-1.cs1cloud.internal:8020; Kind: HDFS_DELEGATION_TOKEN, > Service: 172.18.146.170:8020, Ident: (HDFS_DELEGATION_TOKEN token 82 for > hrt_qa) > 2014-10-18 > 21:21:38,695|beaver.machine|INFO|19267|139929561851648|MainThread|14/10/18 > 21:21:38 INFO input.FileInputFormat: Total input paths to process : 1 > 2014-10-18 > 21:21:38,908|beaver.machine|INFO|19267|139929561851648|MainThread|14/10/18 > 21:21:38 INFO mapreduce.JobSubmitter: number of splits:1 > 2014-10-18 > 21:21:39,273|beaver.machine|INFO|19267|139929561851648|MainThread|14/10/18 > 21:21:39 INFO mapreduce.JobSubmitter: Submitting tokens for job: > job_1413586769062_0027 > 2014-10-18 > 21:21:39,275|beaver.machine|INFO|19267|139929561851648|MainThread|14/10/18 > 21:21:39 INFO mapreduce.JobSubmitter: Kind: HDFS_DELEGATION_TOKEN, Service: > 172.18.146.170:8020, Ident: (HDFS_DELEGATION_TOKEN token 82 for hrt_qa) > 2014-10-18 > 21:21:39,794|beaver.machine|INFO|19267|139929561851648|MainThread|14/10/18 > 21:21:39 INFO impl.YarnClientImpl: Submitted application > application_1413586769062_0027 > 2014-10-18 > 21:21:39,865|beaver.machine|INFO|19267|139929561851648|MainThread|14/10/18 > 21:21:39 INFO mapreduce.Job: The url to track the job: > http://ambari-rerun-su-1.cs1cloud.internal:8088/proxy/application_1413586769062_0027/ > 2014-10-18 > 21:21:39,866|beaver.machine|INFO|19267|139929561851648|MainThread|14/10/18 > 21:21:39 INFO mapreduce.Job: Running job: job_1413586769062_0027 > 2014-10-18 > 21:21:51,064|beaver.machine|INFO|19267|139929561851648|MainThread|14/10/18 > 21:21:51 INFO mapreduce.Job: Job job_1413586769062_0027 running in uber mode > : false > 2014-10-18 > 21:21:51,067|beaver.machine|INFO|19267|139929561851648|MainThread|14/10/18 > 21:21:51 INFO mapreduce.Job: map 0% reduce 0% > 2014-10-18 > 21:21:57,153|beaver.machine|INFO|19267|139929561851648|MainThread|14/10/18 > 21:21:57 INFO mapreduce.Job: map 100% reduce 0% > 2014-10-18 > 21:22:03,201|beaver.machine|INFO|19267|139929561851648|MainThread|14/10/18 > 21:22:03 INFO mapreduce.Job: Task Id : attempt_1413586769062_0027_r_000000_0, > Status : FAILED > 2014-10-18 > 21:22:03,221|beaver.machine|INFO|19267|139929561851648|MainThread|Error: > java.lang.IllegalArgumentException: Compression codec > com.hadoop.compression.lzo.LzoCodec was not found. > 2014-10-18 > 21:22:03,222|beaver.machine|INFO|19267|139929561851648|MainThread|at > org.apache.hadoop.mapreduce.lib.output.FileOutputFormat.getOutputCompressorClass(FileOutputFormat.java:122) > 2014-10-18 > 21:22:03,222|beaver.machine|INFO|19267|139929561851648|MainThread|at > org.apache.hadoop.mapreduce.lib.output.SequenceFileOutputFormat.getSequenceWriter(SequenceFileOutputFormat.java:56) > 2014-10-18 > 21:22:03,222|beaver.machine|INFO|19267|139929561851648|MainThread|at > org.apache.hadoop.mapreduce.lib.output.SequenceFileOutputFormat.getRecordWriter(SequenceFileOutputFormat.java:75) > 2014-10-18 > 21:22:03,222|beaver.machine|INFO|19267|139929561851648|MainThread|at > org.apache.hadoop.mapred.ReduceTask$NewTrackingRecordWriter.<init>(ReduceTask.java:540) > 2014-10-18 > 21:22:03,222|beaver.machine|INFO|19267|139929561851648|MainThread|at > org.apache.hadoop.mapred.ReduceTask.runNewReducer(ReduceTask.java:614) > 2014-10-18 > 21:22:03,222|beaver.machine|INFO|19267|139929561851648|MainThread|at > org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:389) > 2014-10-18 > 21:22:03,223|beaver.machine|INFO|19267|139929561851648|MainThread|at > org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:163) > 2014-10-18 > 21:22:03,223|beaver.machine|INFO|19267|139929561851648|MainThread|at > java.security.AccessController.doPrivileged(Native Method) > 2014-10-18 > 21:22:03,223|beaver.machine|INFO|19267|139929561851648|MainThread|at > javax.security.auth.Subject.doAs(Subject.java:415) > 2014-10-18 > 21:22:03,223|beaver.machine|INFO|19267|139929561851648|MainThread|at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628) > 2014-10-18 > 21:22:03,223|beaver.machine|INFO|19267|139929561851648|MainThread|at > org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:158) > 2014-10-18 > 21:22:03,223|beaver.machine|INFO|19267|139929561851648|MainThread|Caused by: > java.lang.ClassNotFoundException: Class com.hadoop.compression.lzo.LzoCodec > not found > 2014-10-18 > 21:22:03,224|beaver.machine|INFO|19267|139929561851648|MainThread|at > org.apache.hadoop.conf.Configuration.getClassByName(Configuration.java:1954) > 2014-10-18 > 21:22:03,224|beaver.machine|INFO|19267|139929561851648|MainThread|at > org.apache.hadoop.mapreduce.lib.output.FileOutputFormat.getOutputCompressorClass(FileOutputFormat.java:119) > 2014-10-18 > 21:22:03,224|beaver.machine|INFO|19267|139929561851648|MainThread|... 10 more > 2014-10-18 > 21:22:03,224|beaver.machine|INFO|19267|139929561851648|MainThread| > 2014-10-18 > 21:22:09,269|beaver.machine|INFO|19267|139929561851648|MainThread|14/10/18 > 21:22:09 INFO mapreduce.Job: Task Id : attempt_1413586769062_0027_r_000000_1, > Status : FAILED > 2014-10-18 > 21:22:09,273|beaver.machine|INFO|19267|139929561851648|MainThread|Error: > java.lang.IllegalArgumentException: Compression codec > com.hadoop.compression.lzo.LzoCodec was not found. > 2014-10-18 > 21:22:09,273|beaver.machine|INFO|19267|139929561851648|MainThread|at > org.apache.hadoop.mapreduce.lib.output.FileOutputFormat.getOutputCompressorClass(FileOutputFormat.java:122) > 2014-10-18 > 21:22:09,274|beaver.machine|INFO|19267|139929561851648|MainThread|at > org.apache.hadoop.mapreduce.lib.output.SequenceFileOutputFormat.getSequenceWriter(SequenceFileOutputFormat.java:56) > 2014-10-18 > 21:22:09,274|beaver.machine|INFO|19267|139929561851648|MainThread|at > org.apache.hadoop.mapreduce.lib.output.SequenceFileOutputFormat.getRecordWriter(SequenceFileOutputFormat.java:75) > 2014-10-18 > 21:22:09,274|beaver.machine|INFO|19267|139929561851648|MainThread|at > org.apache.hadoop.mapred.ReduceTask$NewTrackingRecordWriter.<init>(ReduceTask.java:540) > 2014-10-18 > 21:22:09,275|beaver.machine|INFO|19267|139929561851648|MainThread|at > org.apache.hadoop.mapred.ReduceTask.runNewReducer(ReduceTask.java:614) > 2014-10-18 > 21:22:09,275|beaver.machine|INFO|19267|139929561851648|MainThread|at > org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:389) > 2014-10-18 > 21:22:09,276|beaver.machine|INFO|19267|139929561851648|MainThread|at > org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:163) > 2014-10-18 > 21:22:09,276|beaver.machine|INFO|19267|139929561851648|MainThread|at > java.security.AccessController.doPrivileged(Native Method) > 2014-10-18 > 21:22:09,277|beaver.machine|INFO|19267|139929561851648|MainThread|at > javax.security.auth.Subject.doAs(Subject.java:415) > 2014-10-18 > 21:22:09,277|beaver.machine|INFO|19267|139929561851648|MainThread|at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628) > 2014-10-18 > 21:22:09,278|beaver.machine|INFO|19267|139929561851648|MainThread|at > org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:158) > 2014-10-18 > 21:22:09,278|beaver.machine|INFO|19267|139929561851648|MainThread|Caused by: > java.lang.ClassNotFoundException: Class com.hadoop.compression.lzo.LzoCodec > not found > 2014-10-18 > 21:22:09,279|beaver.machine|INFO|19267|139929561851648|MainThread|at > org.apache.hadoop.conf.Configuration.getClassByName(Configuration.java:1954) > 2014-10-18 > 21:22:09,279|beaver.machine|INFO|19267|139929561851648|MainThread|at > org.apache.hadoop.mapreduce.lib.output.FileOutputFormat.getOutputCompressorClass(FileOutputFormat.java:119) > 2014-10-18 > 21:22:09,280|beaver.machine|INFO|19267|139929561851648|MainThread|... 10 more > 2014-10-18 > 21:22:09,280|beaver.machine|INFO|19267|139929561851648|MainThread| > 2014-10-18 > 21:22:15,329|beaver.machine|INFO|19267|139929561851648|MainThread|14/10/18 > 21:22:15 INFO mapreduce.Job: Task Id : attempt_1413586769062_0027_r_000000_2, > Status : FAILED > 2014-10-18 > 21:22:15,333|beaver.machine|INFO|19267|139929561851648|MainThread|Error: > java.lang.IllegalArgumentException: Compression codec > com.hadoop.compression.lzo.LzoCodec was not found. > 2014-10-18 > 21:22:15,333|beaver.machine|INFO|19267|139929561851648|MainThread|at > org.apache.hadoop.mapreduce.lib.output.FileOutputFormat.getOutputCompressorClass(FileOutputFormat.java:122) > 2014-10-18 > 21:22:15,334|beaver.machine|INFO|19267|139929561851648|MainThread|at > org.apache.hadoop.mapreduce.lib.output.SequenceFileOutputFormat.getSequenceWriter(SequenceFileOutputFormat.java:56) > 2014-10-18 > 21:22:15,334|beaver.machine|INFO|19267|139929561851648|MainThread|at > org.apache.hadoop.mapreduce.lib.output.SequenceFileOutputFormat.getRecordWriter(SequenceFileOutputFormat.java:75) > 2014-10-18 > 21:22:15,335|beaver.machine|INFO|19267|139929561851648|MainThread|at > org.apache.hadoop.mapred.ReduceTask$NewTrackingRecordWriter.<init>(ReduceTask.java:540) > 2014-10-18 > 21:22:15,336|beaver.machine|INFO|19267|139929561851648|MainThread|at > org.apache.hadoop.mapred.ReduceTask.runNewReducer(ReduceTask.java:614) > 2014-10-18 > 21:22:15,336|beaver.machine|INFO|19267|139929561851648|MainThread|at > org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:389) > 2014-10-18 > 21:22:15,337|beaver.machine|INFO|19267|139929561851648|MainThread|at > org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:163) > 2014-10-18 > 21:22:15,337|beaver.machine|INFO|19267|139929561851648|MainThread|at > java.security.AccessController.doPrivileged(Native Method) > 2014-10-18 > 21:22:15,338|beaver.machine|INFO|19267|139929561851648|MainThread|at > javax.security.auth.Subject.doAs(Subject.java:415) > 2014-10-18 > 21:22:15,339|beaver.machine|INFO|19267|139929561851648|MainThread|at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628) > 2014-10-18 > 21:22:15,339|beaver.machine|INFO|19267|139929561851648|MainThread|at > org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:158) > 2014-10-18 > 21:22:15,340|beaver.machine|INFO|19267|139929561851648|MainThread|Caused by: > java.lang.ClassNotFoundException: Class com.hadoop.compression.lzo.LzoCodec > not found > 2014-10-18 > 21:22:15,341|beaver.machine|INFO|19267|139929561851648|MainThread|at > org.apache.hadoop.conf.Configuration.getClassByName(Configuration.java:1954) > 2014-10-18 > 21:22:15,342|beaver.machine|INFO|19267|139929561851648|MainThread|at > org.apache.hadoop.mapreduce.lib.output.FileOutputFormat.getOutputCompressorClass(FileOutputFormat.java:119) > 2014-10-18 > 21:22:15,343|beaver.machine|INFO|19267|139929561851648|MainThread|... 10 more > 2014-10-18 > 21:22:15,343|beaver.machine|INFO|19267|139929561851648|MainThread| > 2014-10-18 > 21:22:23,398|beaver.machine|INFO|19267|139929561851648|MainThread|14/10/18 > 21:22:23 INFO mapreduce.Job: map 100% reduce 100% > 2014-10-18 > 21:22:23,418|beaver.machine|INFO|19267|139929561851648|MainThread|14/10/18 > 21:22:23 INFO mapreduce.Job: Job job_1413586769062_0027 failed with state > FAILED due to: Task failed task_1413586769062_0027_r_000000 > 2014-10-18 > 21:22:23,419|beaver.machine|INFO|19267|139929561851648|MainThread|Job failed > as tasks failed. failedMaps:0 failedReduces:1 > 2014-10-18 > 21:22:23,419|beaver.machine|INFO|19267|139929561851648|MainThread| > 2014-10-18 > 21:22:23,592|beaver.machine|INFO|19267|139929561851648|MainThread|14/10/18 > 21:22:23 INFO mapreduce.Job: Counters: 37 > 2014-10-18 > 21:22:23,592|beaver.machine|INFO|19267|139929561851648|MainThread|File System > Counters > 2014-10-18 > 21:22:23,593|beaver.machine|INFO|19267|139929561851648|MainThread|FILE: > Number of bytes read=0 > 2014-10-18 > 21:22:23,593|beaver.machine|INFO|19267|139929561851648|MainThread|FILE: > Number of bytes written=115511 > 2014-10-18 > 21:22:23,593|beaver.machine|INFO|19267|139929561851648|MainThread|FILE: > Number of read operations=0 > 2014-10-18 > 21:22:23,593|beaver.machine|INFO|19267|139929561851648|MainThread|FILE: > Number of large read operations=0 > 2014-10-18 > 21:22:23,593|beaver.machine|INFO|19267|139929561851648|MainThread|FILE: > Number of write operations=0 > 2014-10-18 > 21:22:23,593|beaver.machine|INFO|19267|139929561851648|MainThread|HDFS: > Number of bytes read=1767 > 2014-10-18 > 21:22:23,594|beaver.machine|INFO|19267|139929561851648|MainThread|HDFS: > Number of bytes written=0 > 2014-10-18 > 21:22:23,594|beaver.machine|INFO|19267|139929561851648|MainThread|HDFS: > Number of read operations=4 > 2014-10-18 > 21:22:23,594|beaver.machine|INFO|19267|139929561851648|MainThread|HDFS: > Number of large read operations=0 > 2014-10-18 > 21:22:23,594|beaver.machine|INFO|19267|139929561851648|MainThread|HDFS: > Number of write operations=0 > 2014-10-18 > 21:22:23,594|beaver.machine|INFO|19267|139929561851648|MainThread|Job Counters > 2014-10-18 > 21:22:23,594|beaver.machine|INFO|19267|139929561851648|MainThread|Failed > reduce tasks=4 > 2014-10-18 > 21:22:23,594|beaver.machine|INFO|19267|139929561851648|MainThread|Launched > map tasks=1 > 2014-10-18 > 21:22:23,595|beaver.machine|INFO|19267|139929561851648|MainThread|Launched > reduce tasks=4 > 2014-10-18 > 21:22:23,595|beaver.machine|INFO|19267|139929561851648|MainThread|Data-local > map tasks=1 > 2014-10-18 > 21:22:23,595|beaver.machine|INFO|19267|139929561851648|MainThread|Total time > spent by all maps in occupied slots (ms)=3977 > 2014-10-18 > 21:22:23,595|beaver.machine|INFO|19267|139929561851648|MainThread|Total time > spent by all reduces in occupied slots (ms)=15442 > 2014-10-18 > 21:22:23,595|beaver.machine|INFO|19267|139929561851648|MainThread|Total time > spent by all map tasks (ms)=3977 > 2014-10-18 > 21:22:23,596|beaver.machine|INFO|19267|139929561851648|MainThread|Total time > spent by all reduce tasks (ms)=15442 > 2014-10-18 > 21:22:23,596|beaver.machine|INFO|19267|139929561851648|MainThread|Total > vcore-seconds taken by all map tasks=3977 > 2014-10-18 > 21:22:23,596|beaver.machine|INFO|19267|139929561851648|MainThread|Total > vcore-seconds taken by all reduce tasks=15442 > 2014-10-18 > 21:22:23,597|beaver.machine|INFO|19267|139929561851648|MainThread|Total > megabyte-seconds taken by all map tasks=4072448 > 2014-10-18 > 21:22:23,597|beaver.machine|INFO|19267|139929561851648|MainThread|Total > megabyte-seconds taken by all reduce tasks=15812608 > 2014-10-18 > 21:22:23,598|beaver.machine|INFO|19267|139929561851648|MainThread|Map-Reduce > Framework > 2014-10-18 > 21:22:23,598|beaver.machine|INFO|19267|139929561851648|MainThread|Map input > records=2 > 2014-10-18 > 21:22:23,598|beaver.machine|INFO|19267|139929561851648|MainThread|Map output > records=2 > 2014-10-18 > 21:22:23,598|beaver.machine|INFO|19267|139929561851648|MainThread|Map output > bytes=1514 > 2014-10-18 > 21:22:23,598|beaver.machine|INFO|19267|139929561851648|MainThread|Map output > materialized bytes=885 > 2014-10-18 > 21:22:23,598|beaver.machine|INFO|19267|139929561851648|MainThread|Input split > bytes=159 > 2014-10-18 > 21:22:23,599|beaver.machine|INFO|19267|139929561851648|MainThread|Combine > input records=0 > 2014-10-18 > 21:22:23,599|beaver.machine|INFO|19267|139929561851648|MainThread|Spilled > Records=2 > 2014-10-18 > 21:22:23,599|beaver.machine|INFO|19267|139929561851648|MainThread|Failed > Shuffles=0 > 2014-10-18 > 21:22:23,599|beaver.machine|INFO|19267|139929561851648|MainThread|Merged Map > outputs=0 > 2014-10-18 > 21:22:23,599|beaver.machine|INFO|19267|139929561851648|MainThread|GC time > elapsed (ms)=17 > 2014-10-18 > 21:22:23,599|beaver.machine|INFO|19267|139929561851648|MainThread|CPU time > spent (ms)=670 > 2014-10-18 > 21:22:23,599|beaver.machine|INFO|19267|139929561851648|MainThread|Physical > memory (bytes) snapshot=595439616 > 2014-10-18 > 21:22:23,600|beaver.machine|INFO|19267|139929561851648|MainThread|Virtual > memory (bytes) snapshot=1667874816 > 2014-10-18 > 21:22:23,600|beaver.machine|INFO|19267|139929561851648|MainThread|Total > committed heap usage (bytes)=632291328 > 2014-10-18 > 21:22:23,600|beaver.machine|INFO|19267|139929561851648|MainThread|File Input > Format Counters > 2014-10-18 > 21:22:23,600|beaver.machine|INFO|19267|139929561851648|MainThread|Bytes > Read=1608 > 2014-10-18 > 21:22:23,600|beaver.machine|INFO|19267|139929561851648|MainThread|Job ended: > Sat Oct 18 21:22:23 UTC 2014 > 2014-10-18 > 21:22:23,600|beaver.machine|INFO|19267|139929561851648|MainThread|The job > took 45 seconds. > > > **lzo package search in nano** > > > > find / -name "*lzo*" > /usr/hdp/2.2.0.0-908/hadoop/lib/hadoop-lzo-0.6.0.jar > /usr/hdp/current/share/lzo > /usr/hdp/current/share/lzo/0.6.0/lib/hadoop-lzo-0.6.0.jar > > > **lzo package search in ambari** > > > > sudo find / -name "*lzo*" > /usr/share/mime/application/x-lzop.xml > /grid/0/hadoopqe/set_tez_lzo.ps1 > /grid/0/hadoopqe/tests/flume/conf/exec-file-hdfs-lzop.properties > /grid/0/hadoopqe/tests/flume/conf/exec-file-hdfs-lzo.properties > /grid/0/hadoopqe/tests/flume/conf/exec-memory-hdfs-lzop.properties > /grid/0/hadoopqe/tests/flume/conf/exec-memory-hdfs-lzo.properties > /lib/modules/2.6.32-358.el6.x86_64/kernel/crypto/lzo.ko > /lib/modules/2.6.32-358.el6.x86_64/kernel/lib/lzo > /lib/modules/2.6.32-358.el6.x86_64/kernel/lib/lzo/lzo_compress.ko > /lib/modules/2.6.32-358.el6.x86_64/kernel/lib/lzo/lzo_decompress.ko > > > **hadoop version** > > > > hadoop version > Hadoop 2.6.0.2.2.0.0-945 > Subversion [email protected]:hortonworks/hadoop.git -r > 5e72cc2773fc079a72735bd3f4fd347ed24df743 > Compiled by jenkins on 2014-10-16T23:47Z > Compiled with protoc 2.5.0 > From source with checksum af8da4bc9b78bbbd52225cb96f1bd71 > This command was run using > /usr/hdp/2.2.0.0-945/hadoop/hadoop-common-2.6.0.2.2.0.0-945.jar > > > > Diffs > ----- > > > ambari-server/src/main/resources/stacks/HDP/1.3.2/services/HDFS/configuration/core-site.xml > f9d6ebe > > ambari-server/src/main/resources/stacks/HDP/1.3.2/services/HDFS/package/scripts/hdfs.py > 6f24f8e > > ambari-server/src/main/resources/stacks/HDP/1.3.2/services/HDFS/package/scripts/params.py > 504e6ac > > ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HDFS/configuration/core-site.xml > 52764ac > > ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HDFS/package/scripts/hdfs.py > 873aa15 > > ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HDFS/package/scripts/params.py > 404244b > ambari-server/src/test/python/stacks/1.3.2/configs/default.json 74bbb64 > ambari-server/src/test/python/stacks/1.3.2/configs/secured.json d504281 > > Diff: https://reviews.apache.org/r/27052/diff/ > > > Testing > ------- > > mvn clean test > > > Thanks, > > Andrew Onischuk > >
