Figured out. Need <dependency>
<groupId>org.apache.hbase</groupId> <artifactId>hbase-it</artifactId> <version>0.96.1.1-cdh5.0.1</version> </dependency> as well.. On Thu, Jun 19, 2014 at 5:21 PM, Chen Wang <[email protected]> wrote: > Hi folks, > > I am running bulk load with HFileOutputFormat. The reducer throws the > following NoSuchMethodError.Just wondering where this class is? > > My pom looks like this:(0.96.1.1-cdh5.0.1) > > <dependency> > > <groupId>org.apache.hadoop</groupId> > > <artifactId>hadoop-client</artifactId> > > <version>2.3.0-mr1-cdh5.0.1</version> > > </dependency> > > <dependency> > > <groupId>org.apache.hadoop</groupId> > > <artifactId>hadoop-core</artifactId> > > <version>2.3.0-mr1-cdh5.0.1</version> > > </dependency> > > <!-- <dependency> <groupId>com.mapr.hadoop</groupId> > <artifactId>maprfs</artifactId> > > <version>1.0.3-mapr-2.1.3.1</version> </dependency> --> > > <dependency> > > <groupId>org.apache.hbase</groupId> > > <artifactId>hbase</artifactId> > > <version>0.96.1.1-cdh5.0.1</version> > > <type>pom</type> > > </dependency> > > <dependency> > > <groupId>org.apache.hbase</groupId> > > <artifactId>hbase-common</artifactId> > > <version>0.96.1.1-cdh5.0.1</version> > > </dependency> > > <dependency> > > <groupId>org.apache.httpcomponents</groupId> > > <artifactId>httpclient</artifactId> > > <version>4.1.1</version> > > </dependency> > > > > 2014-06-19 17:09:52,496 FATAL [main] org.apache.hadoop.mapred.YarnChild: > Error running child : java.lang.NoSuchMethodError: > org.apache.hadoop.hbase.io.hfile.AbstractHFileWriter.compressionByName(Ljava/lang/String;)Lorg/apache/hadoop/hbase/io/compress/Compression$Algorithm; > at > org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2$1.getNewWriter(HFileOutputFormat2.java:220) > at > org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2$1.write(HFileOutputFormat2.java:174) > at > org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2$1.write(HFileOutputFormat2.java:133) > at > org.apache.hadoop.mapred.ReduceTask$NewTrackingRecordWriter.write(ReduceTask.java:558) > at > org.apache.hadoop.mapreduce.task.TaskInputOutputContextImpl.write(TaskInputOutputContextImpl.java:89) > at > org.apache.hadoop.mapreduce.lib.reduce.WrappedReducer$Context.write(WrappedReducer.java:105) > at > org.apache.hadoop.hbase.mapreduce.PutSortReducer.reduce(PutSortReducer.java:72) > at > org.apache.hadoop.hbase.mapreduce.PutSortReducer.reduce(PutSortReducer.java:40) > at org.apache.hadoop.mapreduce.Reducer.run(Reducer.java:171) > at > org.apache.hadoop.mapred.ReduceTask.runNewReducer(ReduceTask.java:627) > at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:389) > at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:168) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:415) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1548) > at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:163) > > Thanks! > > Chen > >
