Ted suggested to remove the following call
TableMapReduceUtil.initTableReducerJob(otherArgs[0],
null, job);


You are doing 2 things in your earlier code snippet


*TableMapReduceUtil.initTableMapperJob(otherArgs[0],
scan,EntitySearcherMapper.class, ImmutableBytesWritable.class,
Put.class,job);*

and

*TableMapReduceUtil.initTableReducerJob(otherArgs[0],null, job);*


You need to remove the second one and keep the first one. In the first one
you are specifying the data input which is must. Otherwise how would the
job know to read or gets input? The second call (initTableReducerJob) is
not necessary as your output format has changed.


Regards,
Shahab


On Tue, Aug 26, 2014 at 3:31 PM, yeshwanth kumar <[email protected]>
wrote:

> i was doing that earlier
>
> TableMapReduceUtil.initTableMapperJob(otherArgs[0], scan,
> EntitySearcherMapper.class, ImmutableBytesWritable.class, Put.class,
> job);//otherArgs[0]=i1 TableMapReduceUtil.initTableReducerJob(otherArgs[0],
> null, job);
>
> Ted suggested to remove them,
> if u see the first message in this thread, u will know the issue by
> specifying the table.
>
> -yeshwanth
>
>
>
> On Wed, Aug 27, 2014 at 12:54 AM, Shahab Yunus <[email protected]>
> wrote:
>
> > So are you calling or making the following call with the right
> > parameters,specifying the table to read from?
> >
> > TableMapReduceUtil.initTableMapperJob
> >
> >
> > Can you should your whole job setup/driver code?
> >
> >
> > Regards,
> >
> > Shahab
> >
> >
> > On Tue, Aug 26, 2014 at 3:18 PM, yeshwanth kumar <[email protected]>
> > wrote:
> >
> > > that mapreduce job reads data from hbase table,
> > > it doesn't take any explicit input data/file/
> > >
> > > -yeshwanth
> > >
> > >
> > > On Wed, Aug 27, 2014 at 12:44 AM, Shahab Yunus <[email protected]
> >
> > > wrote:
> > >
> > > > Where are you setting the input data/path/format for the job? I don't
> > see
> > > > that in the code below that you just pasted...
> > > >
> > > >
> > > >
> > > >
> > > >
> > > >
> > > >
> > > >
> > > >
> > > >
> > > >
> > > > *Job job = new Job(config,
> > > >
> > > >
> > >
> >
> "GetEntitiesMR");job.setJarByClass(EntitySearcherMR.class);job.setMapperClass(EntitySearcherMapper.class);job.setOutputFormatClass(MultiTableOutputFormat.class);TableMapReduceUtil.addDependencyJars(job);job.setNumReduceTasks(0);
> > > >               boolean b = job.waitForCompletion(true);if (!b) {throw
> > new
> > > > IOException("error with job!");}*
> > > >
> > > > Regards,
> > > > Shahab
> > > >
> > > >
> > > > On Tue, Aug 26, 2014 at 3:11 PM, yeshwanth kumar <
> > [email protected]>
> > > > wrote:
> > > >
> > > > > hi shahab,
> > > > >
> > > > > i tried in that way, by specifying outputformat as
> > > > MultiTableOutputFormat,
> > > > > it is throwing
> > > > >
> > > > >  java.io.IOException: No input paths specified in job
> > > > > at
> > > > >
> > > > >
> > > >
> > >
> >
> org.apache.hadoop.mapreduce.lib.input.FileInputFormat.listStatus(FileInputFormat.java:193)
> > > > > at
> > > > >
> > > > >
> > > >
> > >
> >
> org.apache.hadoop.mapreduce.lib.input.FileInputFormat.getSplits(FileInputFormat.java:252)
> > > > > at
> > > org.apache.hadoop.mapred.JobClient.writeNewSplits(JobClient.java:919)
> > > > > at
> org.apache.hadoop.mapred.JobClient.writeSplits(JobClient.java:936)
> > > > > at
> org.apache.hadoop.mapred.JobClient.access$500(JobClient.java:170)
> > > > > at org.apache.hadoop.mapred.JobClient$2.run(JobClient.java:854)
> > > > > at org.apache.hadoop.mapred.JobClient$2.run(JobClient.java:807)
> > > > > at java.security.AccessController.doPrivileged(Native Method)
> > > > > at javax.security.auth.Subject.doAs(Subject.java:415)
> > > > > at
> > > > >
> > > > >
> > > >
> > >
> >
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1059)
> > > > > at
> > > >
> > org.apache.hadoop.mapred.JobClient.submitJobInternal(JobClient.java:807)
> > > > > at org.apache.hadoop.mapreduce.Job.submit(Job.java:465)
> > > > > at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:495)
> > > > > at
> > > > >
> > > > >
> > > >
> > >
> >
> com.serendio.icvs.analysis.text.EntitySearcherMR.main(EntitySearcherMR.java:161)
> > > > >
> > > > > here's the job config
> > > > >
> > > > >                Job job = new Job(config, "GetEntitiesMR");
> > > > > job.setJarByClass(EntitySearcherMR.class);
> > > > > job.setMapperClass(EntitySearcherMapper.class);
> > > > > job.setOutputFormatClass(MultiTableOutputFormat.class);
> > > > > TableMapReduceUtil.addDependencyJars(job);
> > > > > job.setNumReduceTasks(0);
> > > > >                 boolean b = job.waitForCompletion(true);
> > > > > if (!b) {
> > > > > throw new IOException("error with job!");
> > > > > }
> > > > >
> > > > > i am unable to figure out, what i am missing here,
> > > > >
> > > > > -yeshwanth
> > > > >
> > > > >
> > > > >
> > > > >
> > > > >
> > > > > On Wed, Aug 27, 2014 at 12:23 AM, Shahab Yunus <
> > [email protected]
> > > >
> > > > > wrote:
> > > > >
> > > > > > You don't need to initialize the tables.
> > > > > >
> > > > > > You just need to specify the output format as
> > > MultipleTableOutputFormat
> > > > > > class.
> > > > > >
> > > > > > Something like this:
> > > > > > job.setOutputFormatClass(MultipleTableOutputFormat.class);
> > > > > >
> > > > > >
> > > > > > Because if you see the code for MultipleTableOutputFormat, it
> > creates
> > > > the
> > > > > > table on the fly and stores it in the internal map when you call
> > > > > > context.write.
> > > > > > When context.write is called:
> > > > > >
> > > > > >  @Override <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repository.grepcode.com/java/root/jdk/openjdk/6-b27/java/lang/Override.java#Override
> > > > > > >
> > > > > >
> > > > > > 126 <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hbase/hbase/0.92.1/org/apache/hadoop/hbase/mapreduce/MultiTableOutputFormat.java#126
> > > > > > >
> > > > > >
> > > > > > <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hbase/hbase/0.92.1/org/apache/hadoop/hbase/mapreduce/MultiTableOutputFormat.java#
> > > > > > >
> > > > > >
> > > > > >     public void
> > > > > > <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hbase/hbase/0.92.1/org/apache/hadoop/hbase/mapreduce/MultiTableOutputFormat.java#
> > > > > > >write(ImmutableBytesWritable
> > > > > > <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hbase/hbase/0.92.1/org/apache/hadoop/hbase/io/ImmutableBytesWritable.java#ImmutableBytesWritable
> > > > > > >
> > > > > > tableName, Writable action) throws IOException
> > > > > > <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repository.grepcode.com/java/root/jdk/openjdk/6-b27/java/io/IOException.java#IOException
> > > > > > >
> > > > > > {
> > > > > >
> > > > > > 127 <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hbase/hbase/0.92.1/org/apache/hadoop/hbase/mapreduce/MultiTableOutputFormat.java#127
> > > > > > >
> > > > > >
> > > > > > <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hbase/hbase/0.92.1/org/apache/hadoop/hbase/mapreduce/MultiTableOutputFormat.java#
> > > > > > >
> > > > > >
> > > > > >       HTable <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hbase/hbase/0.92.1/org/apache/hadoop/hbase/client/HTable.java#HTable
> > > > > > >
> > > > > > table = getTable
> > > > > > <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hbase/hbase/0.92.1/org/apache/hadoop/hbase/mapreduce/MultiTableOutputFormat.java#MultiTableOutputFormat.MultiTableRecordWriter.getTable%28org.apache.hadoop.hbase.io.ImmutableBytesWritable%29
> > > > > > >(tableName);
> > > > > >
> > > > > >
> > > > > >
> > > > > > Which calls getTable() shown below which cr
> > > > > >
> > > > > > eates the table on the fly and stores it in the internal map :
> > > > > >
> > > > > >
> > > > > >
> > > > > >  HTable <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hbase/hbase/0.92.1/org/apache/hadoop/hbase/client/HTable.java#HTable
> > > > > > >
> > > > > >  <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hbase/hbase/0.92.1/org/apache/hadoop/hbase/mapreduce/MultiTableOutputFormat.java#
> > > > > > >getTable(ImmutableBytesWritable
> > > > > > <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hbase/hbase/0.92.1/org/apache/hadoop/hbase/io/ImmutableBytesWritable.java#ImmutableBytesWritable
> > > > > > >
> > > > > > tableName) throws IOException
> > > > > > <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repository.grepcode.com/java/root/jdk/openjdk/6-b27/java/io/IOException.java#IOException
> > > > > > >
> > > > > > {
> > > > > >
> > > > > > 99 <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hbase/hbase/0.92.1/org/apache/hadoop/hbase/mapreduce/MultiTableOutputFormat.java#99
> > > > > > >
> > > > > >
> > > > > > <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hbase/hbase/0.92.1/org/apache/hadoop/hbase/mapreduce/MultiTableOutputFormat.java#
> > > > > > >
> > > > > >
> > > > > >        if (!tables
> > > > > > <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hbase/hbase/0.92.1/org/apache/hadoop/hbase/mapreduce/MultiTableOutputFormat.java#MultiTableOutputFormat.MultiTableRecordWriter.0tables
> > > > > > >.containsKey
> > > > > > <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repository.grepcode.com/java/root/jdk/openjdk/6-b27/java/util/Map.java#Map.containsKey%28java.lang.Object%29
> > > > > > >(tableName))
> > > > > > {
> > > > > >
> > > > > > 100 <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hbase/hbase/0.92.1/org/apache/hadoop/hbase/mapreduce/MultiTableOutputFormat.java#100
> > > > > > >
> > > > > >
> > > > > > <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hbase/hbase/0.92.1/org/apache/hadoop/hbase/mapreduce/MultiTableOutputFormat.java#
> > > > > > >
> > > > > >
> > > > > >         LOG <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hbase/hbase/0.92.1/org/apache/hadoop/hbase/mapreduce/MultiTableOutputFormat.java#MultiTableOutputFormat.MultiTableRecordWriter.0LOG
> > > > > > >.debug
> > > > > > <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repo1.maven.org/maven2/commons-logging/commons-logging/1.1.1/org/apache/commons/logging/Log.java#Log.debug%28java.lang.Object%29
> > > > > > >("Opening
> > > > > > HTable \"" + Bytes.toString
> > > > > > <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hbase/hbase/0.92.1/org/apache/hadoop/hbase/util/Bytes.java#Bytes.toString%28byte%5B%5D%29
> > > > > > >(tableName.get
> > > > > > <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hbase/hbase/0.92.1/org/apache/hadoop/hbase/io/ImmutableBytesWritable.java#ImmutableBytesWritable.get%28%29
> > > > > > >())+
> > > > > > "\" for writing");
> > > > > >
> > > > > > 101 <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hbase/hbase/0.92.1/org/apache/hadoop/hbase/mapreduce/MultiTableOutputFormat.java#101
> > > > > > >
> > > > > >
> > > > > > <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hbase/hbase/0.92.1/org/apache/hadoop/hbase/mapreduce/MultiTableOutputFormat.java#
> > > > > > >
> > > > > >
> > > > > >         HTable <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hbase/hbase/0.92.1/org/apache/hadoop/hbase/client/HTable.java#HTable
> > > > > > >
> > > > > > table = new HTable
> > > > > > <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hbase/hbase/0.92.1/org/apache/hadoop/hbase/client/HTable.java#HTable
> > > > > > >(conf
> > > > > > <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hbase/hbase/0.92.1/org/apache/hadoop/hbase/mapreduce/MultiTableOutputFormat.java#MultiTableOutputFormat.MultiTableRecordWriter.0conf
> > > > > > >,
> > > > > > tableName.get <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hbase/hbase/0.92.1/org/apache/hadoop/hbase/io/ImmutableBytesWritable.java#ImmutableBytesWritable.get%28%29
> > > > > > >());
> > > > > >
> > > > > > 102 <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hbase/hbase/0.92.1/org/apache/hadoop/hbase/mapreduce/MultiTableOutputFormat.java#102
> > > > > > >
> > > > > >
> > > > > > <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hbase/hbase/0.92.1/org/apache/hadoop/hbase/mapreduce/MultiTableOutputFormat.java#
> > > > > > >
> > > > > >
> > > > > >         table.setAutoFlush
> > > > > > <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hbase/hbase/0.92.1/org/apache/hadoop/hbase/client/HTable.java#HTable.setAutoFlush%28boolean%29
> > > > > > >(false);
> > > > > >
> > > > > > 103 <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hbase/hbase/0.92.1/org/apache/hadoop/hbase/mapreduce/MultiTableOutputFormat.java#103
> > > > > > >
> > > > > >
> > > > > > <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hbase/hbase/0.92.1/org/apache/hadoop/hbase/mapreduce/MultiTableOutputFormat.java#
> > > > > > >
> > > > > >
> > > > > >         tables <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hbase/hbase/0.92.1/org/apache/hadoop/hbase/mapreduce/MultiTableOutputFormat.java#MultiTableOutputFormat.MultiTableRecordWriter.0tables
> > > > > > >.put
> > > > > > <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repository.grepcode.com/java/root/jdk/openjdk/6-b27/java/util/Map.java#Map.put%28org.apache.hadoop.hbase.io.ImmutableBytesWritable%2Corg.apache.hadoop.hbase.client.HTable%29
> > > > > > >(tableName,
> > > > > > table);
> > > > > >
> > > > > > 104 <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hbase/hbase/0.92.1/org/apache/hadoop/hbase/mapreduce/MultiTableOutputFormat.java#104
> > > > > > >
> > > > > >
> > > > > > <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hbase/hbase/0.92.1/org/apache/hadoop/hbase/mapreduce/MultiTableOutputFormat.java#
> > > > > > >
> > > > > >
> > > > > >       }
> > > > > >
> > > > > > 105 <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hbase/hbase/0.92.1/org/apache/hadoop/hbase/mapreduce/MultiTableOutputFormat.java#105
> > > > > > >
> > > > > >
> > > > > > <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hbase/hbase/0.92.1/org/apache/hadoop/hbase/mapreduce/MultiTableOutputFormat.java#
> > > > > > >
> > > > > >
> > > > > >       return tables
> > > > > > <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hbase/hbase/0.92.1/org/apache/hadoop/hbase/mapreduce/MultiTableOutputFormat.java#MultiTableOutputFormat.MultiTableRecordWriter.0tables
> > > > > > >.get
> > > > > > <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repository.grepcode.com/java/root/jdk/openjdk/6-b27/java/util/Map.java#Map.get%28java.lang.Object%29
> > > > > > >(tableName);
> > > > > >
> > > > > > 106 <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hbase/hbase/0.92.1/org/apache/hadoop/hbase/mapreduce/MultiTableOutputFormat.java#106
> > > > > > >
> > > > > >
> > > > > > <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hbase/hbase/0.92.1/org/apache/hadoop/hbase/mapreduce/MultiTableOutputFormat.java#
> > > > > > >
> > > > > >
> > > > > >     }
> > > > > >
> > > > > > 107 <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hbase/hbase/0.92.1/org/apache/hadoop/hbase/mapreduce/MultiTableOutputFormat.java#107
> > > > > > >
> > > > > >
> > > > > > <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hbase/hbase/0.92.1/org/apache/hadoop/hbase/mapreduce/MultiTableOutputFormat.java#
> > > > > > >
> > > > > >
> > > > > > 108 <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hbase/hbase/0.92.1/org/apache/hadoop/hbase/mapreduce/MultiTableOutputFormat.java#108
> > > > > > >
> > > > > >
> > > > > > <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hbase/hbase/0.92.1/org/apache/hadoop/hbase/mapreduce/MultiTableOutputFormat.java#
> > > > > > >
> > > > > >
> > > > > >     @Override <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repository.grepcode.com/java/root/jdk/openjdk/6-b27/java/lang/Override.java#Override
> > > > > > >
> > > > > >
> > > > > > 109 <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hbase/hbase/0.92.1/org/apache/hadoop/hbase/mapreduce/MultiTableOutputFormat.java#109
> > > > > > >
> > > > > >
> > > > > > <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hbase/hbase/0.92.1/org/apache/hadoop/hbase/mapreduce/MultiTableOutputFormat.java#
> > > > > > >
> > > > > >
> > > > > >     public void
> > > > > > <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hbase/hbase/0.92.1/org/apache/hadoop/hbase/mapreduce/MultiTableOutputFormat.java#
> > > > > > >close(TaskAttemptContext
> > > > > > context) throws IOException
> > > > > > <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repository.grepcode.com/java/root/jdk/openjdk/6-b27/java/io/IOException.java#IOException
> > > > > > >
> > > > > > {
> > > > > >
> > > > > > 110 <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hbase/hbase/0.92.1/org/apache/hadoop/hbase/mapreduce/MultiTableOutputFormat.java#110
> > > > > > >
> > > > > >
> > > > > > <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hbase/hbase/0.92.1/org/apache/hadoop/hbase/mapreduce/MultiTableOutputFormat.java#
> > > > > > >
> > > > > >
> > > > > >       for (HTable
> > > > > > <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hbase/hbase/0.92.1/org/apache/hadoop/hbase/client/HTable.java#HTable
> > > > > > >
> > > > > > table : tables <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hbase/hbase/0.92.1/org/apache/hadoop/hbase/mapreduce/MultiTableOutputFormat.java#MultiTableOutputFormat.MultiTableRecordWriter.0tables
> > > > > > >.values
> > > > > > <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repository.grepcode.com/java/root/jdk/openjdk/6-b27/java/util/Map.java#Map.values%28%29
> > > > > > >())
> > > > > > {
> > > > > >
> > > > > > 111 <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hbase/hbase/0.92.1/org/apache/hadoop/hbase/mapreduce/MultiTableOutputFormat.java#111
> > > > > > >
> > > > > >
> > > > > > <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hbase/hbase/0.92.1/org/apache/hadoop/hbase/mapreduce/MultiTableOutputFormat.java#
> > > > > > >
> > > > > >
> > > > > >         table.flushCommits
> > > > > > <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hbase/hbase/0.92.1/org/apache/hadoop/hbase/client/HTable.java#HTable.flushCommits%28%29
> > > > > > >();
> > > > > >
> > > > > > 112 <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hbase/hbase/0.92.1/org/apache/hadoop/hbase/mapreduce/MultiTableOutputFormat.java#112
> > > > > > >
> > > > > >
> > > > > > <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hbase/hbase/0.92.1/org/apache/hadoop/hbase/mapreduce/MultiTableOutputFormat.java#
> > > > > > >
> > > > > >
> > > > > >       }
> > > > > >
> > > > > > 113 <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hbase/hbase/0.92.1/org/apache/hadoop/hbase/mapreduce/MultiTableOutputFormat.java#113
> > > > > > >
> > > > > >
> > > > > > <
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hbase/hbase/0.92.1/org/apache/hadoop/hbase/mapreduce/MultiTableOutputFormat.java#
> > > > > > >
> > > > > >
> > > > > >     }
> > > > > >
> > > > > >
> > > > > > In fact, I would suggest to go through this code here for the
> whole
> > > > > class:
> > > > > >
> > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> http://grepcode.com/file/repo1.maven.org/maven2/org.apache.hbase/hbase/0.92.1/org/apache/hadoop/hbase/mapreduce/MultiTableOutputFormat.java#MultiTableOutputFormat.MultiTableRecordWriter.getTable%28org.apache.hadoop.hbase.io.ImmutableBytesWritable%29
> > > > > >
> > > > > >
> > > > > >
> > > > > > It is different from TableOutputFormat approach where you do need
> > to
> > > > > > intialize the table by using the Util class.
> > > > > >
> > > > > >
> > > > > >
> > > > > > Regards,
> > > > > >
> > > > > > Shahab
> > > > > >
> > > > > >
> > > > > >
> > > > > > On Tue, Aug 26, 2014 at 2:29 PM, yeshwanth kumar <
> > > > [email protected]>
> > > > > > wrote:
> > > > > >
> > > > > > > hi ted,
> > > > > > >
> > > > > > > i need to process the data in table i1, and then i need to
> write
> > > the
> > > > > > > results to tables i1 and i2
> > > > > > > so input for the mapper in my mapreduce job is from hbase
> table,
> > i1
> > > > > > > whereas in WALPlayer input is HLogInputFormat,
> > > > > > >
> > > > > > > if i remove the statement as you said and specify  the
> > inputformat
> > > > > > > as TableInputFormat it is throwing "No table was provided "
> > > Exception
> > > > > > > if i specify the input table as in the statements
> > > > > > >
> > > > > > > TableMapReduceUtil.initTableMapperJob(otherArgs[0], scan,
> > > > > > > EntitySearcherMapper.class, ImmutableBytesWritable.class,
> > > Put.class,
> > > > > > > job);//otherArgs[0]=i1
> > > > > > >
> > > > > > > mapper is not considering other table,
> > > > > > > any suggestions to resolve  this issue,
> > > > > > >
> > > > > > > thanks,
> > > > > > > yeshwanth
> > > > > > >
> > > > > > >
> > > > > > >
> > > > > > >
> > > > > > > On Tue, Aug 26, 2014 at 10:39 PM, Ted Yu <[email protected]>
> > > > wrote:
> > > > > > >
> > > > > > > > Please take a look at WALPlayer.java in hbase where you can
> > find
> > > > > > example
> > > > > > > of
> > > > > > > > how MultiTableOutputFormat is used.
> > > > > > > >
> > > > > > > > Cheers
> > > > > > > >
> > > > > > > >
> > > > > > > > On Tue, Aug 26, 2014 at 10:04 AM, yeshwanth kumar <
> > > > > > [email protected]
> > > > > > > >
> > > > > > > > wrote:
> > > > > > > >
> > > > > > > > > hi ted,
> > > > > > > > >
> > > > > > > > > how can we intialise the mapper if i comment out those
> lines
> > > > > > > > >
> > > > > > > > >
> > > > > > > > >
> > > > > > > > > On Tue, Aug 26, 2014 at 10:08 PM, Ted Yu <
> > [email protected]>
> > > > > > wrote:
> > > > > > > > >
> > > > > > > > > > TableMapReduceUtil.initTableMapperJob(otherArgs[0], scan,
> > > > > > > > > > EntitySearcherMapper.class, ImmutableBytesWritable.class,
> > > > > > Put.class,
> > > > > > > > > > job);//otherArgs[0]=i1
> > > > > > > > > >
> > > > > > > > > > You're initializing with table 'i1'
> > > > > > > > > > Please remove the above call and try again.
> > > > > > > > > >
> > > > > > > > > > Cheers
> > > > > > > > > >
> > > > > > > > > >
> > > > > > > > > > On Tue, Aug 26, 2014 at 9:18 AM, yeshwanth kumar <
> > > > > > > > [email protected]>
> > > > > > > > > > wrote:
> > > > > > > > > >
> > > > > > > > > > > hi i am running  HBase 0.94.20  on Hadoop 2.2.0
> > > > > > > > > > >
> > > > > > > > > > > i am using MultiTableOutputFormat,
> > > > > > > > > > > for writing processed output to two different tables in
> > > > hbase.
> > > > > > > > > > >
> > > > > > > > > > > here's the code snippet
> > > > > > > > > > >
> > > > > > > > > > > private ImmutableBytesWritable tab_cr = new
> > > > > > ImmutableBytesWritable(
> > > > > > > > > > > Bytes.toBytes("i1")); private ImmutableBytesWritable
> > > tab_cvs
> > > > =
> > > > > > new
> > > > > > > > > > > ImmutableBytesWritable( Bytes.toBytes("i2"));
> > > > > > > > > > >
> > > > > > > > > > > @Override
> > > > > > > > > > > public void map(ImmutableBytesWritable row, final
> Result
> > > > value,
> > > > > > > > > > > final Context context) throws IOException,
> > > > > InterruptedException {
> > > > > > > > > > >
> > > > > > > > > > > -----------------------------------------
> > > > > > > > > > > Put pcvs = new Put(entry.getKey().getBytes());
> > > > > > > > > > >
> > > > pcvs.add("cf".getBytes(),"type".getBytes(),column.getBytes());
> > > > > > > > > > > Put put = new Put(value.getRow());
> > > > > > > > > > > put.add("Entity".getBytes(), "json".getBytes(),
> > > > > > > > > > > entry.getValue().getBytes());
> > > > > > > > > > > context.write(tab_cr, put);// table i1
> > > context.write(tab_cvs,
> > > > > > > > > > pcvs);//table
> > > > > > > > > > > i2
> > > > > > > > > > >
> > > > > > > > > > > }
> > > > > > > > > > >
> > > > > > > > > > > job.setJarByClass(EntitySearcherMR.class);
> > > > > > > > > > > job.setMapperClass(EntitySearcherMapper.class);
> > > > > > > > > > > job.setOutputFormatClass(MultiTableOutputFormat.class);
> > > Scan
> > > > > > scan =
> > > > > > > > new
> > > > > > > > > > > Scan(); scan.setCacheBlocks(false);
> > > > > > > > > > > TableMapReduceUtil.initTableMapperJob(otherArgs[0],
> scan,
> > > > > > > > > > > EntitySearcherMapper.class,
> ImmutableBytesWritable.class,
> > > > > > > Put.class,
> > > > > > > > > > > job);//otherArgs[0]=i1
> > > > > > > > > > TableMapReduceUtil.initTableReducerJob(otherArgs[0],
> > > > > > > > > > > null, job); job.setNumReduceTasks(0);
> > > > > > > > > > >
> > > > > > > > > > > mapreduce job fails by saying nosuchcolumnfamily "cf"
> > > > > exception,
> > > > > > in
> > > > > > > > > table
> > > > > > > > > > > i1
> > > > > > > > > > > i am writing data to two different columnfamilies one
> in
> > > each
> > > > > > > table,
> > > > > > > > cf
> > > > > > > > > > > belongs to table i2.
> > > > > > > > > > > does the columnfamilies should present in both tables??
> > > > > > > > > > > is there anything i am missing
> > > > > > > > > > > can someone point me in the right direction
> > > > > > > > > > >
> > > > > > > > > > > thanks,
> > > > > > > > > > > yeshwanth.
> > > > > > > > > > >
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> > --
> > You received this message because you are subscribed to the Google Groups
> > "NoSQL Databases" group.
> > To unsubscribe from this group and stop receiving emails from it, send an
> > email to [email protected].
> > For more options, visit https://groups.google.com/d/optout.
> >
>

Reply via email to