Hi,
Well, I removed all the spaces between the column names and it worked.
Another problem I'm facing now is that when I use the importtsv.bulk.output
option it throws another error:
INFO mapred.JobClient: Task Id : attempt_201106161041_0003_m_000000_2,
Status : FAILED
java.lang.IllegalArgumentException: Can't read partitions file
at
org.apache.hadoop.hbase.mapreduce.hadoopbackport.TotalOrderPartitioner.setConf(TotalOrderPartitioner.java:111)
at
org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:62)
at
org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:117)
at
org.apache.hadoop.mapred.MapTask$NewOutputCollector.<init>(MapTask.java:527)
at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:613)
at org.apache.hadoop.mapred.MapTask.run(MapTask.java:305)
at org.apache.hadoop.mapred.Child.main(Child.java:170)
Caused by: java.io.FileNotFoundException: File _partition.lst does not
exist.
at
org.apache.hadoop.fs.RawLocalFileSystem.getFileStatus(RawLocalFileSystem.java:361)
at
org.apache.hadoop.fs.FilterFileSystem.getFileStatus(FilterFileSystem.java:245)
at org.apache.hadoop.fs.FileSystem.getLength(FileSystem.java:676)
at
org.apache.hadoop.io.SequenceFile$Reader.<init>(SequenceFile.java:1417)
at
org.apache.hadoop.io.SequenceFile$Reader.<init>(SequenceFile.java:1412)
at
org.apache.hadoop.hbase.mapreduce.hadoopbackport.TotalOrderPartitioner.readPartitions(TotalOrderPartitioner.java:296)
*I* am running a 5 machine cluster and all the mapred.xml files include the
jobtracker address. Also, I have attached the mapred-site.xml with this
mail.
I tried to validate this file with xmllint and I got the error
"mapred-site.xml:4: validity error : Validation failed: no DTD found !"
The mapred-default.xml doesn't look much different from the one I have
attached. The mapred-default.xml doesn't seem to include a dtd either.
Thanks,
JR
On Wed, Jun 15, 2011 at 12:13 PM, Prashant Sharma <[email protected]
> wrote:
> Did you write the table name ? and remove an extra space after hbase row
> key. I think that must be th reason
> ( I am not an expert , but have struggled alot with it. )
> Thanks,
> Prashant
> On Wed, Jun 15, 2011 at 11:59 AM, James Ram <[email protected]> wrote:
>
> > Hi,
> >
> > I'm having trouble with using the importtsv tool.
> > I ran the following command:
> >
> > hadoop jar hadoop_sws/hbase-0.90.0/hbase-0.90.0.jar importtsv
> > -Dimporttsv.columns=HBASE_ROW_KEY ,b_info:name, b_info:contactNo,
> > b_info:dob, b_info:email, b_info:marital_status, b_info:p_address,
> > b_info:photo, demo_info:caste, demo_info:gender, demo_info:religion,
> > beneficiarydetails user/sao_user/UserProfile.tsv
> >
> >
> > And I got the following error:
> > ERROR: One or more columns in addition to the row key are required
> > Usage: importtsv -Dimporttsv.columns=a,b,c <tablename> <inputdir>
> >
> > I can't really find out the problem in the way I wrote the command. I
> have
> > added the column names after the HBASE_ROW_KEY.
> > Where did I go wrong?
> >
> > --
> > With Regards,
> > Jr.
> >
>
--
With Regards,
Jr.
<?xml version="1.0"?>
<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
<!-- Put site-specific property overrides in this file. -->
<configuration>
<property>
<name>mapred.job.tracker</name>
<value>160.110.185.93:8021</value>
<final>true</final>
</property>
<property>
<name>mapred.local.dir</name>
<value>/home/sao_user/hadoop_sws/mapred/local</value>
<final>true</final>
</property>
<property>
<name>mapred.system.dir</name>
<value>/tmp/hadoop-sao_user/mapred/system/</value>
<final>true</final>
</property>
<property>
<name>mapred.tasktracker.map.tasks.maximum</name>
<value>2</value>
<final>true</final>
</property>
<property>
<name>mapred.tasktracker.reduce.tasks.maximum</name>
<value>2</value>
<final>true</final>
</property>
<property>
<name>mapred.child.java.opts</name>
<value>-Xmx512M</value>
</property>
</configuration>