Hi Prasant, You should try to analyze the error message first. It says no delimiter found becuase importtsv expects tsv file ( tab separated values ) and not csv( comma separated values) so if you just replace the commas with tab in the file and then try it should work. Exceptions always (almost) say something about them, so always try to read through them.
Regards, Gaurav ________________________________________ From: Prashant Sharma [[email protected]] Sent: Tuesday, June 14, 2011 7:38 PM To: [email protected] Subject: Re: Problem with importtsv on trsnferring data from HDFS to hbase table: My input file is a CSV with 3 fields.. uniqueID,year,name is there a problem with the format? I have checked it like 10 times.. everything seem fine ... cant figure whats wrong. Any input would be very helpful. Thanks in advance, Prashant On Tue, Jun 14, 2011 at 3:44 PM, Prashant Sharma <[email protected]>wrote: > Hi, > I am getting following errors while trying to transfer data from hdfs to > hbase. > > Table at hbase: > hbase(main):007:0> describe 'movies' > DESCRIPTION ENABLED > {NAME => 'movies', FAMILIES => [{NAME => 'HBASE_ROW true > _KEY', BLOOMFILTER => 'NONE', REPLICATION_SCOPE => > '0', COMPRESSION => 'NONE', VERSIONS => '3', TTL => > '2147483647', BLOCKSIZE => '65536', IN_MEMORY => ' > false', BLOCKCACHE => 'true'}, {NAME => 'name', BLO > OMFILTER => 'NONE', REPLICATION_SCOPE => '0', COMPR > ESSION => 'NONE', VERSIONS => '3', TTL => '21474836 > 47', BLOCKSIZE => '65536', IN_MEMORY => 'false', BL > OCKCACHE => 'true'}, {NAME => 'year', BLOOMFILTER = > > 'NONE', REPLICATION_SCOPE => '0', COMPRESSION => > 'NONE', VERSIONS => '3', TTL => '2147483647', BLOCK > SIZE => '65536', IN_MEMORY => 'false', BLOCKCACHE = > > 'true'}]} > 1 row(s) in 0.1820 seconds > > > hbase(main):006:0> scan 'movies' > ROW COLUMN+CELL > 1 column=name:, timestamp=1308044917482, value=new > 1 column=year:, timestamp=1308044926957, value=2055 > 1 row(s) in 0.0710 seconds > > Command line:hadoop@hadoop:~/work/hadoop/hadoop-0.20.203.0$ bin/hadoop jar > ../../hbase/hbase-0.90.3/hbase-0.90.3.jar importtsv > -Dimporttscolumns=HBASE_ROW_KEY,year,name movies > /user/hadoop/movies/movie.csv -Dimporttsv.separator=',' 2>log > > Output on stderr: > ..some lines ommitted.. > 11/06/14 15:35:21 INFO zookeeper.ZooKeeper: Client > environment:java.library.path=/home/hadoop/work/hadoop/hadoop-0.20.203.0/bin/../lib/native/Linux-i386-32 > 11/06/14 15:35:21 INFO zookeeper.ZooKeeper: Client > environment:java.io.tmpdir=/tmp > 11/06/14 15:35:21 INFO zookeeper.ZooKeeper: Client > environment:java.compiler=<NA> > 11/06/14 15:35:21 INFO zookeeper.ZooKeeper: Client environment:os.name > =Linux > 11/06/14 15:35:21 INFO zookeeper.ZooKeeper: Client environment:os.arch=i386 > 11/06/14 15:35:21 INFO zookeeper.ZooKeeper: Client > environment:os.version=2.6.35-25-generic > 11/06/14 15:35:21 INFO zookeeper.ZooKeeper: Client environment:user.name > =hadoop > 11/06/14 15:35:21 INFO zookeeper.ZooKeeper: Client > environment:user.home=/home/hadoop > 11/06/14 15:35:21 INFO zookeeper.ZooKeeper: Client > environment:user.dir=/home/hadoop/work/hadoop/hadoop-0.20.203.0 > 11/06/14 15:35:21 INFO zookeeper.ZooKeeper: Initiating client connection, > connectString=localhost:2181 sessionTimeout=180000 watcher=hconnection > 11/06/14 15:35:21 INFO zookeeper.ClientCnxn: Opening socket connection to > server localhost/0:0:0:0:0:0:0:1:2181 > 11/06/14 15:35:21 INFO zookeeper.ClientCnxn: Socket connection established > to localhost/0:0:0:0:0:0:0:1:2181, initiating session > 11/06/14 15:35:21 INFO zookeeper.ClientCnxn: Session establishment complete > on server localhost/0:0:0:0:0:0:0:1:2181, sessionid = 0x1308d8861600014, > negotiated timeout = 180000 > 11/06/14 15:35:22 INFO mapreduce.TableOutputFormat: Created table instance > for movies > 11/06/14 15:35:22 INFO input.FileInputFormat: Total input paths to process > : 1 > 11/06/14 15:35:22 INFO mapred.JobClient: Running job: job_201106141233_0042 > 11/06/14 15:35:23 INFO mapred.JobClient: map 0% reduce 0% > 11/06/14 15:38:16 INFO mapred.JobClient: Task Id : > attempt_201106141233_0042_m_000000_0, Status : FAILED > java.lang.NullPointerException > at > org.apache.hadoop.hbase.mapreduce.TableOutputFormat$TableRecordWriter.close(TableOutputFormat.java:107) > at > org.apache.hadoop.mapred.MapTask$NewDirectOutputCollector.close(MapTask.java:650) > at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:765) > at org.apache.hadoop.mapred.MapTask.run(MapTask.java:369) > at org.apache.hadoop.mapred.Child$4.run(Child.java:259) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:396) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1059) > at org.apache.hadoop.mapred.Child.main(Child.java:253) > > attempt_201106141233_0042_m_000000_0: Bad line at offset: 0: > attempt_201106141233_0042_m_000000_0: No delimiter > attempt_201106141233_0042_m_000000_0: Bad line at offset: 34: > attempt_201106141233_0042_m_000000_0: No delimiter > attempt_201106141233_0042_m_000000_0: Bad line at offset: 51: > .......................... x33123 lines > > > ---------------------------------------------------------------- > This message was sent using IMP, the Internet Messaging Program. > > > -- Prashant Sharma Development Engineer Pramati Technologies Begumpet "Hare Krishna"
