Hi Harsh,

If I try without using the HBASE_ROW_KEY, it says must specify the 
HBASE_ROW_KEY and does not goes further.

With the HBASE_ROW_KEY added with column names I get no error but following 
output

12/03/07 02:52:30 INFO mapreduce.HFileOutputFormat: Looking up current regions 
for table org.apache.hadoop.hbase.client.HTable@6bade9
12/03/07 02:52:30 INFO mapreduce.HFileOutputFormat: Configuring 1 reduce 
partitions to match current region count
12/03/07 02:52:30 INFO mapreduce.HFileOutputFormat: Writing partition 
information to hdfs://master/user/hadoop/partitions_1331088750667
12/03/07 02:52:30 INFO util.NativeCodeLoader: Loaded the native-hadoop library
12/03/07 02:52:30 INFO zlib.ZlibFactory: Successfully loaded & initialized 
native-zlib library
12/03/07 02:52:30 INFO compress.CodecPool: Got brand-new compressor
12/03/07 02:52:31 INFO mapreduce.HFileOutputFormat: Incremental table output 
configured.
12/03/07 02:52:31 INFO input.FileInputFormat: Total input paths to process : 1
12/03/07 02:52:32 INFO mapred.JobClient: Running job: job_201203062231_0002
12/03/07 02:52:33 INFO mapred.JobClient:  map 0% reduce 0%
12/03/07 02:52:49 INFO mapred.JobClient:  map 100% reduce 0%
12/03/07 02:53:01 INFO mapred.JobClient:  map 100% reduce 100%
12/03/07 02:53:06 INFO mapred.JobClient: Job complete: job_201203062231_0002
12/03/07 02:53:06 INFO mapred.JobClient: Counters: 25
12/03/07 02:53:06 INFO mapred.JobClient:   Job Counters
12/03/07 02:53:06 INFO mapred.JobClient:     Launched reduce tasks=1
12/03/07 02:53:06 INFO mapred.JobClient:     SLOTS_MILLIS_MAPS=13349
12/03/07 02:53:06 INFO mapred.JobClient:     Total time spent by all reduces 
waiting after reserving slots (ms)=0
12/03/07 02:53:06 INFO mapred.JobClient:     Total time spent by all maps 
waiting after reserving slots (ms)=0
12/03/07 02:53:06 INFO mapred.JobClient:     Rack-local map tasks=1
12/03/07 02:53:06 INFO mapred.JobClient:     Launched map tasks=1
12/03/07 02:53:06 INFO mapred.JobClient:     SLOTS_MILLIS_REDUCES=10467
12/03/07 02:53:06 INFO mapred.JobClient:   ImportTsv
12/03/07 02:53:06 INFO mapred.JobClient:     Bad Lines=6
12/03/07 02:53:06 INFO mapred.JobClient:   File Output Format Counters
12/03/07 02:53:06 INFO mapred.JobClient:     Bytes Written=0
12/03/07 02:53:06 INFO mapred.JobClient:   FileSystemCounters
12/03/07 02:53:06 INFO mapred.JobClient:     FILE_BYTES_READ=282
12/03/07 02:53:06 INFO mapred.JobClient:     HDFS_BYTES_READ=102
12/03/07 02:53:06 INFO mapred.JobClient:     FILE_BYTES_WRITTEN=65025
12/03/07 02:53:06 INFO mapred.JobClient:   File Input Format Counters
12/03/07 02:53:06 INFO mapred.JobClient:     Bytes Read=123
12/03/07 02:53:06 INFO mapred.JobClient:   Map-Reduce Framework
12/03/07 02:53:06 INFO mapred.JobClient:     Reduce input groups=0
12/03/07 02:53:06 INFO mapred.JobClient:     Map output materialized bytes=6
12/03/07 02:53:06 INFO mapred.JobClient:     Combine output records=0
12/03/07 02:53:06 INFO mapred.JobClient:     Map input records=6
12/03/07 02:53:06 INFO mapred.JobClient:     Reduce shuffle bytes=0
12/03/07 02:53:06 INFO mapred.JobClient:     Reduce output records=0
12/03/07 02:53:06 INFO mapred.JobClient:     Spilled Records=0
12/03/07 02:53:06 INFO mapred.JobClient:     Map output bytes=0
12/03/07 02:53:06 INFO mapred.JobClient:     Combine input records=0
12/03/07 02:53:06 INFO mapred.JobClient:     Map output records=0
12/03/07 02:53:06 INFO mapred.JobClient:     SPLIT_RAW_BYTES=102
12/03/07 02:53:06 INFO mapred.JobClient:     Reduce input records=0


Kind regards,
Keshav C Savant 



-----Original Message-----
From: Harsh J [mailto:[email protected]] 
Sent: Tuesday, March 06, 2012 5:29 PM
To: Savant, Keshav
Cc: [email protected]
Subject: Re: Inserting Data from CSV into HBase

Hi,

Can you share the exact error message/stack trace you get?

One observation though: Given that data, I notice it has just 3 columns but you 
are specifying four elements in the columns argument of the importer:

Data: "3,data31,data32" (3 elements)
Cols passed: "-Dimporttsv.columns=HBASE_ROW_KEY,SerialNumber,Column1,Column2"
(4 elements)

Perhaps this may be why your importtsv complains about bad lines.

On Tue, Mar 6, 2012 at 3:32 PM, Savant, Keshav <[email protected]> 
wrote:
> Hi,
>
> I tried bulk uploading and it ran well with TSV files, we first ran importtsv 
> and then completebulkload, after doing these two steps I can scan my HBase 
> table and see the data. I can also see the data when I traverse HDFS of my 
> Hadoop cluster using web browser.
>
> But when I try to upload my CSVs in a folder, I get bad lines for all 
> the lines of my CSV files. I use following command to upload my CSVs 
> on my local file system to HDFS,
>
> HADOOP_CLASSPATH=`hbase classpath` $HADOOP_HOME/bin/hadoop jar 
> /hbase_home/hbase-0.92.0/hbase-0.92.0.jar importtsv  
> -Dimporttsv.bulk.output=/my_output_dir 
> -Dimporttsv.columns=HBASE_ROW_KEY,SerialNumber,Column1,Column2 my_table 
> file:/my_csv/data.txt '-Dimporttsv.separator=,'
>
> my csv file is of following format
>
> 1,data11,data12
> 2,data21,data22
> 3,data31,data32
> .....
> .....
>
> And my HBase table has 3 columns
>
>
> Please let me know what is the exact problem and how this can be resolved?
>
> Kind regards,
> Keshav
>
>
>
> -----Original Message-----
> From: Savant, Keshav
> Sent: Friday, March 02, 2012 7:02 PM
> To: [email protected]
> Cc: '[email protected]'
> Subject: RE: Inserting Data from CSV into HBase
>
> Hi Harsh,
>
> Thanks for your response, I don't get any error using the code mentioned in 
> that URL. I will get back to you after analyzing the tools suggested by you.
> Thanks again.
>
>
> Kind regards,
> Keshav C Savant
>
> -----Original Message-----
> From: Harsh J [mailto:[email protected]]
> Sent: Friday, March 02, 2012 6:51 PM
> To: [email protected]
> Subject: Re: Inserting Data from CSV into HBase
>
> Hi,
>
> You may use the importtsv tool and the bulk-load utilities in HBase to 
> achieve this fast-and-easy.
>
> This is detailed at http://hbase.apache.org/bulk-loads.html (See section 
> about importtsv along the bottom) and also under section "Using the importtsv 
> tool" on Page 460 of Lars George's "HBase: The Definitive Guide" (O'Reilly).
>
> Also when you say something didn't work, please also supply any errors you 
> encountered and the configuration you used. Its hard to help without those.
>
> On Fri, Mar 2, 2012 at 6:24 PM, Savant, Keshav 
> <[email protected]> wrote:
>> Hi All,
>>
>> I am looking for a way so that I can map my existing CSV file to HBase 
>> table, basically for each column family I want only one value (just like 
>> RDBMS).
>>
>> Just to illustrate more suppose I define a HBase table as
>>
>> create 'inventory', 'item', 'supplier', 'quantity'
>> (here table name is inventory and it has three columns named as item, 
>> supplier and quantity)
>>
>> Now I want to load my N number of CSVs in following format into this 
>> HBase table
>>
>> Burger,abc confectionary,100
>> Pizza,xyz bakers,50
>> ...
>> ...
>> ...
>>
>> Here I want to put the data of CSV into my inventory table on HBase, the 
>> number of lines in a CSV and even number of CSVs are dynamic, and this will 
>> be a continuous process.
>>
>> What I want to know that, do we have any way by which we can achieve above 
>> goal, I tried SampleUploader as specified on 
>> http://svn.apache.org/repos/asf/hbase/trunk/src/examples/mapreduce/org/apache/hadoop/hbase/mapreduce/SampleUploader.java,
>>  but it did not worked and data does not gets populated in HBase table 
>> though the program ran successfully.
>>
>> Please suggest on this, any help is appreciated.
>>
>> Kind regards,
>> Keshav C Savant
>>
>> _____________
>> The information contained in this message is proprietary and/or 
>> confidential. If you are not the intended recipient, please: (i) delete the 
>> message and all copies; (ii) do not disclose, distribute or use the message 
>> in any manner; and (iii) notify the sender immediately. In addition, please 
>> be aware that any message addressed to our domain is subject to archiving 
>> and review by persons other than the intended recipient. Thank you.
>
>
>
> --
> Harsh J
>
> _____________
> The information contained in this message is proprietary and/or confidential. 
> If you are not the intended recipient, please: (i) delete the message and all 
> copies; (ii) do not disclose, distribute or use the message in any manner; 
> and (iii) notify the sender immediately. In addition, please be aware that 
> any message addressed to our domain is subject to archiving and review by 
> persons other than the intended recipient. Thank you.



--
Harsh J

_____________
The information contained in this message is proprietary and/or confidential. 
If you are not the intended recipient, please: (i) delete the message and all 
copies; (ii) do not disclose, distribute or use the message in any manner; and 
(iii) notify the sender immediately. In addition, please be aware that any 
message addressed to our domain is subject to archiving and review by persons 
other than the intended recipient. Thank you.

Reply via email to