Hi,

Regarding the sqoop import: I noticed you wrote -table instead of
--table (1 dash instead of 2)

Ruslan

On Wed, Jul 4, 2012 at 12:41 PM, Bejoy Ks <bejoy...@yahoo.com> wrote:
> Hi Yogesh
>
> To add on, looks like the table definition doesn't match with data as well.
>
> Your table defn has 4 columns defined, with 4th column as int
>
> describe formatted letstry;
> OK
> # col_name                data_type               comment
>
> rollno                  int                     None
> name                    string                  None
> numbr                   int                     None
> sno                     int                     None
>
>
> But the data has 5 columns with the 4th column as String
>
> 1,John,123,abc,2
>
>
> Also When you create the table, make sure to specify the right field
> separator
>
> ....
> ROW FORMAT DELIMITED FIELDS TERMINATED BY ','
>  STORED AS TEXTFILE
>
>
> Regards
> Bejoy KS
>
> ________________________________
> From: Bejoy Ks <bejoy...@yahoo.com>
> To: "user@hive.apache.org" <user@hive.apache.org>
> Sent: Wednesday, July 4, 2012 1:59 PM
> Subject: Re: Hive upload
>
> Hi Yogesh
>
> Looks like Sqoop import from rdbms to hdfs is getting successful but is
> failing at hive create table. You are seeing data in hive ware house because
> you have specified that as your target dir in sqoop import (--target-dir
> /user/hive/warehouse/new). It is recommended to use a different target dir
> while doing sqoop import other than the hive warehouse dir.
>
> Can you post in the full console log of sqoop with --verbose logging
> enabled. It can give some clues.
>
>
> With the second issue, You already have your data in
> '/user/hive/warehouse/letstry/' which is the location for the hive table
> 'letstry'. Why you still want to do a LOAD DATA again in there?
>
> If you are doing a SQOOP import of that, Again it is recommended to use a
> different target dir other than hive ware house dir. It'll help you avoid
> some confusions as well.
>
>
> ________________________________
> From: yogesh dhari <yogeshdh...@live.com>
> To: hive request <user@hive.apache.org>
> Sent: Wednesday, July 4, 2012 1:40 PM
> Subject: RE: Hive upload
>
>
> Hi Bejoy,
>
> Thank you very much for your response,
>
> 1)
>
> A) When I run command  show tables it doesn't show  newhive table.
> B) Yes the the newhive directory is present into /user/hive/warehouse and
> also containing the values imported from RDBMS
>
> Please suggest and give me an example for the sqoop import command according
> to you for this case.
>
>
> 2)
>
> A) Here is the command
>
> describe formatted letstry;
> OK
> # col_name                data_type               comment
>
> rollno                  int                     None
> name                    string                  None
> numbr                   int                     None
> sno                     int                     None
>
> # Detailed Table Information
> Database:               default
> Owner:                  mediaadmin
> CreateTime:             Tue Jul 03 17:06:27 GMT+05:30 2012
> LastAccessTime:         UNKNOWN
> Protect Mode:           None
> Retention:              0
> Location:               hdfs://localhost:9000/user/hive/warehouse/letstry
> Table Type:             MANAGED_TABLE
> Table Parameters:
>     transient_lastDdlTime    1341315550
>
> # Storage Information
> SerDe Library:          org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
> InputFormat:            org.apache.hadoop.mapred.TextInputFormat
> OutputFormat:
> org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
> Compressed:             No
> Num Buckets:            -1
> Bucket Columns:         []
> Sort Columns:           []
> Storage Desc Params:
>     serialization.format    1
> Time taken: 0.101 seconds
>
>
> B) hadoop dfs -ls /user/hive/warehouse/letstry/
> Found 1 items
> -rw-r--r--   1 mediaadmin supergroup         17 2012-07-02 12:05
> /user/hive/warehouse/letstry/part-m-00000
>
> hadoop dfs -cat /user/hive/warehouse/letstry/part-m-00000
> 1,John,123,abc,2
>
>
>
>
> Here data is present but when I upload it to Hive it gets deleted from HDFS
> and in Hive value appers NULL instead of  ( 1,John,123,abc,2). and I didn't
> understad your point regarding correct data format? ( this data was imported
> from Mysql table)
> And what kind of confugration neede in sqoop
>
> Please suggest and help
>
>
> Greetings
> Yogesh Kumar
>
>
>
>
>
> ________________________________
> Subject: Re: Hive upload
> To: user@hive.apache.org
> From: bejoy...@yahoo.com
> Date: Wed, 4 Jul 2012 05:58:41 +0000
>
> Hi Yogesh
>
> The first issue (sqoop one).
> 1) Is the table newhive coming when you list tables using 'show table'?
> 2) Are you seeing a directory 'newhive' in your hive warte house dir(usually
> /usr/hive/warehouse)?
>
> If not sqoop is failing to create hive tables /load data into them. Only
> sqoop import to hdfs is getting successful the hive part is failing.
>
> If hive in stand alone mode works as desired you need to check the sqoop
> configurations.
>
> Regarding the second issue, can you check the storage location of NewTable
> and check whether there are files within. If so then do a 'cat' of those
> files and see whether it has the correct data format.
>
> You can get the location of your table from the following command
> describe formatted NewTable;
> Regards
> Bejoy KS
>
> Sent from handheld, please excuse typos.
> ________________________________
> From: yogesh dhari <yogeshdh...@live.com>
> Date: Wed, 4 Jul 2012 11:09:02 +0530
> To: hive request<user@hive.apache.org>
> ReplyTo: user@hive.apache.org
> Subject: Hive upload
>
> Hi all,
>
> I am trying to upload the tables from RDBMS to hive through sqoop, hive
> imports successfully. but i didn't find any table in hive that imported
> table gets uploaded into hdfs idr /user/hive/warehouse
> I want it to be present into hive, I used this command
>
> sqoop import --connect jdbc:mysql://localhost:3306/Demo --username sqoop1
> --password SQOOP1 -table newone --hive-table newhive --create-hive-table
> --hive-import --target-dir /user/hive/warehouse/new
>
>
> And another thing is,
> If I upload any file or table from HDFS or from Local then its uploads but
> data doesn't show in Hive table,
>
> If I run command
> Select * from NewTable;
> it reflects
>
> Null     Null     Null    Null
>
>
> although the real data is
>
> Yogesh    4    Bangalore   1234
>
>
> Please Suggest and help
>
> Regards
> Yogesh Kumar
>
>
>
>



-- 
Best Regards,
Ruslan Al-Fakikh

Reply via email to