Hi Yogesh,
could you provide log output generated with --verbose flag set on?

In meantime - it seems that hive upload was successful and if the data are 
available on the HDFS in expected location, then I would suggest checking 
configured hive metastore (is it default derby?). Are both sqoop and hive 
command using same hive metastore?

Jarcec

On Jul 6, 2012, at 8:45 AM, <[email protected]> 
<[email protected]> wrote:

> Hi all,
> 
> Please help me out regarding this issue.
> 
> I have created a table in Mysql by name Dummy and it has 2 columns, and 1 row 
> of data
> I want to upload that table into Hive using Sqoop tool.
> I used this command  
> 
> 
> sqoop import --connect jdbc:mysql://localhost:3306/Demo --username sqoop1 
> --password SQOOP1 -table Dummy --hive-table dummyhive  --create-hive-table  
> --hive-import --hive-home HADOOP/hive
> 
> 
> The table has been succesfully uploaded into HDFS  /user/hive/warehouse
> but when I run command in Hive 
> 
> Show Tables;
> 
> I don't find dummyhive table in it.
> 
> Please suggest and Help
> 
> 
> Details of the command and output
> 
> mediaadmins-iMac-2:hive mediaadmin$ sqoop import --connect 
> jdbc:mysql://localhost:3306/Demo --username sqoop1 --password SQOOP1 -table 
> Dummy --hive-table dummyhive  --create-hive-table  --hive-import --hive-home 
> HADOOP/hive
> 12/07/05 11:09:15 WARN tool.BaseSqoopTool: Setting your password on the 
> command-line is insecure. Consider using -P instead.
> 12/07/05 11:09:15 INFO tool.BaseSqoopTool: Using Hive-specific delimiters for 
> output. You can override
> 12/07/05 11:09:15 INFO tool.BaseSqoopTool: delimiters with 
> --fields-terminated-by, etc.
> 12/07/05 11:09:15 INFO manager.MySQLManager: Preparing to use a MySQL 
> streaming resultset.
> 12/07/05 11:09:15 INFO tool.CodeGenTool: Beginning code generation
> 12/07/05 11:09:16 INFO manager.SqlManager: Executing SQL statement: SELECT 
> t.* FROM `Dummy` AS t LIMIT 1
> 12/07/05 11:09:16 INFO orm.CompilationManager: HADOOP_HOME is 
> /HADOOP/hadoop-0.20.2/bin/..
> 12/07/05 11:09:16 INFO orm.CompilationManager: Found hadoop core jar at: 
> /HADOOP/hadoop-0.20.2/bin/../hadoop-0.20.2-core.jar
> Note: 
> /tmp/sqoop-mediaadmin/compile/382d1c58323cea76efd197632bebbfcd/Dummy.java 
> uses or overrides a deprecated API.
> Note: Recompile with -Xlint:deprecation for details.
> 12/07/05 11:09:17 INFO orm.CompilationManager: Writing jar file: 
> /tmp/sqoop-mediaadmin/compile/382d1c58323cea76efd197632bebbfcd/Dummy.jar
> 12/07/05 11:09:17 WARN manager.MySQLManager: It looks like you are importing 
> from mysql.
> 12/07/05 11:09:17 WARN manager.MySQLManager: This transfer can be faster! Use 
> the --direct
> 12/07/05 11:09:17 WARN manager.MySQLManager: option to exercise a 
> MySQL-specific fast path.
> 12/07/05 11:09:17 INFO manager.MySQLManager: Setting zero DATETIME behavior 
> to convertToNull (mysql)
> 12/07/05 11:09:17 INFO mapreduce.ImportJobBase: Beginning import of Dummy
> 12/07/05 11:09:18 INFO db.DataDrivenDBInputFormat: BoundingValsQuery: SELECT 
> MIN(`Sno`), MAX(`Sno`) FROM `Dummy`
> 12/07/05 11:09:18 INFO mapred.JobClient: Running job: job_201207051104_0001
> 12/07/05 11:09:19 INFO mapred.JobClient:  map 0% reduce 0%
> 12/07/05 11:09:33 INFO mapred.JobClient:  map 100% reduce 0%
> 12/07/05 11:09:35 INFO mapred.JobClient: Job complete: job_201207051104_0001
> 12/07/05 11:09:35 INFO mapred.JobClient: Counters: 5
> 12/07/05 11:09:35 INFO mapred.JobClient:   Job Counters 
> 12/07/05 11:09:35 INFO mapred.JobClient:     Launched map tasks=1
> 12/07/05 11:09:35 INFO mapred.JobClient:   FileSystemCounters
> 12/07/05 11:09:35 INFO mapred.JobClient:     HDFS_BYTES_WRITTEN=8
> 12/07/05 11:09:35 INFO mapred.JobClient:   Map-Reduce Framework
> 12/07/05 11:09:35 INFO mapred.JobClient:     Map input records=1
> 12/07/05 11:09:35 INFO mapred.JobClient:     Spilled Records=0
> 12/07/05 11:09:35 INFO mapred.JobClient:     Map output records=1
> 12/07/05 11:09:35 INFO mapreduce.ImportJobBase: Transferred 8 bytes in 17.945 
> seconds (0.4458 bytes/sec)
> 12/07/05 11:09:35 INFO mapreduce.ImportJobBase: Retrieved 1 records.
> 12/07/05 11:09:35 INFO hive.HiveImport: Removing temporary files from import 
> process: Dummy/_logs
> 12/07/05 11:09:35 INFO hive.HiveImport: Loading uploaded data into Hive
> 12/07/05 11:09:35 INFO manager.SqlManager: Executing SQL statement: SELECT 
> t.* FROM `Dummy` AS t LIMIT 1
> 12/07/05 11:09:37 INFO hive.HiveImport: Logging initialized using 
> configuration in 
> jar:file:/HADOOP/hive/lib/hive-common-0.8.1.jar!/hive-log4j.properties
> 12/07/05 11:09:37 INFO hive.HiveImport: Hive history 
> file=/tmp/mediaadmin/hive_job_log_mediaadmin_201207051109_1901926452.txt
> 12/07/05 11:09:41 INFO hive.HiveImport: OK
> 12/07/05 11:09:41 INFO hive.HiveImport: Time taken: 3.934 seconds
> 12/07/05 11:09:41 INFO hive.HiveImport: Loading data to table 
> default.dummyhive
> 12/07/05 11:09:41 INFO hive.HiveImport: OK
> 12/07/05 11:09:41 INFO hive.HiveImport: Time taken: 0.262 seconds
> 12/07/05 11:09:41 INFO hive.HiveImport: Hive import complete.
> 
> 
> 
> Why is it so? Please help me out
> 
> Thanks & Regards
> Yogesh Kumar
> 
> 
> From: Yogesh Kumar (WT01 - Communication and Media)
> Sent: Thursday, July 05, 2012 7:31 PM
> To: [email protected]; Bejoy Ks
> Subject: RE: Hive uploading
> 
> there was no any hive-site.xml when I extracted it.
> 
> although I am able to create tables in Hive, uploading data from HDFS to hive 
> by using 
> 
> LOAD DATA INPATH 
> 'hdfs://localhost:9000/user/hive/warehouse/dummyhive/part-m-00000' OVERWRITE 
> INTO TABLE demoo PARTITION (ROLL = 001);
> 
> Loading data to table default.demoo partition (roll=001)
> OK
> Time taken: 0.466 seconds
> 
> and able to see records..
> 
> tmp/mediaadmin/hive_job_log_mediaadmin_201207051849_2011545354.txt contains
> 
> SessionStart SESSION_ID="mediaadmin_201207051849" TIME="1341494353320"
> QueryStart QUERY_STRING="CREATE TABLE `yeshivee` ( `no` INT, `name` STRING) 
> COMMENT 'Imported by sqoop on 2012/07/05 18:49:11' ROW FORMAT DELIMITED 
> FIELDS TERMINATED BY '\001' LINES TERMINATED BY '\012' STORED AS TEXTFILE" 
> QUERY_ID="mediaadmin_20120705184949_24437ec0-3332-49ea-851c-3ba8a3d5c872" 
> TIME="1341494353751"
> Counters 
> plan="{"queryId":"mediaadmin_20120705184949_24437ec0-3332-49ea-851c-3ba8a3d5c872","queryType":null,"queryAttributes":{"queryString":"CREATE
>  TABLE `yeshivee` ( `no` INT, `name` STRING) COMMENT 'Imported by sqoop on 
> 2012/07/05 18:49:11' ROW FORMAT DELIMITED FIELDS TERMINATED BY '\001' LINES 
> TERMINATED BY '\012' STORED AS 
> TEXTFILE"},"queryCounters":"null","stageGraph":{"nodeType":"STAGE","roots":"null","adjacencyList":"]"},"stageList":[{"stageId":"Stage-0","stageType":"DDL","stageAttributes":"null","stageCounters":"}","taskList":[{"taskId":"Stage-0_OTHER","taskType":"OTHER","taskAttributes":"null","taskCounters":"null","operatorGraph":"null","operatorList":"]","done":"false","started":"false"}],"done":"false","started":"false"}],"done":"false","started":"true"}"
>  TIME="1341494353762"
> TaskStart TASK_NAME="org.apache.hadoop.hive.ql.exec.DDLTask" 
> TASK_ID="Stage-0" 
> QUERY_ID="mediaadmin_20120705184949_24437ec0-3332-49ea-851c-3ba8a3d5c872" 
> TIME="1341494353766"
> Counters 
> plan="{"queryId":"mediaadmin_20120705184949_24437ec0-3332-49ea-851c-3ba8a3d5c872","queryType":null,"queryAttributes":{"queryString":"CREATE
>  TABLE `yeshivee` ( `no` INT, `name` STRING) COMMENT 'Imported by sqoop on 
> 2012/07/05 18:49:11' ROW FORMAT DELIMITED FIELDS TERMINATED BY '\001' LINES 
> TERMINATED BY '\012' STORED AS 
> TEXTFILE"},"queryCounters":"null","stageGraph":{"nodeType":"STAGE","roots":"null","adjacencyList":"]"},"stageList":[{"stageId":"Stage-0","stageType":"DDL","stageAttributes":"null","stageCounters":"}","taskList":[{"taskId":"Stage-0_OTHER","taskType":"OTHER","taskAttributes":"null","taskCounters":"null","operatorGraph":"null","operatorList":"]","done":"false","started":"true"}],"done":"false","started":"true"}],"done":"false","started":"true"}"
>  TIME="1341494353768"
> Counters 
> plan="{"queryId":"mediaadmin_20120705184949_24437ec0-3332-49ea-851c-3ba8a3d5c872","queryType":null,"queryAttributes":{"queryString":"CREATE
>  TABLE `yeshivee` ( `no` INT, `name` STRING) COMMENT 'Imported by sqoop on 
> 2012/07/05 18:49:11' ROW FORMAT DELIMITED FIELDS TERMINATED BY '\001' LINES 
> TERMINATED BY '\012' STORED AS 
> TEXTFILE"},"queryCounters":"null","stageGraph":{"nodeType":"STAGE","roots":"null","adjacencyList":"]"},"stageList":[{"stageId":"Stage-0","stageType":"DDL","stageAttributes":"null","stageCounters":"}","taskList":[{"taskId":"Stage-0_OTHER","taskType":"OTHER","taskAttributes":"null","taskCounters":"null","operatorGraph":"null","operatorList":"]","done":"true","started":"true"}],"done":"true","started":"true"}],"done":"false","started":"true"}"
>  TIME="1341494357307"
> TaskEnd TASK_RET_CODE="0" TASK_NAME="org.apache.hadoop.hive.ql.exec.DDLTask" 
> TASK_ID="Stage-0" 
> QUERY_ID="mediaadmin_20120705184949_24437ec0-3332-49ea-851c-3ba8a3d5c872" 
> TIME="1341494357307"
> QueryEnd QUERY_STRING="CREATE TABLE `yeshivee` ( `no` INT, `name` STRING) 
> COMMENT 'Imported by sqoop on 2012/07/05 18:49:11' ROW FORMAT DELIMITED 
> FIELDS TERMINATED BY '\001' LINES TERMINATED BY '\012' STORED AS TEXTFILE" 
> QUERY_ID="mediaadmin_20120705184949_24437ec0-3332-49ea-851c-3ba8a3d5c872" 
> QUERY_RET_CODE="0" QUERY_NUM_TASKS="0" TIME="1341494357363"
> Counters 
> plan="{"queryId":"mediaadmin_20120705184949_24437ec0-3332-49ea-851c-3ba8a3d5c872","queryType":null,"queryAttributes":{"queryString":"CREATE
>  TABLE `yeshivee` ( `no` INT, `name` STRING) COMMENT 'Imported by sqoop on 
> 2012/07/05 18:49:11' ROW FORMAT DELIMITED FIELDS TERMINATED BY '\001' LINES 
> TERMINATED BY '\012' STORED AS 
> TEXTFILE"},"queryCounters":"null","stageGraph":{"nodeType":"STAGE","roots":"null","adjacencyList":"]"},"stageList":[{"stageId":"Stage-0","stageType":"DDL","stageAttributes":"null","stageCounters":"}","taskList":[{"taskId":"Stage-0_OTHER","taskType":"OTHER","taskAttributes":"null","taskCounters":"null","operatorGraph":"null","operatorList":"]","done":"true","started":"true"}],"done":"true","started":"true"}],"done":"true","started":"true"}"
>  TIME="1341494357364"
> QueryStart QUERY_STRING=" LOAD DATA INPATH 
> 'hdfs://localhost:9000/userdata/yogesh/sqoop/imports/yesrdbms1' INTO TABLE 
> `yeshivee`" 
> QUERY_ID="mediaadmin_20120705184949_72d256c3-1698-40db-a8d6-6536512dbd63" 
> TIME="1341494357502"
> Counters 
> plan="{"queryId":"mediaadmin_20120705184949_72d256c3-1698-40db-a8d6-6536512dbd63","queryType":null,"queryAttributes":{"queryString":"
>  LOAD DATA INPATH 
> 'hdfs://localhost:9000/userdata/yogesh/sqoop/imports/yesrdbms1' INTO TABLE 
> `yeshivee`"},"queryCounters":"null","stageGraph":{"nodeType":"STAGE","roots":"null","adjacencyList":"]"},"stageList":[{"stageId":"Stage-0","stageType":"MOVE","stageAttributes":"null","stageCounters":"}","taskList":[{"taskId":"Stage-0_OTHER","taskType":"OTHER","taskAttributes":"null","taskCounters":"null","operatorGraph":"null","operatorList":"]","done":"false","started":"false"}],"done":"false","started":"false"}],"done":"false","started":"true"}"
>  TIME="1341494357502"
> TaskStart TASK_NAME="org.apache.hadoop.hive.ql.exec.MoveTask" 
> TASK_ID="Stage-0" 
> QUERY_ID="mediaadmin_20120705184949_72d256c3-1698-40db-a8d6-6536512dbd63" 
> TIME="1341494357502"
> Counters 
> plan="{"queryId":"mediaadmin_20120705184949_72d256c3-1698-40db-a8d6-6536512dbd63","queryType":null,"queryAttributes":{"queryString":"
>  LOAD DATA INPATH 
> 'hdfs://localhost:9000/userdata/yogesh/sqoop/imports/yesrdbms1' INTO TABLE 
> `yeshivee`"},"queryCounters":"null","stageGraph":{"nodeType":"STAGE","roots":"null","adjacencyList":"]"},"stageList":[{"stageId":"Stage-0","stageType":"MOVE","stageAttributes":"null","stageCounters":"}","taskList":[{"taskId":"Stage-0_OTHER","taskType":"OTHER","taskAttributes":"null","taskCounters":"null","operatorGraph":"null","operatorList":"]","done":"false","started":"true"}],"done":"false","started":"true"}],"done":"false","started":"true"}"
>  TIME="1341494357503"
> Counters 
> plan="{"queryId":"mediaadmin_20120705184949_72d256c3-1698-40db-a8d6-6536512dbd63","queryType":null,"queryAttributes":{"queryString":"
>  LOAD DATA INPATH 
> 'hdfs://localhost:9000/userdata/yogesh/sqoop/imports/yesrdbms1' INTO TABLE 
> `yeshivee`"},"queryCounters":"null","stageGraph":{"nodeType":"STAGE","roots":"null","adjacencyList":"]"},"stageList":[{"stageId":"Stage-0","stageType":"MOVE","stageAttributes":"null","stageCounters":"}","taskList":[{"taskId":"Stage-0_OTHER","taskType":"OTHER","taskAttributes":"null","taskCounters":"null","operatorGraph":"null","operatorList":"]","done":"true","started":"true"}],"done":"true","started":"true"}],"done":"false","started":"true"}"
>  TIME="1341494357611"
> TaskEnd TASK_RET_CODE="0" TASK_NAME="org.apache.hadoop.hive.ql.exec.MoveTask" 
> TASK_ID="Stage-0" 
> QUERY_ID="mediaadmin_20120705184949_72d256c3-1698-40db-a8d6-6536512dbd63" 
> TIME="1341494357611"
> QueryEnd QUERY_STRING=" LOAD DATA INPATH 
> 'hdfs://localhost:9000/userdata/yogesh/sqoop/imports/yesrdbms1' INTO TABLE 
> `yeshivee`" 
> QUERY_ID="mediaadmin_20120705184949_72d256c3-1698-40db-a8d6-6536512dbd63" 
> QUERY_RET_CODE="0" QUERY_NUM_TASKS="0" TIME="1341494357611"
> Counters 
> plan="{"queryId":"mediaadmin_20120705184949_72d256c3-1698-40db-a8d6-6536512dbd63","queryType":null,"queryAttributes":{"queryString":"
>  LOAD DATA INPATH 
> 'hdfs://localhost:9000/userdata/yogesh/sqoop/imports/yesrdbms1' INTO TABLE 
> `yeshivee`"},"queryCounters":"null","stageGraph":{"nodeType":"STAGE","roots":"null","adjacencyList":"]"},"stageList":[{"stageId":"Stage-0","stageType":"MOVE","stageAttributes":"null","stageCounters":"}","taskList":[{"taskId":"Stage-0_OTHER","taskType":"OTHER","taskAttributes":"null","taskCounters":"null","operatorGraph":"null","operatorList":"]","done":"true","started":"true"}],"done":"true","started":"true"}],"done":"true","started":"true"}"
>  TIME="1341494357611"
> 
> 
> 
> /tmp/mediaadmin/hive_job_log_mediaadmin_201207051854_1851664234.txt  contains
> 
> SessionStart SESSION_ID="mediaadmin_201207051854" TIME="1341494660061"
> QueryStart QUERY_STRING="CREATE TABLE `yeshive2` ( `num` INT, `name` STRING) 
> COMMENT 'Imported by sqoop on 2012/07/05 18:54:18' ROW FORMAT DELIMITED 
> FIELDS TERMINATED BY '\001' LINES TERMINATED BY '\012' STORED AS TEXTFILE" 
> QUERY_ID="mediaadmin_20120705185454_d40a27b9-5434-4764-880d-902e91bd3c28" 
> TIME="1341494660513"
> Counters 
> plan="{"queryId":"mediaadmin_20120705185454_d40a27b9-5434-4764-880d-902e91bd3c28","queryType":null,"queryAttributes":{"queryString":"CREATE
>  TABLE `yeshive2` ( `num` INT, `name` STRING) COMMENT 'Imported by sqoop on 
> 2012/07/05 18:54:18' ROW FORMAT DELIMITED FIELDS TERMINATED BY '\001' LINES 
> TERMINATED BY '\012' STORED AS 
> TEXTFILE"},"queryCounters":"null","stageGraph":{"nodeType":"STAGE","roots":"null","adjacencyList":"]"},"stageList":[{"stageId":"Stage-0","stageType":"DDL","stageAttributes":"null","stageCounters":"}","taskList":[{"taskId":"Stage-0_OTHER","taskType":"OTHER","taskAttributes":"null","taskCounters":"null","operatorGraph":"null","operatorList":"]","done":"false","started":"false"}],"done":"false","started":"false"}],"done":"false","started":"true"}"
>  TIME="1341494660523"
> TaskStart TASK_NAME="org.apache.hadoop.hive.ql.exec.DDLTask" 
> TASK_ID="Stage-0" 
> QUERY_ID="mediaadmin_20120705185454_d40a27b9-5434-4764-880d-902e91bd3c28" 
> TIME="1341494660527"
> Counters 
> plan="{"queryId":"mediaadmin_20120705185454_d40a27b9-5434-4764-880d-902e91bd3c28","queryType":null,"queryAttributes":{"queryString":"CREATE
>  TABLE `yeshive2` ( `num` INT, `name` STRING) COMMENT 'Imported by sqoop on 
> 2012/07/05 18:54:18' ROW FORMAT DELIMITED FIELDS TERMINATED BY '\001' LINES 
> TERMINATED BY '\012' STORED AS 
> TEXTFILE"},"queryCounters":"null","stageGraph":{"nodeType":"STAGE","roots":"null","adjacencyList":"]"},"stageList":[{"stageId":"Stage-0","stageType":"DDL","stageAttributes":"null","stageCounters":"}","taskList":[{"taskId":"Stage-0_OTHER","taskType":"OTHER","taskAttributes":"null","taskCounters":"null","operatorGraph":"null","operatorList":"]","done":"false","started":"true"}],"done":"false","started":"true"}],"done":"false","started":"true"}"
>  TIME="1341494660529"
> Counters 
> plan="{"queryId":"mediaadmin_20120705185454_d40a27b9-5434-4764-880d-902e91bd3c28","queryType":null,"queryAttributes":{"queryString":"CREATE
>  TABLE `yeshive2` ( `num` INT, `name` STRING) COMMENT 'Imported by sqoop on 
> 2012/07/05 18:54:18' ROW FORMAT DELIMITED FIELDS TERMINATED BY '\001' LINES 
> TERMINATED BY '\012' STORED AS 
> TEXTFILE"},"queryCounters":"null","stageGraph":{"nodeType":"STAGE","roots":"null","adjacencyList":"]"},"stageList":[{"stageId":"Stage-0","stageType":"DDL","stageAttributes":"null","stageCounters":"}","taskList":[{"taskId":"Stage-0_OTHER","taskType":"OTHER","taskAttributes":"null","taskCounters":"null","operatorGraph":"null","operatorList":"]","done":"true","started":"true"}],"done":"true","started":"true"}],"done":"false","started":"true"}"
>  TIME="1341494664453"
> TaskEnd TASK_RET_CODE="0" TASK_NAME="org.apache.hadoop.hive.ql.exec.DDLTask" 
> TASK_ID="Stage-0" 
> QUERY_ID="mediaadmin_20120705185454_d40a27b9-5434-4764-880d-902e91bd3c28" 
> TIME="1341494664453"
> QueryEnd QUERY_STRING="CREATE TABLE `yeshive2` ( `num` INT, `name` STRING) 
> COMMENT 'Imported by sqoop on 2012/07/05 18:54:18' ROW FORMAT DELIMITED 
> FIELDS TERMINATED BY '\001' LINES TERMINATED BY '\012' STORED AS TEXTFILE" 
> QUERY_ID="mediaadmin_20120705185454_d40a27b9-5434-4764-880d-902e91bd3c28" 
> QUERY_RET_CODE="0" QUERY_NUM_TASKS="0" TIME="1341494664466"
> Counters 
> plan="{"queryId":"mediaadmin_20120705185454_d40a27b9-5434-4764-880d-902e91bd3c28","queryType":null,"queryAttributes":{"queryString":"CREATE
>  TABLE `yeshive2` ( `num` INT, `name` STRING) COMMENT 'Imported by sqoop on 
> 2012/07/05 18:54:18' ROW FORMAT DELIMITED FIELDS TERMINATED BY '\001' LINES 
> TERMINATED BY '\012' STORED AS 
> TEXTFILE"},"queryCounters":"null","stageGraph":{"nodeType":"STAGE","roots":"null","adjacencyList":"]"},"stageList":[{"stageId":"Stage-0","stageType":"DDL","stageAttributes":"null","stageCounters":"}","taskList":[{"taskId":"Stage-0_OTHER","taskType":"OTHER","taskAttributes":"null","taskCounters":"null","operatorGraph":"null","operatorList":"]","done":"true","started":"true"}],"done":"true","started":"true"}],"done":"true","started":"true"}"
>  TIME="1341494664467"
> QueryStart QUERY_STRING=" LOAD DATA INPATH 
> 'hdfs://localhost:9000/userdata/yogesh/sqoop/imports/oldtoy' INTO TABLE 
> `yeshive2`" 
> QUERY_ID="mediaadmin_20120705185454_c1afa8b5-6f97-4be7-9bc6-9a30426a3cbd" 
> TIME="1341494664599"
> Counters 
> plan="{"queryId":"mediaadmin_20120705185454_c1afa8b5-6f97-4be7-9bc6-9a30426a3cbd","queryType":null,"queryAttributes":{"queryString":"
>  LOAD DATA INPATH 
> 'hdfs://localhost:9000/userdata/yogesh/sqoop/imports/oldtoy' INTO TABLE 
> `yeshive2`"},"queryCounters":"null","stageGraph":{"nodeType":"STAGE","roots":"null","adjacencyList":"]"},"stageList":[{"stageId":"Stage-0","stageType":"MOVE","stageAttributes":"null","stageCounters":"}","taskList":[{"taskId":"Stage-0_OTHER","taskType":"OTHER","taskAttributes":"null","taskCounters":"null","operatorGraph":"null","operatorList":"]","done":"false","started":"false"}],"done":"false","started":"false"}],"done":"false","started":"true"}"
>  TIME="1341494664599"
> TaskStart TASK_NAME="org.apache.hadoop.hive.ql.exec.MoveTask" 
> TASK_ID="Stage-0" 
> QUERY_ID="mediaadmin_20120705185454_c1afa8b5-6f97-4be7-9bc6-9a30426a3cbd" 
> TIME="1341494664599"
> Counters 
> plan="{"queryId":"mediaadmin_20120705185454_c1afa8b5-6f97-4be7-9bc6-9a30426a3cbd","queryType":null,"queryAttributes":{"queryString":"
>  LOAD DATA INPATH 
> 'hdfs://localhost:9000/userdata/yogesh/sqoop/imports/oldtoy' INTO TABLE 
> `yeshive2`"},"queryCounters":"null","stageGraph":{"nodeType":"STAGE","roots":"null","adjacencyList":"]"},"stageList":[{"stageId":"Stage-0","stageType":"MOVE","stageAttributes":"null","stageCounters":"}","taskList":[{"taskId":"Stage-0_OTHER","taskType":"OTHER","taskAttributes":"null","taskCounters":"null","operatorGraph":"null","operatorList":"]","done":"false","started":"true"}],"done":"false","started":"true"}],"done":"false","started":"true"}"
>  TIME="1341494664599"
> Counters 
> plan="{"queryId":"mediaadmin_20120705185454_c1afa8b5-6f97-4be7-9bc6-9a30426a3cbd","queryType":null,"queryAttributes":{"queryString":"
>  LOAD DATA INPATH 
> 'hdfs://localhost:9000/userdata/yogesh/sqoop/imports/oldtoy' INTO TABLE 
> `yeshive2`"},"queryCounters":"null","stageGraph":{"nodeType":"STAGE","roots":"null","adjacencyList":"]"},"stageList":[{"stageId":"Stage-0","stageType":"MOVE","stageAttributes":"null","stageCounters":"}","taskList":[{"taskId":"Stage-0_OTHER","taskType":"OTHER","taskAttributes":"null","taskCounters":"null","operatorGraph":"null","operatorList":"]","done":"true","started":"true"}],"done":"true","started":"true"}],"done":"false","started":"true"}"
>  TIME="1341494664744"
> TaskEnd TASK_RET_CODE="0" TASK_NAME="org.apache.hadoop.hive.ql.exec.MoveTask" 
> TASK_ID="Stage-0" 
> QUERY_ID="mediaadmin_20120705185454_c1afa8b5-6f97-4be7-9bc6-9a30426a3cbd" 
> TIME="1341494664744"
> QueryEnd QUERY_STRING=" LOAD DATA INPATH 
> 'hdfs://localhost:9000/userdata/yogesh/sqoop/imports/oldtoy' INTO TABLE 
> `yeshive2`" 
> QUERY_ID="mediaadmin_20120705185454_c1afa8b5-6f97-4be7-9bc6-9a30426a3cbd" 
> QUERY_RET_CODE="0" QUERY_NUM_TASKS="0" TIME="1341494664744"
> Counters 
> plan="{"queryId":"mediaadmin_20120705185454_c1afa8b5-6f97-4be7-9bc6-9a30426a3cbd","queryType":null,"queryAttributes":{"queryString":"
>  LOAD DATA INPATH 
> 'hdfs://localhost:9000/userdata/yogesh/sqoop/imports/oldtoy' INTO TABLE 
> `yeshive2`"},"queryCounters":"null","stageGraph":{"nodeType":"STAGE","roots":"null","adjacencyList":"]"},"stageList":[{"stageId":"Stage-0","stageType":"MOVE","stageAttributes":"null","stageCounters":"}","taskList":[{"taskId":"Stage-0_OTHER","taskType":"OTHER","taskAttributes":"null","taskCounters":"null","operatorGraph":"null","operatorList":"]","done":"true","started":"true"}],"done":"true","started":"true"}],"done":"true","started":"true"}"
>  TIME="1341494664744"
> 
> 
> I dont thik so that its showing any error or is it ?
> 
> Yeah sure I will also raise this issue at sqoop user list
> 
> 
> Regards
> Yogesh Kumar
> From: Bejoy Ks [[email protected]]
> Sent: Thursday, July 05, 2012 7:14 PM
> To: [email protected]
> Subject: Re: Hive uploading
> 
> Hi Yogesh
> 
> The verbose logging is still not getting enabled, some issue with SQOOP 
> installation I guess.
> 
> The console log shows hive table creation and data load is sucess. Are you 
> still not seeing the tables 'yeshivee' and 'yeshive2 ' in hive. Login to your 
> default hive CLI and check. ($HIVE_HOME/bin/hive)
> 
> You can look at the hive history file and see if there are any errors 
> reported in hive. 
> 12/07/05 18:49:13 INFO hive.HiveImport: Hive history 
> file=/tmp/mediaadmin/hive_job_log_mediaadmin_201207051849_2011545354.txt
> 12/07/05 18:54:20 INFO hive.HiveImport: Hive history 
> file=/tmp/mediaadmin/hive_job_log_mediaadmin_201207051854_1851664234.txt
> 
> This is purely a SQOOP issue and  you should be getting better help in Sqoop 
> user group. Please take this conversation to SQOOP user list, we can continue 
> this conversation there.
> 
> Regards
> Bejoy KS
> 
> From: "[email protected]" <[email protected]>
> To: [email protected]; [email protected] 
> Sent: Thursday, July 5, 2012 6:56 PM
> Subject: RE: Hive uploading
> 
> Hi Bejoy,
> 
> I used it. after creating userdata/....   dir in hdfs
>  
> sqoop import --verbose --connect jdbc:mysql://localhost:3306/Demo --username 
> sqoop1 --password SQOOP1 --table yesrdbms1 -num-mappers 1 --warehouse-dir 
> /userdata/yogesh/sqoop/imports --hive-import  --hive-table yeshivee
> 
> outcome is 
> 
> 12/07/05 18:48:56 WARN tool.BaseSqoopTool: Setting your password on the 
> command-line is insecure. Consider using -P instead.
> 12/07/05 18:48:56 INFO tool.BaseSqoopTool: Using Hive-specific delimiters for 
> output. You can override
> 12/07/05 18:48:56 INFO tool.BaseSqoopTool: delimiters with 
> --fields-terminated-by, etc.
> 12/07/05 18:48:56 INFO manager.MySQLManager: Preparing to use a MySQL 
> streaming resultset.
> 12/07/05 18:48:56 INFO tool.CodeGenTool: Beginning code generation
> 12/07/05 18:48:56 INFO manager.SqlManager: Executing SQL statement: SELECT 
> t.* FROM `yesrdbms1` AS t LIMIT 1
> 12/07/05 18:48:56 INFO orm.CompilationManager: HADOOP_HOME is 
> /HADOOP/hadoop-0.20.2/bin/..
> 12/07/05 18:48:56 INFO orm.CompilationManager: Found hadoop core jar at: 
> /HADOOP/hadoop-0.20.2/bin/../hadoop-0.20.2-core.jar
> Note: 
> /tmp/sqoop-mediaadmin/compile/603b6e26b2fd1160693ba0a66786d12e/yesrdbms1.java 
> uses or overrides a deprecated API.
> Note: Recompile with -Xlint:deprecation for details.
> 12/07/05 18:48:57 INFO orm.CompilationManager: Writing jar file: 
> /tmp/sqoop-mediaadmin/compile/603b6e26b2fd1160693ba0a66786d12e/yesrdbms1.jar
> 12/07/05 18:48:57 WARN manager.MySQLManager: It looks like you are importing 
> from mysql.
> 12/07/05 18:48:57 WARN manager.MySQLManager: This transfer can be faster! Use 
> the --direct
> 12/07/05 18:48:57 WARN manager.MySQLManager: option to exercise a 
> MySQL-specific fast path.
> 12/07/05 18:48:57 INFO manager.MySQLManager: Setting zero DATETIME behavior 
> to convertToNull (mysql)
> 12/07/05 18:48:57 INFO mapreduce.ImportJobBase: Beginning import of yesrdbms1
> 12/07/05 18:48:58 INFO mapred.JobClient: Running job: job_201207051104_0011
> 12/07/05 18:48:59 INFO mapred.JobClient:  map 0% reduce 0%
> 12/07/05 18:49:09 INFO mapred.JobClient:  map 100% reduce 0%
> 12/07/05 18:49:11 INFO mapred.JobClient: Job complete: job_201207051104_0011
> 12/07/05 18:49:11 INFO mapred.JobClient: Counters: 5
> 12/07/05 18:49:11 INFO mapred.JobClient:   Job Counters 
> 12/07/05 18:49:11 INFO mapred.JobClient:     Launched map tasks=1
> 12/07/05 18:49:11 INFO mapred.JobClient:   FileSystemCounters
> 12/07/05 18:49:11 INFO mapred.JobClient:     HDFS_BYTES_WRITTEN=9
> 12/07/05 18:49:11 INFO mapred.JobClient:   Map-Reduce Framework
> 12/07/05 18:49:11 INFO mapred.JobClient:     Map input records=1
> 12/07/05 18:49:11 INFO mapred.JobClient:     Spilled Records=0
> 12/07/05 18:49:11 INFO mapred.JobClient:     Map output records=1
> 12/07/05 18:49:11 INFO mapreduce.ImportJobBase: Transferred 9 bytes in 
> 13.6291 seconds (0.6604 bytes/sec)
> 12/07/05 18:49:11 INFO mapreduce.ImportJobBase: Retrieved 1 records.
> 12/07/05 18:49:11 INFO hive.HiveImport: Removing temporary files from import 
> process: /userdata/yogesh/sqoop/imports/yesrdbms1/_logs
> 12/07/05 18:49:11 INFO hive.HiveImport: Loading uploaded data into Hive
> 12/07/05 18:49:11 INFO manager.SqlManager: Executing SQL statement: SELECT 
> t.* FROM `yesrdbms1` AS t LIMIT 1
> 12/07/05 18:49:13 INFO hive.HiveImport: Logging initialized using 
> configuration in 
> jar:file:/HADOOP/hive/lib/hive-common-0.8.1.jar!/hive-log4j.properties
> 12/07/05 18:49:13 INFO hive.HiveImport: Hive history 
> file=/tmp/mediaadmin/hive_job_log_mediaadmin_201207051849_2011545354.txt
> 12/07/05 18:49:17 INFO hive.HiveImport: OK
> 12/07/05 18:49:17 INFO hive.HiveImport: Time taken: 3.864 seconds
> 12/07/05 18:49:17 INFO hive.HiveImport: Loading data to table default.yeshivee
> 12/07/05 18:49:17 INFO hive.HiveImport: OK
> 12/07/05 18:49:17 INFO hive.HiveImport: Time taken: 0.245 seconds
> 12/07/05 18:49:17 INFO hive.HiveImport: Hive import complete.
> 
> ------------------------------------*****************************************------------------------------------
> and also this one 
> 
> sqoop import --connect jdbc:mysql://localhost:3306/Demo --username sqoop1 
> --password SQOOP1 --table oldtoy --verbose  -num-mappers 1 --warehouse-dir 
> /userdata/yogesh/sqoop/imports --hive-import  --hive-table yeshive2 
> --create-hive-table
> 
> outcome is 
> 
> Setting your password on the command-line is insecure. Consider using -P 
> instead.
> 12/07/05 18:53:58 INFO tool.BaseSqoopTool: Using Hive-specific delimiters for 
> output. You can override
> 12/07/05 18:53:58 INFO tool.BaseSqoopTool: delimiters with 
> --fields-terminated-by, etc.
> 12/07/05 18:53:59 INFO manager.MySQLManager: Preparing to use a MySQL 
> streaming resultset.
> 12/07/05 18:53:59 INFO tool.CodeGenTool: Beginning code generation
> 12/07/05 18:53:59 INFO manager.SqlManager: Executing SQL statement: SELECT 
> t.* FROM `oldtoy` AS t LIMIT 1
> 12/07/05 18:53:59 INFO orm.CompilationManager: HADOOP_HOME is 
> /HADOOP/hadoop-0.20.2/bin/..
> 12/07/05 18:53:59 INFO orm.CompilationManager: Found hadoop core jar at: 
> /HADOOP/hadoop-0.20.2/bin/../hadoop-0.20.2-core.jar
> Note: 
> /tmp/sqoop-mediaadmin/compile/c5fab2a1adf9725e1c5d556d0cceefd6/oldtoy.java 
> uses or overrides a deprecated API.
> Note: Recompile with -Xlint:deprecation for details.
> 12/07/05 18:54:00 INFO orm.CompilationManager: Writing jar file: 
> /tmp/sqoop-mediaadmin/compile/c5fab2a1adf9725e1c5d556d0cceefd6/oldtoy.jar
> 12/07/05 18:54:00 WARN manager.MySQLManager: It looks like you are importing 
> from mysql.
> 12/07/05 18:54:00 WARN manager.MySQLManager: This transfer can be faster! Use 
> the --direct
> 12/07/05 18:54:00 WARN manager.MySQLManager: option to exercise a 
> MySQL-specific fast path.
> 12/07/05 18:54:00 INFO manager.MySQLManager: Setting zero DATETIME behavior 
> to convertToNull (mysql)
> 12/07/05 18:54:00 INFO mapreduce.ImportJobBase: Beginning import of oldtoy
> 12/07/05 18:54:01 INFO mapred.JobClient: Running job: job_201207051104_0013
> 12/07/05 18:54:02 INFO mapred.JobClient:  map 0% reduce 0%
> 12/07/05 18:54:16 INFO mapred.JobClient:  map 100% reduce 0%
> 12/07/05 18:54:18 INFO mapred.JobClient: Job complete: job_201207051104_0013
> 12/07/05 18:54:18 INFO mapred.JobClient: Counters: 5
> 12/07/05 18:54:18 INFO mapred.JobClient:   Job Counters 
> 12/07/05 18:54:18 INFO mapred.JobClient:     Launched map tasks=1
> 12/07/05 18:54:18 INFO mapred.JobClient:   FileSystemCounters
> 12/07/05 18:54:18 INFO mapred.JobClient:     HDFS_BYTES_WRITTEN=8
> 12/07/05 18:54:18 INFO mapred.JobClient:   Map-Reduce Framework
> 12/07/05 18:54:18 INFO mapred.JobClient:     Map input records=1
> 12/07/05 18:54:18 INFO mapred.JobClient:     Spilled Records=0
> 12/07/05 18:54:18 INFO mapred.JobClient:     Map output records=1
> 12/07/05 18:54:18 INFO mapreduce.ImportJobBase: Transferred 8 bytes in 
> 17.8107 seconds (0.4492 bytes/sec)
> 12/07/05 18:54:18 INFO mapreduce.ImportJobBase: Retrieved 1 records.
> 12/07/05 18:54:18 INFO hive.HiveImport: Removing temporary files from import 
> process: /userdata/yogesh/sqoop/imports/oldtoy/_logs
> 12/07/05 18:54:18 INFO hive.HiveImport: Loading uploaded data into Hive
> 12/07/05 18:54:18 INFO manager.SqlManager: Executing SQL statement: SELECT 
> t.* FROM `oldtoy` AS t LIMIT 1
> 12/07/05 18:54:20 INFO hive.HiveImport: Logging initialized using 
> configuration in 
> jar:file:/HADOOP/hive/lib/hive-common-0.8.1.jar!/hive-log4j.properties
> 12/07/05 18:54:20 INFO hive.HiveImport: Hive history 
> file=/tmp/mediaadmin/hive_job_log_mediaadmin_201207051854_1851664234.txt
> 12/07/05 18:54:24 INFO hive.HiveImport: OK
> 12/07/05 18:54:24 INFO hive.HiveImport: Time taken: 4.222 seconds
> 12/07/05 18:54:24 INFO hive.HiveImport: Loading data to table default.yeshive2
> 12/07/05 18:54:24 INFO hive.HiveImport: OK
> 12/07/05 18:54:24 INFO hive.HiveImport: Time taken: 0.278 seconds
> 12/07/05 18:54:24 INFO hive.HiveImport: Hive import complete.
> 
> 
> 
> Please suggest
> 
> Greetings 
> Yogesh kumar
> 
> 
> From: Bejoy Ks [[email protected]]
> Sent: Thursday, July 05, 2012 6:28 PM
> To: [email protected]
> Subject: Re: Hive uploading
> 
> Hi Yogesh
> 
> The verbose option didn't work there as there is no DEBUG logging, can you 
> please add the verbose to the beginning of your sqoop command?
> 
> Lemme frame a small sqoop import sample or you, Please run this command and 
> post in the console log
> 
> sqoop import --verbose --connect jdbc:mysql://localhost:3306/Demo --username 
> sqoop1 -P --table troy -num-mappers 1 --warehouse-dir 
> /userdata/yogesh/sqoop/imports --hive-import --hive-table troyhive  
> 
> I haven't tried this on my end, If it poses any issues with the verbose 
> please chnage its position after --table argument.
> 
> Also you need to create the --warehouse-dir in hdfs before running the sqoop 
> import.
> 
> 
> Regards
> Bejoy KS
> 
> From: "[email protected]" <[email protected]>
> To: [email protected]; [email protected] 
> Sent: Thursday, July 5, 2012 6:07 PM
> Subject: RE: Hive uploading
> 
> Hi Bejoy,
> 
> I have created new table called Troy  and for hive its troyhive, as it was 
> showing Outputdirectory already exists
> 
> 
> sqoop import --connect jdbc:mysql://localhost:3306/Demo --username sqoop1 
> --password SQOOP1 -table troy --hive-table troyhive  --create-hive-table  
> --hive-import --hive-home HADOOP/hive --verbose
> 12/07/05 17:57:16 WARN tool.BaseSqoopTool: Setting your password on the 
> command-line is insecure. Consider using -P instead.
> 12/07/05 17:57:16 INFO tool.BaseSqoopTool: Using Hive-specific delimiters for 
> output. You can override
> 12/07/05 17:57:16 INFO tool.BaseSqoopTool: delimiters with 
> --fields-terminated-by, etc.
> 12/07/05 17:57:16 INFO manager.MySQLManager: Preparing to use a MySQL 
> streaming resultset.
> 12/07/05 17:57:16 INFO tool.CodeGenTool: Beginning code generation
> 12/07/05 17:57:17 INFO manager.SqlManager: Executing SQL statement: SELECT 
> t.* FROM `troy` AS t LIMIT 1
> 12/07/05 17:57:17 INFO orm.CompilationManager: HADOOP_HOME is 
> /HADOOP/hadoop-0.20.2/bin/..
> 12/07/05 17:57:17 INFO orm.CompilationManager: Found hadoop core jar at: 
> /HADOOP/hadoop-0.20.2/bin/../hadoop-0.20.2-core.jar
> Note: 
> /tmp/sqoop-mediaadmin/compile/26f5861253b910681eade0bd0e84efb5/troy.java uses 
> or overrides a deprecated API.
> Note: Recompile with -Xlint:deprecation for details.
> 12/07/05 17:57:17 INFO orm.CompilationManager: Writing jar file: 
> /tmp/sqoop-mediaadmin/compile/26f5861253b910681eade0bd0e84efb5/troy.jar
> 12/07/05 17:57:17 WARN manager.MySQLManager: It looks like you are importing 
> from mysql.
> 12/07/05 17:57:17 WARN manager.MySQLManager: This transfer can be faster! Use 
> the --direct
> 12/07/05 17:57:17 WARN manager.MySQLManager: option to exercise a 
> MySQL-specific fast path.
> 12/07/05 17:57:17 INFO manager.MySQLManager: Setting zero DATETIME behavior 
> to convertToNull (mysql)
> 12/07/05 17:57:17 INFO mapreduce.ImportJobBase: Beginning import of troy
> 12/07/05 17:57:18 INFO db.DataDrivenDBInputFormat: BoundingValsQuery: SELECT 
> MIN(`num`), MAX(`num`) FROM `troy`
> 12/07/05 17:57:18 INFO mapred.JobClient: Running job: job_201207051104_0005
> 12/07/05 17:57:19 INFO mapred.JobClient:  map 0% reduce 0%
> 12/07/05 17:57:30 INFO mapred.JobClient:  map 100% reduce 0%
> 12/07/05 17:57:32 INFO mapred.JobClient: Job complete: job_201207051104_0005
> 12/07/05 17:57:32 INFO mapred.JobClient: Counters: 5
> 12/07/05 17:57:32 INFO mapred.JobClient:   Job Counters 
> 12/07/05 17:57:32 INFO mapred.JobClient:     Launched map tasks=1
> 12/07/05 17:57:32 INFO mapred.JobClient:   FileSystemCounters
> 12/07/05 17:57:32 INFO mapred.JobClient:     HDFS_BYTES_WRITTEN=8
> 12/07/05 17:57:32 INFO mapred.JobClient:   Map-Reduce Framework
> 12/07/05 17:57:32 INFO mapred.JobClient:     Map input records=1
> 12/07/05 17:57:32 INFO mapred.JobClient:     Spilled Records=0
> 12/07/05 17:57:32 INFO mapred.JobClient:     Map output records=1
> 12/07/05 17:57:32 INFO mapreduce.ImportJobBase: Transferred 8 bytes in 
> 14.6895 seconds (0.5446 bytes/sec)
> 12/07/05 17:57:32 INFO mapreduce.ImportJobBase: Retrieved 1 records.
> 12/07/05 17:57:32 INFO hive.HiveImport: Removing temporary files from import 
> process: troy/_logs
> 12/07/05 17:57:32 INFO hive.HiveImport: Loading uploaded data into Hive
> 12/07/05 17:57:32 INFO manager.SqlManager: Executing SQL statement: SELECT 
> t.* FROM `troy` AS t LIMIT 1
> 12/07/05 17:57:34 INFO hive.HiveImport: Logging initialized using 
> configuration in 
> jar:file:/HADOOP/hive/lib/hive-common-0.8.1.jar!/hive-log4j.properties
> 12/07/05 17:57:34 INFO hive.HiveImport: Hive history 
> file=/tmp/mediaadmin/hive_job_log_mediaadmin_201207051757_1184599996.txt
> 12/07/05 17:57:39 INFO hive.HiveImport: OK
> 12/07/05 17:57:39 INFO hive.HiveImport: Time taken: 4.249 seconds
> 12/07/05 17:57:39 INFO hive.HiveImport: Loading data to table default.troyhive
> 12/07/05 17:57:39 INFO hive.HiveImport: OK
> 12/07/05 17:57:39 INFO hive.HiveImport: Time taken: 0.257 seconds
> 12/07/05 17:57:39 INFO hive.HiveImport: Hive import complete.
> 
> Regards
> Yogesh Kumar
> 
> From: Bejoy Ks [[email protected]]
> Sent: Thursday, July 05, 2012 6:03 PM
> To: [email protected]
> Subject: Re: Hive uploading
> 
> Hi Yogesh
> 
> Verbose option won't create any difference in operation, but gives more 
> logging information on console which could be helpful to search for any hints.
> 
> So please post in your console dump/log along with the sqoop import command 
> with verbose enabled.
> 
> Regards
> Bejoy KS
> 
> From: "[email protected]" <[email protected]>
> To: [email protected]; [email protected] 
> Sent: Thursday, July 5, 2012 6:00 PM
> Subject: RE: Hive uploading
> 
> Hello Bejoy,
> 
> sqoop import --connect jdbc:mysql://localhost:3306/Demo --username sqoop1 
> --password SQOOP1 -table Dummy --hive-table dummyhive  --create-hive-table  
> --hive-import --hive-home HADOOP/hive --verbose
> 
> Still the same, no table has been created. I am not able to see the dummyhive 
> table in hive by using command 
> Show Tables ;
> 
> although table dummyhive created into HDFS in dir:
> user/hive/warehouse/dummyhive
> 
> 
> Please suggest
> Yogesh Kumar
> 
> From: Bejoy Ks [[email protected]]
> Sent: Thursday, July 05, 2012 5:29 PM
> To: [email protected]
> Subject: Re: Hive uploading
> 
> Hi Yogesh
> 
> Please try out this command
> 
>  sqoop import --connect jdbc:mysql://localhost:3306/Demo --username sqoop1 
> --password SQOOP1 -table Dummy --hive-table dummyhive  --create-hive-table  
> --hive-import --hive-home HADOOP/hive --verbose
> 
> 
> Regards
> Bejoy KS
> 
> From: "[email protected]" <[email protected]>
> To: [email protected]; [email protected] 
> Sent: Thursday, July 5, 2012 5:03 PM
> Subject: RE: Hive uploading
> 
> Hi Bejoy
> 
> I have confirmed hive installation its same for both
> I used command echo $HIVE_HOME on both sqoop terminal and hive terminal
> both result the same Path 
> HADOOP/hive
> 
> I am new to Hive and sqoop, would you please give an example using -verbose 
> option with this command
> 
> 
>  sqoop import --connect jdbc:mysql://localhost:3306/Demo --username sqoop1 
> --password SQOOP1 -table Dummy --hive-table dummyhive  --create-hive-table  
> --hive-import --hive-home HADOOP/hive 
> 
> 
> 
> Please help
> 
> 
> From: Bejoy Ks [[email protected]]
> Sent: Thursday, July 05, 2012 3:14 PM
> To: [email protected]
> Subject: Re: Hive uploading
> 
> Hi Yogesh
> 
> No issues seen on the first look. Can you run the sqoop import with --verbose 
> option and post in the console dump?
> 
> Are you having multiple hive installation? If so please verify whether you 
> are using the same hive for both SQOOP import and then for verifying data 
> using hive cli. (the hive installation @ HADOOP/hive)
> 
> Regards
> Bejoy KS
> 
> From: "[email protected]" <[email protected]>
> To: [email protected] 
> Sent: Thursday, July 5, 2012 2:58 PM
> Subject: Hive uploading
> 
> Hi
> 
> I have created a table in Mysql by name Dummy and it has 2 columns, and 1 row 
> of data
> 
> I want to upload that table into Hive using Sqoop tool.
> I used this command  
> 
> 
> sqoop import --connect jdbc:mysql://localhost:3306/Demo --username sqoop1 
> --password SQOOP1 -table Dummy --hive-table dummyhive  --create-hive-table  
> --hive-import --hive-home HADOOP/hive
> 
> 
> The table has been succesfully uploaded into HDFS  /user/hive/warehouse
> but when I run command in Hive 
> 
> Show Tables;
> 
> I don't find dummyhive table in it.
> 
> Please suggest and Help
> 
> 
> Details of the command and output
> 
> mediaadmins-iMac-2:hive mediaadmin$ sqoop import --connect 
> jdbc:mysql://localhost:3306/Demo --username sqoop1 --password SQOOP1 -table 
> Dummy --hive-table dummyhive  --create-hive-table  --hive-import --hive-home 
> HADOOP/hive
> 12/07/05 11:09:15 WARN tool.BaseSqoopTool: Setting your password on the 
> command-line is insecure. Consider using -P instead.
> 12/07/05 11:09:15 INFO tool.BaseSqoopTool: Using Hive-specific delimiters for 
> output. You can override
> 12/07/05 11:09:15 INFO tool.BaseSqoopTool: delimiters with 
> --fields-terminated-by, etc.
> 12/07/05 11:09:15 INFO manager.MySQLManager: Preparing to use a MySQL 
> streaming resultset.
> 12/07/05 11:09:15 INFO tool.CodeGenTool: Beginning code generation
> 12/07/05 11:09:16 INFO manager.SqlManager: Executing SQL statement: SELECT 
> t.* FROM `Dummy` AS t LIMIT 1
> 12/07/05 11:09:16 INFO orm.CompilationManager: HADOOP_HOME is 
> /HADOOP/hadoop-0.20.2/bin/..
> 12/07/05 11:09:16 INFO orm.CompilationManager: Found hadoop core jar at: 
> /HADOOP/hadoop-0.20.2/bin/../hadoop-0.20.2-core.jar
> Note: 
> /tmp/sqoop-mediaadmin/compile/382d1c58323cea76efd197632bebbfcd/Dummy.java 
> uses or overrides a deprecated API.
> Note: Recompile with -Xlint:deprecation for details.
> 12/07/05 11:09:17 INFO orm.CompilationManager: Writing jar file: 
> /tmp/sqoop-mediaadmin/compile/382d1c58323cea76efd197632bebbfcd/Dummy.jar
> 12/07/05 11:09:17 WARN manager.MySQLManager: It looks like you are importing 
> from mysql.
> 12/07/05 11:09:17 WARN manager.MySQLManager: This transfer can be faster! Use 
> the --direct
> 12/07/05 11:09:17 WARN manager.MySQLManager: option to exercise a 
> MySQL-specific fast path.
> 12/07/05 11:09:17 INFO manager.MySQLManager: Setting zero DATETIME behavior 
> to convertToNull (mysql)
> 12/07/05 11:09:17 INFO mapreduce.ImportJobBase: Beginning import of Dummy
> 12/07/05 11:09:18 INFO db.DataDrivenDBInputFormat: BoundingValsQuery: SELECT 
> MIN(`Sno`), MAX(`Sno`) FROM `Dummy`
> 12/07/05 11:09:18 INFO mapred.JobClient: Running job: job_201207051104_0001
> 12/07/05 11:09:19 INFO mapred.JobClient:  map 0% reduce 0%
> 12/07/05 11:09:33 INFO mapred.JobClient:  map 100% reduce 0%
> 12/07/05 11:09:35 INFO mapred.JobClient: Job complete: job_201207051104_0001
> 12/07/05 11:09:35 INFO mapred.JobClient: Counters: 5
> 12/07/05 11:09:35 INFO mapred.JobClient:   Job Counters 
> 12/07/05 11:09:35 INFO mapred.JobClient:     Launched map tasks=1
> 12/07/05 11:09:35 INFO mapred.JobClient:   FileSystemCounters
> 12/07/05 11:09:35 INFO mapred.JobClient:     HDFS_BYTES_WRITTEN=8
> 12/07/05 11:09:35 INFO mapred.JobClient:   Map-Reduce Framework
> 12/07/05 11:09:35 INFO mapred.JobClient:     Map input records=1
> 12/07/05 11:09:35 INFO mapred.JobClient:     Spilled Records=0
> 12/07/05 11:09:35 INFO mapred.JobClient:     Map output records=1
> 12/07/05 11:09:35 INFO mapreduce.ImportJobBase: Transferred 8 bytes in 17.945 
> seconds (0.4458 bytes/sec)
> 12/07/05 11:09:35 INFO mapreduce.ImportJobBase: Retrieved 1 records.
> 12/07/05 11:09:35 INFO hive.HiveImport: Removing temporary files from import 
> process: Dummy/_logs
> 12/07/05 11:09:35 INFO hive.HiveImport: Loading uploaded data into Hive
> 12/07/05 11:09:35 INFO manager.SqlManager: Executing SQL statement: SELECT 
> t.* FROM `Dummy` AS t LIMIT 1
> 12/07/05 11:09:37 INFO hive.HiveImport: Logging initialized using 
> configuration in 
> jar:file:/HADOOP/hive/lib/hive-common-0.8.1.jar!/hive-log4j.properties
> 12/07/05 11:09:37 INFO hive.HiveImport: Hive history 
> file=/tmp/mediaadmin/hive_job_log_mediaadmin_201207051109_1901926452.txt
> 12/07/05 11:09:41 INFO hive.HiveImport: OK
> 12/07/05 11:09:41 INFO hive.HiveImport: Time taken: 3.934 seconds
> 12/07/05 11:09:41 INFO hive.HiveImport: Loading data to table 
> default.dummyhive
> 12/07/05 11:09:41 INFO hive.HiveImport: OK
> 12/07/05 11:09:41 INFO hive.HiveImport: Time taken: 0.262 seconds
> 12/07/05 11:09:41 INFO hive.HiveImport: Hive import complete.
> 
> 
> 
> Why is it so? Please help me out
> 
> Thanks & Regards
> Yogesh Kumar
> 
> Please do not print this email unless it is absolutely necessary.
> The information contained in this electronic message and any attachments to 
> this message are intended for the exclusive use of the addressee(s) and may 
> contain proprietary, confidential or privileged information. If you are not 
> the intended recipient, you should not disseminate, distribute or copy this 
> e-mail. Please notify the sender immediately and destroy all copies of this 
> message and any attachments.
> WARNING: Computer viruses can be transmitted via email. The recipient should 
> check this email and any attachments for the presence of viruses. The company 
> accepts no liability for any damage caused by any virus transmitted by this 
> email.
> www.wipro.com
> 
> 
> Please do not print this email unless it is absolutely necessary.
> The information contained in this electronic message and any attachments to 
> this message are intended for the exclusive use of the addressee(s) and may 
> contain proprietary, confidential or privileged information. If you are not 
> the intended recipient, you should not disseminate, distribute or copy this 
> e-mail. Please notify the sender immediately and destroy all copies of this 
> message and any attachments.
> WARNING: Computer viruses can be transmitted via email. The recipient should 
> check this email and any attachments for the presence of viruses. The company 
> accepts no liability for any damage caused by any virus transmitted by this 
> email.
> www.wipro.com
> 
> 
> Please do not print this email unless it is absolutely necessary.
> The information contained in this electronic message and any attachments to 
> this message are intended for the exclusive use of the addressee(s) and may 
> contain proprietary, confidential or privileged information. If you are not 
> the intended recipient, you should not disseminate, distribute or copy this 
> e-mail. Please notify the sender immediately and destroy all copies of this 
> message and any attachments.
> WARNING: Computer viruses can be transmitted via email. The recipient should 
> check this email and any attachments for the presence of viruses. The company 
> accepts no liability for any damage caused by any virus transmitted by this 
> email.
> www.wipro.com
> 
> 
> Please do not print this email unless it is absolutely necessary.
> The information contained in this electronic message and any attachments to 
> this message are intended for the exclusive use of the addressee(s) and may 
> contain proprietary, confidential or privileged information. If you are not 
> the intended recipient, you should not disseminate, distribute or copy this 
> e-mail. Please notify the sender immediately and destroy all copies of this 
> message and any attachments.
> WARNING: Computer viruses can be transmitted via email. The recipient should 
> check this email and any attachments for the presence of viruses. The company 
> accepts no liability for any damage caused by any virus transmitted by this 
> email.
> www.wipro.com
> 
> 
> Please do not print this email unless it is absolutely necessary.
> The information contained in this electronic message and any attachments to 
> this message are intended for the exclusive use of the addressee(s) and may 
> contain proprietary, confidential or privileged information. If you are not 
> the intended recipient, you should not disseminate, distribute or copy this 
> e-mail. Please notify the sender immediately and destroy all copies of this 
> message and any attachments.
> WARNING: Computer viruses can be transmitted via email. The recipient should 
> check this email and any attachments for the presence of viruses. The company 
> accepts no liability for any damage caused by any virus transmitted by this 
> email.
> www.wipro.com
> 
> 
> Please do not print this email unless it is absolutely necessary.
> The information contained in this electronic message and any attachments to 
> this message are intended for the exclusive use of the addressee(s) and may 
> contain proprietary, confidential or privileged information. If you are not 
> the intended recipient, you should not disseminate, distribute or copy this 
> e-mail. Please notify the sender immediately and destroy all copies of this 
> message and any attachments.
> WARNING: Computer viruses can be transmitted via email. The recipient should 
> check this email and any attachments for the presence of viruses. The company 
> accepts no liability for any damage caused by any virus transmitted by this 
> email.
> www.wipro.com

Attachment: signature.asc
Description: Message signed with OpenPGP using GPGMail

Reply via email to