hi guys, Can any one help me on this..I got struck up with this error.
Appreciate Your Help Narasimharao On Fri, Jan 20, 2012 at 8:58 PM, chelikani narasimharao < narasimhar...@gmail.com> wrote: > Hi , > I am using sqoop to get extract the data from HPNeoView database and > then want to write onto Hadoop file system.When i am using the following > command i am getting the below error. Here i am providing full error > details what i was received when i am running the below command > > sqoop import --connect jdbc:hpt4jdbc:// > g4n0601a.houston.hp.com:18650/chema=EDW_INT --driver > com.hp.t4jdbc.HPT4Driver --username boyapatr_write -P --verbose --table > EDW_INT.OPPTY_ALNC_PTNR_F --split-by SRC_SYS_KY --target-dir > /home/narasimharao/opptyfact > > DEBUG manager.SqlManager: Using fetchSize for next query: 1000 > 12/01/20 19:32:31 INFO manager.SqlManager: Executing SQL statement: SELECT > t.* FROM EDW_INT.OPPTY_ALNC_PTNR_F AS t WHERE 1=0 > 12/01/20 19:32:40 DEBUG manager.SqlManager: Using fetchSize for next > query: 1000 > 12/01/20 19:32:40 INFO manager.SqlManager: Executing SQL statement: SELECT > t.* FROM EDW_INT.OPPTY_ALNC_PTNR_F AS t WHERE 1=0 > 12/01/20 19:32:47 DEBUG orm.ClassWriter: selected columns: > 12/01/20 19:32:47 DEBUG orm.ClassWriter: OPPTY_ALNC_PTNR_ID > 12/01/20 19:32:47 DEBUG orm.ClassWriter: SRC_SYS_KY > 12/01/20 19:32:47 DEBUG orm.ClassWriter: PTNR_SRC_SYS_KY > 12/01/20 19:32:47 DEBUG orm.ClassWriter: OPPTY_ID > 12/01/20 19:32:47 DEBUG orm.ClassWriter: PTNR_ID > 12/01/20 19:32:47 DEBUG orm.ClassWriter: PTNR_ROLE_CD > 12/01/20 19:32:47 DEBUG orm.ClassWriter: PRIM_PTNR_FG > 12/01/20 19:32:47 DEBUG orm.ClassWriter: INS_GMT_TS > 12/01/20 19:32:47 DEBUG orm.ClassWriter: UPD_GMT_TS > 12/01/20 19:32:47 DEBUG orm.ClassWriter: LOAD_JOB_NR > 12/01/20 19:32:47 DEBUG orm.ClassWriter: REC_ST_NR > 12/01/20 19:32:47 DEBUG orm.ClassWriter: HPQ_RATING_CD > 12/01/20 19:32:47 DEBUG orm.ClassWriter: SRC_SYS_UPD_TS > 12/01/20 19:32:47 DEBUG orm.ClassWriter: Writing source file: > /tmp/sqoop-narasimharao/compile/eb164bb8e2c0031416457bda6b86dd19/EDW_INT_OPPTY_ALNC_PTNR_F.java > 12/01/20 19:32:47 DEBUG orm.ClassWriter: Table name: > EDW_INT.OPPTY_ALNC_PTNR_F > 12/01/20 19:32:47 DEBUG orm.ClassWriter: Columns: OPPTY_ALNC_PTNR_ID:1, > SRC_SYS_KY:-5, PTNR_SRC_SYS_KY:-5, OPPTY_ID:1, PTNR_ID:1, PTNR_ROLE_CD:1, > PRIM_PTNR_FG:1, INS_GMT_TS:93, UPD_GMT_TS:93, LOAD_JOB_NR:2, REC_ST_NR:5, > HPQ_RATING_CD:1, SRC_SYS_UPD_TS:93, > 12/01/20 19:32:47 DEBUG orm.ClassWriter: sourceFilename is > EDW_INT_OPPTY_ALNC_PTNR_F.java > 12/01/20 19:32:47 DEBUG orm.CompilationManager: Found existing > /tmp/sqoop-narasimharao/compile/eb164bb8e2c0031416457bda6b86dd19/ > 12/01/20 19:32:47 INFO orm.CompilationManager: HADOOP_HOME is > /usr/lib/hadoop > 12/01/20 19:32:47 INFO orm.CompilationManager: Found hadoop core jar at: > /usr/lib/hadoop/hadoop-core.jar > 12/01/20 19:32:48 DEBUG orm.CompilationManager: Adding source file: > /tmp/sqoop-narasimharao/compile/eb164bb8e2c0031416457bda6b86dd19/EDW_INT_OPPTY_ALNC_PTNR_F.java > 12/01/20 19:32:48 DEBUG orm.CompilationManager: Invoking javac with args: > 12/01/20 19:32:48 DEBUG orm.CompilationManager: -sourcepath > 12/01/20 19:32:48 DEBUG orm.CompilationManager: > /tmp/sqoop-narasimharao/compile/eb164bb8e2c0031416457bda6b86dd19/ > 12/01/20 19:32:48 DEBUG orm.CompilationManager: -d > 12/01/20 19:32:48 DEBUG orm.CompilationManager: > /tmp/sqoop-narasimharao/compile/eb164bb8e2c0031416457bda6b86dd19/ > 12/01/20 19:32:48 DEBUG orm.CompilationManager: -classpath > > narasimharao/compile/eb164bb8e2c0031416457bda6b86dd19/EDW_INT.OPPTY_ALNC_PTNR_F.jar > 12/01/20 19:32:52 DEBUG orm.CompilationManager: Scanning for .class files > in directory: > /tmp/sqoop-narasimharao/compile/eb164bb8e2c0031416457bda6b86dd19 > 12/01/20 19:32:52 DEBUG orm.CompilationManager: Got classfile: > /tmp/sqoop-narasimharao/compile/eb164bb8e2c0031416457bda6b86dd19/EDW_INT_OPPTY_ALNC_PTNR_F.class > -> EDW_INT_OPPTY_ALNC_PTNR_F.class > 12/01/20 19:32:52 DEBUG orm.CompilationManager: Finished writing jar file > /tmp/sqoop-narasimharao/compile/eb164bb8e2c0031416457bda6b86dd19/EDW_INT.OPPTY_ALNC_PTNR_F.jar > 12/01/20 19:32:52 INFO mapreduce.ImportJobBase: Beginning import of > EDW_INT.OPPTY_ALNC_PTNR_F > 12/01/20 19:32:55 DEBUG manager.SqlManager: Using fetchSize for next > query: 1000 > 12/01/20 19:32:55 INFO manager.SqlManager: Executing SQL statement: SELECT > t.* FROM EDW_INT.OPPTY_ALNC_PTNR_F AS t WHERE 1=0 > 12/01/20 19:33:05 DEBUG mapreduce.DataDrivenImportJob: Using table class: > EDW_INT_OPPTY_ALNC_PTNR_F > 12/01/20 19:33:05 DEBUG mapreduce.DataDrivenImportJob: Using InputFormat: > class com.cloudera.sqoop.mapreduce.db.DataDrivenDBInputFormat > 12/01/20 19:33:06 DEBUG mapreduce.JobBase: Adding to job classpath: > file:/usr/lib/sqoop/sqoop-1.3.0-cdh3u2.jar > 12/01/20 19:33:06 DEBUG mapreduce.JobBase: Adding to job classpath: > file:/usr/lib/sqoop/lib/hpt4jdbc.jar > 12/01/20 19:33:06 DEBUG mapreduce.JobBase: Adding to job classpath: > file:/usr/lib/sqoop/sqoop-1.3.0-cdh3u2.jar > 12/01/20 19:33:06 DEBUG mapreduce.JobBase: Adding to job classpath: > file:/usr/lib/sqoop/sqoop-1.3.0-cdh3u2.jar > 12/01/20 19:33:06 DEBUG mapreduce.JobBase: Adding to job classpath: > file:/usr/lib/sqoop/lib/hadoop-mrunit-0.20.2-CDH3b2-SNAPSHOT.jar > 12/01/20 19:33:06 DEBUG mapreduce.JobBase: Adding to job classpath: > file:/usr/lib/sqoop/lib/hpt4jdbc.jar > 12/01/20 19:33:06 DEBUG mapreduce.JobBase: Adding to job classpath: > file:/usr/lib/sqoop/lib/jackson-mapper-asl-1.7.3.jar > 12/01/20 19:33:06 DEBUG mapreduce.JobBase: Adding to job classpath: > file:/usr/lib/sqoop/lib/paranamer-2.3.jar > 12/01/20 19:33:06 DEBUG mapreduce.JobBase: Adding to job classpath: > file:/usr/lib/sqoop/lib/jackson-core-asl-1.7.3.jar > 12/01/20 19:33:06 DEBUG mapreduce.JobBase: Adding to job classpath: > file:/usr/lib/sqoop/lib/avro-mapred-1.5.4.jar > 12/01/20 19:33:06 DEBUG mapreduce.JobBase: Adding to job classpath: > file:/usr/lib/sqoop/lib/ant-contrib-1.0b3.jar > 12/01/20 19:33:06 DEBUG mapreduce.JobBase: Adding to job classpath: > file:/usr/lib/sqoop/lib/commons-io-1.4.jar > 12/01/20 19:33:06 DEBUG mapreduce.JobBase: Adding to job classpath: > file:/usr/lib/sqoop/lib/jopt-simple-3.2.jar > 12/01/20 19:33:06 DEBUG mapreduce.JobBase: Adding to job classpath: > file:/usr/lib/sqoop/lib/avro-ipc-1.5.4.jar > 12/01/20 19:33:06 DEBUG mapreduce.JobBase: Adding to job classpath: > file:/usr/lib/sqoop/lib/snappy-java-1.0.3.2.jar > 12/01/20 19:33:06 DEBUG mapreduce.JobBase: Adding to job classpath: > file:/usr/lib/sqoop/lib/ant-eclipse-1.0-jvm1.2.jar > 12/01/20 19:33:06 DEBUG mapreduce.JobBase: Adding to job classpath: > file:/usr/lib/sqoop/lib/avro-1.5.4.jar > 12/01/20 19:33:28 INFO db.DataDrivenDBInputFormat: BoundingValsQuery: > SELECT MIN(SRC_SYS_KY), MAX(SRC_SYS_KY) FROM EDW_INT.OPPTY_ALNC_PTNR_F > 12/01/20 19:33:29 DEBUG db.IntegerSplitter: Splits: [ > 113 to 126] into 4 parts > 12/01/20 19:33:29 DEBUG db.IntegerSplitter: 113 > 12/01/20 19:33:29 DEBUG db.IntegerSplitter: 117 > 12/01/20 19:33:29 DEBUG db.IntegerSplitter: 120 > 12/01/20 19:33:29 DEBUG db.IntegerSplitter: 123 > 12/01/20 19:33:29 DEBUG db.IntegerSplitter: 126 > 12/01/20 19:33:35 INFO mapred.JobClient: Running job: job_201201201838_0001 > 12/01/20 19:33:36 INFO mapred.JobClient: map 0% reduce 0% > 12/01/20 19:34:18 INFO mapred.JobClient: Task Id : > attempt_201201201838_0001_m_000000_0, Status : FAILED > java.io.IOException: SQLException in nextKeyValue > at > com.cloudera.sqoop.mapreduce.db.DBRecordReader.nextKeyValue(DBRecordReader.java:251) > at > org.apache.hadoop.mapred.MapTask$NewTrackingRecordReader.nextKeyValue(MapTask.java:456) > at org.apache.hadoop.mapreduce.MapContext.nextKeyValue(MapContext.java:67) > at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:143) > at > com.cloudera.sqoop.mapreduce.AutoProgressMapper.run(AutoProgressMapper.java:189) > at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:647) > at org.apache.hadoop.mapred.MapTask.run(MapTask.java:323) > at org.apache.hadoop.mapred.Child$4.run(Child.java:270) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:396) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1127) > at org.apache.hadoop.mapred.Child.main(Child.java:264) > Caused by: com.hp.t4jdbc.HPT4Exception: *** ERROR[15001] A syntax error > occurred at or before: > SELECT OPPTY_ALNC_PTNR_ID, SR > attempt_201201201838_0001_m_000000_0: log4j:WARN No appenders could be > found for logger (org.apache.hadoop.hdfs.DFSClient). > attempt_201201201838_0001_m_000000_0: log4j:WARN Please initialize the > log4j system properly. > 12/01/20 19:34:28 INFO mapred.JobClient: Task Id : > attempt_201201201838_0001_m_000001_0, Status : FAILED > java.io.IOException: SQLException in nextKeyValue > at > com.cloudera.sqoop.mapreduce.db.DBRecordReader.nextKeyValue(DBRecordReader.java:251) > at > org.apache.hadoop.mapred.MapTask$NewTrackingRecordReader.nextKeyValue(MapTask.java:456) > at org.apache.hadoop.mapreduce.MapContext.nextKeyValue(MapContext.java:67) > at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:143) > at > com.cloudera.sqoop.mapreduce.AutoProgressMapper.run(AutoProgressMapper.java:189) > at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:647) > at org.apache.hadoop.mapred.MapTask.run(MapTask.java:323) > at org.apache.hadoop.mapred.Child$4.run(Child.java:270) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:396) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1127) > at org.apache.hadoop.mapred.Child.main(Child.java:264) > Caused by: com.hp.t4jdbc.HPT4Exception: *** ERROR[15001] A syntax error > occurred at or before: > SELECT OPPTY_ALNC_PTNR_ID, SR > attempt_201201201838_0001_m_000001_0: log4j:WARN No appenders could be > found for logger (org.apache.hadoop.hdfs.DFSClient). > attempt_201201201838_0001_m_000001_0: log4j:WARN Please initialize the > log4j system properly. > 12/01/20 19:35:13 INFO mapred.JobClient: Task Id : > attempt_201201201838_0001_m_000000_1, Status : FAILED > java.io.IOException: SQLException in nextKeyValue > at > com.cloudera.sqoop.mapreduce.db.DBRecordReader.nextKeyValue(DBRecordReader.java:251) > at > org.apache.hadoop.mapred.MapTask$NewTrackingRecordReader.nextKeyValue(MapTask.java:456) > at org.apache.hadoop.mapreduce.MapContext.nextKeyValue(MapContext.java:67) > at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:143) > at > com.cloudera.sqoop.mapreduce.AutoProgressMapper.run(AutoProgressMapper.java:189) > at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:647) > at org.apache.hadoop.mapred.MapTask.run(MapTask.java:323) > at org.apache.hadoop.mapred.Child$4.run(Child.java:270) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:396) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1127) > at org.apache.hadoop.mapred.Child.main(Child.java:264) > Caused by: com.hp.t4jdbc.HPT4Exception: *** ERROR[15001] A syntax error > occurred at or before: > SELECT OPPTY_ALNC_PTNR_ID, SR > attempt_201201201838_0001_m_000000_1: log4j:WARN No appenders could be > found for logger (org.apache.hadoop.hdfs.DFSClient). > attempt_201201201838_0001_m_000000_1: log4j:WARN Please initialize the > log4j system properly. > 12/01/20 19:35:32 INFO mapred.JobClient: Task Id : > attempt_201201201838_0001_m_000000_2, Status : FAILED > *java.io.IOException: SQLException in nextKeyValue* > at > com.cloudera.sqoop.mapreduce.db.DBRecordReader.nextKeyValue(DBRecordReader.java:251) > at > org.apache.hadoop.mapred.MapTask$NewTrackingRecordReader.nextKeyValue(MapTask.java:456) > at org.apache.hadoop.mapreduce.MapContext.nextKeyValue(MapContext.java:67) > at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:143) > at > com.cloudera.sqoop.mapreduce.AutoProgressMapper.run(AutoProgressMapper.java:189) > at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:647) > at org.apache.hadoop.mapred.MapTask.run(MapTask.java:323) > at org.apache.hadoop.mapred.Child$4.run(Child.java:270) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:396) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1127) > at org.apache.hadoop.mapred.Child.main(Child.java:264) > *Caused by: com.hp.t4jdbc.HPT4Exception: *** ERROR[15001] A syntax error > occurred at or before: > SELECT OPPTY_ALNC_PTNR_ID, SR > attempt_201201201838_0001_m_000000_2: log4j:WARN No appenders could be > found for logger (org.apache.hadoop.hdfs.DFSClient). > attempt_201201201838_0001_m_000000_2: log4j:WARN Please initialize the > log4j system properly.* > 12/01/20 19:35:52 INFO mapred.JobClient: Job complete: > job_201201201838_0001 > 12/01/20 19:35:52 INFO mapred.JobClient: Counters: 6 > 12/01/20 19:35:52 INFO mapred.JobClient: Job Counters > 12/01/20 19:35:52 INFO mapred.JobClient: SLOTS_MILLIS_MAPS=245116 > 12/01/20 19:35:52 INFO mapred.JobClient: Total time spent by all > reduces waiting after reserving slots (ms)=0 > 12/01/20 19:35:52 INFO mapred.JobClient: Total time spent by all maps > waiting after reserving slots (ms)=0 > 12/01/20 19:35:52 INFO mapred.JobClient: Launched map tasks=6 > 12/01/20 19:35:52 INFO mapred.JobClient: SLOTS_MILLIS_REDUCES=0 > 12/01/20 19:35:52 INFO mapred.JobClient: Failed map tasks=1 > 12/01/20 19:35:52 INFO mapreduce.ImportJobBase: Transferred 0 bytes in > 165.7817 seconds (0 bytes/sec) > 12/01/20 19:35:52 INFO mapreduce.ImportJobBase: Retrieved 0 records. > *12/01/20 19:35:52 ERROR tool.ImportTool: Error during import: Import job > failed* > * > * > * > * > *Please help me if you have any thoughts.* > * > * > *Advanced thanks,* > *Narasimharao* >