Sent from my iPhone
On Aug 31, 2011, at 1:44 PM, Arvind Prabhakar <arv...@apache.org> wrote: > Srini, > > This is happening because the GSS config Jar is not getting put in > Distributed Cache. Sqoop only puts certain jars in the cache as > opposed to putting every jar that exists in its classpath. In order to > force any Jar to be put in the Distributed Cache, you must copy it > over to Sqoop's lib directory. > > Thanks, > Arvind > > On Tue, Aug 30, 2011 at 9:36 PM, SRINIVAS SURASANI <vas...@gmail.com> wrote: >> Getting the error while exporting .. And from my observation while compiling >> .java, sets the classpath for terajdbc4.jar and tdgssconfig.jar ( as I >> marked in bold letters below) but just before launching map-reduce adding >> jar classpath of tdgssconfig.jar. >> I set the HADOOP_CLASSPTH=<path to> terajdbc4.jar:<path to>tdgssconfig.jar >> Any Help Appreciated. >> $ sqoop export --verbose --driver com.teradata.jdbc.TeraDriver --connect >> jdbc:teradata://TD/DB --username WBD -P --table DB.Temp_Table --export-dir >> /user/hadoop/sqoop_test.txt --fields-terminated-by , --lines-terminated-by >> \n -m 1> >> 11/08/30 22:59:43 DEBUG tool.BaseSqoopTool: Enabled debug logging. >> Enter password: >> 11/08/30 22:59:50 DEBUG sqoop.ConnFactory: Loaded manager factory: >> com.cloudera.sqoop.manager.DefaultManagerFactory >> 11/08/30 22:59:50 DEBUG sqoop.ConnFactory: Trying ManagerFactory: >> com.cloudera.sqoop.manager.DefaultManagerFactory >> 11/08/30 22:59:50 INFO manager.SqlManager: Using default fetchSize of 1000 >> 11/08/30 22:59:50 DEBUG sqoop.ConnFactory: Instantiated ConnManager >> com.cloudera.sqoop.manager.GenericJdbcManager@2b76e552 >> 11/08/30 22:59:50 INFO tool.CodeGenTool: Beginning code generation >> 11/08/30 22:59:51 DEBUG manager.SqlManager: Using fetchSize for next query: >> 1000 >> 11/08/30 22:59:51 INFO manager.SqlManager: Executing SQL statement: SELECT >> t.* FROM DB.Temp_Table AS t WHERE 1=0 >> 11/08/30 22:59:51 DEBUG manager.SqlManager: Using fetchSize for next query: >> 1000 >> 11/08/30 22:59:51 INFO manager.SqlManager: Executing SQL statement: SELECT >> t.* FROM DB.Temp_Table AS t WHERE 1=0 >> 11/08/30 22:59:51 DEBUG orm.ClassWriter: selected columns: >> 11/08/30 22:59:51 DEBUG orm.ClassWriter: NAME >> 11/08/30 22:59:51 DEBUG orm.ClassWriter: SALARY >> 11/08/30 22:59:51 DEBUG orm.ClassWriter: Writing source file: >> /tmp/sqoop-hadoop/compile/cc68f6e38603d6705ad75a3d8ea1bf3a/DB_Temp_Table.java >> 11/08/30 22:59:51 DEBUG orm.ClassWriter: Table name:DB.Temp_Table >> 11/08/30 22:59:51 DEBUG orm.ClassWriter: Columns: NAME:12, SALARY:3, >> 11/08/30 22:59:51 DEBUG orm.ClassWriter: sourceFilename is >> DB_Temp_Table.java >> 11/08/30 22:59:51 DEBUG orm.CompilationManager: Found existing >> /tmp/sqoop-hadoop/compile/cc68f6e38603d6705ad75a3d8ea1bf3a/ >> 11/08/30 22:59:51 INFO orm.CompilationManager: HADOOP_HOME is >> /usr/lib/hadoop >> 11/08/30 22:59:51 INFO orm.CompilationManager: Found hadoop core jar at: >> /usr/lib/hadoop/hadoop-0.20.2-cdh3u0-core.jar >> 11/08/30 22:59:51 DEBUG orm.CompilationManager: Adding source file: >> /tmp/sqoop-haoop/compile/cc68f6e38603d6705ad75a3d8ea1bf3a/DB_Temp_Table.java >> 11/08/30 22:59:51 DEBUG orm.CompilationManager: Invoking javac with args: >> 11/08/30 22:59:51 DEBUG orm.CompilationManager: -sourcepath >> 11/08/30 22:59:51 DEBUG orm.CompilationManager: >> /tmp/sqoop-hadoop/compile/cc68f6e38603d6705ad75a3d8ea1bf3a/ >> 11/08/30 22:59:51 DEBUG orm.CompilationManager: -d >> 11/08/30 22:59:51 DEBUG orm.CompilationManager: >> /tmp/sqoop-hadoop/compile/cc68f6e38603d6705ad75a3d8ea1bf3a/ >> 11/08/30 22:59:51 DEBUG orm.CompilationManager: -classpath >> 11/08/30 22:59:51 DEBUG orm.CompilationManager: >> /usr/lib/hadoop:/usr/lib/hadoop/hadoop-core-0.20.2-cdh3u0.jar:/usr/lib/hadoop/lib/ant-contrib-1.0b3.jar:/usr/lib/hadoop/lib/aspectjrt-1.6.5.jar:/usr/lib/hadoop/lib/aspectjtools-1.6.5.jar:/usr/lib/hadoop/lib/commons-cli-1.2.jar:/usr/lib/hadoop/lib/commons-codec-1.4.jar:/usr/lib/hadoop/lib/commons-daemon-1.0.1.jar:/usr/lib/hadoop/lib/commons-el-1.0.jar:/usr/lib/hadoop/lib/commons-httpclient-3.0.1.jar:/usr/lib/hadoop/lib/commons-logging-1.0.4.jar:/usr/lib/hadoop/lib/commons-logging-api-1.0.4.jar:/usr/lib/hadoop/lib/commons-net-1.4.1.jar:/usr/lib/hadoop/lib/core-3.1.1.jar:/usr/lib/hadoop/lib/elephant-bird-1.0.jar:/usr/lib/hadoop/lib/hadoop-fairscheduler-0.20.2-cdh3u0.jar:/usr/lib/hadoop/lib/hadoop-lzo-0.4.8.jar:/usr/lib/hadoop/lib/hsqldb-1.8.0.10.jar:/usr/lib/hadoop/lib/jackson-core-asl-1.5.2.jar:/usr/lib/hadoop/lib/jackson-mapper-asl-1.5.2.jar:/usr/lib/hadoop/lib/jasper-compiler-5.5.12.jar:/usr/lib/hadoop/lib/jasper-runtime-5.5.12.jar:/usr/lib/hadoop/lib/jets3t-0.6.1.jar:/usr/lib/hadoop/lib/jetty-6.1.26.jar:/usr/lib/hadoop/lib/jetty-servlet-tester-6.1.26.jar:/usr/lib/hadoop/lib/jetty-util-6.1.26.jar:/usr/lib/hadoop/lib/jsch-0.1.42.jar:/usr/lib/hadoop/lib/junit-4.5.jar:/usr/lib/hadoop/lib/kfs-0.2.2.jar:/usr/lib/hadoop/lib/log4j-1.2.15.jar:/usr/lib/hadoop/lib/mockito-all-1.8.2.jar:/usr/lib/hadoop/lib/oro-2.0.8.jar:/usr/lib/hadoop/lib/protobuf-java-2.3.0.jar:/usr/lib/hadoop/lib/servlet-api-2.5-20081211.jar:/usr/lib/hadoop/lib/servlet-api-2.5-6.1.14.jar:/usr/lib/hadoop/lib/slf4j-api-1.4.3.jar:/usr/lib/hadoop/lib/slf4j-api-1.5.8.jar:/usr/lib/hadoop/lib/slf4j-log4j12-1.4.3.jar:/usr/lib/hadoop/lib/slf4j-log4j12-1.5.10.jar:/usr/lib/hadoop/lib/xmlenc-0.52.jar:/usr/lib/hadoop/lib/yamlbeans-0.9.3.jar:/usr/lib/hadoop/lib/jsp-2.1/jsp-2.1.jar:/usr/lib/hadoop/lib/jsp-2.1/jsp-api-2.1.jar:/usr/lib/sqoop/conf:/usr/lib/hbase/conf::/usr/lib/sqoop/lib/ant-contrib-1.0b3.jar:/usr/lib/sqoop/lib/ant-eclipse-1.0-jvm1.2.jar:/usr/lib/sqoop/lib/commons-io-1.4.jar:/usr/lib/sqoop/lib/hadoop-mrunit-0.20.2-CDH3b2-SNAPSHOT.jar:/usr/lib/sqoop/lib/ivy-2.0.0-rc2.jar:/usr/lib/sqoop/lib/sqljdbc4.jar:/usr/lib/hbase/hbase-0.90.1-cdh3u0-tests.jar:/usr/lib/hbase/hbase-0.90.1-cdh3u0.jar:/usr/lib/hbase/lib/activation-1.1.jar:/usr/lib/hbase/lib/asm-3.1.jar:/usr/lib/hbase/lib/avro-1.3.3.jar:/usr/lib/hbase/lib/commons-cli-1.2.jar:/usr/lib/hbase/lib/commons-codec-1.4.jar:/usr/lib/hbase/lib/commons-el-1.0.jar:/usr/lib/hbase/lib/commons-httpclient-3.1.jar:/usr/lib/hbase/lib/commons-lang-2.5.jar:/usr/lib/hbase/lib/commons-logging-1.1.1.jar:/usr/lib/hbase/lib/commons-net-1.4.1.jar:/usr/lib/hbase/lib/core-3.1.1.jar:/usr/lib/hbase/lib/guava-r06.jar:/usr/lib/hbase/lib/hadoop-core.jar:/usr/lib/hbase/lib/hbase-0.90.1-cdh3u0.jar:/usr/lib/hbase/lib/jackson-core-asl-1.5.2.jar:/usr/lib/hbase/lib/jackson-jaxrs-1.5.5.jar:/usr/lib/hbase/lib/jackson-mapper-asl-1.5.2.jar:/usr/lib/hbase/lib/jackson-xc-1.5.5.jar:/usr/lib/hbase/lib/jasper-compiler-5.5.23.jar:/usr/lib/hbase/lib/jasper-runtime-5.5.23.jar:/usr/lib/hbase/lib/jaxb-api-2.1.jar:/usr/lib/hbase/lib/jaxb-impl-2.1.12.jar:/usr/lib/hbase/lib/jersey-core-1.4.jar:/usr/lib/hbase/lib/jersey-json-1.4.jar:/usr/lib/hbase/lib/jersey-server-1.4.jar:/usr/lib/hbase/lib/jettison-1.1.jar:/usr/lib/hbase/lib/jetty-6.1.26.jar:/usr/lib/hbase/lib/jetty-util-6.1.26.jar:/usr/lib/hbase/lib/jruby-complete-1.0.3.jar:/usr/lib/hbase/lib/jsp-2.1-6.1.14.jar:/usr/lib/hbase/lib/jsp-api-2.1-6.1.14.jar:/usr/lib/hbase/lib/jsp-api-2.1.jar:/usr/lib/hbase/lib/jsr311-api-1.1.1.jar:/usr/lib/hbase/lib/log4j-1.2.16.jar:/usr/lib/hbase/lib/protobuf-java-2.3.0.jar:/usr/lib/hbase/lib/servlet-api-2.5-6.1.14.jar:/usr/lib/hbase/lib/servlet-api-2.5.jar:/usr/lib/hbase/lib/slf4j-api-1.5.8.jar:/usr/lib/hbase/lib/slf4j-log4j12-1.5.8.jar:/usr/lib/hbase/lib/stax-api-1.0.1.jar:/usr/lib/hbase/lib/thrift-0.2.0.jar:/usr/lib/hbase/lib/xmlenc-0.52.jar:/usr/lib/hbase/lib/zookeeper.jar:/usr/lib/zookeeper/zookeeper-3.3.3-cdh3u0.jar:/usr/lib/zookeeper/zookeeper.jar:/usr/lib/zookeeper/lib/jline-0.9.94.jar:/usr/lib/zookeeper/lib/log4j-1.2.15.jar:/usr/lib/sqoop/sqoop-1.2.0-cdh3u0.jar:/usr/lib/sqoop/sqoop-test-1.2.0-cdh3u0.jar:<somepath>/lib/tdgssconfig.jar:<somepath>/lib/terajdbc4.jar:/usr/lib/hadoop/hadoop-0.20.2-cdh3u0-core.jar:/usr/lib/sqoop/sqoop-1.2.0-cdh3u0.jar >> Note: >> /tmp/sqoop-hadoop/compile/cc68f6e38603d6705ad75a3d8ea1bf3a/DB_Temp_Table.java >> uses or overrides a deprecated API. >> Note: Recompile with -Xlint:deprecation for details. >> 11/08/30 22:59:52 INFO orm.CompilationManager: Writing jar file: >> /tmp/sqoop-hadoop/compile/cc68f6e38603d6705ad75a3d8ea1bf3a/DB.Temp_Table.jar >> 11/08/30 22:59:52 DEBUG orm.CompilationManager: Scanning for .class files in >> directory: /tmp/sqoop-hadoop/compile/cc68f6e38603d6705ad75a3d8ea1bf3a >> 11/08/30 22:59:52 DEBUG orm.CompilationManager: Got classfile: >> /tmp/sqoop-hadoop/compile/cc68f6e38603d6705ad75a3d8ea1bf3a/DBTemp_Table.class >> -> DB_Temp_Table.class >> 11/08/30 22:59:52 DEBUG orm.CompilationManager: Finished writing jar file >> /tmp/sqoop-hadrdev/compile/cc68f6e38603d6705ad75a3d8ea1bf3a/DB.Temp_Table.jar >> 11/08/30 22:59:52 INFO mapreduce.ExportJobBase: Beginning export of >> DB.Temp_Table >> 11/08/30 22:59:52 DEBUG mapreduce.JobBase: Using InputFormat: class >> com.cloudera.sqoop.mapreduce.ExportInputFormat >> 11/08/30 22:59:52 DEBUG manager.SqlManager: Using fetchSize for next query: >> 1000 >> 11/08/30 22:59:52 INFO manager.SqlManager: Executing SQL statement: SELECT >> t.* FROM DB.Temp_Table AS t WHERE 1=0 >> 11/08/30 22:59:53 DEBUG mapreduce.JobBase: Adding to job classpath: >> file:/usr/lib/sqoop/sqoop-1.2.0-cdh3u0.jar >> 11/08/30 22:59:53 DEBUG mapreduce.JobBase: Adding to job classpath: >> file:<somepath>/lib/terajdbc4.jar >> 11/08/30 22:59:53 DEBUG mapreduce.JobBase: Adding to job classpath: >> file:/usr/lib/sqoop/sqoop-1.2.0-cdh3u0.jar >> 11/08/30 22:59:53 DEBUG mapreduce.JobBase: Adding to job classpath: >> file:/usr/lib/sqoop/sqoop-1.2.0-cdh3u0.jar >> 11/08/30 22:59:53 DEBUG mapreduce.JobBase: Adding to job classpath: >> file:/usr/lib/sqoop/lib/ivy-2.0.0-rc2.jar >> 11/08/30 22:59:53 DEBUG mapreduce.JobBase: Adding to job classpath: >> file:/usr/lib/sqoop/lib/commons-io-1.4.jar >> 11/08/30 22:59:53 DEBUG mapreduce.JobBase: Adding to job classpath: >> file:/usr/lib/sqoop/lib/ant-eclipse-1.0-jvm1.2.jar >> 11/08/30 22:59:53 DEBUG mapreduce.JobBase: Adding to job classpath: >> file:/usr/lib/sqoop/lib/sqljdbc4.jar >> 11/08/30 22:59:53 DEBUG mapreduce.JobBase: Adding to job classpath: >> file:/usr/lib/sqoop/lib/ant-contrib-1.0b3.jar >> 11/08/30 22:59:53 DEBUG mapreduce.JobBase: Adding to job classpath: >> file:/usr/lib/sqoop/lib/hadoop-mrunit-0.20.2-CDH3b2-SNAPSHOT.jar >> 11/08/30 22:59:53 INFO hdfs.DFSClient: Created HDFS_DELEGATION_TOKEN token >> 795 for hadoop >> 11/08/30 22:59:53 INFO security.TokenCache: Got dt for >> hdfs://<cname>:9000/tmp/hadoop-mapred/mapred/staging/hadoop/.staging/job_201107010928_0398/libjars/hadoop-mrunit-0.20.2-CDH3b2-SNAPSHOT.jar;uri=xx.xxx.xx.xx:9000;t.service=xx.xxx.xx.xx:9000 >> 11/08/30 22:59:53 INFO input.FileInputFormat: Total input paths to process : >> 1 >> 11/08/30 22:59:53 DEBUG mapreduce.ExportInputFormat: Target numMapTasks=1 >> 11/08/30 22:59:53 DEBUG mapreduce.ExportInputFormat: Total input bytes=18 >> 11/08/30 22:59:53 DEBUG mapreduce.ExportInputFormat: maxSplitSize=18 >> 11/08/30 22:59:53 INFO input.FileInputFormat: Total input paths to process : >> 1 >> 11/08/30 22:59:53 DEBUG mapreduce.ExportInputFormat: Generated splits: >> 11/08/30 22:59:53 DEBUG mapreduce.ExportInputFormat: >> Paths:/user/hadrdev/sqoop_test.txt:0+18 Locations: >> 11/08/30 22:59:53 INFO mapred.JobClient: Running job: job_201107010928_0398 >> 11/08/30 22:59:54 INFO mapred.JobClient: map 0% reduce 0% >> 11/08/30 23:00:01 INFO mapred.JobClient: Task Id : >> attempt_201107010928_0398_m_000000_0, Status : FAILED >> java.io.IOException: java.lang.NullPointerException >> at >> com.cloudera.sqoop.mapreduce.ExportOutputFormat.getRecordWriter(ExportOutputFormat.java:80) >> at >> org.apache.hadoop.mapred.MapTask$NewDirectOutputCollector.<init>(MapTask.java:520) >> at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:635) >> at org.apache.hadoop.mapred.MapTask.run(MapTask.java:322) >> at org.apache.hadoop.mapred.Child$4.run(Child.java:268) >> at java.security.AccessController.doPrivileged(Native Method) >> at javax.security.auth.Subject.doAs(Subject.java:396) >> at >> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1115) >> at org.apache.hadoop.mapred.Child.main(Child.java:262) >> Caused by: java.lang.NullPointerException >> at >> com.teradata.tdgss.jtdgss.TdgssConfigApi.GetMechanisms(DashoA1*..) >> at com.teradata.tdgss.jtdgss.TdgssManager.<init>(DashoA1*..) >> at com.teradata.tdgss.jtdgss.TdgssManager.getInstance(DashoA1*..) >> at >> com.teradata.jdbc.jdbc.GenericTeraEncrypt.getGSSM(GenericTeraEncrypt.java:612) >> at >> com.teradata.jdbc.jdbc.GenericTeraEncrypt.getConfig(GenericTeraEncrypt.java:630) >> at >> com.teradata.jdbc.jdbc.GenericTeraEncrypt.getUserNameForOid(GenericTeraEncrypt.java:723) >> at com.teradata.jdbc.AuthMechanism.<init>(AuthMechanism.java:50) >> at >> com.teradata.jdbc.jdbc.GenericInitDBConfigState.action(GenericInitDBConfigState.java:105) >> at >> com.teradata.jdbc.jdbc.GenericLogonController.run(GenericLogonController.java:49) >> at com.teradata.jdbc.jdbc_4.TDSession.<init>(TDSession.java:199) >> at >> com.teradata.jdbc.jdbc_3.ifjdbc_4.TeraLocalConnection.<init>(TeraLocalConnection.java:95) >> at >> com.teradata.jdbc.jdbc.ConnectionFactory.createConnection(ConnectionFactory.java:54) >> at com.teradata.jdbc.TeraDriver.doConnect(TeraDriver.java:217) >> at com.teradata.jdbc.TeraDriver.connect(TeraDriver.java:150) >> at java.sql.DriverManager.getConnection(DriverManager.java:582) >> at java.sql.DriverManager.getConnection(DriverManager.java:185) >> at >> com.cloudera.sqoop.mapreduce.db.DBConfiguration.getConnection(DBConfiguration.java:184) >> at >> com.cloudera.sqoop.mapreduce.AsyncSqlRecordWriter.<init>(AsyncSqlRecordWriter.java:73) >> at >> com.cloudera.sqoop.mapreduce.ExportOutputFormat$ExportRecordWriter.<init>(ExportOutputFormat.java:96) >> at >> com.cloudera.sqoop.mapreduce.ExportOutputFormat.getRecordWriter(ExportOutputFormat.java:78) >> ... 8 more >> >> attempt_201107010928_0398_m_000000_0: GSSException: Failure unspecified at >> GSS-API level (Mechanism level: UserFile parameter null) >> attempt_201107010928_0398_m_000000_0: at >> com.teradata.tdgss.jtdgss.TdgssParseXml.<init>(DashoA1*..) >> attempt_201107010928_0398_m_000000_0: at >> com.teradata.tdgss.jtdgss.TdgssConfigApi.<init>(DashoA1*..) >> attempt_201107010928_0398_m_000000_0: at >> com.teradata.tdgss.jtdgss.TdgssManager.<init>(DashoA1*..) >> attempt_201107010928_0398_m_000000_0: at >> com.teradata.tdgss.jtdgss.TdgssManager.getInstance(DashoA1*..) >> attempt_201107010928_0398_m_000000_0: at >> com.teradata.jdbc.jdbc.GenericTeraEncrypt.getGSSM(GenericTeraEncrypt.java:612) >> attempt_201107010928_0398_m_000000_0: at >> com.teradata.jdbc.jdbc.GenericTeraEncrypt.getConfig(GenericTeraEncrypt.java:630) >> attempt_201107010928_0398_m_000000_0: at >> com.teradata.jdbc.jdbc.GenericTeraEncrypt.getUserNameForOid( >> On Mon, Aug 29, 2011 at 2:19 PM, SRINIVAS SURASANI <vas...@gmail.com> wrote: >>> >>> Arvind, >>> >>> I have subscribed at sqoop-user@incubator.apache.org and posted Question . >>> Sorry for the inconvinence from my end, since I'm close to deadline I'm >>> taking your valuable time. >>> >>> sqoop list-tables --driver com.teradata.jdbc.TeraDriver --connect >>> jdbc:teradata://PKTD/E1_CMS_WORK --username srini -P >>> Iam getting the following error: >>> 11/08/29 13:08:03 INFO manager.SqlManager: Using default fetchSize of 1000 >>> GSSException: Failure unspecified at GSS-API level (Mechanism level: >>> UserFile parameter null) >>> at com.teradata.tdgss.jtdgss.TdgssParseXml.<init>(DashoA1*..) >>> at com.teradata.tdgss.jtdgss.TdgssConfigApi.<init>(DashoA1*..) >>> at com.teradata.tdgss.jtdgss.TdgssManager.<init>(DashoA1*..) >>> at com.teradata.tdgss.jtdgss.TdgssManager.getInstance(DashoA1*..) >>> at >>> com.teradata.jdbc.jdbc.GenericTeraEncrypt.getGSSM(GenericTeraEncrypt.java:612) >>> at >>> com.teradata.jdbc.jdbc.GenericTeraEncrypt.getConfig(GenericTeraEncrypt.java:630) >>> at >>> com.teradata.jdbc.jdbc.GenericTeraEncrypt.getUserNameForOid(GenericTeraEncrypt.java:723) >>> at com.teradata.jdbc.AuthMechanism.<init>(AuthMechanism.java:50) >>> at >>> com.teradata.jdbc.jdbc.GenericInitDBConfigState.action(GenericInitDBConfigState.java:105) >>> at >>> com.teradata.jdbc.jdbc.GenericLogonController.run(GenericLogonController.java:49) >>> at com.teradata.jdbc.jdbc_4.TDSession.<init>(TDSession.java:199) >>> at >>> com.teradata.jdbc.jdbc_3.ifjdbc_4.TeraLocalConnection.<init>(TeraLocalConnection.java:95) >>> at >>> com.teradata.jdbc.jdbc.ConnectionFactory.createConnection(ConnectionFactory.java:54) >>> at com.teradata.jdbc.TeraDriver.doConnect(TeraDriver.java:217) >>> at com.teradata.jdbc.TeraDriver.connect(TeraDriver.java:150) >>> at java.sql.DriverManager.getConnection(DriverManager.java:582) >>> at java.sql.DriverManager.getConnection(DriverManager.java:185) >>> at >>> com.cloudera.sqoop.manager.SqlManager.makeConnection(SqlManager.java:643) >>> at >>> com.cloudera.sqoop.manager.GenericJdbcManager.getConnection(GenericJdbcManager.java:51) >>> at >>> com.cloudera.sqoop.manager.SqlManager.listTables(SqlManager.java:270) >>> at >>> com.cloudera.sqoop.tool.ListTablesTool.run(ListTablesTool.java:49) >>> at com.cloudera.sqoop.Sqoop.run(Sqoop.java:144) >>> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65) >>> at com.cloudera.sqoop.Sqoop.runSqoop(Sqoop.java:180) >>> at com.cloudera.sqoop.Sqoop.runTool(Sqoop.java:219) >>> at com.cloudera.sqoop.Sqoop.runTool(Sqoop.java:228) >>> at com.cloudera.sqoop.Sqoop.main(Sqoop.java:237) >>> 11/08/29 13:08:04 ERROR sqoop.Sqoop: Got exception running Sqoop: >>> java.lang.NullPointerException >>> java.lang.NullPointerException >>> at >>> com.teradata.tdgss.jtdgss.TdgssConfigApi.GetMechanisms(DashoA1*..) >>> at com.teradata.tdgss.jtdgss.TdgssManager.<init>(DashoA1*..) >>> at com.teradata.tdgss.jtdgss.TdgssManager.getInstance(DashoA1*..) >>> at >>> com.teradata.jdbc.jdbc.GenericTeraEncrypt.getGSSM(GenericTeraEncrypt.java:612) >>> at >>> com.teradata.jdbc.jdbc.GenericTeraEncrypt.getConfig(GenericTeraEncrypt.java:630) >>> at >>> com.teradata.jdbc.jdbc.GenericTeraEncrypt.getUserNameForOid(GenericTeraEncrypt.java:723) >>> at com.teradata.jdbc.AuthMechanism.<init>(AuthMechanism.java:50) >>> at >>> com.teradata.jdbc.jdbc.GenericInitDBConfigState.action(GenericInitDBConfigState.java:105) >>> at >>> com.teradata.jdbc.jdbc.GenericLogonController.run(GenericLogonController.java:49) >>> at com.teradata.jdbc.jdbc_4.TDSession.<init>(TDSession.java:199) >>> at >>> com.teradata.jdbc.jdbc_3.ifjdbc_4.TeraLocalConnection.<init>(TeraLocalConnection.java:95) >>> at >>> com.teradata.jdbc.jdbc.ConnectionFactory.createConnection(ConnectionFactory.java:54) >>> at com.teradata.jdbc.TeraDriver.doConnect(TeraDriver.java:217) >>> at com.teradata.jdbc.TeraDriver.connect(TeraDriver.java:150) >>> at java.sql.DriverManager.getConnection(DriverManager.java:582) >>> at java.sql.DriverManager.getConnection(DriverManager.java:185) >>> at >>> com.cloudera.sqoop.manager.SqlManager.makeConnection(SqlManager.java:643) >>> at >>> com.cloudera.sqoop.manager.GenericJdbcManager.getConnection(GenericJdbcManager.java:51) >>> at >>> com.cloudera.sqoop.manager.SqlManager.listTables(SqlManager.java:270) >>> at >>> com.cloudera.sqoop.tool.ListTablesTool.run(ListTablesTool.java:49) >>> at com.cloudera.sqoop.Sqoop.run(Sqoop.java:144) >>> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65) >>> at com.cloudera.sqoop.Sqoop.runSqoop(Sqoop.java:180) >>> at com.cloudera.sqoop.Sqoop.runTool(Sqoop.java:219) >>> at com.cloudera.sqoop.Sqoop.runTool(Sqoop.java:228) >>> at com.cloudera.sqoop.Sqoop.main(Sqoop.java:237) >>> Thanks, >>> Srini >>> >>> >>> >>> On Mon, Aug 29, 2011 at 12:58 PM, arv...@cloudera.com >>> <arv...@cloudera.com> wrote: >>>> >>>> [Please subscribe and respond to sqoop-user@incubator.apache.org] >>>> >>>> Please use HADOOP_CLASSPATH instead of CLASSPATH. Also, in order to >>>> use the generic JDBC connector, you will have to specify the driver >>>> class explicitly via the command line option --driver >>>> com.teradata.jdbc.TeraDriver. >>>> >>>> Thanks, >>>> Arvind >>>> >>>> On Mon, Aug 29, 2011 at 9:53 AM, SRINIVAS SURASANI <vas...@gmail.com> >>>> wrote: >>>>> Arvind, >>>>> I have set the classpath to teradata4.jar [ not placed the >>>>> teradata4.jar in >>>>> sqoop lib, as I dont have permissions]. >>>>> I'm getting the following error >>>>> >>>>> sqoop list-tables --connect jdbc:teradata://PKTD/E1_CMS_WORK --username >>>>> srini -P >>>>> ERROR: tool.BaseSqoopTool: Got error creating database manager: >>>>> java.io.IOexception: No manager for connect string: >>>>> jdbc:teradata:///PKTD/E1_CMS_WORK >>>>> at com.cloudera.sqoop.ConnFactory.getManager(ConnFactory.java:119) >>>>> at >>>>> com.cloudera.sqoop.tool.BaseSqoopTool.init(BaseSqoopTool.java:186) >>>>> ... >>>>> ... >>>>> Thanks, >>>>> Srini >>>>> >>>>> >>>>> >>>>> On Mon, Aug 29, 2011 at 8:52 AM, SRINIVAS SURASANI <vas...@gmail.com> >>>>> wrote: >>>>>> >>>>>> Thanks-a-lot Arvind. >>>>>> >>>>>> On Mon, Aug 29, 2011 at 8:45 AM, arv...@cloudera.com >>>>>> <arv...@cloudera.com> >>>>>> wrote: >>>>>>> >>>>>>> [Moving the thread to sqoop-user@incubator.apache.org] >>>>>>> >>>>>>> Hi Srini, >>>>>>> >>>>>>> You should be able to use the generic JDBC connector to import/export >>>>>>> from Teradata. There is also a specialized connector that is >>>>>>> available >>>>>>> for use with Teradata if you are interested. This connector is not a >>>>>>> part of Sqoop and can be obtained from Cloudera by going to: >>>>>>> >>>>>>> http://www.cloudera.com/partners/connectors/ >>>>>>> >>>>>>> Thanks, >>>>>>> Arvind >>>>>>> >>>>>>> On Mon, Aug 29, 2011 at 8:17 AM, SRINIVAS SURASANI <vas...@gmail.com> >>>>>>> wrote: >>>>>>>> I have csv file in hadoop and looking to load into Teradata. I was >>>>>>>> wondering does the sqoop works with Terradata.(with JDBC jar >>>>>>>> placing >>>>>>>> in sqoop lib dir). >>>>>>>> >>>>>>>> Regards >>>>>>>> Srini >>>>>>>> >>>>>>>> -- >>>>>>>> NOTE: The mailing list sqoop-u...@cloudera.org is deprecated in >>>>>>>> favor >>>>>>>> of Apache Sqoop mailing list sqoop-user@incubator.apache.org. >>>>>>>> Please >>>>>>>> subscribe to it by sending an email to >>>>>>>> incubator-sqoop-user-subscr...@apache.org. >>>>>>>> >>>>>>> >>>>>>> -- >>>>>>> NOTE: The mailing list sqoop-u...@cloudera.org is deprecated in favor >>>>>>> of >>>>>>> Apache Sqoop mailing list sqoop-user@incubator.apache.org. Please >>>>>>> subscribe >>>>>>> to it by sending an email to >>>>>>> incubator-sqoop-user-subscr...@apache.org. >>>>>> >>>>> >>>>> -- >>>>> NOTE: The mailing list sqoop-u...@cloudera.org is deprecated in favor >>>>> of >>>>> Apache Sqoop mailing list sqoop-user@incubator.apache.org. Please >>>>> subscribe >>>>> to it by sending an email to incubator-sqoop-user-subscr...@apache.org. >>>>> >>>> >>>> -- >>>> NOTE: The mailing list sqoop-u...@cloudera.org is deprecated in favor of >>>> Apache Sqoop mailing list sqoop-user@incubator.apache.org. Please subscribe >>>> to it by sending an email to incubator-sqoop-user-subscr...@apache.org. >>> >> >>