Arvind, I understand to place the GSS config jar to be placed in Sqoop lib directory. I was wondering is there any alternative way to achieve this [ meaning , how terajdbc4.jar is added to is distributed cache automatically before launching Map-Reduce ].
Thanks, Srini On Wed, Aug 31, 2011 at 1:44 PM, Arvind Prabhakar <arv...@apache.org> wrote: > Srini, > > This is happening because the GSS config Jar is not getting put in > Distributed Cache. Sqoop only puts certain jars in the cache as > opposed to putting every jar that exists in its classpath. In order to > force any Jar to be put in the Distributed Cache, you must copy it > over to Sqoop's lib directory. > > Thanks, > Arvind > > On Tue, Aug 30, 2011 at 9:36 PM, SRINIVAS SURASANI <vas...@gmail.com> > wrote: > > Getting the error while exporting .. And from my observation while > compiling > > .java, sets the classpath for terajdbc4.jar and tdgssconfig.jar ( as I > > marked in bold letters below) but just before launching map-reduce adding > > jar classpath of tdgssconfig.jar. > > I set the HADOOP_CLASSPTH=<path to> terajdbc4.jar:<path > to>tdgssconfig.jar > > Any Help Appreciated. > > $ sqoop export --verbose --driver com.teradata.jdbc.TeraDriver --connect > > jdbc:teradata://TD/DB --username WBD -P --table DB.Temp_Table > --export-dir > > /user/hadoop/sqoop_test.txt --fields-terminated-by , > --lines-terminated-by > > \n -m 1> > > 11/08/30 22:59:43 DEBUG tool.BaseSqoopTool: Enabled debug logging. > > Enter password: > > 11/08/30 22:59:50 DEBUG sqoop.ConnFactory: Loaded manager factory: > > com.cloudera.sqoop.manager.DefaultManagerFactory > > 11/08/30 22:59:50 DEBUG sqoop.ConnFactory: Trying ManagerFactory: > > com.cloudera.sqoop.manager.DefaultManagerFactory > > 11/08/30 22:59:50 INFO manager.SqlManager: Using default fetchSize of > 1000 > > 11/08/30 22:59:50 DEBUG sqoop.ConnFactory: Instantiated ConnManager > > com.cloudera.sqoop.manager.GenericJdbcManager@2b76e552 > > 11/08/30 22:59:50 INFO tool.CodeGenTool: Beginning code generation > > 11/08/30 22:59:51 DEBUG manager.SqlManager: Using fetchSize for next > query: > > 1000 > > 11/08/30 22:59:51 INFO manager.SqlManager: Executing SQL statement: > SELECT > > t.* FROM DB.Temp_Table AS t WHERE 1=0 > > 11/08/30 22:59:51 DEBUG manager.SqlManager: Using fetchSize for next > query: > > 1000 > > 11/08/30 22:59:51 INFO manager.SqlManager: Executing SQL statement: > SELECT > > t.* FROM DB.Temp_Table AS t WHERE 1=0 > > 11/08/30 22:59:51 DEBUG orm.ClassWriter: selected columns: > > 11/08/30 22:59:51 DEBUG orm.ClassWriter: NAME > > 11/08/30 22:59:51 DEBUG orm.ClassWriter: SALARY > > 11/08/30 22:59:51 DEBUG orm.ClassWriter: Writing source file: > > > /tmp/sqoop-hadoop/compile/cc68f6e38603d6705ad75a3d8ea1bf3a/DB_Temp_Table.java > > 11/08/30 22:59:51 DEBUG orm.ClassWriter: Table name:DB.Temp_Table > > 11/08/30 22:59:51 DEBUG orm.ClassWriter: Columns: NAME:12, SALARY:3, > > 11/08/30 22:59:51 DEBUG orm.ClassWriter: sourceFilename is > > DB_Temp_Table.java > > 11/08/30 22:59:51 DEBUG orm.CompilationManager: Found existing > > /tmp/sqoop-hadoop/compile/cc68f6e38603d6705ad75a3d8ea1bf3a/ > > 11/08/30 22:59:51 INFO orm.CompilationManager: HADOOP_HOME is > > /usr/lib/hadoop > > 11/08/30 22:59:51 INFO orm.CompilationManager: Found hadoop core jar at: > > /usr/lib/hadoop/hadoop-0.20.2-cdh3u0-core.jar > > 11/08/30 22:59:51 DEBUG orm.CompilationManager: Adding source file: > > > /tmp/sqoop-haoop/compile/cc68f6e38603d6705ad75a3d8ea1bf3a/DB_Temp_Table.java > > 11/08/30 22:59:51 DEBUG orm.CompilationManager: Invoking javac with args: > > 11/08/30 22:59:51 DEBUG orm.CompilationManager: -sourcepath > > 11/08/30 22:59:51 DEBUG orm.CompilationManager: > > /tmp/sqoop-hadoop/compile/cc68f6e38603d6705ad75a3d8ea1bf3a/ > > 11/08/30 22:59:51 DEBUG orm.CompilationManager: -d > > 11/08/30 22:59:51 DEBUG orm.CompilationManager: > > /tmp/sqoop-hadoop/compile/cc68f6e38603d6705ad75a3d8ea1bf3a/ > > 11/08/30 22:59:51 DEBUG orm.CompilationManager: -classpath > > 11/08/30 22:59:51 DEBUG orm.CompilationManager: > > > /usr/lib/hadoop:/usr/lib/hadoop/hadoop-core-0.20.2-cdh3u0.jar:/usr/lib/hadoop/lib/ant-contrib-1.0b3.jar:/usr/lib/hadoop/lib/aspectjrt-1.6.5.jar:/usr/lib/hadoop/lib/aspectjtools-1.6.5.jar:/usr/lib/hadoop/lib/commons-cli-1.2.jar:/usr/lib/hadoop/lib/commons-codec-1.4.jar:/usr/lib/hadoop/lib/commons-daemon-1.0.1.jar:/usr/lib/hadoop/lib/commons-el-1.0.jar:/usr/lib/hadoop/lib/commons-httpclient-3.0.1.jar:/usr/lib/hadoop/lib/commons-logging-1.0.4.jar:/usr/lib/hadoop/lib/commons-logging-api-1.0.4.jar:/usr/lib/hadoop/lib/commons-net-1.4.1.jar:/usr/lib/hadoop/lib/core-3.1.1.jar:/usr/lib/hadoop/lib/elephant-bird-1.0.jar:/usr/lib/hadoop/lib/hadoop-fairscheduler-0.20.2-cdh3u0.jar:/usr/lib/hadoop/lib/hadoop-lzo-0.4.8.jar:/usr/lib/hadoop/lib/hsqldb-1.8.0.10.jar:/usr/lib/hadoop/lib/jackson-core-asl-1.5.2.jar:/usr/lib/hadoop/lib/jackson-mapper-asl-1.5.2.jar:/usr/lib/hadoop/lib/jasper-compiler-5.5.12.jar:/usr/lib/hadoop/lib/jasper-runtime-5.5.12.jar:/usr/lib/hadoop/lib/jets3t-0.6.1.jar:/usr/lib/hadoop/lib/jetty-6.1.26.jar:/usr/lib/hadoop/lib/jetty-servlet-tester-6.1.26.jar:/usr/lib/hadoop/lib/jetty-util-6.1.26.jar:/usr/lib/hadoop/lib/jsch-0.1.42.jar:/usr/lib/hadoop/lib/junit-4.5.jar:/usr/lib/hadoop/lib/kfs-0.2.2.jar:/usr/lib/hadoop/lib/log4j-1.2.15.jar:/usr/lib/hadoop/lib/mockito-all-1.8.2.jar:/usr/lib/hadoop/lib/oro-2.0.8.jar:/usr/lib/hadoop/lib/protobuf-java-2.3.0.jar:/usr/lib/hadoop/lib/servlet-api-2.5-20081211.jar:/usr/lib/hadoop/lib/servlet-api-2.5-6.1.14.jar:/usr/lib/hadoop/lib/slf4j-api-1.4.3.jar:/usr/lib/hadoop/lib/slf4j-api-1.5.8.jar:/usr/lib/hadoop/lib/slf4j-log4j12-1.4.3.jar:/usr/lib/hadoop/lib/slf4j-log4j12-1.5.10.jar:/usr/lib/hadoop/lib/xmlenc-0.52.jar:/usr/lib/hadoop/lib/yamlbeans-0.9.3.jar:/usr/lib/hadoop/lib/jsp-2.1/jsp-2.1.jar:/usr/lib/hadoop/lib/jsp-2.1/jsp-api-2.1.jar:/usr/lib/sqoop/conf:/usr/lib/hbase/conf::/usr/lib/sqoop/lib/ant-contrib-1.0b3.jar:/usr/lib/sqoop/lib/ant-eclipse-1.0-jvm1.2.jar:/usr/lib/sqoop/lib/commons-io-1.4.jar:/usr/lib/sqoop/lib/hadoop-mrunit-0.20.2-CDH3b2-SNAPSHOT.jar:/usr/lib/sqoop/lib/ivy-2.0.0-rc2.jar:/usr/lib/sqoop/lib/sqljdbc4.jar:/usr/lib/hbase/hbase-0.90.1-cdh3u0-tests.jar:/usr/lib/hbase/hbase-0.90.1-cdh3u0.jar:/usr/lib/hbase/lib/activation-1.1.jar:/usr/lib/hbase/lib/asm-3.1.jar:/usr/lib/hbase/lib/avro-1.3.3.jar:/usr/lib/hbase/lib/commons-cli-1.2.jar:/usr/lib/hbase/lib/commons-codec-1.4.jar:/usr/lib/hbase/lib/commons-el-1.0.jar:/usr/lib/hbase/lib/commons-httpclient-3.1.jar:/usr/lib/hbase/lib/commons-lang-2.5.jar:/usr/lib/hbase/lib/commons-logging-1.1.1.jar:/usr/lib/hbase/lib/commons-net-1.4.1.jar:/usr/lib/hbase/lib/core-3.1.1.jar:/usr/lib/hbase/lib/guava-r06.jar:/usr/lib/hbase/lib/hadoop-core.jar:/usr/lib/hbase/lib/hbase-0.90.1-cdh3u0.jar:/usr/lib/hbase/lib/jackson-core-asl-1.5.2.jar:/usr/lib/hbase/lib/jackson-jaxrs-1.5.5.jar:/usr/lib/hbase/lib/jackson-mapper-asl-1.5.2.jar:/usr/lib/hbase/lib/jackson-xc-1.5.5.jar:/usr/lib/hbase/lib/jasper-compiler-5.5.23.jar:/usr/lib/hbase/lib/jasper-runtime-5.5.23.jar:/usr/lib/hbase/lib/jaxb-api-2.1.jar:/usr/lib/hbase/lib/jaxb-impl-2.1.12.jar:/usr/lib/hbase/lib/jersey-core-1.4.jar:/usr/lib/hbase/lib/jersey-json-1.4.jar:/usr/lib/hbase/lib/jersey-server-1.4.jar:/usr/lib/hbase/lib/jettison-1.1.jar:/usr/lib/hbase/lib/jetty-6.1.26.jar:/usr/lib/hbase/lib/jetty-util-6.1.26.jar:/usr/lib/hbase/lib/jruby-complete-1.0.3.jar:/usr/lib/hbase/lib/jsp-2.1-6.1.14.jar:/usr/lib/hbase/lib/jsp-api-2.1-6.1.14.jar:/usr/lib/hbase/lib/jsp-api-2.1.jar:/usr/lib/hbase/lib/jsr311-api-1.1.1.jar:/usr/lib/hbase/lib/log4j-1.2.16.jar:/usr/lib/hbase/lib/protobuf-java-2.3.0.jar:/usr/lib/hbase/lib/servlet-api-2.5-6.1.14.jar:/usr/lib/hbase/lib/servlet-api-2.5.jar:/usr/lib/hbase/lib/slf4j-api-1.5.8.jar:/usr/lib/hbase/lib/slf4j-log4j12-1.5.8.jar:/usr/lib/hbase/lib/stax-api-1.0.1.jar:/usr/lib/hbase/lib/thrift-0.2.0.jar:/usr/lib/hbase/lib/xmlenc-0.52.jar:/usr/lib/hbase/lib/zookeeper.jar:/usr/lib/zookeeper/zookeeper-3.3.3-cdh3u0.jar:/usr/lib/zookeeper/zookeeper.jar:/usr/lib/zookeeper/lib/jline-0.9.94.jar:/usr/lib/zookeeper/lib/log4j-1.2.15.jar:/usr/lib/sqoop/sqoop-1.2.0-cdh3u0.jar:/usr/lib/sqoop/sqoop-test-1.2.0-cdh3u0.jar:<somepath>/lib/tdgssconfig.jar:<somepath>/lib/terajdbc4.jar:/usr/lib/hadoop/hadoop-0.20.2-cdh3u0-core.jar:/usr/lib/sqoop/sqoop-1.2.0-cdh3u0.jar > > Note: > > > /tmp/sqoop-hadoop/compile/cc68f6e38603d6705ad75a3d8ea1bf3a/DB_Temp_Table.java > > uses or overrides a deprecated API. > > Note: Recompile with -Xlint:deprecation for details. > > 11/08/30 22:59:52 INFO orm.CompilationManager: Writing jar file: > > > /tmp/sqoop-hadoop/compile/cc68f6e38603d6705ad75a3d8ea1bf3a/DB.Temp_Table.jar > > 11/08/30 22:59:52 DEBUG orm.CompilationManager: Scanning for .class files > in > > directory: /tmp/sqoop-hadoop/compile/cc68f6e38603d6705ad75a3d8ea1bf3a > > 11/08/30 22:59:52 DEBUG orm.CompilationManager: Got classfile: > > > /tmp/sqoop-hadoop/compile/cc68f6e38603d6705ad75a3d8ea1bf3a/DBTemp_Table.class > > -> DB_Temp_Table.class > > 11/08/30 22:59:52 DEBUG orm.CompilationManager: Finished writing jar file > > > /tmp/sqoop-hadrdev/compile/cc68f6e38603d6705ad75a3d8ea1bf3a/DB.Temp_Table.jar > > 11/08/30 22:59:52 INFO mapreduce.ExportJobBase: Beginning export of > > DB.Temp_Table > > 11/08/30 22:59:52 DEBUG mapreduce.JobBase: Using InputFormat: class > > com.cloudera.sqoop.mapreduce.ExportInputFormat > > 11/08/30 22:59:52 DEBUG manager.SqlManager: Using fetchSize for next > query: > > 1000 > > 11/08/30 22:59:52 INFO manager.SqlManager: Executing SQL statement: > SELECT > > t.* FROM DB.Temp_Table AS t WHERE 1=0 > > 11/08/30 22:59:53 DEBUG mapreduce.JobBase: Adding to job classpath: > > file:/usr/lib/sqoop/sqoop-1.2.0-cdh3u0.jar > > 11/08/30 22:59:53 DEBUG mapreduce.JobBase: Adding to job classpath: > > file:<somepath>/lib/terajdbc4.jar > > 11/08/30 22:59:53 DEBUG mapreduce.JobBase: Adding to job classpath: > > file:/usr/lib/sqoop/sqoop-1.2.0-cdh3u0.jar > > 11/08/30 22:59:53 DEBUG mapreduce.JobBase: Adding to job classpath: > > file:/usr/lib/sqoop/sqoop-1.2.0-cdh3u0.jar > > 11/08/30 22:59:53 DEBUG mapreduce.JobBase: Adding to job classpath: > > file:/usr/lib/sqoop/lib/ivy-2.0.0-rc2.jar > > 11/08/30 22:59:53 DEBUG mapreduce.JobBase: Adding to job classpath: > > file:/usr/lib/sqoop/lib/commons-io-1.4.jar > > 11/08/30 22:59:53 DEBUG mapreduce.JobBase: Adding to job classpath: > > file:/usr/lib/sqoop/lib/ant-eclipse-1.0-jvm1.2.jar > > 11/08/30 22:59:53 DEBUG mapreduce.JobBase: Adding to job classpath: > > file:/usr/lib/sqoop/lib/sqljdbc4.jar > > 11/08/30 22:59:53 DEBUG mapreduce.JobBase: Adding to job classpath: > > file:/usr/lib/sqoop/lib/ant-contrib-1.0b3.jar > > 11/08/30 22:59:53 DEBUG mapreduce.JobBase: Adding to job classpath: > > file:/usr/lib/sqoop/lib/hadoop-mrunit-0.20.2-CDH3b2-SNAPSHOT.jar > > 11/08/30 22:59:53 INFO hdfs.DFSClient: Created HDFS_DELEGATION_TOKEN > token > > 795 for hadoop > > 11/08/30 22:59:53 INFO security.TokenCache: Got dt for > > > hdfs://<cname>:9000/tmp/hadoop-mapred/mapred/staging/hadoop/.staging/job_201107010928_0398/libjars/hadoop-mrunit-0.20.2-CDH3b2-SNAPSHOT.jar;uri=xx.xxx.xx.xx:9000;t.service=xx.xxx.xx.xx:9000 > > 11/08/30 22:59:53 INFO input.FileInputFormat: Total input paths to > process : > > 1 > > 11/08/30 22:59:53 DEBUG mapreduce.ExportInputFormat: Target numMapTasks=1 > > 11/08/30 22:59:53 DEBUG mapreduce.ExportInputFormat: Total input bytes=18 > > 11/08/30 22:59:53 DEBUG mapreduce.ExportInputFormat: maxSplitSize=18 > > 11/08/30 22:59:53 INFO input.FileInputFormat: Total input paths to > process : > > 1 > > 11/08/30 22:59:53 DEBUG mapreduce.ExportInputFormat: Generated splits: > > 11/08/30 22:59:53 DEBUG mapreduce.ExportInputFormat: > > Paths:/user/hadrdev/sqoop_test.txt:0+18 Locations: > > 11/08/30 22:59:53 INFO mapred.JobClient: Running job: > job_201107010928_0398 > > 11/08/30 22:59:54 INFO mapred.JobClient: map 0% reduce 0% > > 11/08/30 23:00:01 INFO mapred.JobClient: Task Id : > > attempt_201107010928_0398_m_000000_0, Status : FAILED > > java.io.IOException: java.lang.NullPointerException > > at > > > com.cloudera.sqoop.mapreduce.ExportOutputFormat.getRecordWriter(ExportOutputFormat.java:80) > > at > > > org.apache.hadoop.mapred.MapTask$NewDirectOutputCollector.<init>(MapTask.java:520) > > at > org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:635) > > at org.apache.hadoop.mapred.MapTask.run(MapTask.java:322) > > at org.apache.hadoop.mapred.Child$4.run(Child.java:268) > > at java.security.AccessController.doPrivileged(Native Method) > > at javax.security.auth.Subject.doAs(Subject.java:396) > > at > > > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1115) > > at org.apache.hadoop.mapred.Child.main(Child.java:262) > > Caused by: java.lang.NullPointerException > > at > > com.teradata.tdgss.jtdgss.TdgssConfigApi.GetMechanisms(DashoA1*..) > > at com.teradata.tdgss.jtdgss.TdgssManager.<init>(DashoA1*..) > > at com.teradata.tdgss.jtdgss.TdgssManager.getInstance(DashoA1*..) > > at > > > com.teradata.jdbc.jdbc.GenericTeraEncrypt.getGSSM(GenericTeraEncrypt.java:612) > > at > > > com.teradata.jdbc.jdbc.GenericTeraEncrypt.getConfig(GenericTeraEncrypt.java:630) > > at > > > com.teradata.jdbc.jdbc.GenericTeraEncrypt.getUserNameForOid(GenericTeraEncrypt.java:723) > > at com.teradata.jdbc.AuthMechanism.<init>(AuthMechanism.java:50) > > at > > > com.teradata.jdbc.jdbc.GenericInitDBConfigState.action(GenericInitDBConfigState.java:105) > > at > > > com.teradata.jdbc.jdbc.GenericLogonController.run(GenericLogonController.java:49) > > at com.teradata.jdbc.jdbc_4.TDSession.<init>(TDSession.java:199) > > at > > > com.teradata.jdbc.jdbc_3.ifjdbc_4.TeraLocalConnection.<init>(TeraLocalConnection.java:95) > > at > > > com.teradata.jdbc.jdbc.ConnectionFactory.createConnection(ConnectionFactory.java:54) > > at com.teradata.jdbc.TeraDriver.doConnect(TeraDriver.java:217) > > at com.teradata.jdbc.TeraDriver.connect(TeraDriver.java:150) > > at java.sql.DriverManager.getConnection(DriverManager.java:582) > > at java.sql.DriverManager.getConnection(DriverManager.java:185) > > at > > > com.cloudera.sqoop.mapreduce.db.DBConfiguration.getConnection(DBConfiguration.java:184) > > at > > > com.cloudera.sqoop.mapreduce.AsyncSqlRecordWriter.<init>(AsyncSqlRecordWriter.java:73) > > at > > > com.cloudera.sqoop.mapreduce.ExportOutputFormat$ExportRecordWriter.<init>(ExportOutputFormat.java:96) > > at > > > com.cloudera.sqoop.mapreduce.ExportOutputFormat.getRecordWriter(ExportOutputFormat.java:78) > > ... 8 more > > > > attempt_201107010928_0398_m_000000_0: GSSException: Failure unspecified > at > > GSS-API level (Mechanism level: UserFile parameter null) > > attempt_201107010928_0398_m_000000_0: at > > com.teradata.tdgss.jtdgss.TdgssParseXml.<init>(DashoA1*..) > > attempt_201107010928_0398_m_000000_0: at > > com.teradata.tdgss.jtdgss.TdgssConfigApi.<init>(DashoA1*..) > > attempt_201107010928_0398_m_000000_0: at > > com.teradata.tdgss.jtdgss.TdgssManager.<init>(DashoA1*..) > > attempt_201107010928_0398_m_000000_0: at > > com.teradata.tdgss.jtdgss.TdgssManager.getInstance(DashoA1*..) > > attempt_201107010928_0398_m_000000_0: at > > > com.teradata.jdbc.jdbc.GenericTeraEncrypt.getGSSM(GenericTeraEncrypt.java:612) > > attempt_201107010928_0398_m_000000_0: at > > > com.teradata.jdbc.jdbc.GenericTeraEncrypt.getConfig(GenericTeraEncrypt.java:630) > > attempt_201107010928_0398_m_000000_0: at > > com.teradata.jdbc.jdbc.GenericTeraEncrypt.getUserNameForOid( > > On Mon, Aug 29, 2011 at 2:19 PM, SRINIVAS SURASANI <vas...@gmail.com> > wrote: > >> > >> Arvind, > >> > >> I have subscribed at sqoop-user@incubator.apache.org and posted > Question . > >> Sorry for the inconvinence from my end, since I'm close to deadline I'm > >> taking your valuable time. > >> > >> sqoop list-tables --driver com.teradata.jdbc.TeraDriver --connect > >> jdbc:teradata://PKTD/E1_CMS_WORK --username srini -P > >> Iam getting the following error: > >> 11/08/29 13:08:03 INFO manager.SqlManager: Using default fetchSize of > 1000 > >> GSSException: Failure unspecified at GSS-API level (Mechanism level: > >> UserFile parameter null) > >> at com.teradata.tdgss.jtdgss.TdgssParseXml.<init>(DashoA1*..) > >> at com.teradata.tdgss.jtdgss.TdgssConfigApi.<init>(DashoA1*..) > >> at com.teradata.tdgss.jtdgss.TdgssManager.<init>(DashoA1*..) > >> at > com.teradata.tdgss.jtdgss.TdgssManager.getInstance(DashoA1*..) > >> at > >> > com.teradata.jdbc.jdbc.GenericTeraEncrypt.getGSSM(GenericTeraEncrypt.java:612) > >> at > >> > com.teradata.jdbc.jdbc.GenericTeraEncrypt.getConfig(GenericTeraEncrypt.java:630) > >> at > >> > com.teradata.jdbc.jdbc.GenericTeraEncrypt.getUserNameForOid(GenericTeraEncrypt.java:723) > >> at com.teradata.jdbc.AuthMechanism.<init>(AuthMechanism.java:50) > >> at > >> > com.teradata.jdbc.jdbc.GenericInitDBConfigState.action(GenericInitDBConfigState.java:105) > >> at > >> > com.teradata.jdbc.jdbc.GenericLogonController.run(GenericLogonController.java:49) > >> at com.teradata.jdbc.jdbc_4.TDSession.<init>(TDSession.java:199) > >> at > >> > com.teradata.jdbc.jdbc_3.ifjdbc_4.TeraLocalConnection.<init>(TeraLocalConnection.java:95) > >> at > >> > com.teradata.jdbc.jdbc.ConnectionFactory.createConnection(ConnectionFactory.java:54) > >> at com.teradata.jdbc.TeraDriver.doConnect(TeraDriver.java:217) > >> at com.teradata.jdbc.TeraDriver.connect(TeraDriver.java:150) > >> at java.sql.DriverManager.getConnection(DriverManager.java:582) > >> at java.sql.DriverManager.getConnection(DriverManager.java:185) > >> at > >> > com.cloudera.sqoop.manager.SqlManager.makeConnection(SqlManager.java:643) > >> at > >> > com.cloudera.sqoop.manager.GenericJdbcManager.getConnection(GenericJdbcManager.java:51) > >> at > >> com.cloudera.sqoop.manager.SqlManager.listTables(SqlManager.java:270) > >> at > >> com.cloudera.sqoop.tool.ListTablesTool.run(ListTablesTool.java:49) > >> at com.cloudera.sqoop.Sqoop.run(Sqoop.java:144) > >> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65) > >> at com.cloudera.sqoop.Sqoop.runSqoop(Sqoop.java:180) > >> at com.cloudera.sqoop.Sqoop.runTool(Sqoop.java:219) > >> at com.cloudera.sqoop.Sqoop.runTool(Sqoop.java:228) > >> at com.cloudera.sqoop.Sqoop.main(Sqoop.java:237) > >> 11/08/29 13:08:04 ERROR sqoop.Sqoop: Got exception running Sqoop: > >> java.lang.NullPointerException > >> java.lang.NullPointerException > >> at > >> com.teradata.tdgss.jtdgss.TdgssConfigApi.GetMechanisms(DashoA1*..) > >> at com.teradata.tdgss.jtdgss.TdgssManager.<init>(DashoA1*..) > >> at > com.teradata.tdgss.jtdgss.TdgssManager.getInstance(DashoA1*..) > >> at > >> > com.teradata.jdbc.jdbc.GenericTeraEncrypt.getGSSM(GenericTeraEncrypt.java:612) > >> at > >> > com.teradata.jdbc.jdbc.GenericTeraEncrypt.getConfig(GenericTeraEncrypt.java:630) > >> at > >> > com.teradata.jdbc.jdbc.GenericTeraEncrypt.getUserNameForOid(GenericTeraEncrypt.java:723) > >> at com.teradata.jdbc.AuthMechanism.<init>(AuthMechanism.java:50) > >> at > >> > com.teradata.jdbc.jdbc.GenericInitDBConfigState.action(GenericInitDBConfigState.java:105) > >> at > >> > com.teradata.jdbc.jdbc.GenericLogonController.run(GenericLogonController.java:49) > >> at com.teradata.jdbc.jdbc_4.TDSession.<init>(TDSession.java:199) > >> at > >> > com.teradata.jdbc.jdbc_3.ifjdbc_4.TeraLocalConnection.<init>(TeraLocalConnection.java:95) > >> at > >> > com.teradata.jdbc.jdbc.ConnectionFactory.createConnection(ConnectionFactory.java:54) > >> at com.teradata.jdbc.TeraDriver.doConnect(TeraDriver.java:217) > >> at com.teradata.jdbc.TeraDriver.connect(TeraDriver.java:150) > >> at java.sql.DriverManager.getConnection(DriverManager.java:582) > >> at java.sql.DriverManager.getConnection(DriverManager.java:185) > >> at > >> > com.cloudera.sqoop.manager.SqlManager.makeConnection(SqlManager.java:643) > >> at > >> > com.cloudera.sqoop.manager.GenericJdbcManager.getConnection(GenericJdbcManager.java:51) > >> at > >> com.cloudera.sqoop.manager.SqlManager.listTables(SqlManager.java:270) > >> at > >> com.cloudera.sqoop.tool.ListTablesTool.run(ListTablesTool.java:49) > >> at com.cloudera.sqoop.Sqoop.run(Sqoop.java:144) > >> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65) > >> at com.cloudera.sqoop.Sqoop.runSqoop(Sqoop.java:180) > >> at com.cloudera.sqoop.Sqoop.runTool(Sqoop.java:219) > >> at com.cloudera.sqoop.Sqoop.runTool(Sqoop.java:228) > >> at com.cloudera.sqoop.Sqoop.main(Sqoop.java:237) > >> Thanks, > >> Srini > >> > >> > >> > >> On Mon, Aug 29, 2011 at 12:58 PM, arv...@cloudera.com > >> <arv...@cloudera.com> wrote: > >>> > >>> [Please subscribe and respond to sqoop-user@incubator.apache.org] > >>> > >>> Please use HADOOP_CLASSPATH instead of CLASSPATH. Also, in order to > >>> use the generic JDBC connector, you will have to specify the driver > >>> class explicitly via the command line option --driver > >>> com.teradata.jdbc.TeraDriver. > >>> > >>> Thanks, > >>> Arvind > >>> > >>> On Mon, Aug 29, 2011 at 9:53 AM, SRINIVAS SURASANI <vas...@gmail.com> > >>> wrote: > >>> > Arvind, > >>> > I have set the classpath to teradata4.jar [ not placed the > >>> > teradata4.jar in > >>> > sqoop lib, as I dont have permissions]. > >>> > I'm getting the following error > >>> > > >>> > sqoop list-tables --connect jdbc:teradata://PKTD/E1_CMS_WORK > --username > >>> > srini -P > >>> > ERROR: tool.BaseSqoopTool: Got error creating database manager: > >>> > java.io.IOexception: No manager for connect string: > >>> > jdbc:teradata:///PKTD/E1_CMS_WORK > >>> > at com.cloudera.sqoop.ConnFactory.getManager(ConnFactory.java:119) > >>> > at > >>> > com.cloudera.sqoop.tool.BaseSqoopTool.init(BaseSqoopTool.java:186) > >>> > ... > >>> > ... > >>> > Thanks, > >>> > Srini > >>> > > >>> > > >>> > > >>> > On Mon, Aug 29, 2011 at 8:52 AM, SRINIVAS SURASANI <vas...@gmail.com > > > >>> > wrote: > >>> >> > >>> >> Thanks-a-lot Arvind. > >>> >> > >>> >> On Mon, Aug 29, 2011 at 8:45 AM, arv...@cloudera.com > >>> >> <arv...@cloudera.com> > >>> >> wrote: > >>> >>> > >>> >>> [Moving the thread to sqoop-user@incubator.apache.org] > >>> >>> > >>> >>> Hi Srini, > >>> >>> > >>> >>> You should be able to use the generic JDBC connector to > import/export > >>> >>> from Teradata. There is also a specialized connector that is > >>> >>> available > >>> >>> for use with Teradata if you are interested. This connector is not > a > >>> >>> part of Sqoop and can be obtained from Cloudera by going to: > >>> >>> > >>> >>> http://www.cloudera.com/partners/connectors/ > >>> >>> > >>> >>> Thanks, > >>> >>> Arvind > >>> >>> > >>> >>> On Mon, Aug 29, 2011 at 8:17 AM, SRINIVAS SURASANI < > vas...@gmail.com> > >>> >>> wrote: > >>> >>> > I have csv file in hadoop and looking to load into Teradata. I > was > >>> >>> > wondering does the sqoop works with Terradata.(with JDBC jar > >>> >>> > placing > >>> >>> > in sqoop lib dir). > >>> >>> > > >>> >>> > Regards > >>> >>> > Srini > >>> >>> > > >>> >>> > -- > >>> >>> > NOTE: The mailing list sqoop-u...@cloudera.org is deprecated in > >>> >>> > favor > >>> >>> > of Apache Sqoop mailing list sqoop-user@incubator.apache.org. > >>> >>> > Please > >>> >>> > subscribe to it by sending an email to > >>> >>> > incubator-sqoop-user-subscr...@apache.org. > >>> >>> > > >>> >>> > >>> >>> -- > >>> >>> NOTE: The mailing list sqoop-u...@cloudera.org is deprecated in > favor > >>> >>> of > >>> >>> Apache Sqoop mailing list sqoop-user@incubator.apache.org. Please > >>> >>> subscribe > >>> >>> to it by sending an email to > >>> >>> incubator-sqoop-user-subscr...@apache.org. > >>> >> > >>> > > >>> > -- > >>> > NOTE: The mailing list sqoop-u...@cloudera.org is deprecated in > favor > >>> > of > >>> > Apache Sqoop mailing list sqoop-user@incubator.apache.org. Please > >>> > subscribe > >>> > to it by sending an email to > incubator-sqoop-user-subscr...@apache.org. > >>> > > >>> > >>> -- > >>> NOTE: The mailing list sqoop-u...@cloudera.org is deprecated in favor > of > >>> Apache Sqoop mailing list sqoop-user@incubator.apache.org. Please > subscribe > >>> to it by sending an email to incubator-sqoop-user-subscr...@apache.org > . > >> > > > > >