Thanks in advance!
The disk is not full: [ptoole@ptoole hadoop-sqoop]$ df -h Filesystem Size Used Avail Use% Mounted on /dev/sda1 455G 75G 358G 18% / tmpfs 3.9G 744K 3.9G 1% /dev/shm Here are my versions: [ptoole@ptoole hadoop-sqoop]$ sqoop version Sqoop 1.3.0-cdh3u1 git commit id 3a60cc809b14d538dd1eb0e90ffa9767e8d06a43 Compiled by jenkins@ubuntu-slave01 on Mon Jul 18 08:38:49 PDT 2011 [ptoole@ptoole hadoop-sqoop]$ hadoop version Hadoop 0.20.2-cdh3u1 Subversion file:///tmp/topdir/BUILD/hadoop-0.20.2-cdh3u1 -r bdafb1dbffd0d5f2fbc6ee022e1c8df6500fd638 Compiled by root on Mon Jul 18 09:40:26 PDT 2011 >From source with checksum 3127e3d410455d2bacbff7673bf3284c Additional info: [ptoole@ptoole hadoop-sqoop]$ uname -a Linux ptoole 2.6.32-131.0.15.el6.x86_64 #1 SMP Tue May 10 15:42:40 EDT 2011 x86_64 x86_64 x86_64 GNU/Linux [ptoole@ptoole hadoop-sqoop]$ id ptoole uid=502(ptoole) gid=503(ptoole) groups=503(ptoole),486(hadoop) [ptoole@ptoole hadoop-sqoop]$ hadoop --config /home/ptoole/Desktop/work/opensource/hadoop-sqoop-vertica/hadoop-conf fs -ls /user/ptoole Found 1 items drwxr-xr-x - ptoole hadoop 0 2011-10-14 12:10 /user/ptoole/files Here's the verbose command: [ptoole@ptoole hadoop-sqoop]$ sqoop import --connect jdbc:mysql://localhost/sqoop_test --table sample_data --username root --password <passwd> --m 1 --verbose 11/10/14 12:05:37 DEBUG tool.BaseSqoopTool: Enabled debug logging. 11/10/14 12:05:37 WARN tool.BaseSqoopTool: Setting your password on the command-line is insecure. Consider using -P instead. 11/10/14 12:05:37 DEBUG sqoop.ConnFactory: Loaded manager factory: com.cloudera.sqoop.manager.DefaultManagerFactory 11/10/14 12:05:37 DEBUG sqoop.ConnFactory: Trying ManagerFactory: com.cloudera.sqoop.manager.DefaultManagerFactory 11/10/14 12:05:37 DEBUG manager.DefaultManagerFactory: Trying with scheme: jdbc:mysql: 11/10/14 12:05:37 INFO manager.MySQLManager: Preparing to use a MySQL streaming resultset. 11/10/14 12:05:37 DEBUG sqoop.ConnFactory: Instantiated ConnManager com.cloudera.sqoop.manager.MySQLManager@3c1d332b 11/10/14 12:05:37 INFO tool.CodeGenTool: Beginning code generation 11/10/14 12:05:37 DEBUG manager.SqlManager: No connection paramenters specified. Using regular API for making connection. 11/10/14 12:05:37 DEBUG manager.SqlManager: Using fetchSize for next query: -2147483648 11/10/14 12:05:37 INFO manager.SqlManager: Executing SQL statement: SELECT t.* FROM `sample_data` AS t LIMIT 1 11/10/14 12:05:37 DEBUG manager.SqlManager: Using fetchSize for next query: -2147483648 11/10/14 12:05:37 INFO manager.SqlManager: Executing SQL statement: SELECT t.* FROM `sample_data` AS t LIMIT 1 11/10/14 12:05:37 DEBUG orm.ClassWriter: selected columns: 11/10/14 12:05:37 DEBUG orm.ClassWriter: example_col 11/10/14 12:05:37 DEBUG orm.ClassWriter: Writing source file: /tmp/sqoop-ptoole/compile/cb8a4d7e585bc91be77605c4032afe07/sample_data.java 11/10/14 12:05:37 DEBUG orm.ClassWriter: Table name: sample_data 11/10/14 12:05:37 DEBUG orm.ClassWriter: Columns: example_col:12, 11/10/14 12:05:37 DEBUG orm.ClassWriter: sourceFilename is sample_data.java 11/10/14 12:05:37 DEBUG orm.CompilationManager: Found existing /tmp/sqoop-ptoole/compile/cb8a4d7e585bc91be77605c4032afe07/ 11/10/14 12:05:37 INFO orm.CompilationManager: HADOOP_HOME is /usr/lib/hadoop 11/10/14 12:05:37 INFO orm.CompilationManager: Found hadoop core jar at: /usr/lib/hadoop/hadoop-core.jar 11/10/14 12:05:37 DEBUG orm.CompilationManager: Adding source file: /tmp/sqoop-ptoole/compile/cb8a4d7e585bc91be77605c4032afe07/sample_data.java 11/10/14 12:05:37 DEBUG orm.CompilationManager: Invoking javac with args: 11/10/14 12:05:37 DEBUG orm.CompilationManager: -sourcepath 11/10/14 12:05:37 DEBUG orm.CompilationManager: /tmp/sqoop-ptoole/compile/cb8a4d7e585bc91be77605c4032afe07/ 11/10/14 12:05:37 DEBUG orm.CompilationManager: -d 11/10/14 12:05:37 DEBUG orm.CompilationManager: /tmp/sqoop-ptoole/compile/cb8a4d7e585bc91be77605c4032afe07/ 11/10/14 12:05:37 DEBUG orm.CompilationManager: -classpath 11/10/14 12:05:37 DEBUG orm.CompilationManager: /usr/lib/hadoop/conf:/usr/java/jdk1.6.0_21/lib/tools.jar:/usr/lib/hadoop:/usr/lib/hadoop/hadoop-core-0.20.2-cdh3u1.jar:/usr/lib/hadoop/lib/ant-contrib-1.0b3.jar:/usr/lib/hadoop/lib/aspectjrt-1.6.5.jar:/usr/lib/hadoop/lib/aspectjtools-1.6.5.jar:/usr/lib/hadoop/lib/commons-cli-1.2.jar:/usr/lib/hadoop/lib/commons-codec-1.4.jar:/usr/lib/hadoop/lib/commons-daemon-1.0.1.jar:/usr/lib/hadoop/lib/commons-el-1.0.jar:/usr/lib/hadoop/lib/commons-httpclient-3.0.1.jar:/usr/lib/hadoop/lib/commons-logging-1.0.4.jar:/usr/lib/hadoop/lib/commons-logging-api-1.0.4.jar:/usr/lib/hadoop/lib/commons-net-1.4.1.jar:/usr/lib/hadoop/lib/core-3.1.1.jar:/usr/lib/hadoop/lib/hadoop-fairscheduler-0.20.2-cdh3u1.jar:/usr/lib/hadoop/lib/hsqldb-1.8.0.10.jar:/usr/lib/hadoop/lib/hue-plugins-1.2.0-cdh3u1.jar:/usr/lib/hadoop/lib/jackson-core-asl-1.5.2.jar:/usr/lib/hadoop/lib/jackson-mapper-asl-1.5.2.jar:/usr/lib/hadoop/lib/jasper-compiler-5.5.12.jar:/usr/lib/hadoop/lib/jasper-runtime-5.5.12.jar:/usr/lib/hadoop/lib/jets3t-0.6.1.jar:/usr/lib/hadoop/lib/jetty-6.1.26.jar:/usr/lib/hadoop/lib/jetty-servlet-tester-6.1.26.jar:/usr/lib/hadoop/lib/jetty-util-6.1.26.jar:/usr/lib/hadoop/lib/jsch-0.1.42.jar:/usr/lib/hadoop/lib/junit-4.5.jar:/usr/lib/hadoop/lib/kfs-0.2.2.jar:/usr/lib/hadoop/lib/log4j-1.2.15.jar:/usr/lib/hadoop/lib/mockito-all-1.8.2.jar:/usr/lib/hadoop/lib/oro-2.0.8.jar:/usr/lib/hadoop/lib/servlet-api-2.5-20081211.jar:/usr/lib/hadoop/lib/servlet-api-2.5-6.1.14.jar:/usr/lib/hadoop/lib/slf4j-api-1.4.3.jar:/usr/lib/hadoop/lib/slf4j-log4j12-1.4.3.jar:/usr/lib/hadoop/lib/xmlenc-0.52.jar:/usr/lib/hadoop/lib/jsp-2.1/jsp-2.1.jar:/usr/lib/hadoop/lib/jsp-2.1/jsp-api-2.1.jar:/usr/lib/sqoop/conf:/etc/zookeeper::/usr/lib/sqoop/lib/ant-contrib-1.0b3.jar:/usr/lib/sqoop/lib/ant-eclipse-1.0-jvm1.2.jar:/usr/lib/sqoop/lib/avro-1.5.1.jar:/usr/lib/sqoop/lib/avro-ipc-1.5.1.jar:/usr/lib/sqoop/lib/avro-mapred-1.5.1.jar:/usr/lib/sqoop/lib/commons-io-1.4.jar:/usr/lib/sqoop/lib/ivy-2.0.0-rc2.jar:/usr/lib/sqoop/lib/jackson-core-asl-1.7.3.jar:/usr/lib/sqoop/lib/jackson-mapper-asl-1.7.3.jar:/usr/lib/sqoop/lib/jopt-simple-3.2.jar:/usr/lib/sqoop/lib/mysql-connector-java-5.1.18-bin.jar:/usr/lib/sqoop/lib/paranamer-2.3.jar:/usr/lib/sqoop/lib/snappy-java-1.0.3-rc2.jar:/usr/lib/hbase/bin/../conf:/usr/java/jdk1.6.0_21/lib/tools.jar:/usr/lib/hbase/bin/..:/usr/lib/hbase/bin/../hbase-0.90.3-cdh3u1.jar:/usr/lib/hbase/bin/../hbase-0.90.3-cdh3u1-tests.jar:/usr/lib/hbase/bin/../lib/activation-1.1.jar:/usr/lib/hbase/bin/../lib/asm-3.1.jar:/usr/lib/hbase/bin/../lib/avro-1.3.3.jar:/usr/lib/hbase/bin/../lib/commons-cli-1.2.jar:/usr/lib/hbase/bin/../lib/commons-codec-1.4.jar:/usr/lib/hbase/bin/../lib/commons-el-1.0.jar:/usr/lib/hbase/bin/../lib/commons-httpclient-3.1.jar:/usr/lib/hbase/bin/../lib/commons-lang-2.5.jar:/usr/lib/hbase/bin/../lib/commons-logging-1.1.1.jar:/usr/lib/hbase/bin/../lib/commons-net-1.4.1.jar:/usr/lib/hbase/bin/../lib/core-3.1.1.jar:/usr/lib/hbase/bin/../lib/guava-r06.jar:/usr/lib/hbase/bin/../lib/hadoop-core.jar:/usr/lib/hbase/bin/../lib/jackson-core-asl-1.5.2.jar:/usr/lib/hbase/bin/../lib/jackson-jaxrs-1.5.5.jar:/usr/lib/hbase/bin/../lib/jackson-mapper-asl-1.5.2.jar:/usr/lib/hbase/bin/../lib/jackson-xc-1.5.5.jar:/usr/lib/hbase/bin/../lib/jasper-compiler-5.5.23.jar:/usr/lib/hbase/bin/../lib/jasper-runtime-5.5.23.jar:/usr/lib/hbase/bin/../lib/jaxb-api-2.1.jar:/usr/lib/hbase/bin/../lib/jaxb-impl-2.1.12.jar:/usr/lib/hbase/bin/../lib/jersey-core-1.4.jar:/usr/lib/hbase/bin/../lib/jersey-json-1.4.jar:/usr/lib/hbase/bin/../lib/jersey-server-1.4.jar:/usr/lib/hbase/bin/../lib/jettison-1.1.jar:/usr/lib/hbase/bin/../lib/jetty-6.1.26.jar:/usr/lib/hbase/bin/../lib/jetty-util-6.1.26.jar:/usr/lib/hbase/bin/../lib/jruby-complete-1.6.0.jar:/usr/lib/hbase/bin/../lib/jsp-2.1-6.1.14.jar:/usr/lib/hbase/bin/../lib/jsp-api-2.1-6.1.14.jar:/usr/lib/hbase/bin/../lib/jsp-api-2.1.jar:/usr/lib/hbase/bin/../lib/jsr311-api-1.1.1.jar:/usr/lib/hbase/bin/../lib/log4j-1.2.16.jar:/usr/lib/hbase/bin/../lib/protobuf-java-2.3.0.jar:/usr/lib/hbase/bin/../lib/servlet-api-2.5-6.1.14.jar:/usr/lib/hbase/bin/../lib/servlet-api-2.5.jar:/usr/lib/hbase/bin/../lib/slf4j-api-1.5.8.jar:/usr/lib/hbase/bin/../lib/slf4j-log4j12-1.5.8.jar:/usr/lib/hbase/bin/../lib/stax-api-1.0.1.jar:/usr/lib/hbase/bin/../lib/thrift-0.2.0.jar:/usr/lib/hbase/bin/../lib/xmlenc-0.52.jar:/usr/lib/hbase/bin/../lib/zookeeper.jar:/usr/lib/sqoop/sqoop-1.3.0-cdh3u1.jar:/usr/lib/sqoop/sqoop-test-1.3.0-cdh3u1.jar::/usr/lib/hadoop/hadoop-core.jar:/usr/lib/sqoop/sqoop-1.3.0-cdh3u1.jar 11/10/14 12:05:38 INFO orm.CompilationManager: Writing jar file: /tmp/sqoop-ptoole/compile/cb8a4d7e585bc91be77605c4032afe07/sample_data.jar 11/10/14 12:05:38 DEBUG orm.CompilationManager: Scanning for .class files in directory: /tmp/sqoop-ptoole/compile/cb8a4d7e585bc91be77605c4032afe07 11/10/14 12:05:38 DEBUG orm.CompilationManager: Got classfile: /tmp/sqoop-ptoole/compile/cb8a4d7e585bc91be77605c4032afe07/sample_data.class -> sample_data.class 11/10/14 12:05:38 DEBUG orm.CompilationManager: Finished writing jar file /tmp/sqoop-ptoole/compile/cb8a4d7e585bc91be77605c4032afe07/sample_data.jar 11/10/14 12:05:38 WARN manager.MySQLManager: It looks like you are importing from mysql. 11/10/14 12:05:38 WARN manager.MySQLManager: This transfer can be faster! Use the --direct 11/10/14 12:05:38 WARN manager.MySQLManager: option to exercise a MySQL-specific fast path. 11/10/14 12:05:38 INFO manager.MySQLManager: Setting zero DATETIME behavior to convertToNull (mysql) 11/10/14 12:05:38 DEBUG manager.MySQLManager: Rewriting connect string to jdbc:mysql://localhost/sqoop_test?zeroDateTimeBehavior=convertToNull 11/10/14 12:05:38 INFO mapreduce.ImportJobBase: Beginning import of sample_data 11/10/14 12:05:38 DEBUG util.ClassLoaderStack: Checking for existing class: sample_data 11/10/14 12:05:38 DEBUG util.ClassLoaderStack: Attempting to load jar through URL: jar:file:///tmp/sqoop-ptoole/compile/cb8a4d7e585bc91be77605c4032afe07/sample_data.jar!/ 11/10/14 12:05:38 DEBUG util.ClassLoaderStack: Previous classloader is sun.misc.Launcher$AppClassLoader@6d6f0472 11/10/14 12:05:38 DEBUG util.ClassLoaderStack: Testing class in jar: sample_data 11/10/14 12:05:38 DEBUG util.ClassLoaderStack: Loaded jar into current JVM: jar:file:///tmp/sqoop-ptoole/compile/cb8a4d7e585bc91be77605c4032afe07/sample_data.jar!/ 11/10/14 12:05:38 DEBUG util.ClassLoaderStack: Added classloader for jar /tmp/sqoop-ptoole/compile/cb8a4d7e585bc91be77605c4032afe07/sample_data.jar: java.net.FactoryURLClassLoader@3caa4b 11/10/14 12:05:38 DEBUG manager.SqlManager: Using fetchSize for next query: -2147483648 11/10/14 12:05:38 INFO manager.SqlManager: Executing SQL statement: SELECT t.* FROM `sample_data` AS t LIMIT 1 11/10/14 12:05:38 DEBUG mapreduce.DataDrivenImportJob: Using table class: sample_data 11/10/14 12:05:38 DEBUG mapreduce.DataDrivenImportJob: Using InputFormat: class com.cloudera.sqoop.mapreduce.db.DataDrivenDBInputFormat 11/10/14 12:05:38 DEBUG mapreduce.JobBase: Adding to job classpath: file:/usr/lib/sqoop/sqoop-1.3.0-cdh3u1.jar 11/10/14 12:05:38 DEBUG mapreduce.JobBase: Adding to job classpath: file:/usr/lib/sqoop/lib/mysql-connector-java-5.1.18-bin.jar 11/10/14 12:05:38 DEBUG mapreduce.JobBase: Adding to job classpath: file:/usr/lib/sqoop/sqoop-1.3.0-cdh3u1.jar 11/10/14 12:05:38 DEBUG mapreduce.JobBase: Adding to job classpath: file:/usr/lib/sqoop/sqoop-1.3.0-cdh3u1.jar 11/10/14 12:05:38 DEBUG mapreduce.JobBase: Adding to job classpath: file:/usr/lib/sqoop/lib/paranamer-2.3.jar 11/10/14 12:05:38 DEBUG mapreduce.JobBase: Adding to job classpath: file:/usr/lib/sqoop/lib/commons-io-1.4.jar 11/10/14 12:05:38 DEBUG mapreduce.JobBase: Adding to job classpath: file:/usr/lib/sqoop/lib/jackson-core-asl-1.7.3.jar 11/10/14 12:05:38 DEBUG mapreduce.JobBase: Adding to job classpath: file:/usr/lib/sqoop/lib/avro-mapred-1.5.1.jar 11/10/14 12:05:38 DEBUG mapreduce.JobBase: Adding to job classpath: file:/usr/lib/sqoop/lib/jopt-simple-3.2.jar 11/10/14 12:05:38 DEBUG mapreduce.JobBase: Adding to job classpath: file:/usr/lib/sqoop/lib/ant-eclipse-1.0-jvm1.2.jar 11/10/14 12:05:38 DEBUG mapreduce.JobBase: Adding to job classpath: file:/usr/lib/sqoop/lib/jackson-mapper-asl-1.7.3.jar 11/10/14 12:05:38 DEBUG mapreduce.JobBase: Adding to job classpath: file:/usr/lib/sqoop/lib/snappy-java-1.0.3-rc2.jar 11/10/14 12:05:38 DEBUG mapreduce.JobBase: Adding to job classpath: file:/usr/lib/sqoop/lib/ivy-2.0.0-rc2.jar 11/10/14 12:05:38 DEBUG mapreduce.JobBase: Adding to job classpath: file:/usr/lib/sqoop/lib/avro-ipc-1.5.1.jar 11/10/14 12:05:38 DEBUG mapreduce.JobBase: Adding to job classpath: file:/usr/lib/sqoop/lib/mysql-connector-java-5.1.18-bin.jar 11/10/14 12:05:38 DEBUG mapreduce.JobBase: Adding to job classpath: file:/usr/lib/sqoop/lib/ant-contrib-1.0b3.jar 11/10/14 12:05:38 DEBUG mapreduce.JobBase: Adding to job classpath: file:/usr/lib/sqoop/lib/avro-1.5.1.jar 11/10/14 12:05:38 INFO jvm.JvmMetrics: Initializing JVM Metrics with processName=JobTracker, sessionId= 11/10/14 12:05:38 INFO util.NativeCodeLoader: Loaded the native-hadoop library 11/10/14 12:05:38 DEBUG util.ClassLoaderStack: Restoring classloader: sun.misc.Launcher$AppClassLoader@6d6f0472 11/10/14 12:05:38 ERROR tool.ImportTool: Encountered IOException running import job: ENOENT: No such file or directory at org.apache.hadoop.io.nativeio.NativeIO.chmod(Native Method) at org.apache.hadoop.fs.RawLocalFileSystem.setPermission(RawLocalFileSystem.java:496) at org.apache.hadoop.fs.RawLocalFileSystem.mkdirs(RawLocalFileSystem.java:319) at org.apache.hadoop.fs.FilterFileSystem.mkdirs(FilterFileSystem.java:189) at org.apache.hadoop.mapreduce.JobSubmissionFiles.getStagingDir(JobSubmissionFiles.java:126) at org.apache.hadoop.mapred.JobClient$2.run(JobClient.java:839) at org.apache.hadoop.mapred.JobClient$2.run(JobClient.java:833) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:396) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1127) at org.apache.hadoop.mapred.JobClient.submitJobInternal(JobClient.java:833) at org.apache.hadoop.mapreduce.Job.submit(Job.java:476) at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:506) at com.cloudera.sqoop.mapreduce.ImportJobBase.runJob(ImportJobBase.java:121) at com.cloudera.sqoop.mapreduce.ImportJobBase.runImport(ImportJobBase.java:181) at com.cloudera.sqoop.manager.SqlManager.importTable(SqlManager.java:405) at com.cloudera.sqoop.manager.MySQLManager.importTable(MySQLManager.java:132) at com.cloudera.sqoop.tool.ImportTool.importTable(ImportTool.java:350) at com.cloudera.sqoop.tool.ImportTool.run(ImportTool.java:423) at com.cloudera.sqoop.Sqoop.run(Sqoop.java:144) at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65) at com.cloudera.sqoop.Sqoop.runSqoop(Sqoop.java:180) at com.cloudera.sqoop.Sqoop.runTool(Sqoop.java:219) at com.cloudera.sqoop.Sqoop.runTool(Sqoop.java:228) at com.cloudera.sqoop.Sqoop.main(Sqoop.java:237) On Fri, Oct 14, 2011 at 11:36 AM, arv...@cloudera.com <arv...@cloudera.com>wrote: > [Moving conversation to sqoop-user@incubator.apache.org. Please subscribe > to this list.] > > Please check if the disk partition that holds temporary files or log files > is not full on the task node. Also - please tell us which exact version of > Sqoop and Hadoop you are using, and the output of the command with the > --verbose flag. > > Thanks, > Arvind > > > On Fri, Oct 14, 2011 at 7:45 AM, Patrick <pto...@gmail.com> wrote: > >> I'm running a very simple command: >> sqoop import --connect jdbc:mysql://localhost/sqoop_test --table >> sample_data --username root --password <psswd> --m 1 --verbose >> >> And getting this error: >> >> 11/10/14 10:42:11 INFO jvm.JvmMetrics: Initializing JVM Metrics with >> processName=JobTracker, sessionId= >> 11/10/14 10:42:11 INFO util.NativeCodeLoader: Loaded the native-hadoop >> library >> 11/10/14 10:42:11 DEBUG util.ClassLoaderStack: Restoring classloader: >> sun.misc.Launcher$AppClassLoader@1a45a877 >> 11/10/14 10:42:11 ERROR tool.ImportTool: Encountered IOException >> running import job: ENOENT: No such file or directory >> at org.apache.hadoop.io.nativeio.NativeIO.chmod(Native Method) >> at >> >> org.apache.hadoop.fs.RawLocalFileSystem.setPermission(RawLocalFileSystem.java: >> 496) >> at >> org.apache.hadoop.fs.RawLocalFileSystem.mkdirs(RawLocalFileSystem.java: >> 319) >> at >> org.apache.hadoop.fs.FilterFileSystem.mkdirs(FilterFileSystem.java: >> 189) >> at >> >> org.apache.hadoop.mapreduce.JobSubmissionFiles.getStagingDir(JobSubmissionFiles.java: >> 126) >> at org.apache.hadoop.mapred.JobClient$2.run(JobClient.java:839) >> at org.apache.hadoop.mapred.JobClient$2.run(JobClient.java:833) >> at java.security.AccessController.doPrivileged(Native Method) >> at javax.security.auth.Subject.doAs(Subject.java:396) >> at >> >> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java: >> 1127) >> at >> org.apache.hadoop.mapred.JobClient.submitJobInternal(JobClient.java: >> 833) >> at org.apache.hadoop.mapreduce.Job.submit(Job.java:476) >> at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:506) >> at >> com.cloudera.sqoop.mapreduce.ImportJobBase.runJob(ImportJobBase.java: >> 124) >> at >> com.cloudera.sqoop.mapreduce.ImportJobBase.runImport(ImportJobBase.java: >> 185) >> at >> com.cloudera.sqoop.manager.SqlManager.importTable(SqlManager.java: >> 413) >> at >> com.cloudera.sqoop.manager.MySQLManager.importTable(MySQLManager.java: >> 98) >> at com.cloudera.sqoop.tool.ImportTool.importTable(ImportTool.java: >> 383) >> at com.cloudera.sqoop.tool.ImportTool.run(ImportTool.java:456) >> at com.cloudera.sqoop.Sqoop.run(Sqoop.java:146) >> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65) >> at com.cloudera.sqoop.Sqoop.runSqoop(Sqoop.java:182) >> at com.cloudera.sqoop.Sqoop.runTool(Sqoop.java:221) >> at com.cloudera.sqoop.Sqoop.runTool(Sqoop.java:230) >> at com.cloudera.sqoop.Sqoop.main(Sqoop.java:239) >> >> >> Anyone seen this before or know how to solve it? >> >> I cant tell if its having permission issues with HDFS or it cant write >> temporary files to my local FS. This is on RHEL6 on a CDH3 install >> express edition and the user executing is in the hadoop group (root). >> Sqoop appears to be sucessfully generating and compiling the *.java >> file in the /temp directory too....just no data. >> >> -- >> NOTE: The mailing list sqoop-u...@cloudera.org is deprecated in favor of >> Apache Sqoop mailing list sqoop-user@incubator.apache.org. Please >> subscribe to it by sending an email to >> incubator-sqoop-user-subscr...@apache.org. >> > > -- > NOTE: The mailing list sqoop-u...@cloudera.org is deprecated in favor of > Apache Sqoop mailing list sqoop-user@incubator.apache.org. Please > subscribe to it by sending an email to > incubator-sqoop-user-subscr...@apache.org. >