Hi Stephan , Please use the following desc extended to see where is the table' s directory on hdfs. Here is an example.
hive -e "desc extended hcatsmokeid0b0abc02_date252113 ;" WARNING: org.apache.hadoop.metrics.jvm.EventCounter is deprecated. Please use org.apache.hadoop.log.metrics.EventCounter in all the log4j.properties files. Logging initialized using configuration in jar:file:/usr/lib/hive/lib/hive-common-0.10.0.21.jar!/hive-log4j.properties Hive history file=/tmp/root/hive_job_log_root_201301211932_612003297.txt OK id int name string Detailed Table Information Table(tableName:hcatsmokeid0b0abc02_date252113, dbName:default, owner:ambari_qa, createTime:1358814367, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:id, type:int, comment:null), FieldSchema(name:name, type:string, comment:null)], location:hdfs://ambari1:8020/apps/hive/warehouse/hcatsmokeid0b0abc02_date252113, inputFormat:org.apache.hadoop.hive.ql.io.RCFileInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.RCFileOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe, parameters:{serialization.format=1}), bucketCols:[], sortCols:[], parameters:{}, skewedInfo:SkewedInfo(skewedColNames:[], skewedColValues:[], skewedColValueLocationMaps:{}), storedAsSubDirectories:false), partitionKeys:[], parameters:{transient_lastDdlTime=1358814367}, viewOriginalText:null, viewExpandedText:null, tableType:MANAGED_TABLE) Time taken: 2.965 seconds Hope this helps. Hortonworks, Inc. Technical Support Engineer Abdelrahman Shettia ashet...@hortonworks.com Office phone: (708) 689-9609 How am I doing? Please feel free to provide feedback to my manager Rick Morris at r...@hortonworks.com On Mar 2, 2013, at 1:59 AM, Stephen Boesch <java...@gmail.com> wrote: > > I am struggling with a "no such file or directory exception " when running a > simple query in hive. It is unfortunate that the actual path were not > included with the stacktrace: the following is all that is provided. > > I have a query that fails with the following error when done as hive -e > "select * from <table>'". But it works properly when done within the hive > shell. But at the same time, doing hive> select * from <table2>;" fails with > the same error message. > > I am also seeing this error both for hdfs files and for s3 files. Without > any path information it is very difficult and time consuming to track this > down. > > Any pointers appreciated. > > > Automatically selecting local only mode for query > Total MapReduce jobs = 1 > Launching Job 1 out of 1 > Number of reduce tasks determined at compile time: 1 > In order to change the average load for a reducer (in bytes): > set hive.exec.reducers.bytes.per.reducer=<number> > In order to limit the maximum number of reducers: > set hive.exec.reducers.max=<number> > In order to set a constant number of reducers: > set mapred.reduce.tasks=<number> > WARNING: org.apache.hadoop.metrics.jvm.EventCounter is deprecated. Please use > org.apache.hadoop.log.metrics.EventCounter in all the log4j.properties files. > Execution log at: > /tmp/impala/impala_20130302095252_79ce9404-6af7-405b-8b06-849fe6c5328d.log > ENOENT: No such file or directory > at org.apache.hadoop.io.nativeio.NativeIO.chmod(Native Method) > at > org.apache.hadoop.fs.RawLocalFileSystem.setPermission(RawLocalFileSystem.java:568) > at > org.apache.hadoop.fs.FilterFileSystem.setPermission(FilterFileSystem.java:411) > at org.apache.hadoop.fs.FileSystem.mkdirs(FileSystem.java:501) > at > org.apache.hadoop.mapred.JobClient.copyAndConfigureFiles(JobClient.java:733) > at > org.apache.hadoop.mapred.JobClient.copyAndConfigureFiles(JobClient.java:692) > at org.apache.hadoop.mapred.JobClient.access$400(JobClient.java:172) > at org.apache.hadoop.mapred.JobClient$2.run(JobClient.java:910) > at org.apache.hadoop.mapred.JobClient$2.run(JobClient.java:895) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:416) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1332) > at > org.apache.hadoop.mapred.JobClient.submitJobInternal(JobClient.java:895) > at org.apache.hadoop.mapred.JobClient.submitJob(JobClient.java:869) > at > org.apache.hadoop.hive.ql.exec.ExecDriver.execute(ExecDriver.java:435) > at org.apache.hadoop.hive.ql.exec.ExecDriver.main(ExecDriver.java:677) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:616) > at org.apache.hadoop.util.RunJar.main(RunJar.java:208) > Job Submission failed with exception > 'org.apache.hadoop.io.nativeio.NativeIOException(No such file or directory)' > Execution failed with exit status: 1 > Obtaining error information >