Some suggestion :
1. Chown of the hive folder 2. change permission of hive folder to 755 3. Set this to hive-site.xml <property> <name>hive.exec.scratchdir</name> <value>/home/yourusername/mydir</value> <description>Scratch space for Hive jobs</description> </property> 3. put hadoop-0.20-core.jar hive/lib/hive-exec-0.7.1.jar hive/lib/hive-jdbc-0.7.1.jar hive/lib/hive-metastore-0.7.1.jar hive/lib/hive-service-0.7.1.jar hive/lib/libfb303.jar lib/commons-logging-1.0.4.jar slf4j-api-1.6.1.jar slf4j-log4j12-1.6.1.jar these file to hive lib folder Define external jar in hive site xml file as follows : IN hive-env.sh change following: # Set HADOOP_HOME to point to a specific hadoop install directory #here instead of path what i have given you give your own path where hadoop #is there export HADOOP_HOME=/home/shashwat/Hadoop/hadoop-0.20.205 # Hive Configuration Directory can be controlled by: #here you specify the conf directory path of hive export HIVE_CONF_DIR=/home/shashwat/Hadoop/hive-0.7.1/conf #Folder containing extra ibraries required for hive compilation/execution #can be controlled by: #here you specify the lib file directory, here you can specify the lib #directory of hadoop lib folder too like follows: export HIVE_AUX_JARS_PATH=/home/shashwat/Hadoop/hive-0.7.1/lib:/home/shashwat/Hadoop/hadoop-0.20.205/lib:/home/shashwat/Hadoop/hbase/lib and also in hive-site.xml <property> <name>hive.aux.jars.path</name> <value>file:///home/shashwat/Hadoop/hive-0.7.1/lib/hive-hbase-handler-0.7.1.jar,file:///home/shashwat/Hadoop/hive-0.7.1/lib/hbase-0.90.4.jar,file:///home/shashwat/Hadoop/hive-0.7.1/lib/zookeeper-3.3.1.jar</value> </property> by default if you have just downloaded the hive there will not be hive-site.xml or hive-env.sh by default in conf dir there will be hive-site.xml.template and same for the hive-env.sh so make a backupe of both and rename hive-site.xml.template to hive-site.xml and hive-env.sh.template to hive-env.sh and make the changes what i suggested... Note : when ever you download something and extract in linux dont forget to change owner and change mode using chown and chmod for the folder recursively Hope this will help you.. Cheer ∞ Shashwat Shriparv On Thu, May 10, 2012 at 9:59 PM, Mahsa Mofidpoor <mofidp...@gmail.com>wrote: > Hi, > > When I want to join two tables, I receive the following error: > > 12/05/10 12:03:31 WARN conf.HiveConf: hive-site.xml not found on CLASSPATH > WARNING: org.apache.hadoop.metrics.jvm.EventCounter is deprecated. Please > use org.apache.hadoop.log.metrics.EventCounter in all the log4j.properties > files. > Execution log at: > /tmp/umroot/umroot_20120510120303_4d0145bb-27fa-4d4a-8cbc-95d8353fccaf.log > ENOENT: No such file or directory > at org.apache.hadoop.io.nativeio.NativeIO.chmod(Native Method) > at org.apache.hadoop.fs.FileUtil.execSetPermission(FileUtil.java:692) > at org.apache.hadoop.fs.FileUtil.setPermission(FileUtil.java:647) > at > org.apache.hadoop.fs.RawLocalFileSystem.setPermission(RawLocalFileSystem.java:509) > at > org.apache.hadoop.fs.RawLocalFileSystem.mkdirs(RawLocalFileSystem.java:344) > at > org.apache.hadoop.fs.FilterFileSystem.mkdirs(FilterFileSystem.java:189) > at > org.apache.hadoop.mapreduce.JobSubmissionFiles.getStagingDir(JobSubmissionFiles.java:116) > at org.apache.hadoop.mapred.JobClient$2.run(JobClient.java:856) > at org.apache.hadoop.mapred.JobClient$2.run(JobClient.java:850) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:396) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1093) > at org.apache.hadoop.mapred.JobClient.submitJobInternal(JobClient.java:850) > at org.apache.hadoop.mapred.JobClient.submitJob(JobClient.java:824) > at org.apache.hadoop.hive.ql.exec.ExecDriver.execute(ExecDriver.java:435) > at org.apache.hadoop.hive.ql.exec.ExecDriver.main(ExecDriver.java:693) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) > at java.lang.reflect.Method.invoke(Method.java:597) > at org.apache.hadoop.util.RunJar.main(RunJar.java:156) > Job Submission failed with exception > 'org.apache.hadoop.io.nativeio.NativeIOException(No such file or directory)' > Execution failed with exit status: 2 > Obtaining error information > > Task failed! > Task ID: > Stage-1 > > Logs: > > /tmp/umroot/hive.log > FAILED: Execution Error, return code 2 from > org.apache.hadoop.hive.ql.exec.MapRedTask > > > I use hadoop-0.20.2 (single-node setup) and I have build Hive through the > latest source code. > > > Thank you in advance for your help, > > Mahsa > -- ∞ Shashwat Shriparv