I am now getting this exception,  all my hadoop and hbase processes are running 
(nn,dnn,dn,jt,tt,  hm, hr, hq)

java.net.ConnectException: Call to localhost/127.0.0.1:9000 failed on 
connection exception: java.net.ConnectException: Connection refused
at org.apache.hadoop.ipc.Client.wrapException(Client.java:1131)
        at org.apache.hadoop.ipc.Client.call(Client.java:1107)
        at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:226)
        at $Proxy4.getProtocolVersion(Unknown Source)
        at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:398)
        at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:384)
        at 
org.apache.hadoop.hdfs.DFSClient.createRPCNamenode(DFSClient.java:111)
        at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:213)
        at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:180)
        at 
org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:89)
        at 
org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:1514)
        at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:67)
        at 
org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:1548)
        at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:1530)
        at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:228)
        at org.apache.hadoop.fs.Path.getFileSystem(Path.java:183)
        at 
org.apache.hadoop.hive.ql.exec.Utilities.isEmptyPath(Utilities.java:1475)
        at 
org.apache.hadoop.hive.ql.exec.ExecDriver.addInputPaths(ExecDriver.java:1253)
        at 
org.apache.hadoop.hive.ql.exec.ExecDriver.execute(ExecDriver.java:632)
        at 
org.apache.hadoop.hive.ql.exec.MapRedTask.execute(MapRedTask.java:123)
        at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:130)
        at 
org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:57)
        at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1063)
        at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:900)
        at org.apache.hadoop.hive.ql.Driver.run(Driver.java:748)
        at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:164)
        at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:241)
        at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:456)
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
        at java.lang.reflect.Method.invoke(Method.java:597)
        at org.apache.hadoop.util.RunJar.main(RunJar.java:186)
Caused by: java.net.ConnectException: Connection refused
        at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
        at 
sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:567)
        at 
org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
        at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:408)
        at 
org.apache.hadoop.ipc.Client$Connection.setupConnection(Client.java:425)
        at 
org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:532)
        at org.apache.hadoop.ipc.Client$Connection.access$2300(Client.java:210)
        at org.apache.hadoop.ipc.Client.getConnection(Client.java:1244)
        at org.apache.hadoop.ipc.Client.call(Client.java:1075)


On Sep 21, 2011, at 3:01 PM, Ayon Sinha wrote:

> I'm a bit concerned about port 9000 for the HDFS location. Is your namenode 
> at port 9000? Can you run
> hadoop dfs -ls  hdfs://localhost:9000/user/hive/warehouse/supplier     
>  
> -Ayon
> See My Photos on Flickr
> Also check out my Blog for answers to commonly asked questions.
> 
> From: Krish Khambadkone <kkhambadk...@apple.com>
> To: user@hive.apache.org
> Sent: Wednesday, September 21, 2011 2:45 PM
> Subject: Re: Exception when joining HIVE tables
> 
> Here is the table info,   and the query is "select acctbal, availqty, partkey 
> from partsupp JOIN supplier ON (partsupp.suppkey == supplier.suppkey); "
> 
>  desc formatted supplier;
> OK
> # col_name                    data_type               comment             
>                
> key                   string                  None                
> acctbal               string                  None                
> address               string                  None                
> name                  string                  None                
> nationkey             bigint                  None                
> phone                 string                  None                
> suppkey               bigint                  None                
>                
> # Detailed Table Information           
> Database:             default                  
> Owner:                hadoop                   
> CreateTime:           Wed Sep 21 13:05:50 PDT 2011     
> LastAccessTime:       UNKNOWN                  
> Protect Mode:         None                     
> Retention:            0                        
> Location:             hdfs://localhost:9000/user/hive/warehouse/supplier      
>  
> Table Type:           MANAGED_TABLE            
> Table Parameters:              
>       transient_lastDdlTime   1316635649          
>                
> # Storage Information  
>        
> SerDe Library:        org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe      
>  
> InputFormat:          org.apache.hadoop.mapred.TextInputFormat         
> OutputFormat:         
> org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat       
> Compressed:           No                       
> Num Buckets:          -1                       
> Bucket Columns:       []                       
> Sort Columns:         []                       
> Storage Desc Params:           
>       field.delim             ,                   
>       serialization.format    ,                   
> Time taken: 0.213 seconds
> 
> 
> desc formatted partsupp;
> OK
> # col_name                    data_type               comment             
>                
> key                   string                  None                
> availqty              int                     None                
> partkey               bigint                  None                
> suppkey               bigint                  None                
> supplycost            double                  None                
>                
> # Detailed Table Information           
> Database:             default                  
> Owner:                hadoop                   
> CreateTime:           Wed Sep 21 13:05:37 PDT 2011     
> LastAccessTime:       UNKNOWN                  
> Protect Mode:         None                     
> Retention:            0                   
>        
> Location:             hdfs://localhost:9000/user/hive/warehouse/partsupp      
>  
> Table Type:           MANAGED_TABLE            
> Table Parameters:              
>       transient_lastDdlTime   1316635698          
>                
> # Storage Information          
> SerDe Library:        org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe      
>  
> InputFormat:          org.apache.hadoop.mapred.TextInputFormat         
> OutputFormat:       
>       org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat       
> Compressed:           No                       
> Num Buckets:          -1                       
> Bucket Columns:       []                       
> Sort Columns:         []                       
> Storage Desc Params:           
>       field.delim             ,                   
>       serialization.format    ,                   
> Time taken: 2.192 seconds
> 
> 
> On Sep 21, 2011, at 2:10 PM, Ayon Sinha wrote:
> 
>> If you can share details of your tables and query we might be able to help. 
>> Do a desc formatted <tablename>
>>  
>> -Ayon
>> See My Photos on Flickr
>> Also check out my Blog for answers to commonly asked questions.
>> 
>> From: Krish Khambadkone <kkhambadk...@apple.com>
>> To: user@hive.apache.org
>> Sent: Wednesday, September 21, 2011 1:11 PM
>> Subject: Exception when joining HIVE tables
>> 
>> Hi,  I get this exception when I try to join two hive tables or even when I 
>> use a specific WHERE clause.  "SELECT *" from any individual table seems to 
>> work fine.  Any idea what is missing here.  I am on hive version 
>> hive-0.7.0-cdh3u0.
>> 
>> java.lang.IllegalArgumentException: Can not create a Path from an empty 
>> string
>>     at org.apache.hadoop.fs.Path.checkPathArg(Path.java:82)
>>     at org.apache.hadoop.fs.Path.<init>(Path.java:90)
>>     at org.apache.hadoop.fs.Path.<init>(Path.java:50)
>>     at org.apache.hadoop.mapred.JobClient.copyRemoteFiles(JobClient.java:608)
>>     at 
>> org.apache.hadoop.mapred.JobClient.copyAndConfigureFiles(JobClient.java:713)
>>     at 
>> org.apache.hadoop.mapred.JobClient.copyAndConfigureFiles(JobClient.java:637)
>>     at org.apache.hadoop.mapred.JobClient.access$300(JobClient.java:170)
>>     at org.apache.hadoop.mapred.JobClient$2.run(JobClient.java:848)
>>     at org.apache.hadoop.mapred.JobClient$2.run(JobClient.java:833)
>>     at java.security.AccessController.doPrivileged(Native Method)
>>     at javax.security.auth.Subject.doAs(Subject.java:396)
>>     at 
>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1115)
>>     at 
>> org.apache.hadoop.mapred.JobClient.submitJobInternal(JobClient.java:833)
>>     at org.apache.hadoop.mapred.JobClient.submitJob(JobClient.java:807)
>>     at org.apache.hadoop.hive.ql.exec.ExecDriver.execute(ExecDriver.java:657)
>>     at org.apache.hadoop.hive.ql.exec.MapRedTask.execute(MapRedTask.java:123)
>>     at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:130)
>>     at 
>> org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:57)
>>     at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1063)
>>     at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:900)
>>     at org.apache.hadoop.hive.ql.Driver.run(Driver.java:748)
>>     at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:164)
>>     at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:241)
>>     at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:456)
>>     at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>     at 
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>>     at 
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>>     at java.lang.reflect.Method.invoke(Method.java:597)
>>     at org.apache.hadoop.util.RunJar.main(RunJar.java:186)
>> Job Submission failed with exception 'java.lang.IllegalArgumentException(Can 
>> not create a Path from an empty string)'
>> FAILED: Execution Error, return code 1 from 
>> org.apache.hadoop.hive.ql.exec.MapRedTask
>> 
>> 
> 
> 
> 

Reply via email to