Thanks,

There was no entry in the log. But as you mentioned a datanode and I wasn't
running any datanode on that machine, I checked the port settings again and
it turned out I was connecting to the wrong daemon! (50010 instead on port
50070 with the default settings) Silly me :(

There should be a better error message though ;)

Thanks,
Thibaut



Andrew Purtell wrote:
> 
> Hello Thibaut,
> 
> What does the log in the corresponding data node show?
> 
> It does look like a jar version mismatch problem to me (EOF
> or socket timeout exceptions during RPC are a strong
> indicator of that), but as you say you have already checked
> into this. All the same perhaps you should verify again
> that things are running from the paths/directories you 
> expect and that every instance of the Hadoop and HBase jar
> files are exactly the same, using md5sum or similar. 
> 
> Thanks,
> 
>    - Andy
> 
> 
>> From: Thibaut_ <[EMAIL PROTECTED]>
>> Subject: Problems running hbase on dfs
>> To: [email protected]
>> Date: Thursday, November 27, 2008, 6:17 AM
>> Hi,
>> 
>> I'm having some problems running hbase on a small
>> cluster of 8 nodes running hadoop. I already worked
>> once, but I must have changed something and it
>> doesn't work anymore... I already spent hours trying to
>> figure out what it was, but I couldn't find the cause.
>> 
>> I'm using the hbase 18.1 release version and tried it
>> with the 18.0,18.1 and 18.2 release of hadoop. My
>> normal java applications can access the
>> distributed file system without any problems
>> 
>> Below is the exception I get when I try to start the hbase
>> mater node. (The hadoop namenode is running on server1 and
>> I'm also starting the hbase master node on server1). I
>> suspected some type of jar version conflict, but I made
>> sure that I uploaded the original files of the
>> distributions to the cluster, while only keeping the conf
>> directory with my settings.
>> 
>> Not using a shared directory on the dfs but a temporary
>> local directory (file:///...) brings the master up without
>> any problems.
>> 
>> 
>> server1:/software/hbase-0.18.1# bin/hbase master start
>> 08/11/27 14:51:22 ERROR master.HMaster: Can not start
>> master
>> java.lang.reflect.InvocationTargetException
>>         at
>> sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native
>> Method)
>>         at
>> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstruct
>>                                                             
>>                                                             
>>     
>> orAccessorImpl.java:39)
>>         at
>> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingC
>>                                                             
>>                                                             
>>     
>> onstructorAccessorImpl.java:27)
>>         at
>> java.lang.reflect.Constructor.newInstance(Constructor.java:513)
>>         at
>> org.apache.hadoop.hbase.master.HMaster.doMain(HMaster.java:784)
>>         at
>> org.apache.hadoop.hbase.master.HMaster.main(HMaster.java:818)
>> Caused by: java.io.IOException: Call failed on local
>> exception
>>         at
>> org.apache.hadoop.ipc.Client.call(Client.java:718)
>>         at
>> org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:216)
>>         at
>> org.apache.hadoop.dfs.$Proxy0.getProtocolVersion(Unknown
>> Source)
>>         at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:319)
>>         at
>> org.apache.hadoop.dfs.DFSClient.createRPCNamenode(DFSClient.java:103)
>>         at
>> org.apache.hadoop.dfs.DFSClient.<init>(DFSClient.java:173)
>>         at
>> org.apache.hadoop.dfs.DistributedFileSystem.initialize(DistributedFil
>>                                                             
>>                                                             
>>     
>> eSystem.java:67)
>>         at
>> org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:1339
>>                                                             
>>                                                             
>>     
>> )
>>         at
>> org.apache.hadoop.fs.FileSystem.access$300(FileSystem.java:56)
>>         at
>> org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:1351)
>>         at
>> org.apache.hadoop.fs.FileSystem.get(FileSystem.java:213)
>>         at
>> org.apache.hadoop.fs.FileSystem.get(FileSystem.java:118)
>>         at
>> org.apache.hadoop.hbase.master.HMaster.<init>(HMaster.java:177)
>>         at
>> org.apache.hadoop.hbase.master.HMaster.<init>(HMaster.java:147)
>>         ... 6 more
>> Caused by: java.io.EOFException
>>         at
>> java.io.DataInputStream.readInt(DataInputStream.java:375)
>>         at
>> org.apache.hadoop.ipc.Client$Connection.receiveResponse(Client.java:4
>>                                                             
>>                                                             
>>     
>> 99)
>>         at
>> org.apache.hadoop.ipc.Client$Connection.run(Client.java:441)
>> 
>> 
>> Here is my hbase-site configuration:
>> 
>>    <property>
>>     <name>hbase.master</name>
>>     <value>server1:60000</value>
>>   </property>
>>   <property>
>>     <name>hbase.rootdir</name>
>>     <value>hdfs://server1:50070/hbase</value>
>>   </property>
>>   <property>
>>    
>> <name>hbase.regionserver.dns.interface</name>
>>     <value>eth0</value>
>>   </property>
>> 
>> 
>> Would appreciate any help!
>> Thanks,
>> Thibaut
> 
> 
>       
> 
> 

-- 
View this message in context: 
http://www.nabble.com/Problems-running-hbase-on-dfs-tp20720031p20726135.html
Sent from the HBase User mailing list archive at Nabble.com.

Reply via email to