Fourie,

Good to know. Just for the record, do you remember the version of
OpenJDK that was installed? So that we can mark that as unusable in
our docs/wiki.

On Thu, May 10, 2012 at 12:22 PM, Fourie Joubert
<fourie.joub...@up.ac.za> wrote:
> Hi
>
> Yes - that was indeed the problem...
>
> I cleaned up the Java's on all the nodes, did a clean reinstall of Sun
> jdk1.6.0_23 and the problem is gone.
>
> Many thanks and regards!
>
>
> Fourie
>
> On 05/09/2012 05:47 PM, Harsh J wrote:
>>
>> You may be hitting https://issues.apache.org/jira/browse/HDFS-1115?
>> Have you ensured Sun JDK is the only JDK available in the machines and
>> your services aren't using OpenJDK accidentally?
>>
>> On Wed, May 9, 2012 at 8:44 PM, Fourie Joubert<fourie.joub...@up.ac.za>
>>  wrote:
>>>
>>> Hi
>>>
>>> I am running Hadoop-1.0.1 with Sun jdk1.6.0_23.
>>>
>>> My system is a head node with 14 compute blades
>>>
>>> When trying to start hadoop, I get the following message in the logs for
>>> each data node:
>>>
>>>
>>> 2012-05-09 16:53:35,548 ERROR
>>> org.apache.hadoop.hdfs.server.datanode.DataNode:
>>> DatanodeRegistration(137.215.75.201:50010,
>>> storageID=DS-2067460883-137.215.75.201-50010-1336575105195,
>>> infoPort=50075,
>>> ipcPort=50020):DataXceiver
>>>
>>> java.net.SocketException: Protocol not available
>>> ...
>>> ...
>>>
>>> The full log is shown below.
>>>
>>> I can't seem to get past this problem - any help or advice would be
>>> sincerely appreciated.
>>>
>>> Kindest regards!
>>>
>>> Fourie
>>>
>>>
>>>
>>>
>>>
>>> 2012-05-09 16:53:31,800 INFO
>>> org.apache.hadoop.hdfs.server.datanode.DataNode: STARTUP_MSG:
>>> /************************************************************
>>> STARTUP_MSG: Starting DataNode
>>> STARTUP_MSG:   host = wonko1/137.215.75.201
>>> STARTUP_MSG:   args = []
>>> STARTUP_MSG:   version = 1.0.1
>>> STARTUP_MSG:   build =
>>> https://svn.apache.org/repos/asf/hadoop/common/branches/branch-1.0 -r
>>> 1243785; compiled by 'hortonfo' on Tue Feb 14 08:15:38 UTC 2012
>>> ************************************************************/
>>> 2012-05-09 16:53:31,934 INFO
>>> org.apache.hadoop.metrics2.impl.MetricsConfig:
>>> loaded properties from hadoop-metrics2.properties
>>> 2012-05-09 16:53:31,945 INFO
>>> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source
>>> MetricsSystem,sub=Stats registered.
>>> 2012-05-09 16:53:31,946 INFO
>>> org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Scheduled snapshot
>>> period
>>> at 10 second(s).
>>> 2012-05-09 16:53:31,946 INFO
>>> org.apache.hadoop.metrics2.impl.MetricsSystemImpl: DataNode metrics
>>> system
>>> started
>>> 2012-05-09 16:53:32,022 INFO
>>> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source
>>> ugi
>>> registered.
>>> 2012-05-09 16:53:32,232 INFO
>>> org.apache.hadoop.hdfs.server.datanode.DataNode: Registered
>>> FSDatasetStatusMBean
>>> 2012-05-09 16:53:32,242 INFO
>>> org.apache.hadoop.hdfs.server.datanode.DataNode: Opened info server at
>>> 50010
>>> 2012-05-09 16:53:32,244 INFO
>>> org.apache.hadoop.hdfs.server.datanode.DataNode: Balancing bandwith is
>>> 1048576 bytes/s
>>> 2012-05-09 16:53:32,291 INFO org.mortbay.log: Logging to
>>> org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via
>>> org.mortbay.log.Slf4jLog
>>> 2012-05-09 16:53:32,347 INFO org.apache.hadoop.http.HttpServer: Added
>>> global
>>> filtersafety (class=org.apache.hadoop.http.HttpServer$QuotingInputFilter)
>>> 2012-05-09 16:53:32,359 INFO
>>> org.apache.hadoop.hdfs.server.datanode.DataNode: dfs.webhdfs.enabled =
>>> false
>>> 2012-05-09 16:53:32,359 INFO org.apache.hadoop.http.HttpServer: Port
>>> returned by webServer.getConnectors()[0].getLocalPort() before open() is
>>> -1.
>>> Opening the listener on 50075
>>> 2012-05-09 16:53:32,359 INFO org.apache.hadoop.http.HttpServer:
>>> listener.getLocalPort() returned 50075
>>> webServer.getConnectors()[0].getLocalPort() returned 50075
>>> 2012-05-09 16:53:32,360 INFO org.apache.hadoop.http.HttpServer: Jetty
>>> bound
>>> to port 50075
>>> 2012-05-09 16:53:32,360 INFO org.mortbay.log: jetty-6.1.26
>>> 2012-05-09 16:53:32,590 INFO org.mortbay.log: Started
>>> SelectChannelConnector@0.0.0.0:50075
>>> 2012-05-09 16:53:32,594 INFO
>>> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source
>>> jvm
>>> registered.
>>> 2012-05-09 16:53:32,595 INFO
>>> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source
>>> DataNode registered.
>>> 2012-05-09 16:53:32,614 INFO org.apache.hadoop.ipc.Server: Starting
>>> SocketReader
>>> 2012-05-09 16:53:32,616 INFO
>>> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source
>>> RpcDetailedActivityForPort50020 registered.
>>> 2012-05-09 16:53:32,616 INFO
>>> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source
>>> RpcActivityForPort50020 registered.
>>> 2012-05-09 16:53:32,618 INFO
>>> org.apache.hadoop.hdfs.server.datanode.DataNode: dnRegistration =
>>> DatanodeRegistration(wonko1.bi.up.ac.za:50010,
>>> storageID=DS-2067460883-137.215.75.201-50010-1336575105195,
>>> infoPort=50075,
>>> ipcPort=50020)
>>> 2012-05-09 16:53:32,620 INFO
>>> org.apache.hadoop.hdfs.server.datanode.DataNode: Starting asynchronous
>>> block
>>> report scan
>>> 2012-05-09 16:53:32,620 INFO
>>> org.apache.hadoop.hdfs.server.datanode.DataNode:
>>> DatanodeRegistration(137.215.75.201:50010,
>>> storageID=DS-2067460883-137.215.75.201-50010-1336575105195,
>>> infoPort=50075,
>>> ipcPort=50020)In DataNode.run, data =
>>> FSDataset{dirpath='/hadooplocal/datadir/current'}
>>> 2012-05-09 16:53:32,620 INFO
>>> org.apache.hadoop.hdfs.server.datanode.DataNode: Finished asynchronous
>>> block
>>> report scan in 0ms
>>> 2012-05-09 16:53:32,621 INFO org.apache.hadoop.ipc.Server: IPC Server
>>> Responder: starting
>>> 2012-05-09 16:53:32,621 INFO org.apache.hadoop.ipc.Server: IPC Server
>>> listener on 50020: starting
>>> 2012-05-09 16:53:32,623 INFO org.apache.hadoop.ipc.Server: IPC Server
>>> handler 0 on 50020: starting
>>> 2012-05-09 16:53:32,623 INFO org.apache.hadoop.ipc.Server: IPC Server
>>> handler 1 on 50020: starting
>>> 2012-05-09 16:53:32,623 INFO
>>> org.apache.hadoop.hdfs.server.datanode.DataNode: using
>>> BLOCKREPORT_INTERVAL
>>> of 3600000msec Initial delay: 0msec
>>> 2012-05-09 16:53:32,623 INFO org.apache.hadoop.ipc.Server: IPC Server
>>> handler 2 on 50020: starting
>>> 2012-05-09 16:53:32,626 INFO
>>> org.apache.hadoop.hdfs.server.datanode.DataNode: Reconciled asynchronous
>>> block report against current state in 0 ms
>>> 2012-05-09 16:53:32,628 INFO
>>> org.apache.hadoop.hdfs.server.datanode.DataNode: BlockReport of 0 blocks
>>> took 0 msec to generate and 2 msecs for RPC and NN processing
>>> 2012-05-09 16:53:32,628 INFO
>>> org.apache.hadoop.hdfs.server.datanode.DataNode: Starting Periodic block
>>> scanner.
>>> 2012-05-09 16:53:32,629 INFO
>>> org.apache.hadoop.hdfs.server.datanode.DataNode: Generated rough
>>> (lockless)
>>> block report in 0 ms
>>> 2012-05-09 16:53:32,629 INFO
>>> org.apache.hadoop.hdfs.server.datanode.DataNode: Reconciled asynchronous
>>> block report against current state in 0 ms
>>> 2012-05-09 16:53:35,548 ERROR
>>> org.apache.hadoop.hdfs.server.datanode.DataNode:
>>> DatanodeRegistration(137.215.75.201:50010,
>>> storageID=DS-2067460883-137.215.75.201-50010-1336575105195,
>>> infoPort=50075,
>>> ipcPort=50020):DataXceiver
>>> java.net.SocketException: Protocol not available
>>>        at sun.nio.ch.Net.getIntOption0(Native Method)
>>>        at sun.nio.ch.Net.getIntOption(Net.java:181)
>>>        at
>>> sun.nio.ch.SocketChannelImpl$1.getInt(SocketChannelImpl.java:419)
>>>        at sun.nio.ch.SocketOptsImpl.getInt(SocketOptsImpl.java:60)
>>>        at
>>> sun.nio.ch.SocketOptsImpl.receiveBufferSize(SocketOptsImpl.java:142)
>>>        at
>>>
>>> sun.nio.ch.SocketOptsImpl$IP$TCP.receiveBufferSize(SocketOptsImpl.java:286)
>>>        at
>>> sun.nio.ch.OptionAdaptor.getReceiveBufferSize(OptionAdaptor.java:148)
>>>        at
>>> sun.nio.ch.SocketAdaptor.getReceiveBufferSize(SocketAdaptor.java:336)
>>>        at
>>>
>>> org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:238)
>>>        at
>>>
>>> org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:107)
>>>        at java.lang.Thread.run(Thread.java:636)
>>> 2012-05-09 16:54:36,377 ERROR
>>> org.apache.hadoop.hdfs.server.datanode.DataNode:
>>> DatanodeRegistration(137.215.75.201:50010,
>>> storageID=DS-2067460883-137.215.75.201-50010-1336575105195,
>>> infoPort=50075,
>>> ipcPort=50020):DataXceiver
>>> java.net.SocketException: Protocol not available
>>>        at sun.nio.ch.Net.getIntOption0(Native Method)
>>>        at sun.nio.ch.Net.getIntOption(Net.java:181)
>>>        at
>>> sun.nio.ch.SocketChannelImpl$1.getInt(SocketChannelImpl.java:419)
>>>        at sun.nio.ch.SocketOptsImpl.getInt(SocketOptsImpl.java:60)
>>>        at
>>> sun.nio.ch.SocketOptsImpl.receiveBufferSize(SocketOptsImpl.java:142)
>>>        at
>>>
>>> sun.nio.ch.SocketOptsImpl$IP$TCP.receiveBufferSize(SocketOptsImpl.java:286)
>>>        at
>>> sun.nio.ch.OptionAdaptor.getReceiveBufferSize(OptionAdaptor.java:148)
>>>        at
>>> sun.nio.ch.SocketAdaptor.getReceiveBufferSize(SocketAdaptor.java:336)
>>>        at
>>>
>>> org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:238)
>>>        at
>>>
>>> org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:107)
>>>        at java.lang.Thread.run(Thread.java:636)
>>> 2012-05-09 16:54:46,427 ERROR
>>> org.apache.hadoop.hdfs.server.datanode.DataNode:
>>> DatanodeRegistration(137.215.75.201:50010,
>>> storageID=DS-2067460883-137.215.75.201-50010-1336575105195,
>>> infoPort=50075,
>>> ipcPort=50020):DataXceiver
>>> java.net.SocketException: Protocol not available
>>>        at sun.nio.ch.Net.getIntOption0(Native Method)
>>>
>>>
>>>
>>> --
>>> --------------
>>> Prof Fourie Joubert
>>> Bioinformatics and Computational Biology Unit
>>> Department of Biochemistry
>>> University of Pretoria
>>> fourie.joub...@up.ac.za
>>> http://www.bi.up.ac.za
>>> Tel. +27-12-420-5825
>>> Fax. +27-12-420-5800
>>>
>>> -------------------------------------------------------------------------
>>> This message and attachments are subject to a disclaimer. Please refer
>>> to www.it.up.ac.za/documentation/governance/disclaimer/ for full details.
>>>
>>
>>
>
>
> --
> --------------
> Prof Fourie Joubert
> Bioinformatics and Computational Biology Unit
> Department of Biochemistry
> University of Pretoria
> fourie.joub...@up.ac.za
> http://www.bi.up.ac.za
> Tel. +27-12-420-5825
> Fax. +27-12-420-5800
>
> -------------------------------------------------------------------------
> This message and attachments are subject to a disclaimer. Please refer
> to www.it.up.ac.za/documentation/governance/disclaimer/ for full details.
>



-- 
Harsh J

Reply via email to