Hi

Yes - that was indeed the problem...

I cleaned up the Java's on all the nodes, did a clean reinstall of Sun jdk1.6.0_23 and the problem is gone.

Many thanks and regards!

Fourie

On 05/09/2012 05:47 PM, Harsh J wrote:
You may be hitting https://issues.apache.org/jira/browse/HDFS-1115?
Have you ensured Sun JDK is the only JDK available in the machines and
your services aren't using OpenJDK accidentally?

On Wed, May 9, 2012 at 8:44 PM, Fourie Joubert<fourie.joub...@up.ac.za>  wrote:
Hi

I am running Hadoop-1.0.1 with Sun jdk1.6.0_23.

My system is a head node with 14 compute blades

When trying to start hadoop, I get the following message in the logs for
each data node:


2012-05-09 16:53:35,548 ERROR
org.apache.hadoop.hdfs.server.datanode.DataNode:
DatanodeRegistration(137.215.75.201:50010,
storageID=DS-2067460883-137.215.75.201-50010-1336575105195, infoPort=50075,
ipcPort=50020):DataXceiver

java.net.SocketException: Protocol not available
...
...

The full log is shown below.

I can't seem to get past this problem - any help or advice would be
sincerely appreciated.

Kindest regards!

Fourie





2012-05-09 16:53:31,800 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: STARTUP_MSG:
/************************************************************
STARTUP_MSG: Starting DataNode
STARTUP_MSG:   host = wonko1/137.215.75.201
STARTUP_MSG:   args = []
STARTUP_MSG:   version = 1.0.1
STARTUP_MSG:   build =
https://svn.apache.org/repos/asf/hadoop/common/branches/branch-1.0 -r
1243785; compiled by 'hortonfo' on Tue Feb 14 08:15:38 UTC 2012
************************************************************/
2012-05-09 16:53:31,934 INFO org.apache.hadoop.metrics2.impl.MetricsConfig:
loaded properties from hadoop-metrics2.properties
2012-05-09 16:53:31,945 INFO
org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source
MetricsSystem,sub=Stats registered.
2012-05-09 16:53:31,946 INFO
org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Scheduled snapshot period
at 10 second(s).
2012-05-09 16:53:31,946 INFO
org.apache.hadoop.metrics2.impl.MetricsSystemImpl: DataNode metrics system
started
2012-05-09 16:53:32,022 INFO
org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source ugi
registered.
2012-05-09 16:53:32,232 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: Registered
FSDatasetStatusMBean
2012-05-09 16:53:32,242 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: Opened info server at 50010
2012-05-09 16:53:32,244 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: Balancing bandwith is
1048576 bytes/s
2012-05-09 16:53:32,291 INFO org.mortbay.log: Logging to
org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via
org.mortbay.log.Slf4jLog
2012-05-09 16:53:32,347 INFO org.apache.hadoop.http.HttpServer: Added global
filtersafety (class=org.apache.hadoop.http.HttpServer$QuotingInputFilter)
2012-05-09 16:53:32,359 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: dfs.webhdfs.enabled = false
2012-05-09 16:53:32,359 INFO org.apache.hadoop.http.HttpServer: Port
returned by webServer.getConnectors()[0].getLocalPort() before open() is -1.
Opening the listener on 50075
2012-05-09 16:53:32,359 INFO org.apache.hadoop.http.HttpServer:
listener.getLocalPort() returned 50075
webServer.getConnectors()[0].getLocalPort() returned 50075
2012-05-09 16:53:32,360 INFO org.apache.hadoop.http.HttpServer: Jetty bound
to port 50075
2012-05-09 16:53:32,360 INFO org.mortbay.log: jetty-6.1.26
2012-05-09 16:53:32,590 INFO org.mortbay.log: Started
SelectChannelConnector@0.0.0.0:50075
2012-05-09 16:53:32,594 INFO
org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source jvm
registered.
2012-05-09 16:53:32,595 INFO
org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source
DataNode registered.
2012-05-09 16:53:32,614 INFO org.apache.hadoop.ipc.Server: Starting
SocketReader
2012-05-09 16:53:32,616 INFO
org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source
RpcDetailedActivityForPort50020 registered.
2012-05-09 16:53:32,616 INFO
org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source
RpcActivityForPort50020 registered.
2012-05-09 16:53:32,618 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: dnRegistration =
DatanodeRegistration(wonko1.bi.up.ac.za:50010,
storageID=DS-2067460883-137.215.75.201-50010-1336575105195, infoPort=50075,
ipcPort=50020)
2012-05-09 16:53:32,620 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: Starting asynchronous block
report scan
2012-05-09 16:53:32,620 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode:
DatanodeRegistration(137.215.75.201:50010,
storageID=DS-2067460883-137.215.75.201-50010-1336575105195, infoPort=50075,
ipcPort=50020)In DataNode.run, data =
FSDataset{dirpath='/hadooplocal/datadir/current'}
2012-05-09 16:53:32,620 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: Finished asynchronous block
report scan in 0ms
2012-05-09 16:53:32,621 INFO org.apache.hadoop.ipc.Server: IPC Server
Responder: starting
2012-05-09 16:53:32,621 INFO org.apache.hadoop.ipc.Server: IPC Server
listener on 50020: starting
2012-05-09 16:53:32,623 INFO org.apache.hadoop.ipc.Server: IPC Server
handler 0 on 50020: starting
2012-05-09 16:53:32,623 INFO org.apache.hadoop.ipc.Server: IPC Server
handler 1 on 50020: starting
2012-05-09 16:53:32,623 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: using BLOCKREPORT_INTERVAL
of 3600000msec Initial delay: 0msec
2012-05-09 16:53:32,623 INFO org.apache.hadoop.ipc.Server: IPC Server
handler 2 on 50020: starting
2012-05-09 16:53:32,626 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: Reconciled asynchronous
block report against current state in 0 ms
2012-05-09 16:53:32,628 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: BlockReport of 0 blocks
took 0 msec to generate and 2 msecs for RPC and NN processing
2012-05-09 16:53:32,628 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: Starting Periodic block
scanner.
2012-05-09 16:53:32,629 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: Generated rough (lockless)
block report in 0 ms
2012-05-09 16:53:32,629 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: Reconciled asynchronous
block report against current state in 0 ms
2012-05-09 16:53:35,548 ERROR
org.apache.hadoop.hdfs.server.datanode.DataNode:
DatanodeRegistration(137.215.75.201:50010,
storageID=DS-2067460883-137.215.75.201-50010-1336575105195, infoPort=50075,
ipcPort=50020):DataXceiver
java.net.SocketException: Protocol not available
        at sun.nio.ch.Net.getIntOption0(Native Method)
        at sun.nio.ch.Net.getIntOption(Net.java:181)
        at sun.nio.ch.SocketChannelImpl$1.getInt(SocketChannelImpl.java:419)
        at sun.nio.ch.SocketOptsImpl.getInt(SocketOptsImpl.java:60)
        at
sun.nio.ch.SocketOptsImpl.receiveBufferSize(SocketOptsImpl.java:142)
        at
sun.nio.ch.SocketOptsImpl$IP$TCP.receiveBufferSize(SocketOptsImpl.java:286)
        at
sun.nio.ch.OptionAdaptor.getReceiveBufferSize(OptionAdaptor.java:148)
        at
sun.nio.ch.SocketAdaptor.getReceiveBufferSize(SocketAdaptor.java:336)
        at
org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:238)
        at
org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:107)
        at java.lang.Thread.run(Thread.java:636)
2012-05-09 16:54:36,377 ERROR
org.apache.hadoop.hdfs.server.datanode.DataNode:
DatanodeRegistration(137.215.75.201:50010,
storageID=DS-2067460883-137.215.75.201-50010-1336575105195, infoPort=50075,
ipcPort=50020):DataXceiver
java.net.SocketException: Protocol not available
        at sun.nio.ch.Net.getIntOption0(Native Method)
        at sun.nio.ch.Net.getIntOption(Net.java:181)
        at sun.nio.ch.SocketChannelImpl$1.getInt(SocketChannelImpl.java:419)
        at sun.nio.ch.SocketOptsImpl.getInt(SocketOptsImpl.java:60)
        at
sun.nio.ch.SocketOptsImpl.receiveBufferSize(SocketOptsImpl.java:142)
        at
sun.nio.ch.SocketOptsImpl$IP$TCP.receiveBufferSize(SocketOptsImpl.java:286)
        at
sun.nio.ch.OptionAdaptor.getReceiveBufferSize(OptionAdaptor.java:148)
        at
sun.nio.ch.SocketAdaptor.getReceiveBufferSize(SocketAdaptor.java:336)
        at
org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:238)
        at
org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:107)
        at java.lang.Thread.run(Thread.java:636)
2012-05-09 16:54:46,427 ERROR
org.apache.hadoop.hdfs.server.datanode.DataNode:
DatanodeRegistration(137.215.75.201:50010,
storageID=DS-2067460883-137.215.75.201-50010-1336575105195, infoPort=50075,
ipcPort=50020):DataXceiver
java.net.SocketException: Protocol not available
        at sun.nio.ch.Net.getIntOption0(Native Method)



--
--------------
Prof Fourie Joubert
Bioinformatics and Computational Biology Unit
Department of Biochemistry
University of Pretoria
fourie.joub...@up.ac.za
http://www.bi.up.ac.za
Tel. +27-12-420-5825
Fax. +27-12-420-5800

-------------------------------------------------------------------------
This message and attachments are subject to a disclaimer. Please refer
to www.it.up.ac.za/documentation/governance/disclaimer/ for full details.





--
--------------
Prof Fourie Joubert
Bioinformatics and Computational Biology Unit
Department of Biochemistry
University of Pretoria
fourie.joub...@up.ac.za
http://www.bi.up.ac.za
Tel. +27-12-420-5825
Fax. +27-12-420-5800

-------------------------------------------------------------------------
This message and attachments are subject to a disclaimer. Please refer
to www.it.up.ac.za/documentation/governance/disclaimer/ for full details.

Reply via email to