Re: distcp port for 0.17.2

2008-10-23 Thread Aaron Kimball
Hi,

The dfs.http.address is for human use, not program interoperability. You can
visit http://whatever.address.your.namenode.has:50070 in a web browser and
see statistics about your filesystem.

The address of cluster 2 is in its fs.default.name. This should be set to
something like hdfs://cluster2.master.name:9000/

The file:// protocol only refers to paths on the current machine in its
real (non-DFS) filesystem.
- Aaron

On Wed, Oct 22, 2008 at 3:47 PM, bzheng [EMAIL PROTECTED] wrote:


 Thanks.  The fs.default.name is file:/// and dfs.http.address is
 0.0.0.0:50070.  I tried:

 hadoop dfs -ls /path/file to make sure file exists on cluster1
 hadoop distcp file:///cluster1_master_node_ip:50070/path/file
 file:///cluster2_master_node_ip:50070/path/file

 It gives this error message:
 08/10/22 15:43:47 INFO util.CopyFiles:
 srcPaths=[file:/cluster1_master_node_ip:50070/path/file]
 08/10/22 15:43:47 INFO util.CopyFiles:
 destPath=file:/cluster2_master_node_ip:50070/path/file
 With failures, global counters are inaccurate; consider running with -i
 Copy failed: org.apache.hadoop.mapred.InvalidInputException: Input source
 file:/cluster1_master_node_ip:50070/path/file does not exist.
at org.apache.hadoop.util.CopyFiles.checkSrcPath(CopyFiles.java:578)
at org.apache.hadoop.util.CopyFiles.copy(CopyFiles.java:594)
at org.apache.hadoop.util.CopyFiles.run(CopyFiles.java:743)
at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65)
at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:79)
at org.apache.hadoop.util.CopyFiles.main(CopyFiles.java:763)


 If I use hdfs:// instead of file:///, I get:
 Copy failed: java.net.SocketTimeoutException: timed out waiting for rpc
 response
at org.apache.hadoop.ipc.Client.call(Client.java:559)
at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:212)
at org.apache.hadoop.dfs.$Proxy0.getProtocolVersion(Unknown Source)
at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:313)
at
 org.apache.hadoop.dfs.DFSClient.createRPCNamenode(DFSClient.java:102)
at org.apache.hadoop.dfs.DFSClient.init(DFSClient.java:178)
at

 org.apache.hadoop.dfs.DistributedFileSystem.initialize(DistributedFileSystem.java:68)
at
 org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:1280)
at org.apache.hadoop.fs.FileSystem.access$300(FileSystem.java:56)
at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:1291)
at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:203)
at org.apache.hadoop.fs.Path.getFileSystem(Path.java:175)
at org.apache.hadoop.util.CopyFiles.checkSrcPath(CopyFiles.java:572)
at org.apache.hadoop.util.CopyFiles.copy(CopyFiles.java:594)
at org.apache.hadoop.util.CopyFiles.run(CopyFiles.java:743)
at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65)
at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:79)
at org.apache.hadoop.util.CopyFiles.main(CopyFiles.java:763)



 s29752-hadoopuser wrote:
 
  Hi,
 
  There is no such thing called distcp port.  distcp uses (generic) file
  system API and so it does not care about the file system implementation
  details like port number.
 
  It is common to use distcp with HDFS or HFTP.  The urls will look like
  hdfs://namenode:port/path and hftp://namenode:port/path for HDFS and
 HFTP,
  respectively.   The HDFS and HFTP ports are specified by fs.default.name
  and dfs.http.address, respectively.
 
  Nicholas Sze
 
 
 
 
  - Original Message 
  From: bzheng [EMAIL PROTECTED]
  To: core-user@hadoop.apache.org
  Sent: Wednesday, October 22, 2008 11:57:43 AM
  Subject: distcp port for 0.17.2
 
 
  What's the port number for distcp in 0.17.2?  I can't find any
  documentation
  on distcp for version 0.17.2.  For version 0.18, the documentation says
  it's
  8020.
 
  I'm using a standard install and the only open ports associated with
  hadoop
  are 50030, 50070, and 50090.  None of them work with distcp.  So, how do
  you
  use distcp in 0.17.2?  are there any extra setup/configuration needed?
 
  Thanks in advance for your help.
  --
  View this message in context:
  http://www.nabble.com/distcp-port-for-0.17.2-tp20117463p20117463.html
  Sent from the Hadoop core-user mailing list archive at Nabble.com.
 
 
 

 --
 View this message in context:
 http://www.nabble.com/distcp-port-for-0.17.2-tp20117463p20121246.html
 Sent from the Hadoop core-user mailing list archive at Nabble.com.




Re: distcp port for 0.17.2

2008-10-23 Thread bzheng

It's working for me now.  Turns out the cluster have multiple network
interfaces and I was using the wrong one.  Thanks.



Aaron Kimball-3 wrote:
 
 Hi,
 
 The dfs.http.address is for human use, not program interoperability. You
 can
 visit http://whatever.address.your.namenode.has:50070 in a web browser and
 see statistics about your filesystem.
 
 The address of cluster 2 is in its fs.default.name. This should be set to
 something like hdfs://cluster2.master.name:9000/
 
 The file:// protocol only refers to paths on the current machine in its
 real (non-DFS) filesystem.
 - Aaron
 
 On Wed, Oct 22, 2008 at 3:47 PM, bzheng [EMAIL PROTECTED] wrote:
 

 Thanks.  The fs.default.name is file:/// and dfs.http.address is
 0.0.0.0:50070.  I tried:

 hadoop dfs -ls /path/file to make sure file exists on cluster1
 hadoop distcp file:///cluster1_master_node_ip:50070/path/file
 file:///cluster2_master_node_ip:50070/path/file

 It gives this error message:
 08/10/22 15:43:47 INFO util.CopyFiles:
 srcPaths=[file:/cluster1_master_node_ip:50070/path/file]
 08/10/22 15:43:47 INFO util.CopyFiles:
 destPath=file:/cluster2_master_node_ip:50070/path/file
 With failures, global counters are inaccurate; consider running with -i
 Copy failed: org.apache.hadoop.mapred.InvalidInputException: Input source
 file:/cluster1_master_node_ip:50070/path/file does not exist.
at
 org.apache.hadoop.util.CopyFiles.checkSrcPath(CopyFiles.java:578)
at org.apache.hadoop.util.CopyFiles.copy(CopyFiles.java:594)
at org.apache.hadoop.util.CopyFiles.run(CopyFiles.java:743)
at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65)
at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:79)
at org.apache.hadoop.util.CopyFiles.main(CopyFiles.java:763)


 If I use hdfs:// instead of file:///, I get:
 Copy failed: java.net.SocketTimeoutException: timed out waiting for rpc
 response
at org.apache.hadoop.ipc.Client.call(Client.java:559)
at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:212)
at org.apache.hadoop.dfs.$Proxy0.getProtocolVersion(Unknown
 Source)
at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:313)
at
 org.apache.hadoop.dfs.DFSClient.createRPCNamenode(DFSClient.java:102)
at org.apache.hadoop.dfs.DFSClient.init(DFSClient.java:178)
at

 org.apache.hadoop.dfs.DistributedFileSystem.initialize(DistributedFileSystem.java:68)
at
 org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:1280)
at org.apache.hadoop.fs.FileSystem.access$300(FileSystem.java:56)
at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:1291)
at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:203)
at org.apache.hadoop.fs.Path.getFileSystem(Path.java:175)
at
 org.apache.hadoop.util.CopyFiles.checkSrcPath(CopyFiles.java:572)
at org.apache.hadoop.util.CopyFiles.copy(CopyFiles.java:594)
at org.apache.hadoop.util.CopyFiles.run(CopyFiles.java:743)
at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65)
at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:79)
at org.apache.hadoop.util.CopyFiles.main(CopyFiles.java:763)



 s29752-hadoopuser wrote:
 
  Hi,
 
  There is no such thing called distcp port.  distcp uses (generic) file
  system API and so it does not care about the file system implementation
  details like port number.
 
  It is common to use distcp with HDFS or HFTP.  The urls will look like
  hdfs://namenode:port/path and hftp://namenode:port/path for HDFS and
 HFTP,
  respectively.   The HDFS and HFTP ports are specified by
 fs.default.name
  and dfs.http.address, respectively.
 
  Nicholas Sze
 
 
 
 
  - Original Message 
  From: bzheng [EMAIL PROTECTED]
  To: core-user@hadoop.apache.org
  Sent: Wednesday, October 22, 2008 11:57:43 AM
  Subject: distcp port for 0.17.2
 
 
  What's the port number for distcp in 0.17.2?  I can't find any
  documentation
  on distcp for version 0.17.2.  For version 0.18, the documentation
 says
  it's
  8020.
 
  I'm using a standard install and the only open ports associated with
  hadoop
  are 50030, 50070, and 50090.  None of them work with distcp.  So, how
 do
  you
  use distcp in 0.17.2?  are there any extra setup/configuration needed?
 
  Thanks in advance for your help.
  --
  View this message in context:
  http://www.nabble.com/distcp-port-for-0.17.2-tp20117463p20117463.html
  Sent from the Hadoop core-user mailing list archive at Nabble.com.
 
 
 

 --
 View this message in context:
 http://www.nabble.com/distcp-port-for-0.17.2-tp20117463p20121246.html
 Sent from the Hadoop core-user mailing list archive at Nabble.com.


 
 

-- 
View this message in context: 
http://www.nabble.com/distcp-port-for-0.17.2-tp20117463p20137577.html
Sent from the Hadoop core-user mailing list archive at Nabble.com.



distcp port for 0.17.2

2008-10-22 Thread bzheng

What's the port number for distcp in 0.17.2?  I can't find any documentation
on distcp for version 0.17.2.  For version 0.18, the documentation says it's
8020.  

I'm using a standard install and the only open ports associated with hadoop
are 50030, 50070, and 50090.  None of them work with distcp.  So, how do you
use distcp in 0.17.2?  are there any extra setup/configuration needed?  

Thanks in advance for your help.
-- 
View this message in context: 
http://www.nabble.com/distcp-port-for-0.17.2-tp20117463p20117463.html
Sent from the Hadoop core-user mailing list archive at Nabble.com.



Re: distcp port for 0.17.2

2008-10-22 Thread bzheng

Thanks.  The fs.default.name is file:/// and dfs.http.address is
0.0.0.0:50070.  I tried:

hadoop dfs -ls /path/file to make sure file exists on cluster1
hadoop distcp file:///cluster1_master_node_ip:50070/path/file
file:///cluster2_master_node_ip:50070/path/file

It gives this error message:
08/10/22 15:43:47 INFO util.CopyFiles:
srcPaths=[file:/cluster1_master_node_ip:50070/path/file]
08/10/22 15:43:47 INFO util.CopyFiles:
destPath=file:/cluster2_master_node_ip:50070/path/file
With failures, global counters are inaccurate; consider running with -i
Copy failed: org.apache.hadoop.mapred.InvalidInputException: Input source
file:/cluster1_master_node_ip:50070/path/file does not exist.
at org.apache.hadoop.util.CopyFiles.checkSrcPath(CopyFiles.java:578)
at org.apache.hadoop.util.CopyFiles.copy(CopyFiles.java:594)
at org.apache.hadoop.util.CopyFiles.run(CopyFiles.java:743)
at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65)
at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:79)
at org.apache.hadoop.util.CopyFiles.main(CopyFiles.java:763)


If I use hdfs:// instead of file:///, I get:
Copy failed: java.net.SocketTimeoutException: timed out waiting for rpc
response
at org.apache.hadoop.ipc.Client.call(Client.java:559)
at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:212)
at org.apache.hadoop.dfs.$Proxy0.getProtocolVersion(Unknown Source)
at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:313)
at
org.apache.hadoop.dfs.DFSClient.createRPCNamenode(DFSClient.java:102)
at org.apache.hadoop.dfs.DFSClient.init(DFSClient.java:178)
at
org.apache.hadoop.dfs.DistributedFileSystem.initialize(DistributedFileSystem.java:68)
at
org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:1280)
at org.apache.hadoop.fs.FileSystem.access$300(FileSystem.java:56)
at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:1291)
at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:203)
at org.apache.hadoop.fs.Path.getFileSystem(Path.java:175)
at org.apache.hadoop.util.CopyFiles.checkSrcPath(CopyFiles.java:572)
at org.apache.hadoop.util.CopyFiles.copy(CopyFiles.java:594)
at org.apache.hadoop.util.CopyFiles.run(CopyFiles.java:743)
at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65)
at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:79)
at org.apache.hadoop.util.CopyFiles.main(CopyFiles.java:763)



s29752-hadoopuser wrote:
 
 Hi,
 
 There is no such thing called distcp port.  distcp uses (generic) file
 system API and so it does not care about the file system implementation
 details like port number.
 
 It is common to use distcp with HDFS or HFTP.  The urls will look like
 hdfs://namenode:port/path and hftp://namenode:port/path for HDFS and HFTP,
 respectively.   The HDFS and HFTP ports are specified by fs.default.name
 and dfs.http.address, respectively.
 
 Nicholas Sze
 
 
 
 
 - Original Message 
 From: bzheng [EMAIL PROTECTED]
 To: core-user@hadoop.apache.org
 Sent: Wednesday, October 22, 2008 11:57:43 AM
 Subject: distcp port for 0.17.2
 
 
 What's the port number for distcp in 0.17.2?  I can't find any
 documentation
 on distcp for version 0.17.2.  For version 0.18, the documentation says
 it's
 8020.  
 
 I'm using a standard install and the only open ports associated with
 hadoop
 are 50030, 50070, and 50090.  None of them work with distcp.  So, how do
 you
 use distcp in 0.17.2?  are there any extra setup/configuration needed?  
 
 Thanks in advance for your help.
 -- 
 View this message in context: 
 http://www.nabble.com/distcp-port-for-0.17.2-tp20117463p20117463.html
 Sent from the Hadoop core-user mailing list archive at Nabble.com.
 
 
 

-- 
View this message in context: 
http://www.nabble.com/distcp-port-for-0.17.2-tp20117463p20121246.html
Sent from the Hadoop core-user mailing list archive at Nabble.com.