Your program may be trying to run locally on your windows box instead of
talking to the ubuntu micro cluster.
On 10/13/07 11:43 AM, "dcave555" <[EMAIL PROTECTED]> wrote:
>
> Thx for your reply
>
> But i think i don't understand something.
> i have installed hadoop single node on ubuntu.
> my test is running on windows.
>
> i changed property in hadoop-site.xml in my application to
>
> <property>
> <name>hadoop.tmp.dir</name>
> <value>//172.16.50.13/usr/local/hadoop-datastore/hadoop-hadoop</value>
> <description>A base for other temporary directories.</description>
> </property>
>
> amd i am getting Login failed.
>
> If i want to save file from windows to hadoop on ubuntu i need to import ssh
> public key?
> how can i tell my test application on windows to use this pulic key?
>
> Or i need to do something else?
>
> Thx
>
>
>
>
>
> Ted Dunning-3 wrote:
>>
>>
>> Hadoop is having trouble spawning a process to measure how much disk space
>> you have available.
>>
>> It really is much easier to run hadoop on Linux. In case you can't do
>> that,
>> make sure that you really do have cygwin installed in a standard way on
>> each
>> hadoop node.
>>
>>
>> On 10/13/07 8:05 AM, "dcave555" <[EMAIL PROTECTED]> wrote:
>>
>>>
>>> Hello all
>>>
>>> I am new to hadoop .
>>>
>>> I am trying to write file to single cluster and getting this exception
>>> when
>>> i am trying to close output stream
>>>
>>> java.io.IOException: CreateProcess: df -k
>>> C:\usr\local\hadoop-datastore\hadoop-hadoop\dfs\tmp error=2
>>> at java.lang.ProcessImpl.create(Native Method)
>>> at java.lang.ProcessImpl.<init>(Unknown Source)
>>> at java.lang.ProcessImpl.start(Unknown Source)
>>> at java.lang.ProcessBuilder.start(Unknown Source)
>>> at java.lang.Runtime.exec(Unknown Source)
>>> at java.lang.Runtime.exec(Unknown Source)
>>> at org.apache.hadoop.fs.DF.doDF(DF.java:60)
>>> at org.apache.hadoop.fs.DF.<init>(DF.java:53)
>>> at
>>> org.apache.hadoop.fs.LocalDirAllocator$AllocatorPerContext.confChanged(Local
>>> Di
>>> rAllocator.java:198)
>>> at
>>> org.apache.hadoop.fs.LocalDirAllocator$AllocatorPerContext.getLocalPathForWr
>>> it
>>> e(LocalDirAllocator.java:235)
>>> at
>>> org.apache.hadoop.fs.LocalDirAllocator$AllocatorPerContext.createTmpFileForW
>>> ri
>>> te(LocalDirAllocator.java:276)
>>> at
>>> org.apache.hadoop.fs.LocalDirAllocator.createTmpFileForWrite(LocalDirAllocat
>>> or
>>> .java:155)
>>> at
>>> org.apache.hadoop.dfs.DFSClient$DFSOutputStream.newBackupFile(DFSClient.java
>>> :1
>>> 475)
>>> at
>>> org.apache.hadoop.dfs.DFSClient$DFSOutputStream.openBackupStream(DFSClient.j
>>> av
>>> a:1442)
>>> at
>>>
>> org.apache.hadoop.dfs.DFSClient$DFSOutputStream.writeChunk(DFSClient.java:160
>> 0>
>> )
>>> at
>>>
>> org.apache.hadoop.fs.FSOutputSummer.writeChecksumChunk(FSOutputSummer.java:14
>> 0>
>> )
>>> at
>>> org.apache.hadoop.fs.FSOutputSummer.flushBuffer(FSOutputSummer.java:122)
>>> at
>>> org.apache.hadoop.dfs.DFSClient$DFSOutputStream.close(DFSClient.java:1739)
>>> at
>>> org.apache.hadoop.fs.FSDataOutputStream$PositionCache.close(FSDataOutputStre
>>> am
>>> .java:49)
>>> at
>>> org.apache.hadoop.fs.FSDataOutputStream.close(FSDataOutputStream.java:64)
>>> at Test1.main(Test1.java:23)
>>>
>>>
>>>
>>> My test is:
>>>
>>> Configuration configuration = new
>>> Configuration();
>>> FileSystem fileSystem =
>>> FileSystem.get(configuration);
>>> Path path = new Path("/testfile");
>>> //writing:
>>> FSDataOutputStream dataOutputStream = fileSystem.create(path);
>>> dataOutputStream.writeUTF("hello world");
>>> dataOutputStream.close();
>>> //reading
>>> FSDataInputStream dataInputStream = fileSystem.open(path);
>>> System.out.println(dataInputStream.readUTF());
>>> dataInputStream.close();
>>> fileSystem.close();
>>>
>>> i added hadoop-site.xml to classpath :
>>>
>>>
>>> <configuration>
>>>
>>> <property>
>>> <name>hadoop.tmp.dir</name>
>>> <value>/usr/local/hadoop-datastore/hadoop-hadoop</value>
>>> <description>A base for other temporary directories.</description>
>>> </property>
>>>
>>> <property>
>>> <name>fs.default.name</name>
>>> <value>hdfs://172.16.50.13:54310</value>
>>> <description>The name of the default file system. A URI whose
>>> scheme and authority determine the FileSystem implementation. The
>>> uri's scheme determines the config property (fs.SCHEME.impl) naming
>>> the FileSystem implementation class. The uri's authority is used to
>>> determine the host, port, etc. for a filesystem.</description>
>>> </property>
>>>
>>> <property>
>>> <name>mapred.job.tracker</name>
>>> <value>172.16.50.13:54311</value>
>>> <description>The host and port that the MapReduce job tracker runs
>>> at. If "local", then jobs are run in-process as a single map
>>> and reduce task.
>>> </description>
>>> </property>
>>>
>>> <property>
>>> <name>dfs.replication</name>
>>> <value>1</value>
>>> <description>Default block replication.
>>> The actual number of replications can be specified when the file is
>>> created.
>>> The default is used if replication is not specified in create time.
>>> </description>
>>> </property>
>>> </configuration>
>>>
>>>
>>>
>>> Please help me
>>> thx
>>
>>
>>