Thx for your reply
But i think i don't understand something.
i have installed hadoop single node on ubuntu.
my test is running on windows.
i changed property in hadoop-site.xml in my application to
<property>
<name>hadoop.tmp.dir</name>
<value>//172.16.50.13/usr/local/hadoop-datastore/hadoop-hadoop</value>
<description>A base for other temporary directories.</description>
</property>
amd i am getting Login failed.
If i want to save file from windows to hadoop on ubuntu i need to import ssh
public key?
how can i tell my test application on windows to use this pulic key?
Or i need to do something else?
Thx
Ted Dunning-3 wrote:
>
>
> Hadoop is having trouble spawning a process to measure how much disk space
> you have available.
>
> It really is much easier to run hadoop on Linux. In case you can't do
> that,
> make sure that you really do have cygwin installed in a standard way on
> each
> hadoop node.
>
>
> On 10/13/07 8:05 AM, "dcave555" <[EMAIL PROTECTED]> wrote:
>
>>
>> Hello all
>>
>> I am new to hadoop .
>>
>> I am trying to write file to single cluster and getting this exception
>> when
>> i am trying to close output stream
>>
>> java.io.IOException: CreateProcess: df -k
>> C:\usr\local\hadoop-datastore\hadoop-hadoop\dfs\tmp error=2
>> at java.lang.ProcessImpl.create(Native Method)
>> at java.lang.ProcessImpl.<init>(Unknown Source)
>> at java.lang.ProcessImpl.start(Unknown Source)
>> at java.lang.ProcessBuilder.start(Unknown Source)
>> at java.lang.Runtime.exec(Unknown Source)
>> at java.lang.Runtime.exec(Unknown Source)
>> at org.apache.hadoop.fs.DF.doDF(DF.java:60)
>> at org.apache.hadoop.fs.DF.<init>(DF.java:53)
>> at
>> org.apache.hadoop.fs.LocalDirAllocator$AllocatorPerContext.confChanged(LocalDi
>> rAllocator.java:198)
>> at
>> org.apache.hadoop.fs.LocalDirAllocator$AllocatorPerContext.getLocalPathForWrit
>> e(LocalDirAllocator.java:235)
>> at
>> org.apache.hadoop.fs.LocalDirAllocator$AllocatorPerContext.createTmpFileForWri
>> te(LocalDirAllocator.java:276)
>> at
>> org.apache.hadoop.fs.LocalDirAllocator.createTmpFileForWrite(LocalDirAllocator
>> .java:155)
>> at
>> org.apache.hadoop.dfs.DFSClient$DFSOutputStream.newBackupFile(DFSClient.java:1
>> 475)
>> at
>> org.apache.hadoop.dfs.DFSClient$DFSOutputStream.openBackupStream(DFSClient.jav
>> a:1442)
>> at
>>
> org.apache.hadoop.dfs.DFSClient$DFSOutputStream.writeChunk(DFSClient.java:1600>
> )
>> at
>>
> org.apache.hadoop.fs.FSOutputSummer.writeChecksumChunk(FSOutputSummer.java:140>
> )
>> at
>> org.apache.hadoop.fs.FSOutputSummer.flushBuffer(FSOutputSummer.java:122)
>> at
>> org.apache.hadoop.dfs.DFSClient$DFSOutputStream.close(DFSClient.java:1739)
>> at
>> org.apache.hadoop.fs.FSDataOutputStream$PositionCache.close(FSDataOutputStream
>> .java:49)
>> at
>> org.apache.hadoop.fs.FSDataOutputStream.close(FSDataOutputStream.java:64)
>> at Test1.main(Test1.java:23)
>>
>>
>>
>> My test is:
>>
>> Configuration configuration = new
>> Configuration();
>> FileSystem fileSystem =
>> FileSystem.get(configuration);
>> Path path = new Path("/testfile");
>> //writing:
>> FSDataOutputStream dataOutputStream = fileSystem.create(path);
>> dataOutputStream.writeUTF("hello world");
>> dataOutputStream.close();
>> //reading
>> FSDataInputStream dataInputStream = fileSystem.open(path);
>> System.out.println(dataInputStream.readUTF());
>> dataInputStream.close();
>> fileSystem.close();
>>
>> i added hadoop-site.xml to classpath :
>>
>>
>> <configuration>
>>
>> <property>
>> <name>hadoop.tmp.dir</name>
>> <value>/usr/local/hadoop-datastore/hadoop-hadoop</value>
>> <description>A base for other temporary directories.</description>
>> </property>
>>
>> <property>
>> <name>fs.default.name</name>
>> <value>hdfs://172.16.50.13:54310</value>
>> <description>The name of the default file system. A URI whose
>> scheme and authority determine the FileSystem implementation. The
>> uri's scheme determines the config property (fs.SCHEME.impl) naming
>> the FileSystem implementation class. The uri's authority is used to
>> determine the host, port, etc. for a filesystem.</description>
>> </property>
>>
>> <property>
>> <name>mapred.job.tracker</name>
>> <value>172.16.50.13:54311</value>
>> <description>The host and port that the MapReduce job tracker runs
>> at. If "local", then jobs are run in-process as a single map
>> and reduce task.
>> </description>
>> </property>
>>
>> <property>
>> <name>dfs.replication</name>
>> <value>1</value>
>> <description>Default block replication.
>> The actual number of replications can be specified when the file is
>> created.
>> The default is used if replication is not specified in create time.
>> </description>
>> </property>
>> </configuration>
>>
>>
>>
>> Please help me
>> thx
>
>
>
--
View this message in context:
http://www.nabble.com/writing-file-tf4618565.html#a13192358
Sent from the Hadoop Users mailing list archive at Nabble.com.