Ok

I am running my Test on Eclipse IDE

tomorow at work  i will try to copy pulic key to output directory and then
try again to run my application

maybe it will work.

Thx for your reply.




Aaron Kimball wrote:
> 
> That depends on your ssh implementation. Most will look at the file 
> ~/.ssh/id_dsa.pub or id_rsa.pub in the same directory. You can generate 
> your keys on the linux machine ('man ssh-keygen' for more info), which 
> will generate files named id_Xsa and id_Xsa.pub (Depending on the key 
> type). The private key file goes in ~/.ssh/ on the Linux machine. The 
> public key file (should also be copied there), and goes in the specified 
> directory on your Windows machine depending on what ssh client you're 
> using. (If you're using cygwin's, ~/.ssh/ should be the correct one.)
> 
> If you have a sane cygwin install, 'man ssh' for more.
> 
> - Aaron
> 
> dcave555 wrote:
>> Thx again for your reply
>> 
>> Ok maybe you right
>> 
>> I have my ubuntu hadoop node on ip 168.13.51.13
>> 
>> So i write in hadoop-site.xml (windows application)
>> 
>> <configuration>  
>>  <property>
>>    <name>hadoop.tmp.dir</name>
>>    <value>//168.13.51.13/usr/local/hadoop-datastore/hadoop-hadoop</value>
>>    <description>A base for other temporary directories.</description>
>>  </property>
>>  
>>  <property>
>>    <name>fs.default.name</name>
>>    <value>hdfs://168.13.51.13:54310</value>  
>>  </property>
>>   
>>  <property>
>>    <name>mapred.job.tracker</name>
>>    <value>168.13.51.13:54311</value>
>>  </property>
>>   
>>    <property>
>>       <name>dfs.replication</name>
>>       <value>1</value>
>>    </property>
>> </configuration>
>> 
>> 
>> The same xml i have on ubuntu machine.
>> 
>> 
>> after running my application i have exception:
>> 
>> Login failded.
>> 
>> i suspect that i have problem with ssh public key
>> 
>> so i think i need to copy ssh public key to windows machine
>> 
>> but how can i tell to my application to use this key?
>> 
>> Thx again
>> 
>> 
>> Ted Dunning-3 wrote:
>>>
>>> Your program may be trying to run locally on your windows box instead of
>>> talking to the ubuntu micro cluster.
>>>
>>>
>>> On 10/13/07 11:43 AM, "dcave555" <[EMAIL PROTECTED]> wrote:
>>>
>>>> Thx for your reply
>>>>
>>>> But i think i don't understand something.
>>>> i have installed hadoop single node on ubuntu.
>>>> my test is running on windows.
>>>>
>>>> i changed property in hadoop-site.xml in my application to
>>>>
>>>> <property>
>>>>    <name>hadoop.tmp.dir</name>
>>>>   
>>>> <value>//172.16.50.13/usr/local/hadoop-datastore/hadoop-hadoop</value>
>>>>    <description>A base for other temporary directories.</description>
>>>> </property>
>>>>
>>>> amd i am getting Login failed.
>>>>
>>>> If i want to save file from windows to hadoop on ubuntu i need to
>>>> import
>>>> ssh
>>>> public key?
>>>> how can i tell my test application on windows to use this pulic key?
>>>>
>>>> Or i need to do something else?
>>>>
>>>> Thx
>>>>
>>>>
>>>>
>>>>
>>>>
>>>> Ted Dunning-3 wrote:
>>>>>
>>>>> Hadoop is having trouble spawning a process to measure how much disk
>>>>> space
>>>>> you have available.
>>>>>
>>>>> It really is much easier to run hadoop on Linux.  In case you can't do
>>>>> that,
>>>>> make sure that you really do have cygwin installed in a standard way
>>>>> on
>>>>> each
>>>>> hadoop node.
>>>>>
>>>>>
>>>>> On 10/13/07 8:05 AM, "dcave555" <[EMAIL PROTECTED]> wrote:
>>>>>
>>>>>> Hello all
>>>>>>
>>>>>> I am new to hadoop .
>>>>>>
>>>>>> I am trying to write file to single cluster and getting this
>>>>>> exception
>>>>>> when
>>>>>> i am trying to close output stream
>>>>>>
>>>>>> java.io.IOException: CreateProcess: df -k
>>>>>> C:\usr\local\hadoop-datastore\hadoop-hadoop\dfs\tmp error=2
>>>>>> at java.lang.ProcessImpl.create(Native Method)
>>>>>> at java.lang.ProcessImpl.<init>(Unknown Source)
>>>>>> at java.lang.ProcessImpl.start(Unknown Source)
>>>>>> at java.lang.ProcessBuilder.start(Unknown Source)
>>>>>> at java.lang.Runtime.exec(Unknown Source)
>>>>>> at java.lang.Runtime.exec(Unknown Source)
>>>>>> at org.apache.hadoop.fs.DF.doDF(DF.java:60)
>>>>>> at org.apache.hadoop.fs.DF.<init>(DF.java:53)
>>>>>> at
>>>>>> org.apache.hadoop.fs.LocalDirAllocator$AllocatorPerContext.confChanged(Local
>>>>>> Di
>>>>>> rAllocator.java:198)
>>>>>> at
>>>>>> org.apache.hadoop.fs.LocalDirAllocator$AllocatorPerContext.getLocalPathForWr
>>>>>> it
>>>>>> e(LocalDirAllocator.java:235)
>>>>>> at
>>>>>> org.apache.hadoop.fs.LocalDirAllocator$AllocatorPerContext.createTmpFileForW
>>>>>> ri
>>>>>> te(LocalDirAllocator.java:276)
>>>>>> at
>>>>>> org.apache.hadoop.fs.LocalDirAllocator.createTmpFileForWrite(LocalDirAllocat
>>>>>> or
>>>>>> .java:155)
>>>>>> at
>>>>>> org.apache.hadoop.dfs.DFSClient$DFSOutputStream.newBackupFile(DFSClient.java
>>>>>> :1
>>>>>> 475)
>>>>>> at
>>>>>> org.apache.hadoop.dfs.DFSClient$DFSOutputStream.openBackupStream(DFSClient.j
>>>>>> av
>>>>>> a:1442)
>>>>>> at
>>>>>>
>>>>> org.apache.hadoop.dfs.DFSClient$DFSOutputStream.writeChunk(DFSClient.java:160
>>>>> 0>
>>>>> )
>>>>>> at
>>>>>>
>>>>> org.apache.hadoop.fs.FSOutputSummer.writeChecksumChunk(FSOutputSummer.java:14
>>>>> 0>
>>>>> )
>>>>>> at
>>>>>> org.apache.hadoop.fs.FSOutputSummer.flushBuffer(FSOutputSummer.java:122)
>>>>>> at
>>>>>> org.apache.hadoop.dfs.DFSClient$DFSOutputStream.close(DFSClient.java:1739)
>>>>>> at
>>>>>> org.apache.hadoop.fs.FSDataOutputStream$PositionCache.close(FSDataOutputStre
>>>>>> am
>>>>>> .java:49)
>>>>>> at
>>>>>> org.apache.hadoop.fs.FSDataOutputStream.close(FSDataOutputStream.java:64)
>>>>>> at Test1.main(Test1.java:23)
>>>>>>
>>>>>>
>>>>>>
>>>>>> My test is:
>>>>>>
>>>>>>                         Configuration configuration = new
>>>>>> Configuration();
>>>>>>                         FileSystem fileSystem =
>>>>>> FileSystem.get(configuration);
>>>>>>                         Path path = new Path("/testfile");
>>>>>> //writing:
>>>>>> FSDataOutputStream dataOutputStream = fileSystem.create(path);
>>>>>> dataOutputStream.writeUTF("hello world");
>>>>>> dataOutputStream.close();
>>>>>> //reading 
>>>>>> FSDataInputStream dataInputStream = fileSystem.open(path);
>>>>>> System.out.println(dataInputStream.readUTF());
>>>>>> dataInputStream.close();
>>>>>> fileSystem.close();
>>>>>>
>>>>>> i added hadoop-site.xml to classpath :
>>>>>>
>>>>>>
>>>>>> <configuration>
>>>>>>  
>>>>>> <property>
>>>>>>   <name>hadoop.tmp.dir</name>
>>>>>>   <value>/usr/local/hadoop-datastore/hadoop-hadoop</value>
>>>>>>   <description>A base for other temporary directories.</description>
>>>>>> </property>
>>>>>>
>>>>>> <property>
>>>>>>   <name>fs.default.name</name>
>>>>>>   <value>hdfs://172.16.50.13:54310</value>
>>>>>>   <description>The name of the default file system.  A URI whose
>>>>>>   scheme and authority determine the FileSystem implementation.  The
>>>>>>   uri's scheme determines the config property (fs.SCHEME.impl) naming
>>>>>>   the FileSystem implementation class.  The uri's authority is used
>>>>>> to
>>>>>>   determine the host, port, etc. for a filesystem.</description>
>>>>>> </property>
>>>>>>  
>>>>>> <property>
>>>>>>   <name>mapred.job.tracker</name>
>>>>>>   <value>172.16.50.13:54311</value>
>>>>>>   <description>The host and port that the MapReduce job tracker runs
>>>>>>   at.  If "local", then jobs are run in-process as a single map
>>>>>>   and reduce task.
>>>>>>   </description>
>>>>>> </property>
>>>>>>  
>>>>>> <property>
>>>>>>   <name>dfs.replication</name>
>>>>>>   <value>1</value>
>>>>>>   <description>Default block replication.
>>>>>>   The actual number of replications can be specified when the file is
>>>>>> created.
>>>>>>   The default is used if replication is not specified in create time.
>>>>>>   </description>
>>>>>> </property>
>>>>>> </configuration>
>>>>>>
>>>>>>
>>>>>>
>>>>>> Please help me
>>>>>> thx
>>>>>
>>>>>
>>>
>>>
>> 
> 
> 

-- 
View this message in context: 
http://www.nabble.com/writing-file-tf4618565.html#a13194049
Sent from the Hadoop Users mailing list archive at Nabble.com.

Reply via email to