I take steps below:

1. New a project from the exist ant file "build.xml"
2.try to compile the project , its done well.
3.find NameNode.java and configure as a Java App to run.
4.Told me that NameNode not formatted , then i do it with -format argument
5.Then , Exceptions as "webapps" not found in classpath
6.so i try to configure the src/webapps folder as Build->Use as source
folder
7.Build the project again. But i can find the webapps output to
build_output_path
8.Then i just copy the "webapps" to the bin/ path , as my build output path
is Hadoop/bin.
9.Then Exceptions like these:
----------------------------------------------------------------------------------------------------------------------
07/06/22 12:42:22 INFO dfs.StateChange: STATE* Network topology has 0 racks
and 0 datanodes
07/06/22 12:42:22 INFO dfs.StateChange: STATE* UnderReplicatedBlocks has 0
blocks
07/06/22 12:42:22 INFO util.Credential: Checking Resource aliases
07/06/22 12:42:22 INFO http.HttpServer: Version Jetty/5.1.4
07/06/22 12:42:22 INFO util.Container: Started HttpContext[/static,/static]
07/06/22 12:42:23 INFO util.Container: Started
[EMAIL PROTECTED]
07/06/22 12:42:23 INFO http.SocketListener: Started SocketListener on
0.0.0.0:50070
07/06/22 12:42:23 ERROR dfs.NameNode: java.io.IOException: Problem starting
http server
        at
org.apache.hadoop.mapred.StatusHttpServer.start(StatusHttpServer.java:211)
        at org.apache.hadoop.dfs.FSNamesystem.<init>(FSNamesystem.java:274)
        at org.apache.hadoop.dfs.NameNode.init(NameNode.java:178)
        at org.apache.hadoop.dfs.NameNode.<init>(NameNode.java:195)
        at org.apache.hadoop.dfs.NameNode.main(NameNode.java:728)
Caused by: org.mortbay.util.MultiException[java.lang.ClassNotFoundException:
org.apache.hadoop.dfs.dfshealth_jsp, java.lang.ClassNotFoundException:
org.apache.hadoop.dfs.nn_005fbrowsedfscontent_jsp]
        at org.mortbay.http.HttpServer.doStart(HttpServer.java:731)
        at org.mortbay.util.Container.start(Container.java:72)
        at
org.apache.hadoop.mapred.StatusHttpServer.start(StatusHttpServer.java:188)
        ... 4 more
-------------------------------------------------------------------------------------------------------------------------
I tried configure here and there and try and try , but , this exception is
still there.
what the problem this exception might be?

                                                                     Thanks
a lot
                                                     KrzyCube


Konstantin Shvachko wrote:
> 
> I run entire one node cluster in eclipse by just executing main() (run 
> or debug menus) for each component.
> You need to configure eclipse correctly in order to do that. Can you 
> compile the whole thing under eclipse?
> NameNode example:
> = Open NameNode.java in the editor.
> = Run / Run
> = New Java Application -> will create an entry under "Java Application" 
> named NameNode
> = Select NameNode, go to tab Arguments and enter the following arguments 
> under "VM Arguments":
>   -Dhadoop.log.dir=./logs  
>   -Xmx500m
>   -ea
>     The first one is required, can point to your log directory, the 
> other two are optional
> = go to the "Classpath" tab, add "hadoop/build" path under "User entries"
> by
>     Advanced / New Folder / select "hadoop/build"
> That should be it, if the default classpath is configured correctly, and 
> if I am not forgetting anything.
> Let me know if that helped, I 'll send you screenshots of my 
> configuration if not.
> 
> --Konstantin
> 
> 
> Mahajan, Neeraj wrote:
> 
>>There are two sepearete issues you are asking here:
>>1. How to modify/add to haddop code and execute the changed -
>>Eclipse is just an IDE, it doesn't matter whether you use eclipse or
>>some other editor.
>>I have been using eclipse. What I do is modify the code using eclipse
>>and then run "ant jar" in the root folder of hadoop (you could also
>>configure this to work directly from eclipse). This would regenerate the
>>jars and put them in build/ folder. Now you can either copy these jars
>>into hadoop root folder (removing "dev" in their name) so that they
>>replace the original jars or modify the scripts in bin/ to point to the
>>newly generated jars.
>>
>>2. How to debug using a IDE -
>>This page gives a high-level intro to debugging hadoop -
>>http://wiki.apache.org/lucene-hadoop/HowToDebugMapReducePrograms
>>According to me, there are two ways you can debug hadoop programs: Run
>>hadoop in local mode and debug in process in the IDE or run hadoop in
>>distributed mode and remote debug using IDE.
>>
>>The first way is easy. In the bin/hadoop script at the end there is a
>>exec command, instead of that put a echo command and run your program.
>>You can see what the paramters the script passes while starting hadoop.
>>Use these same parameters in the IDE and you can debug hadoop. Remember
>>to make change to the conf files so that hadoop runs in local mode. To
>>be more specific, you will have to set the program arguemnts, VM
>>arguments and add an entry in the classpath pointing to the conf folder.
>>
>>The second method is compilcated. You will have to modify the scripts
>>and put in some extra params like "-Xdebug
>>-Xrunjdwp:transport=dt_socket,server=y,suspend=n,address=<port>" for the
>>java command. Specify the <port> of you choice in it. On the server
>>where you are running both the namenode/jobnode there will be a conflict
>>as the same port would be specified. So you will have to do some
>>intelligent scripting to take care of this. Once the java processes
>>start you can attach eclipse debugger to that machine's <port> and set
>>breakpoints. Till this part you can debug all the things before map
>>reduce tasks. Mapp reduce tasks run in separate process, for debugging
>>them you will have to figure out yourself.
>>
>>The best way is to debug using the first approach (as the above link
>>says). I think by that approach you can fix any map-reduce related
>>problems and for other purely distributed kind of problems you can
>>follow the second approach.
>>
>>~ Neeraj
>>
>>-----Original Message-----
>>From: KrzyCube [mailto:[EMAIL PROTECTED] 
>>Sent: Thursday, June 21, 2007 2:08 AM
>>To: hadoop-user@lucene.apache.org
>>Subject: How to Start Hadoop Cluster from source code in Eclipse
>>
>>
>>Hi,all:
>>
>>I am using Eclipse to View Hadoop source code , and i want to trace to
>>see how it works, I code a few code to call the FSClient  and when i
>>call into the RPC  object, it can not to be deep more .
>>
>>So i just want to start cluster from source code , which i am holding
>>them in Eclipse now. 
>>I browse the start-*.sh , and find that it must start several threads ,
>>such as namenode , datanode,secondnamenode. i just don't know how to
>>figure out.
>>
>>or is there any way to attach my code to a running process , just as the
>>gdb while we are debug c code 
>>
>>Does any body ever use Eclipse to debug these source code , please give
>>some tip.
>>
>> 
>>
>>Thanks .
>> 
>>
>>KrzyCube
>>--
>>View this message in context:
>>http://www.nabble.com/How-to-Start-Hadoop-Cluster-from-source-code-in-Ec
>>lipse-tf3957457.html#a11229322
>>Sent from the Hadoop Users mailing list archive at Nabble.com.
>>
>>  
>>
> 
> 
> 

-- 
View this message in context: 
http://www.nabble.com/How-to-Start-Hadoop-Cluster-from-source-code-in-Eclipse-tf3957457.html#a11246246
Sent from the Hadoop Users mailing list archive at Nabble.com.

Reply via email to