I think you should make sure that the HDFS is running normally by checking its log files.
And, firewalls? (just wild guesses) -Kevin On 5月27日, 下午2时20分, Harshada <[email protected]> wrote: > Thank you for the reply. > > First of all I am sorry for posting this query to the wrong thread. It > you can, please migrate it to the -user mailing list. > > I checked the log file for DfsBroker.hadoop, it says: > > Num CPUs=2 > HdfsBroker.Port=38030 > HdfsBroker.Reactors=2 > HdfsBroker.Workers=20 > HdfsBroker.Server.fs.default.name=hdfs://localhost:54310 > 10/05/27 05:01:58 INFO ipc.Client: Retrying connect to server: > localhost/127.0.0.1:54310. Already tried 0 time(s). > 10/05/27 05:01:59 INFO ipc.Client: Retrying connect to server: > localhost/127.0.0.1:54310. Already tried 1 time(s). > 10/05/27 05:02:00 INFO ipc.Client: Retrying connect to server: > localhost/127.0.0.1:54310. Already tried 2 time(s). > 10/05/27 05:02:01 INFO ipc.Client: Retrying connect to server: > localhost/127.0.0.1:54310. Already tried 3 time(s). > 10/05/27 05:02:02 INFO ipc.Client: Retrying connect to server: > localhost/127.0.0.1:54310. Already tried 4 time(s). > 10/05/27 05:02:03 INFO ipc.Client: Retrying connect to server: > localhost/127.0.0.1:54310. Already tried 5 time(s). > 10/05/27 05:02:04 INFO ipc.Client: Retrying connect to server: > localhost/127.0.0.1:54310. Already tried 6 time(s). > 10/05/27 05:02:05 INFO ipc.Client: Retrying connect to server: > localhost/127.0.0.1:54310. Already tried 7 time(s). > 10/05/27 05:02:06 INFO ipc.Client: Retrying connect to server: > localhost/127.0.0.1:54310. Already tried 8 time(s). > 10/05/27 05:02:07 INFO ipc.Client: Retrying connect to server: > localhost/127.0.0.1:54310. Already tried 9 time(s). > 27 May, 2010 5:02:07 AM org.hypertable.DfsBroker.hadoop.HdfsBroker > <init> > SEVERE: ERROR: Unable to establish connection to HDFS. > ShutdownHook called > Exception in thread "Thread-1" java.lang.NullPointerException > at org.hypertable.DfsBroker.hadoop.main > $ShutdownHook.run(main.java:69) > > --------------------------- > > but hdfs is running. because jps on master gives me following output: > > e...@erts-server:~$ jps > 32538 SecondaryNameNode > 32270 NameNode > 32388 DataNode > 310 TaskTracker > 32671 JobTracker > 21233 Jps > ----------------------------------------------------- > > > Is there a reason you're using 0.9.2.8 and not 0.9.3.1 (the latest and > > greatest) ? > > Oh.. ok. thanks for the info. But since 0.9.2.8 was successfully > installed, I ll continue with it for the moment. > > > Do you have HDFS running and if so make sure the permissions for the > > /hypertable dir are set correctly. > > yes. I followedhttp://code.google.com/p/hypertable/wiki/UpAndRunningWithHadoop > andhttp://code.google.com/p/hypertable/wiki/DeployingHypertable. > > A doubt: do I always need the owner of slave and master machine to be > the same? coz currently I have 'erts' as the user for master and one > slave (which are on the same machine) and harshada as the user on > slave. So what happens is, whenever I use '$cap dist' or '$cap shell > cap>date' it asks for password of e...@slave which does not exist > hence authentication fails. I am in the process of getting same user > on all the machines, but till then I thought of getting max things up. > Is this the reason why DfsBroker.hadoop is also failing? > > If yes, then I should better wait and have same user on all the > machines. > > PS: though hadoop require same installation paths on all the machines, > I managed it with symbolic links, though my users (and hence their > $HOME s ) were different. > > > > > Beyond that, try taking a look at <HT_INSTALL_DIR>/log/DfsBroker.hadoop.log > > to figure out whats going on. > > > -Sanjit > > > On Wed, May 26, 2010 at 4:35 PM, Harshada <[email protected]> wrote: > > > Hi, > > > > I am installing Hypertable 0.9.2.8 on Hadoop. I have successfully set > > > up Hadoop and its working. When I start servers using 'cap start', DFS > > > Server doesn't come up. The output of cap start is: > > > > e...@erts-server:~/hypertable/hypertable-0.9.2.8-alpha/conf$ cap start > > > * executing `start' > > > ** transaction: start > > > * executing `start_hyperspace' > > > * executing "/home/erts/hypertable/hypertable-0.9.2.8-alpha/0.9.2.8/ > > > bin/start-hyperspace.sh <http://0.9.2.8/%0Abin/start-hyperspace.sh> > > > --config=/home/erts/hypertable/ > > > hypertable-0.9.2.8-alpha/0.9.2.8/conf/hypertable.cfg" > > > servers: ["127.0.0.1"] > > > [127.0.0.1] executing command > > > ** [out :: 127.0.0.1] Hyperspace appears to be running (12170): > > > ** [out :: 127.0.0.1] erts 12170 1 0 04:40 ? 00:00:00 /home/erts/ > > > hypertable/hypertable-0.9.2.8-alpha/0.9.2.8/bin/Hyperspace.Master -- > > > pidfile /home/erts/hypertable/hypertable-0.9.2.8-alpha/0.9.2.8/run/ > > > Hyperspace.pid --verbose --config=/home/erts/hypertable/ > > > hypertable-0.9.2.8-alpha/0.9.2.8/conf/hypertable.cfg > > > command finished > > > * executing `start_master' > > > * executing "/home/erts/hypertable/hypertable-0.9.2.8-alpha/0.9.2.8/ > > > bin/start-dfsbroker.sh <http://0.9.2.8/%0Abin/start-dfsbroker.sh> hadoop > > > --config=/home/erts/hypertable/ > > > hypertable-0.9.2.8-alpha/0.9.2.8/conf/hypertable.cfg &&\\\n /home/ > > > erts/hypertable/hypertable-0.9.2.8-alpha/0.9.2.8/bin/start-master.sh -- > > > config=/home/erts/hypertable/hypertable-0.9.2.8-alpha/0.9.2.8/conf/ > > > hypertable.cfg" > > > servers: ["127.0.0.1"] > > > [127.0.0.1] executing command > > > ** [out :: 127.0.0.1] DFS broker: available file descriptors: 1024 > > > ** [out :: 127.0.0.1] Waiting for DFS Broker (hadoop) to come up... > > > ** [out :: 127.0.0.1] Waiting for DFS Broker (hadoop) to come up... > > > ** [out :: 127.0.0.1] Waiting for DFS Broker (hadoop) to come up... > > > ** [out :: 127.0.0.1] Waiting for DFS Broker (hadoop) to come up... > > > ** [out :: 127.0.0.1] Waiting for DFS Broker (hadoop) to come up... > > > ** [out :: 127.0.0.1] Waiting for DFS Broker (hadoop) to come up... > > > ** [out :: 127.0.0.1] Waiting for DFS Broker (hadoop) to come up... > > > ** [out :: 127.0.0.1] Waiting for DFS Broker (hadoop) to come up... > > > ** [out :: 127.0.0.1] ERROR: DFS Broker (hadoop) did not come up > > > command finished > > > failed: "sh -c '/home/erts/hypertable/hypertable-0.9.2.8-alpha/0.9.2.8/ > > > bin/start-dfsbroker.sh <http://0.9.2.8/%0Abin/start-dfsbroker.sh> hadoop > > > --config=/home/erts/hypertable/ > > > hypertable-0.9.2.8-alpha/0.9.2.8/conf/hypertable.cfg &&\\\n /home/ > > > erts/hypertable/hypertable-0.9.2.8-alpha/0.9.2.8/bin/start-master.sh -- > > > config=/home/erts/hypertable/hypertable-0.9.2.8-alpha/0.9.2.8/conf/ > > > hypertable.cfg'" on 127.0.0.1 > > > > --------------------------------------------------------------- > > > > Here, 127.0.0.1 is my master. > > > > My hypertable.cfg looks like: > > > > # > > > # hypertable.cfg > > > # > > > > # HDFS Broker > > > HdfsBroker.Port=38030 > > > HdfsBroker.fs.default.name=hdfs://localhost:54310 > > > HdfsBroker.Workers=20 > > > > # Ceph Broker > > > CephBroker.Port=38030 > > > CephBroker.Workers=20 > > > CephBroker.MonAddr=10.0.1.245:6789 > > > > # Local Broker > > > DfsBroker.Local.Port=38030 > > > DfsBroker.Local.Root=fs/local > > > > # DFS Broker - for clients > > > DfsBroker.Host=localhost > > > DfsBroker.Port=38030 > > > > # Hyperspace > > > Hyperspace.Replica.Host=localhost > > > Hyperspace.Replica.Port=38040 > > > Hyperspace.Replica.Dir=hyperspace > > > Hyperspace.Replica.Workers=20 > > > > # Hypertable.Master > > > Hypertable.Master.Host=localhost > > > Hypertable.Master.Port=38050 > > > Hypertable.Master.Workers=20 > > > > # Hypertable.RangeServer > > > Hypertable.RangeServer.Port=38060 > > > > Hyperspace.KeepAlive.Interval=30000 > > > Hyperspace.Lease.Interval=1000000 > > > Hyperspace.GracePeriod=200000 > > > > # ThriftBroker > > > ThriftBroker.Port=38080 > > > ------------------------------------------ > > > > Note: It does not have Hyperspace.Master.Host=localhost property. > > > > Capfile: > > > > set :source_machine, "127.0.0.1" > > > set :install_dir, "/home/erts/hypertable/hypertable-0.9.2.8-alpha" > > > set :hypertable_version, "0.9.2.8" > > > set :default_dfs, "hadoop" > > > set :default_config, "/home/erts/hypertable/hypertable.cfg" > > > > role :master, "127.0.0.1" > > > role :hyperspace, "127.0.0.1" > > > role :slave, "127.0.0.1", "10.129.125.12" > > > role :localhost, "127.0.0.1" > > > ------------------------------------ > > > > Any idea why DFS Broker is failing? > > > > Thanks, > > > Harshada > > > > -- > > > You received this message because you are subscribed to the Google Groups > > > "Hypertable Development" group. > > > To post to this group, send email to [email protected]. > > > To unsubscribe from this group, send email to > > > [email protected]<hypertable-dev%2bunsubscr...@go > > > oglegroups.com> > > > . > > > For more options, visit this group at > > >http://groups.google.com/group/hypertable-dev?hl=en. -- You received this message because you are subscribed to the Google Groups "Hypertable Development" group. To post to this group, send email to [email protected]. To unsubscribe from this group, send email to [email protected]. For more options, visit this group at http://groups.google.com/group/hypertable-dev?hl=en.
