Can you take a screenshot of your RM UI and send it over? It is usually available in a URI similar to http://c6410.ambari.apache.org:8088/cluster. I am specifically interested in seeing the Cluster Metrics table.
-Gour On 4/7/15, 10:17 AM, "Jon Maron" <[email protected]> wrote: > >> On Apr 7, 2015, at 1:14 PM, Jon Maron <[email protected]> wrote: >> >> >>> On Apr 7, 2015, at 1:08 PM, Chackravarthy Esakkimuthu >>><[email protected]> wrote: >>> >>> Thanks for the reply guys! >>> Contianer allocation happened successfully. >>> >>> *RoleStatus{name='slider-appmaster', key=0, minimum=0, maximum=1, >>> desired=1, actual=1,* >>> *RoleStatus{name='STORM_UI_SERVER', key=2, minimum=0, maximum=1, >>>desired=1, >>> actual=1, * >>> *RoleStatus{name='NIMBUS', key=1, minimum=0, maximum=1, desired=1, >>> actual=1, * >>> *RoleStatus{name='DRPC_SERVER', key=3, minimum=0, maximum=1, desired=1, >>> actual=1, * >>> *RoleStatus{name='SUPERVISOR', key=4, minimum=0, maximum=1, desired=1, >>> actual=1,* >>> >>> Also, have put some logs specific to a container.. (nimbus) Same set of >>> logs available for other Roles also (except Supervisor which has only >>>first >>> 2 lines of below logs) >>> >>> *Installing NIMBUS on container_e04_1427882795362_0070_01_000002.* >>> *Starting NIMBUS on container_e04_1427882795362_0070_01_000002.* >>> *Registering component container_e04_1427882795362_0070_01_000002* >>> *Requesting applied config for NIMBUS on >>> container_e04_1427882795362_0070_01_000002.* >>> *Received and processed config for >>> container_e04_1427882795362_0070_01_000002___NIMBUS* >>> >>> Does this result in any intermediate state? >>> >>> @Maron, I didn't configure any port specifically.. do I need to to? >>>Also, i >>> don't see any error msg in AM logs wrt port conflict. >> >> My only concern was whether you were actually accession the web UIs at >>the correct host and port. If you are then the next step is probably to >>look at the actual storm/hbase logs. you can use the ³yarn logs >>-applicationid ..² command. > >*accessing* ;) > >> >>> >>> Thanks, >>> Chackra >>> >>> >>> >>> On Tue, Apr 7, 2015 at 9:02 PM, Jon Maron <[email protected]> >>>wrote: >>> >>>> >>>>> On Apr 7, 2015, at 11:03 AM, Billie Rinaldi >>>>><[email protected]> >>>> wrote: >>>>> >>>>> One thing you can check is whether your system has enough resources >>>>>to >>>>> allocate all the containers the app needs. You will see info like >>>>>the >>>>> following in the AM log (it will be logged multiple times over the >>>>>life >>>> of >>>>> the AM). In this case, the master I requested was allocated but the >>>>> tservers were not. >>>>> RoleStatus{name='ACCUMULO_TSERVER', key=2, desired=2, actual=0, >>>>> requested=2, releasing=0, failed=0, started=0, startFailed=0, >>>> completed=0, >>>>> failureMessage=''} >>>>> RoleStatus{name='ACCUMULO_MASTER', key=1, desired=1, actual=1, >>>> requested=0, >>>>> releasing=0, failed=0, started=0, startFailed=0, completed=0, >>>>> failureMessage=Œ'} >>>> >>>> You can also check the ³Scheduler² link on the RM Web UI to get a >>>>sense of >>>> whether you are resource constrained. >>>> >>>> Are you certain that you are attempting to invoke the correct port? >>>>The >>>> listening ports are dynamically allocated by Slider. >>>> >>>>> >>>>> >>>>> On Tue, Apr 7, 2015 at 3:29 AM, Chackravarthy Esakkimuthu < >>>>> [email protected]> wrote: >>>>> >>>>>> Hi All, >>>>>> >>>>>> I am new to Apache slider and would like to contribute. >>>>>> >>>>>> Just to start with, I am trying out running "storm" and "hbase" on >>>>>>yarn >>>>>> using slider following the guide : >>>>>> >>>>>> >>>>>> >>>> >>>>http://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.2.0/YARN_RM_v22/run >>>>ning_applications_on_slider/index.html#Item1.1 >>>>>> >>>>>> In both (storm and hbase) the cases, the ApplicationMaster gets >>>>>>launched >>>>>> and still running, but the ApplicationMaster link not working, and >>>>>>from >>>> AM >>>>>> logs, I don't see any errors. >>>>>> >>>>>> How do I debug from this? Please help me. >>>>>> Incase if there is any other mail thread with respect this, please >>>>>>point >>>>>> out to me. Thanks in advance. >>>>>> >>>>>> Thanks, >>>>>> Chackra >>>>>> >>>> >>>> >> >
