Chackra sent the attachment directly to me. From what I see the cluster resources (memory and cores) are abundant.
But I also see that only 1 app is running which is the one we are trying to debug and 5 containers are running. So definitely more containers that just the AM is running. Can you click on the app master link and copy paste the content of that page? No need for screen shot. Also please send your resources JSON file. -Gour - Sent from my iPhone > On Apr 7, 2015, at 11:01 AM, "Jon Maron" <[email protected]> wrote: > > > On Apr 7, 2015, at 1:36 PM, Chackravarthy Esakkimuthu > <[email protected]<mailto:[email protected]>> wrote: > > @Maron, I could not get the logs even though the application is still running. > It's a 10 node cluster and I logged into one of the node and executed the > command : > > sudo -u hdfs yarn logs -applicationId application_1427882795362_0070 > 15/04/07 22:56:09 INFO impl.TimelineClientImpl: Timeline service address: > http://$HOST:PORT/ws/v1/timeline/ > 15/04/07 22:56:09 INFO client.ConfiguredRMFailoverProxyProvider: Failing over > to rm2 > /app-logs/hdfs/logs/application_1427882795362_0070does not have any log files. > > Can you login to the cluster node and look at the logs directory (e.g. in HDP > install it would be under /hadoop/yarn/logs IIRC)? > > > > @Gour, Please find the attachment. > > On Tue, Apr 7, 2015 at 10:57 PM, Gour Saha > <[email protected]<mailto:[email protected]>> wrote: > Can you take a screenshot of your RM UI and send it over? It is usually > available in a URI similar to http://c6410.ambari.apache.org:8088/cluster. > I am specifically interested in seeing the Cluster Metrics table. > > -Gour > >> On 4/7/15, 10:17 AM, "Jon Maron" >> <[email protected]<mailto:[email protected]>> wrote: >> >> >>> On Apr 7, 2015, at 1:14 PM, Jon Maron >>> <[email protected]<mailto:[email protected]>> wrote: >>> >>> >>>> On Apr 7, 2015, at 1:08 PM, Chackravarthy Esakkimuthu >>>> <[email protected]<mailto:[email protected]>> wrote: >>>> >>>> Thanks for the reply guys! >>>> Contianer allocation happened successfully. >>>> >>>> *RoleStatus{name='slider-appmaster', key=0, minimum=0, maximum=1, >>>> desired=1, actual=1,* >>>> *RoleStatus{name='STORM_UI_SERVER', key=2, minimum=0, maximum=1, >>>> desired=1, >>>> actual=1, * >>>> *RoleStatus{name='NIMBUS', key=1, minimum=0, maximum=1, desired=1, >>>> actual=1, * >>>> *RoleStatus{name='DRPC_SERVER', key=3, minimum=0, maximum=1, desired=1, >>>> actual=1, * >>>> *RoleStatus{name='SUPERVISOR', key=4, minimum=0, maximum=1, desired=1, >>>> actual=1,* >>>> >>>> Also, have put some logs specific to a container.. (nimbus) Same set of >>>> logs available for other Roles also (except Supervisor which has only >>>> first >>>> 2 lines of below logs) >>>> >>>> *Installing NIMBUS on container_e04_1427882795362_0070_01_000002.* >>>> *Starting NIMBUS on container_e04_1427882795362_0070_01_000002.* >>>> *Registering component container_e04_1427882795362_0070_01_000002* >>>> *Requesting applied config for NIMBUS on >>>> container_e04_1427882795362_0070_01_000002.* >>>> *Received and processed config for >>>> container_e04_1427882795362_0070_01_000002___NIMBUS* >>>> >>>> Does this result in any intermediate state? >>>> >>>> @Maron, I didn't configure any port specifically.. do I need to to? >>>> Also, i >>>> don't see any error msg in AM logs wrt port conflict. >>> >>> My only concern was whether you were actually accession the web UIs at >>> the correct host and port. If you are then the next step is probably to >>> look at the actual storm/hbase logs. you can use the ³yarn logs >>> -applicationid ..² command. >> >> *accessing* ;) >> >>> >>>> >>>> Thanks, >>>> Chackra >>>> >>>> >>>> >>>> On Tue, Apr 7, 2015 at 9:02 PM, Jon Maron >>>> <[email protected]<mailto:[email protected]>> >>>> wrote: >>>> >>>>> >>>>>> On Apr 7, 2015, at 11:03 AM, Billie Rinaldi >>>>>> <[email protected]<mailto:[email protected]>> >>>>> wrote: >>>>>> >>>>>> One thing you can check is whether your system has enough resources >>>>>> to >>>>>> allocate all the containers the app needs. You will see info like >>>>>> the >>>>>> following in the AM log (it will be logged multiple times over the >>>>>> life >>>>> of >>>>>> the AM). In this case, the master I requested was allocated but the >>>>>> tservers were not. >>>>>> RoleStatus{name='ACCUMULO_TSERVER', key=2, desired=2, actual=0, >>>>>> requested=2, releasing=0, failed=0, started=0, startFailed=0, >>>>> completed=0, >>>>>> failureMessage=''} >>>>>> RoleStatus{name='ACCUMULO_MASTER', key=1, desired=1, actual=1, >>>>> requested=0, >>>>>> releasing=0, failed=0, started=0, startFailed=0, completed=0, >>>>>> failureMessage=Œ'} >>>>> >>>>> You can also check the ³Scheduler² link on the RM Web UI to get a >>>>> sense of >>>>> whether you are resource constrained. >>>>> >>>>> Are you certain that you are attempting to invoke the correct port? >>>>> The >>>>> listening ports are dynamically allocated by Slider. >>>>> >>>>>> >>>>>> >>>>>> On Tue, Apr 7, 2015 at 3:29 AM, Chackravarthy Esakkimuthu < >>>>>> [email protected]<mailto:[email protected]>> wrote: >>>>>> >>>>>>> Hi All, >>>>>>> >>>>>>> I am new to Apache slider and would like to contribute. >>>>>>> >>>>>>> Just to start with, I am trying out running "storm" and "hbase" on >>>>>>> yarn >>>>>>> using slider following the guide : >>>>> >>>>> http://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.2.0/YARN_RM_v22/run >>>>> ning_applications_on_slider/index.html#Item1.1 >>>>>>> >>>>>>> In both (storm and hbase) the cases, the ApplicationMaster gets >>>>>>> launched >>>>>>> and still running, but the ApplicationMaster link not working, and >>>>>>> from >>>>> AM >>>>>>> logs, I don't see any errors. >>>>>>> >>>>>>> How do I debug from this? Please help me. >>>>>>> Incase if there is any other mail thread with respect this, please >>>>>>> point >>>>>>> out to me. Thanks in advance. >>>>>>> >>>>>>> Thanks, >>>>>>> Chackra > > >
