Since 5 containers are running, which means that Storm daemons are already
up and running?
Actually the ApplicationMaster link is not working. It just blanks out
printing the following :
This is standby RM. Redirecting to the current active RM:
http://<host-name>:8088/proxy/application_1427882795362_0070/slideram
And for resources.json, I dint make any change and used the copy of
resources-default.json as follows:
{
"schema" : "http://example.org/specification/v2.0.0",
"metadata" : {
},
"global" : {
"yarn.log.include.patterns": "",
"yarn.log.exclude.patterns": ""
},
"components": {
"slider-appmaster": {
"yarn.memory": "512"
},
"NIMBUS": {
"yarn.role.priority": "1",
"yarn.component.instances": "1",
"yarn.memory": "2048"
},
"STORM_UI_SERVER": {
"yarn.role.priority": "2",
"yarn.component.instances": "1",
"yarn.memory": "1278"
},
"DRPC_SERVER": {
"yarn.role.priority": "3",
"yarn.component.instances": "1",
"yarn.memory": "1278"
},
"SUPERVISOR": {
"yarn.role.priority": "4",
"yarn.component.instances": "1",
"yarn.memory": "3072"
}
}
}
On Tue, Apr 7, 2015 at 11:52 PM, Gour Saha <[email protected]> wrote:
> Chackra sent the attachment directly to me. From what I see the cluster
> resources (memory and cores) are abundant.
>
> But I also see that only 1 app is running which is the one we are trying
> to debug and 5 containers are running. So definitely more containers that
> just the AM is running.
>
> Can you click on the app master link and copy paste the content of that
> page? No need for screen shot. Also please send your resources JSON file.
>
> -Gour
>
> - Sent from my iPhone
>
> > On Apr 7, 2015, at 11:01 AM, "Jon Maron" <[email protected]> wrote:
> >
> >
> > On Apr 7, 2015, at 1:36 PM, Chackravarthy Esakkimuthu <
> [email protected]<mailto:[email protected]>> wrote:
> >
> > @Maron, I could not get the logs even though the application is still
> running.
> > It's a 10 node cluster and I logged into one of the node and executed
> the command :
> >
> > sudo -u hdfs yarn logs -applicationId application_1427882795362_0070
> > 15/04/07 22:56:09 INFO impl.TimelineClientImpl: Timeline service
> address: http://$HOST:PORT/ws/v1/timeline/
> > 15/04/07 22:56:09 INFO client.ConfiguredRMFailoverProxyProvider: Failing
> over to rm2
> > /app-logs/hdfs/logs/application_1427882795362_0070does not have any log
> files.
> >
> > Can you login to the cluster node and look at the logs directory (e.g.
> in HDP install it would be under /hadoop/yarn/logs IIRC)?
> >
> >
> >
> > @Gour, Please find the attachment.
> >
> > On Tue, Apr 7, 2015 at 10:57 PM, Gour Saha <[email protected]
> <mailto:[email protected]>> wrote:
> > Can you take a screenshot of your RM UI and send it over? It is usually
> > available in a URI similar to
> http://c6410.ambari.apache.org:8088/cluster.
> > I am specifically interested in seeing the Cluster Metrics table.
> >
> > -Gour
> >
> >> On 4/7/15, 10:17 AM, "Jon Maron" <[email protected]<mailto:
> [email protected]>> wrote:
> >>
> >>
> >>> On Apr 7, 2015, at 1:14 PM, Jon Maron <[email protected]<mailto:
> [email protected]>> wrote:
> >>>
> >>>
> >>>> On Apr 7, 2015, at 1:08 PM, Chackravarthy Esakkimuthu
> >>>> <[email protected]<mailto:[email protected]>> wrote:
> >>>>
> >>>> Thanks for the reply guys!
> >>>> Contianer allocation happened successfully.
> >>>>
> >>>> *RoleStatus{name='slider-appmaster', key=0, minimum=0, maximum=1,
> >>>> desired=1, actual=1,*
> >>>> *RoleStatus{name='STORM_UI_SERVER', key=2, minimum=0, maximum=1,
> >>>> desired=1,
> >>>> actual=1, *
> >>>> *RoleStatus{name='NIMBUS', key=1, minimum=0, maximum=1, desired=1,
> >>>> actual=1, *
> >>>> *RoleStatus{name='DRPC_SERVER', key=3, minimum=0, maximum=1,
> desired=1,
> >>>> actual=1, *
> >>>> *RoleStatus{name='SUPERVISOR', key=4, minimum=0, maximum=1, desired=1,
> >>>> actual=1,*
> >>>>
> >>>> Also, have put some logs specific to a container.. (nimbus) Same set
> of
> >>>> logs available for other Roles also (except Supervisor which has only
> >>>> first
> >>>> 2 lines of below logs)
> >>>>
> >>>> *Installing NIMBUS on container_e04_1427882795362_0070_01_000002.*
> >>>> *Starting NIMBUS on container_e04_1427882795362_0070_01_000002.*
> >>>> *Registering component container_e04_1427882795362_0070_01_000002*
> >>>> *Requesting applied config for NIMBUS on
> >>>> container_e04_1427882795362_0070_01_000002.*
> >>>> *Received and processed config for
> >>>> container_e04_1427882795362_0070_01_000002___NIMBUS*
> >>>>
> >>>> Does this result in any intermediate state?
> >>>>
> >>>> @Maron, I didn't configure any port specifically.. do I need to to?
> >>>> Also, i
> >>>> don't see any error msg in AM logs wrt port conflict.
> >>>
> >>> My only concern was whether you were actually accession the web UIs at
> >>> the correct host and port. If you are then the next step is probably
> to
> >>> look at the actual storm/hbase logs. you can use the ³yarn logs
> >>> -applicationid ..² command.
> >>
> >> *accessing* ;)
> >>
> >>>
> >>>>
> >>>> Thanks,
> >>>> Chackra
> >>>>
> >>>>
> >>>>
> >>>> On Tue, Apr 7, 2015 at 9:02 PM, Jon Maron <[email protected]
> <mailto:[email protected]>>
> >>>> wrote:
> >>>>
> >>>>>
> >>>>>> On Apr 7, 2015, at 11:03 AM, Billie Rinaldi
> >>>>>> <[email protected]<mailto:[email protected]>>
> >>>>> wrote:
> >>>>>>
> >>>>>> One thing you can check is whether your system has enough resources
> >>>>>> to
> >>>>>> allocate all the containers the app needs. You will see info like
> >>>>>> the
> >>>>>> following in the AM log (it will be logged multiple times over the
> >>>>>> life
> >>>>> of
> >>>>>> the AM). In this case, the master I requested was allocated but the
> >>>>>> tservers were not.
> >>>>>> RoleStatus{name='ACCUMULO_TSERVER', key=2, desired=2, actual=0,
> >>>>>> requested=2, releasing=0, failed=0, started=0, startFailed=0,
> >>>>> completed=0,
> >>>>>> failureMessage=''}
> >>>>>> RoleStatus{name='ACCUMULO_MASTER', key=1, desired=1, actual=1,
> >>>>> requested=0,
> >>>>>> releasing=0, failed=0, started=0, startFailed=0, completed=0,
> >>>>>> failureMessage=Œ'}
> >>>>>
> >>>>> You can also check the ³Scheduler² link on the RM Web UI to get a
> >>>>> sense of
> >>>>> whether you are resource constrained.
> >>>>>
> >>>>> Are you certain that you are attempting to invoke the correct port?
> >>>>> The
> >>>>> listening ports are dynamically allocated by Slider.
> >>>>>
> >>>>>>
> >>>>>>
> >>>>>> On Tue, Apr 7, 2015 at 3:29 AM, Chackravarthy Esakkimuthu <
> >>>>>> [email protected]<mailto:[email protected]>> wrote:
> >>>>>>
> >>>>>>> Hi All,
> >>>>>>>
> >>>>>>> I am new to Apache slider and would like to contribute.
> >>>>>>>
> >>>>>>> Just to start with, I am trying out running "storm" and "hbase" on
> >>>>>>> yarn
> >>>>>>> using slider following the guide :
> >>>>>
> >>>>>
> http://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.2.0/YARN_RM_v22/run
> >>>>> ning_applications_on_slider/index.html#Item1.1
> >>>>>>>
> >>>>>>> In both (storm and hbase) the cases, the ApplicationMaster gets
> >>>>>>> launched
> >>>>>>> and still running, but the ApplicationMaster link not working, and
> >>>>>>> from
> >>>>> AM
> >>>>>>> logs, I don't see any errors.
> >>>>>>>
> >>>>>>> How do I debug from this? Please help me.
> >>>>>>> Incase if there is any other mail thread with respect this, please
> >>>>>>> point
> >>>>>>> out to me. Thanks in advance.
> >>>>>>>
> >>>>>>> Thanks,
> >>>>>>> Chackra
> >
> >
> >
>