Tried running with 'yarn' user, but it remains in same state.
AM link not working, and AM logs are similar.

On Wed, Apr 8, 2015 at 2:14 AM, Gour Saha <[email protected]> wrote:

> In a non-secured cluster you should run as yarn. Can you do that and let
> us know how it goes?
>
> Also you can stop your existing storm instance in hdfs user (run as hdfs
> user) by running stop first -
> slider stop storm1
>
> -Gour
>
> On 4/7/15, 1:39 PM, "Chackravarthy Esakkimuthu" <[email protected]>
> wrote:
>
> >This is not a secured cluster.
> >And yes, I used 'hdfs' user while running slider create.
> >
> >On Wed, Apr 8, 2015 at 2:03 AM, Gour Saha <[email protected]> wrote:
> >
> >> Which user are you running the slider create command as? Seems like you
> >> are running as hdfs user. Is this a secured cluster?
> >>
> >> -Gour
> >>
> >> On 4/7/15, 1:06 PM, "Chackravarthy Esakkimuthu" <[email protected]>
> >> wrote:
> >>
> >> >yes, RM HA has been setup in this cluster.
> >> >
> >> >Active : zs-aaa-001.nm.flipkart.com
> >> >Standby : zs-aaa-002.nm.flipkart.com
> >> >
> >> >RM Link : http://zs-aaa-001.nm.flipkart.com:8088/cluster/scheduler
> >> ><http://zs-exp-01.nm.flipkart.com:8088/cluster/scheduler>
> >> >
> >> >AM Link :
> >> >
> >>
> >>
> http://zs-aaa-001.nm.flipkart.com:8088/proxy/application_1427882795362_00
> >>7
> >> >0/slideram
> >> ><
> >>
> >>
> http://zs-exp-01.nm.flipkart.com:8088/proxy/application_1427882795362_007
> >> >0/slideram>
> >> >
> >> >On Wed, Apr 8, 2015 at 1:05 AM, Gour Saha <[email protected]>
> >>wrote:
> >> >
> >> >> Sorry forgot that the AM link not working was the original issue.
> >> >>
> >> >> Few more things -
> >> >> - Seems like you have RM HA setup, right?
> >> >> - Can you copy paste the complete link of the RM UI and the URL of
> >> >> ApplicationMaster (the link which is broken) with actual hostnames?
> >> >>
> >> >>
> >> >> -Gour
> >> >>
> >> >> On 4/7/15, 11:43 AM, "Chackravarthy Esakkimuthu"
> >><[email protected]
> >> >
> >> >> wrote:
> >> >>
> >> >> >Since 5 containers are running, which means that Storm daemons are
> >> >>already
> >> >> >up and running?
> >> >> >
> >> >> >
> >> >> >Actually the ApplicationMaster link is not working. It just blanks
> >>out
> >> >> >printing the following :
> >> >> >
> >> >> >This is standby RM. Redirecting to the current active RM:
> >> >>
> >>>http://<host-name>:8088/proxy/application_1427882795362_0070/slideram
> >> >> >
> >> >> >
> >> >> >And for resources.json, I dint make any change and used the copy of
> >> >> >resources-default.json as follows:
> >> >> >
> >> >> >
> >> >> >{
> >> >> >
> >> >> >  "schema" : "http://example.org/specification/v2.0.0";,
> >> >> >
> >> >> >  "metadata" : {
> >> >> >
> >> >> >  },
> >> >> >
> >> >> >  "global" : {
> >> >> >
> >> >> >    "yarn.log.include.patterns": "",
> >> >> >
> >> >> >    "yarn.log.exclude.patterns": ""
> >> >> >
> >> >> >  },
> >> >> >
> >> >> >  "components": {
> >> >> >
> >> >> >    "slider-appmaster": {
> >> >> >
> >> >> >      "yarn.memory": "512"
> >> >> >
> >> >> >    },
> >> >> >
> >> >> >    "NIMBUS": {
> >> >> >
> >> >> >      "yarn.role.priority": "1",
> >> >> >
> >> >> >      "yarn.component.instances": "1",
> >> >> >
> >> >> >      "yarn.memory": "2048"
> >> >> >
> >> >> >    },
> >> >> >
> >> >> >    "STORM_UI_SERVER": {
> >> >> >
> >> >> >      "yarn.role.priority": "2",
> >> >> >
> >> >> >      "yarn.component.instances": "1",
> >> >> >
> >> >> >      "yarn.memory": "1278"
> >> >> >
> >> >> >    },
> >> >> >
> >> >> >    "DRPC_SERVER": {
> >> >> >
> >> >> >      "yarn.role.priority": "3",
> >> >> >
> >> >> >      "yarn.component.instances": "1",
> >> >> >
> >> >> >      "yarn.memory": "1278"
> >> >> >
> >> >> >    },
> >> >> >
> >> >> >    "SUPERVISOR": {
> >> >> >
> >> >> >      "yarn.role.priority": "4",
> >> >> >
> >> >> >      "yarn.component.instances": "1",
> >> >> >
> >> >> >      "yarn.memory": "3072"
> >> >> >
> >> >> >    }
> >> >> >
> >> >> >  }
> >> >> >
> >> >> >}
> >> >> >
> >> >> >
> >> >> >
> >> >> >On Tue, Apr 7, 2015 at 11:52 PM, Gour Saha <[email protected]>
> >> >>wrote:
> >> >> >
> >> >> >> Chackra sent the attachment directly to me. From what I see the
> >> >>cluster
> >> >> >> resources (memory and cores) are abundant.
> >> >> >>
> >> >> >> But I also see that only 1 app is running which is the one we are
> >> >>trying
> >> >> >> to debug and 5 containers are running. So definitely more
> >>containers
> >> >> >>that
> >> >> >> just the AM is running.
> >> >> >>
> >> >> >> Can you click on the app master link and copy paste the content of
> >> >>that
> >> >> >> page? No need for screen shot. Also please send your resources
> >>JSON
> >> >> >>file.
> >> >> >>
> >> >> >> -Gour
> >> >> >>
> >> >> >> - Sent from my iPhone
> >> >> >>
> >> >> >> > On Apr 7, 2015, at 11:01 AM, "Jon Maron"
> >><[email protected]>
> >> >> >>wrote:
> >> >> >> >
> >> >> >> >
> >> >> >> > On Apr 7, 2015, at 1:36 PM, Chackravarthy Esakkimuthu <
> >> >> >> [email protected]<mailto:[email protected]>> wrote:
> >> >> >> >
> >> >> >> > @Maron, I could not get the logs even though the application is
> >> >>still
> >> >> >> running.
> >> >> >> > It's a 10 node cluster and I logged into one of the node and
> >> >>executed
> >> >> >> the command :
> >> >> >> >
> >> >> >> > sudo -u hdfs yarn logs -applicationId
> >> >>application_1427882795362_0070
> >> >> >> > 15/04/07 22:56:09 INFO impl.TimelineClientImpl: Timeline service
> >> >> >> address: http://$HOST:PORT/ws/v1/timeline/
> >> >> >> > 15/04/07 22:56:09 INFO client.ConfiguredRMFailoverProxyProvider:
> >> >> >>Failing
> >> >> >> over to rm2
> >> >> >> > /app-logs/hdfs/logs/application_1427882795362_0070does not have
> >>any
> >> >> >>log
> >> >> >> files.
> >> >> >> >
> >> >> >> > Can you login to the cluster node and look at the logs directory
> >> >>(e.g.
> >> >> >> in HDP install it would be under /hadoop/yarn/logs IIRC)?
> >> >> >> >
> >> >> >> >
> >> >> >> >
> >> >> >> > @Gour, Please find the attachment.
> >> >> >> >
> >> >> >> > On Tue, Apr 7, 2015 at 10:57 PM, Gour Saha
> >><[email protected]
> >> >> >> <mailto:[email protected]>> wrote:
> >> >> >> > Can you take a screenshot of your RM UI and send it over? It is
> >> >> >>usually
> >> >> >> > available in a URI similar to
> >> >> >> http://c6410.ambari.apache.org:8088/cluster.
> >> >> >> > I am specifically interested in seeing the Cluster Metrics
> >>table.
> >> >> >> >
> >> >> >> > -Gour
> >> >> >> >
> >> >> >> >> On 4/7/15, 10:17 AM, "Jon Maron"
> >><[email protected]<mailto:
> >> >> >> [email protected]>> wrote:
> >> >> >> >>
> >> >> >> >>
> >> >> >> >>> On Apr 7, 2015, at 1:14 PM, Jon Maron
> >> >> >><[email protected]<mailto:
> >> >> >> [email protected]>> wrote:
> >> >> >> >>>
> >> >> >> >>>
> >> >> >> >>>> On Apr 7, 2015, at 1:08 PM, Chackravarthy Esakkimuthu
> >> >> >> >>>> <[email protected]<mailto:[email protected]>> wrote:
> >> >> >> >>>>
> >> >> >> >>>> Thanks for the reply guys!
> >> >> >> >>>> Contianer allocation happened successfully.
> >> >> >> >>>>
> >> >> >> >>>> *RoleStatus{name='slider-appmaster', key=0, minimum=0,
> >> >>maximum=1,
> >> >> >> >>>> desired=1, actual=1,*
> >> >> >> >>>> *RoleStatus{name='STORM_UI_SERVER', key=2, minimum=0,
> >>maximum=1,
> >> >> >> >>>> desired=1,
> >> >> >> >>>> actual=1, *
> >> >> >> >>>> *RoleStatus{name='NIMBUS', key=1, minimum=0, maximum=1,
> >> >>desired=1,
> >> >> >> >>>> actual=1, *
> >> >> >> >>>> *RoleStatus{name='DRPC_SERVER', key=3, minimum=0, maximum=1,
> >> >> >> desired=1,
> >> >> >> >>>> actual=1, *
> >> >> >> >>>> *RoleStatus{name='SUPERVISOR', key=4, minimum=0, maximum=1,
> >> >> >>desired=1,
> >> >> >> >>>> actual=1,*
> >> >> >> >>>>
> >> >> >> >>>> Also, have put some logs specific to a container.. (nimbus)
> >>Same
> >> >> >>set
> >> >> >> of
> >> >> >> >>>> logs available for other Roles also (except Supervisor which
> >>has
> >> >> >>only
> >> >> >> >>>> first
> >> >> >> >>>> 2 lines of below logs)
> >> >> >> >>>>
> >> >> >> >>>> *Installing NIMBUS on
> >> >>container_e04_1427882795362_0070_01_000002.*
> >> >> >> >>>> *Starting NIMBUS on
> >>container_e04_1427882795362_0070_01_000002.*
> >> >> >> >>>> *Registering component
> >> >>container_e04_1427882795362_0070_01_000002*
> >> >> >> >>>> *Requesting applied config for NIMBUS on
> >> >> >> >>>> container_e04_1427882795362_0070_01_000002.*
> >> >> >> >>>> *Received and processed config for
> >> >> >> >>>> container_e04_1427882795362_0070_01_000002___NIMBUS*
> >> >> >> >>>>
> >> >> >> >>>> Does this result in any intermediate state?
> >> >> >> >>>>
> >> >> >> >>>> @Maron, I didn't configure any port specifically.. do I need
> >>to
> >> >>to?
> >> >> >> >>>> Also, i
> >> >> >> >>>> don't see any error msg in AM logs wrt port conflict.
> >> >> >> >>>
> >> >> >> >>> My only concern was whether you were actually accession the
> >>web
> >> >>UIs
> >> >> >>at
> >> >> >> >>> the correct host and port.  If you are then the next step is
> >> >> >>probably
> >> >> >> to
> >> >> >> >>> look at the actual storm/hbase logs.  you can use the ³yarn
> >>logs
> >> >> >> >>> -applicationid ..² command.
> >> >> >> >>
> >> >> >> >> *accessing* ;)
> >> >> >> >>
> >> >> >> >>>
> >> >> >> >>>>
> >> >> >> >>>> Thanks,
> >> >> >> >>>> Chackra
> >> >> >> >>>>
> >> >> >> >>>>
> >> >> >> >>>>
> >> >> >> >>>> On Tue, Apr 7, 2015 at 9:02 PM, Jon Maron
> >> >><[email protected]
> >> >> >> <mailto:[email protected]>>
> >> >> >> >>>> wrote:
> >> >> >> >>>>
> >> >> >> >>>>>
> >> >> >> >>>>>> On Apr 7, 2015, at 11:03 AM, Billie Rinaldi
> >> >> >> >>>>>> <[email protected]<mailto:[email protected]
> >>
> >> >> >> >>>>> wrote:
> >> >> >> >>>>>>
> >> >> >> >>>>>> One thing you can check is whether your system has enough
> >> >> >>resources
> >> >> >> >>>>>> to
> >> >> >> >>>>>> allocate all the containers the app needs.  You will see
> >>info
> >> >> >>like
> >> >> >> >>>>>> the
> >> >> >> >>>>>> following in the AM log (it will be logged multiple times
> >>over
> >> >> >>the
> >> >> >> >>>>>> life
> >> >> >> >>>>> of
> >> >> >> >>>>>> the AM).  In this case, the master I requested was
> >>allocated
> >> >>but
> >> >> >>the
> >> >> >> >>>>>> tservers were not.
> >> >> >> >>>>>> RoleStatus{name='ACCUMULO_TSERVER', key=2, desired=2,
> >> >>actual=0,
> >> >> >> >>>>>> requested=2, releasing=0, failed=0, started=0,
> >>startFailed=0,
> >> >> >> >>>>> completed=0,
> >> >> >> >>>>>> failureMessage=''}
> >> >> >> >>>>>> RoleStatus{name='ACCUMULO_MASTER', key=1, desired=1,
> >>actual=1,
> >> >> >> >>>>> requested=0,
> >> >> >> >>>>>> releasing=0, failed=0, started=0, startFailed=0,
> >>completed=0,
> >> >> >> >>>>>> failureMessage=Œ'}
> >> >> >> >>>>>
> >> >> >> >>>>> You can also check the ³Scheduler² link on the RM Web UI to
> >> >>get a
> >> >> >> >>>>> sense of
> >> >> >> >>>>> whether you are resource constrained.
> >> >> >> >>>>>
> >> >> >> >>>>> Are you certain that you are attempting to invoke the
> >>correct
> >> >> >>port?
> >> >> >> >>>>> The
> >> >> >> >>>>> listening ports are dynamically allocated by Slider.
> >> >> >> >>>>>
> >> >> >> >>>>>>
> >> >> >> >>>>>>
> >> >> >> >>>>>> On Tue, Apr 7, 2015 at 3:29 AM, Chackravarthy Esakkimuthu <
> >> >> >> >>>>>> [email protected]<mailto:[email protected]>>
> wrote:
> >> >> >> >>>>>>
> >> >> >> >>>>>>> Hi All,
> >> >> >> >>>>>>>
> >> >> >> >>>>>>> I am new to Apache slider and would like to contribute.
> >> >> >> >>>>>>>
> >> >> >> >>>>>>> Just to start with, I am trying out running "storm" and
> >> >> >>"hbase" on
> >> >> >> >>>>>>> yarn
> >> >> >> >>>>>>> using slider following the guide :
> >> >> >> >>>>>
> >> >> >> >>>>>
> >> >> >>
> >>
> >>>>
> http://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.2.0/YARN_RM_v22/run
> >> >> >> >>>>> ning_applications_on_slider/index.html#Item1.1
> >> >> >> >>>>>>>
> >> >> >> >>>>>>> In both (storm and hbase) the cases, the ApplicationMaster
> >> >>gets
> >> >> >> >>>>>>> launched
> >> >> >> >>>>>>> and still running, but the ApplicationMaster link not
> >> >>working,
> >> >> >>and
> >> >> >> >>>>>>> from
> >> >> >> >>>>> AM
> >> >> >> >>>>>>> logs, I don't see any errors.
> >> >> >> >>>>>>>
> >> >> >> >>>>>>> How do I debug from this? Please help me.
> >> >> >> >>>>>>> Incase if there is any other mail thread with respect
> >>this,
> >> >> >>please
> >> >> >> >>>>>>> point
> >> >> >> >>>>>>> out to me. Thanks in advance.
> >> >> >> >>>>>>>
> >> >> >> >>>>>>> Thanks,
> >> >> >> >>>>>>> Chackra
> >> >> >> >
> >> >> >> >
> >> >> >> >
> >> >> >>
> >> >>
> >> >>
> >>
> >>
>
>

Reply via email to