Thank you Ram. I will try it out. Best regards, Sanal
On Mon, Sep 19, 2016 at 9:50 AM, Munagala Ramanath <[email protected]> wrote: > * <source>1.8</source>* > * <target>1.8</target>* > Could you change the source and target java versions to 1.7 in the above > lines of your > pom.xml and try again ? > > With this change I was able to run the application built with your pom.xml. > > Ram > > On Mon, Sep 12, 2016 at 8:48 PM, Sanal Vasudevan <[email protected]> > wrote: > >> Hi, >> >> Please find attached the pom.xml file. >> I added the following to resolve build issue. >> >> <dependency> >> <groupId>org.codehaus.jettison</groupId> >> <artifactId>jettison</artifactId> >> <version>1.3.8</version> >> </dependency> >> >> It is a single-node setup. >> Haddop Version:2.7.2, rb165c4fe8a74265c792ce23f546c64604acf0e41 >> Yes I have used the apex command to lauch the apa file. >> >> >> Thanks >> Sanal >> >> On Tue, Sep 13, 2016 at 1:29 AM, Munagala Ramanath <[email protected]> >> wrote: >> >>> Looks like this happens if jersey-server is not present >>> (e.g. http://stackoverflow.com/questions/8662919/jersey-no-w >>> ebapplication-provider-is-present-when-jersey-json-dependency-added) >>> >>> Have you made any changes to the pom.xml ? If so, can you post it here ? >>> >>> Also, can you tell us a bit more about your deployment environment ? Is >>> it a single or multi-node cluster ? >>> What version of Hadoop ? Are you using the "apex" command line tool to >>> deploy ? >>> >>> Ram >>> >>> On Mon, Sep 12, 2016 at 1:29 AM, Sanal Vasudevan <[email protected]> >>> wrote: >>> >>>> Hi there, >>>> >>>> >>>> >>>> I am unable to get started with Apache Apex. >>>> >>>> >>>> >>>> I used the below mvn command to create the project. >>>> >>>> >>>> >>>> #!/bin/bash >>>> >>>> # script to create a new project >>>> >>>> >>>> >>>> # change project name and archetype version as needed >>>> >>>> name=myapexapp >>>> >>>> version=3.4.0 >>>> >>>> >>>> >>>> mvn -B archetype:generate \ >>>> >>>> -DarchetypeGroupId=org.apache.apex \ >>>> >>>> -DarchetypeArtifactId=apex-app-archetype \ >>>> >>>> -DarchetypeVersion=$version \ >>>> >>>> -DgroupId=com.example \ >>>> >>>> -Dpackage=com.example.$name \ >>>> >>>> -DartifactId=$name \ >>>> >>>> -Dversion=1.0-SNAPSHOT >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> I am trying to execute the sample SortedWordCount application example. >>>> >>>> >>>> >>>> I see that the status of yarn application is FAILED. >>>> >>>> >>>> >>>> Below error in the log file: >>>> >>>> 2016-09-12 00:47:12,284 INFO com.datatorrent.stram.StreamingAppMaster: >>>> Initializing Application Master. >>>> 2016-09-12 00:47:12,385 INFO >>>> com.datatorrent.stram.StreamingAppMasterService: Application master, >>>> appId=3, clustertimestamp=1473662193222, attemptId=1 >>>> 2016-09-12 00:47:13,432 INFO >>>> com.datatorrent.common.util.AsyncFSStorageAgent: using >>>> /scratch/sanav/view_storage/projects/downloads/hadoop/tmp/hadoop-sanav/nm-local-dir/usercache/sanav/appcache/application_1473662193222_0003/container_1473662193222_0003_01_000001/tmp/chkp6399702117805507910 >>>> as the basepath for checkpointing. >>>> 2016-09-12 00:47:15,223 INFO com.datatorrent.stram.FSRecoveryHandler: >>>> Creating >>>> hdfs://den00spj:9000/user/sanav/datatorrent/apps/application_1473662193222_0003/recovery/log >>>> 2016-09-12 00:47:15,272 INFO >>>> com.datatorrent.stram.StreamingAppMasterService: Starting application with >>>> 5 operators in 5 containers >>>> 2016-09-12 00:47:15,286 INFO >>>> org.apache.hadoop.yarn.client.api.async.impl.NMClientAsyncImpl: Upper >>>> bound of the thread pool size is 500 >>>> 2016-09-12 00:47:15,288 INFO >>>> org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy: >>>> yarn.client.max-cached-nodemanagers-proxies : 0 >>>> 2016-09-12 00:47:15,316 INFO org.apache.hadoop.yarn.client.RMProxy: >>>> Connecting to ResourceManager at /0.0.0.0:8030 >>>> 2016-09-12 00:47:15,345 INFO >>>> com.datatorrent.stram.StreamingContainerParent: Config: Configuration: >>>> core-default.xml, core-site.xml, yarn-default.xml, yarn-site.xml, >>>> hdfs-default.xml, hdfs-site.xml >>>> 2016-09-12 00:47:15,345 INFO >>>> com.datatorrent.stram.StreamingContainerParent: Listener thread count 30 >>>> 2016-09-12 00:47:15,351 INFO org.apache.hadoop.ipc.CallQueueManager: Using >>>> callQueue class java.util.concurrent.LinkedBlockingQueue >>>> 2016-09-12 00:47:15,356 INFO org.apache.hadoop.ipc.Server: Starting Socket >>>> Reader #1 for port 55906 >>>> 2016-09-12 00:47:15,369 INFO org.apache.hadoop.ipc.Server: IPC Server >>>> listener on 55906: starting >>>> 2016-09-12 00:47:15,370 INFO org.apache.hadoop.ipc.Server: IPC Server >>>> Responder: starting >>>> 2016-09-12 00:47:15,385 INFO >>>> com.datatorrent.stram.StreamingContainerParent: Container callback server >>>> listening at den00spj/10.196.38.196:55906 >>>> 2016-09-12 00:47:15,440 INFO org.mortbay.log: Logging to >>>> org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via >>>> org.mortbay.log.Slf4jLog >>>> 2016-09-12 00:47:15,507 INFO >>>> org.apache.hadoop.security.authentication.server.AuthenticationFilter: >>>> Unable to initialize FileSignerSecretProvider, falling back to use random >>>> secrets. >>>> 2016-09-12 00:47:15,512 INFO org.apache.hadoop.http.HttpRequestLog: Http >>>> request log for http.requests.stram is not defined >>>> 2016-09-12 00:47:15,518 INFO org.apache.hadoop.http.HttpServer2: Added >>>> global filter 'safety' >>>> (class=org.apache.hadoop.http.HttpServer2$QuotingInputFilter) >>>> 2016-09-12 00:47:15,520 INFO org.apache.hadoop.http.HttpServer2: Added >>>> filter static_user_filter >>>> (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to >>>> context stram >>>> 2016-09-12 00:47:15,520 INFO org.apache.hadoop.http.HttpServer2: Added >>>> filter static_user_filter >>>> (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to >>>> context static >>>> 2016-09-12 00:47:15,520 INFO org.apache.hadoop.http.HttpServer2: Added >>>> filter static_user_filter >>>> (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to >>>> context logs >>>> 2016-09-12 00:47:15,523 INFO org.apache.hadoop.http.HttpServer2: adding >>>> path spec: /stram/* >>>> 2016-09-12 00:47:15,523 INFO org.apache.hadoop.http.HttpServer2: adding >>>> path spec: /ws/* >>>> 2016-09-12 00:47:15,875 INFO org.apache.hadoop.yarn.webapp.WebApps: >>>> Registered webapp guice modules >>>> 2016-09-12 00:47:15,876 INFO org.apache.hadoop.http.HttpServer2: Jetty >>>> bound to port 21097 >>>> >>>> 2016-09-12 00:47:16,147 ERROR >>>> com.datatorrent.stram.StreamingAppMasterService: >>>> Webapps failed to start. Ignoring for now: >>>> >>>> org.apache.hadoop.yarn.webapp.WebAppException: Error starting http >>>> server >>>> >>>> at org.apache.hadoop.yarn.webapp.WebApps$Builder.start(WebApps. >>>> java:310) >>>> >>>> at com.datatorrent.stram.StreamingAppMasterService.serviceStart >>>> (StreamingAppMasterService.java:616) >>>> >>>> at org.apache.hadoop.service.AbstractService.start(AbstractServ >>>> ice.java:193) >>>> >>>> at com.datatorrent.stram.StreamingAppMaster.main(StreamingAppMa >>>> ster.java:103) >>>> >>>> Caused by: java.io.IOException: Unable to initialize WebAppContext >>>> >>>> at org.apache.hadoop.http.HttpServer2.start(HttpServer2.java:87 >>>> 9) >>>> >>>> at org.apache.hadoop.yarn.webapp.WebApps$Builder.start(WebApps. >>>> java:306) >>>> >>>> ... 3 more >>>> >>>> Caused by: com.sun.jersey.api.container.ContainerException: No >>>> WebApplication provider is present >>>> >>>> at com.sun.jersey.spi.container.WebApplicationFactory.createWeb >>>> Application(WebApplicationFactory.java:69) >>>> >>>> at com.sun.jersey.spi.container.servlet.ServletContainer.create >>>> (ServletContainer.java:391) >>>> >>>> at com.sun.jersey.spi.container.servlet.ServletContainer$Intern >>>> alWebComponent.create(ServletContainer.java:306) >>>> >>>> at com.sun.jersey.spi.container.servlet.WebComponent.load(WebCo >>>> mponent.java:607) >>>> >>>> at com.sun.jersey.spi.container.servlet.WebComponent.init(WebCo >>>> mponent.java:210) >>>> >>>> at com.sun.jersey.spi.container.servlet.ServletContainer.init(S >>>> ervletContainer.java:373) >>>> >>>> at com.sun.jersey.spi.container.servlet.ServletContainer.init(S >>>> ervletContainer.java:710) >>>> >>>> at com.google.inject.servlet.FilterDefinition.init(FilterDefini >>>> tion.java:114) >>>> >>>> at com.google.inject.servlet.ManagedFilterPipeline.initPipeline >>>> (ManagedFilterPipeline.java:98) >>>> >>>> at com.google.inject.servlet.GuiceFilter.init(GuiceFilter.java: >>>> 172) >>>> >>>> at org.mortbay.jetty.servlet.FilterHolder.doStart(FilterHolder. >>>> java:97) >>>> >>>> at org.mortbay.component.AbstractLifeCycle.start(AbstractLifeCy >>>> cle.java:50) >>>> >>>> at org.mortbay.jetty.servlet.ServletHandler.initialize(ServletH >>>> andler.java:713) >>>> >>>> at org.mortbay.jetty.servlet.Context.startContext(Context.java: >>>> 140) >>>> >>>> at org.mortbay.jetty.webapp.WebAppContext.startContext(WebAppCo >>>> ntext.java:1282) >>>> >>>> at org.mortbay.jetty.handler.ContextHandler.doStart(ContextHand >>>> ler.java:518) >>>> >>>> at org.mortbay.jetty.webapp.WebAppContext.doStart(WebAppContext >>>> .java:499) >>>> >>>> at org.mortbay.component.AbstractLifeCycle.start(AbstractLifeCy >>>> cle.java:50) >>>> >>>> at org.mortbay.jetty.handler.HandlerCollection.doStart(HandlerC >>>> ollection.java:152) >>>> >>>> at org.mortbay.jetty.handler.ContextHandlerCollection.doStart(C >>>> ontextHandlerCollection.java:156) >>>> >>>> at org.mortbay.component.AbstractLifeCycle.start(AbstractLifeCy >>>> cle.java:50) >>>> >>>> at org.mortbay.jetty.handler.HandlerWrapper.doStart(HandlerWrap >>>> per.java:130) >>>> >>>> at org.mortbay.jetty.Server.doStart(Server.java:224) >>>> >>>> at org.mortbay.component.AbstractLifeCycle.start(AbstractLifeCy >>>> cle.java:50) >>>> >>>> at org.apache.hadoop.http.HttpServer2.start(HttpServer2.java:85 >>>> 7) >>>> >>>> >>>> >>>> >>>> >>>> Apart from this error I don’t see any other errors. >>>> >>>> >>>> >>>> Please advise. >>>> >>>> >>>> -- >>>> Thanks >>>> Sanal >>>> >>> >>> >> >> >> -- >> Sanal Vasudevan Nair >> > > -- Sanal Vasudevan Nair
