Thank you Ram. I will try it out.

Best regards,
Sanal

On Mon, Sep 19, 2016 at 9:50 AM, Munagala Ramanath <r...@datatorrent.com>
wrote:

> *           <source>1.8</source>*
> *           <target>1.8</target>*
> Could you change the source and target java versions to 1.7 in the above
> lines of your
> pom.xml and try again ?
>
> With this change I was able to run the application built with your pom.xml.
>
> Ram
>
> On Mon, Sep 12, 2016 at 8:48 PM, Sanal Vasudevan <get2sa...@gmail.com>
> wrote:
>
>> Hi,
>>
>> Please find attached the pom.xml file.
>> I added the following to resolve build issue.
>>
>> <dependency>
>>   <groupId>org.codehaus.jettison</groupId>
>>   <artifactId>jettison</artifactId>
>>   <version>1.3.8</version>
>> </dependency>
>>
>> It is a single-node setup.
>> Haddop Version:2.7.2, rb165c4fe8a74265c792ce23f546c64604acf0e41
>> Yes I have used the apex command to lauch the apa file.
>>
>>
>> Thanks
>> Sanal
>>
>> On Tue, Sep 13, 2016 at 1:29 AM, Munagala Ramanath <r...@datatorrent.com>
>> wrote:
>>
>>> Looks like this happens if jersey-server is not present
>>> (e.g. http://stackoverflow.com/questions/8662919/jersey-no-w
>>> ebapplication-provider-is-present-when-jersey-json-dependency-added)
>>>
>>> Have you made any changes to the pom.xml ? If so, can you post it here ?
>>>
>>> Also, can you tell us a bit more about your deployment environment ? Is
>>> it a single or multi-node cluster ?
>>> What version of Hadoop ? Are you using the "apex" command line tool to
>>> deploy ?
>>>
>>> Ram
>>>
>>> On Mon, Sep 12, 2016 at 1:29 AM, Sanal Vasudevan <get2sa...@gmail.com>
>>> wrote:
>>>
>>>> Hi there,
>>>>
>>>>
>>>>
>>>> I am unable to get started with Apache Apex.
>>>>
>>>>
>>>>
>>>> I used the below mvn command to create the project.
>>>>
>>>>
>>>>
>>>> #!/bin/bash
>>>>
>>>> # script to create a new project
>>>>
>>>>
>>>>
>>>> # change project name and archetype version as needed
>>>>
>>>> name=myapexapp
>>>>
>>>> version=3.4.0
>>>>
>>>>
>>>>
>>>> mvn -B archetype:generate \
>>>>
>>>>   -DarchetypeGroupId=org.apache.apex \
>>>>
>>>>   -DarchetypeArtifactId=apex-app-archetype \
>>>>
>>>>   -DarchetypeVersion=$version  \
>>>>
>>>>   -DgroupId=com.example \
>>>>
>>>>   -Dpackage=com.example.$name \
>>>>
>>>>   -DartifactId=$name \
>>>>
>>>>   -Dversion=1.0-SNAPSHOT
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>> I am trying to execute the sample SortedWordCount application example.
>>>>
>>>>
>>>>
>>>> I see that the status of yarn application is FAILED.
>>>>
>>>>
>>>>
>>>> Below error in the log file:
>>>>
>>>> 2016-09-12 00:47:12,284 INFO com.datatorrent.stram.StreamingAppMaster: 
>>>> Initializing Application Master.
>>>> 2016-09-12 00:47:12,385 INFO 
>>>> com.datatorrent.stram.StreamingAppMasterService: Application master, 
>>>> appId=3, clustertimestamp=1473662193222, attemptId=1
>>>> 2016-09-12 00:47:13,432 INFO 
>>>> com.datatorrent.common.util.AsyncFSStorageAgent: using 
>>>> /scratch/sanav/view_storage/projects/downloads/hadoop/tmp/hadoop-sanav/nm-local-dir/usercache/sanav/appcache/application_1473662193222_0003/container_1473662193222_0003_01_000001/tmp/chkp6399702117805507910
>>>>  as the basepath for checkpointing.
>>>> 2016-09-12 00:47:15,223 INFO com.datatorrent.stram.FSRecoveryHandler: 
>>>> Creating 
>>>> hdfs://den00spj:9000/user/sanav/datatorrent/apps/application_1473662193222_0003/recovery/log
>>>> 2016-09-12 00:47:15,272 INFO 
>>>> com.datatorrent.stram.StreamingAppMasterService: Starting application with 
>>>> 5 operators in 5 containers
>>>> 2016-09-12 00:47:15,286 INFO 
>>>> org.apache.hadoop.yarn.client.api.async.impl.NMClientAsyncImpl: Upper 
>>>> bound of the thread pool size is 500
>>>> 2016-09-12 00:47:15,288 INFO 
>>>> org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy: 
>>>> yarn.client.max-cached-nodemanagers-proxies : 0
>>>> 2016-09-12 00:47:15,316 INFO org.apache.hadoop.yarn.client.RMProxy: 
>>>> Connecting to ResourceManager at /0.0.0.0:8030
>>>> 2016-09-12 00:47:15,345 INFO 
>>>> com.datatorrent.stram.StreamingContainerParent: Config: Configuration: 
>>>> core-default.xml, core-site.xml, yarn-default.xml, yarn-site.xml, 
>>>> hdfs-default.xml, hdfs-site.xml
>>>> 2016-09-12 00:47:15,345 INFO 
>>>> com.datatorrent.stram.StreamingContainerParent: Listener thread count 30
>>>> 2016-09-12 00:47:15,351 INFO org.apache.hadoop.ipc.CallQueueManager: Using 
>>>> callQueue class java.util.concurrent.LinkedBlockingQueue
>>>> 2016-09-12 00:47:15,356 INFO org.apache.hadoop.ipc.Server: Starting Socket 
>>>> Reader #1 for port 55906
>>>> 2016-09-12 00:47:15,369 INFO org.apache.hadoop.ipc.Server: IPC Server 
>>>> listener on 55906: starting
>>>> 2016-09-12 00:47:15,370 INFO org.apache.hadoop.ipc.Server: IPC Server 
>>>> Responder: starting
>>>> 2016-09-12 00:47:15,385 INFO 
>>>> com.datatorrent.stram.StreamingContainerParent: Container callback server 
>>>> listening at den00spj/10.196.38.196:55906
>>>> 2016-09-12 00:47:15,440 INFO org.mortbay.log: Logging to 
>>>> org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via 
>>>> org.mortbay.log.Slf4jLog
>>>> 2016-09-12 00:47:15,507 INFO 
>>>> org.apache.hadoop.security.authentication.server.AuthenticationFilter: 
>>>> Unable to initialize FileSignerSecretProvider, falling back to use random 
>>>> secrets.
>>>> 2016-09-12 00:47:15,512 INFO org.apache.hadoop.http.HttpRequestLog: Http 
>>>> request log for http.requests.stram is not defined
>>>> 2016-09-12 00:47:15,518 INFO org.apache.hadoop.http.HttpServer2: Added 
>>>> global filter 'safety' 
>>>> (class=org.apache.hadoop.http.HttpServer2$QuotingInputFilter)
>>>> 2016-09-12 00:47:15,520 INFO org.apache.hadoop.http.HttpServer2: Added 
>>>> filter static_user_filter 
>>>> (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to 
>>>> context stram
>>>> 2016-09-12 00:47:15,520 INFO org.apache.hadoop.http.HttpServer2: Added 
>>>> filter static_user_filter 
>>>> (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to 
>>>> context static
>>>> 2016-09-12 00:47:15,520 INFO org.apache.hadoop.http.HttpServer2: Added 
>>>> filter static_user_filter 
>>>> (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to 
>>>> context logs
>>>> 2016-09-12 00:47:15,523 INFO org.apache.hadoop.http.HttpServer2: adding 
>>>> path spec: /stram/*
>>>> 2016-09-12 00:47:15,523 INFO org.apache.hadoop.http.HttpServer2: adding 
>>>> path spec: /ws/*
>>>> 2016-09-12 00:47:15,875 INFO org.apache.hadoop.yarn.webapp.WebApps: 
>>>> Registered webapp guice modules
>>>> 2016-09-12 00:47:15,876 INFO org.apache.hadoop.http.HttpServer2: Jetty 
>>>> bound to port 21097
>>>>
>>>> 2016-09-12 00:47:16,147 ERROR 
>>>> com.datatorrent.stram.StreamingAppMasterService:
>>>> Webapps failed to start. Ignoring for now:
>>>>
>>>> org.apache.hadoop.yarn.webapp.WebAppException: Error starting http
>>>> server
>>>>
>>>>         at org.apache.hadoop.yarn.webapp.WebApps$Builder.start(WebApps.
>>>> java:310)
>>>>
>>>>         at com.datatorrent.stram.StreamingAppMasterService.serviceStart
>>>> (StreamingAppMasterService.java:616)
>>>>
>>>>         at org.apache.hadoop.service.AbstractService.start(AbstractServ
>>>> ice.java:193)
>>>>
>>>>         at com.datatorrent.stram.StreamingAppMaster.main(StreamingAppMa
>>>> ster.java:103)
>>>>
>>>> Caused by: java.io.IOException: Unable to initialize WebAppContext
>>>>
>>>>         at org.apache.hadoop.http.HttpServer2.start(HttpServer2.java:87
>>>> 9)
>>>>
>>>>         at org.apache.hadoop.yarn.webapp.WebApps$Builder.start(WebApps.
>>>> java:306)
>>>>
>>>>         ... 3 more
>>>>
>>>> Caused by: com.sun.jersey.api.container.ContainerException: No
>>>> WebApplication provider is present
>>>>
>>>>         at com.sun.jersey.spi.container.WebApplicationFactory.createWeb
>>>> Application(WebApplicationFactory.java:69)
>>>>
>>>>         at com.sun.jersey.spi.container.servlet.ServletContainer.create
>>>> (ServletContainer.java:391)
>>>>
>>>>         at com.sun.jersey.spi.container.servlet.ServletContainer$Intern
>>>> alWebComponent.create(ServletContainer.java:306)
>>>>
>>>>         at com.sun.jersey.spi.container.servlet.WebComponent.load(WebCo
>>>> mponent.java:607)
>>>>
>>>>         at com.sun.jersey.spi.container.servlet.WebComponent.init(WebCo
>>>> mponent.java:210)
>>>>
>>>>         at com.sun.jersey.spi.container.servlet.ServletContainer.init(S
>>>> ervletContainer.java:373)
>>>>
>>>>         at com.sun.jersey.spi.container.servlet.ServletContainer.init(S
>>>> ervletContainer.java:710)
>>>>
>>>>         at com.google.inject.servlet.FilterDefinition.init(FilterDefini
>>>> tion.java:114)
>>>>
>>>>         at com.google.inject.servlet.ManagedFilterPipeline.initPipeline
>>>> (ManagedFilterPipeline.java:98)
>>>>
>>>>         at com.google.inject.servlet.GuiceFilter.init(GuiceFilter.java:
>>>> 172)
>>>>
>>>>         at org.mortbay.jetty.servlet.FilterHolder.doStart(FilterHolder.
>>>> java:97)
>>>>
>>>>         at org.mortbay.component.AbstractLifeCycle.start(AbstractLifeCy
>>>> cle.java:50)
>>>>
>>>>         at org.mortbay.jetty.servlet.ServletHandler.initialize(ServletH
>>>> andler.java:713)
>>>>
>>>>         at org.mortbay.jetty.servlet.Context.startContext(Context.java:
>>>> 140)
>>>>
>>>>         at org.mortbay.jetty.webapp.WebAppContext.startContext(WebAppCo
>>>> ntext.java:1282)
>>>>
>>>>         at org.mortbay.jetty.handler.ContextHandler.doStart(ContextHand
>>>> ler.java:518)
>>>>
>>>>         at org.mortbay.jetty.webapp.WebAppContext.doStart(WebAppContext
>>>> .java:499)
>>>>
>>>>         at org.mortbay.component.AbstractLifeCycle.start(AbstractLifeCy
>>>> cle.java:50)
>>>>
>>>>         at org.mortbay.jetty.handler.HandlerCollection.doStart(HandlerC
>>>> ollection.java:152)
>>>>
>>>>         at org.mortbay.jetty.handler.ContextHandlerCollection.doStart(C
>>>> ontextHandlerCollection.java:156)
>>>>
>>>>         at org.mortbay.component.AbstractLifeCycle.start(AbstractLifeCy
>>>> cle.java:50)
>>>>
>>>>         at org.mortbay.jetty.handler.HandlerWrapper.doStart(HandlerWrap
>>>> per.java:130)
>>>>
>>>>         at org.mortbay.jetty.Server.doStart(Server.java:224)
>>>>
>>>>         at org.mortbay.component.AbstractLifeCycle.start(AbstractLifeCy
>>>> cle.java:50)
>>>>
>>>>         at org.apache.hadoop.http.HttpServer2.start(HttpServer2.java:85
>>>> 7)
>>>>
>>>>
>>>>
>>>>
>>>>
>>>> Apart from this error I don’t see any other errors.
>>>>
>>>>
>>>>
>>>> Please advise.
>>>>
>>>>
>>>> --
>>>> Thanks
>>>> Sanal
>>>>
>>>
>>>
>>
>>
>> --
>> Sanal Vasudevan Nair
>>
>
>


-- 
Sanal Vasudevan Nair

Reply via email to