We have a Hbase client dependency and do not have any Hadoop dependencies in project.
From: Ashwin Chandra Putta [mailto:[email protected]] Sent: Wednesday, October 19, 2016 2:40 PM To: [email protected] Subject: Re: Retry functionality in Datatorrent Jaspal, Ensure that the app package is not including any hadoop jars as dependencies. Check target/deps/ You can exclude those dependencies in the maven pom file. Regards, Ashwin. On Wed, Oct 19, 2016 at 12:09 PM, Jaspal Singh <[email protected]<mailto:[email protected]>> wrote: Hi Ashwin, That's correct flow !! Team - ANother thing we are receiving below exception in the logs and application is FAILING when we are trying to make an update to HBase table. Any idea what could be the reason ? Container: container_e32_1476503307399_0172_02_000001 on dbslt0079.uhc.com_8091 ================================================================================= LogType:AppMaster.stderr Log Upload Time:Wed Oct 19 11:35:36 -0500 2016 LogLength:1246 Log Contents: SLF4J: Class path contains multiple SLF4J bindings. SLF4J: Found binding in [jar:file:/opt/mapr/tmp/hadoop-mapr/nm-local-dir/usercache/mapr/appcache/application_1476503307399_0172/filecache/26/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class] SLF4J: Found binding in [jar:file:/opt/mapr/lib/slf4j-log4j12-1.7.12.jar!/org/slf4j/impl/StaticLoggerBinder.class] SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation. SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory] Exception in thread "main" java.lang.IllegalArgumentException: Invalid ContainerId: container_e32_1476503307399_0172_02_000001 at org.apache.hadoop.yarn.util.ConverterUtils.toContainerId(ConverterUtils.java:182) at com.datatorrent.stram.StreamingAppMaster.main(StreamingAppMaster.java:91) Caused by: java.lang.NumberFormatException: For input string: "e32" at java.lang.NumberFormatException.forInputString(NumberFormatException.java:65) at java.lang.Long.parseLong(Long.java:589) at java.lang.Long.parseLong(Long.java:631) at org.apache.hadoop.yarn.util.ConverterUtils.toApplicationAttemptId(ConverterUtils.java:137) at org.apache.hadoop.yarn.util.ConverterUtils.toContainerId(ConverterUtils.java:177) ... 1 more End of LogType:AppMaster.stderr Thanks!! On Tue, Oct 18, 2016 at 7:47 PM, Ashwin Chandra Putta <[email protected]<mailto:[email protected]>> wrote: Jaspal, If I understand this correctly, the flow looks like this. REST API --> Kafka topic --> Apex Kafka Input Operator. If this is the flow, then the kafka input operator should be reading messages from Kafka without losing them. There is no retry attempts necessary. Let me know if the understanding of the flow is incorrect. Regards, Ashwin. On Tue, Oct 18, 2016 at 2:49 PM, Jaspal Singh <[email protected]<mailto:[email protected]>> wrote: Hi Team, We are pushing messages from Kafka to Datatorrent application using REST API service as a producer. If say due to some issue with the service the message couldn't be pushed/processed so we want to "retry it for n times" before it is dropped. Is there any retry functionality built within Datatorrent or we have write some code logic for the same ?? Thanks!! -- Regards, Ashwin. -- Regards, Ashwin. This e-mail, including attachments, may include confidential and/or proprietary information, and may be used only by the person or entity to which it is addressed. If the reader of this e-mail is not the intended recipient or his or her authorized agent, the reader is hereby notified that any dissemination, distribution or copying of this e-mail is prohibited. If you have received this e-mail in error, please notify the sender by replying to this message and delete this e-mail immediately.
