I think i get it, so i need to define the output type of the message as well. I'm will take a look at this now.
On Thu, Aug 7, 2014 at 4:12 PM, Telles Nobrega <[email protected]> wrote: > Here it goes: http://pastebin.com/hhswEnyW > > > On Thu, Aug 7, 2014 at 4:10 PM, Chris Riccomini < > [email protected]> wrote: > >> Hey Telles, >> >> Can you paste the code for your StreamTask? If it's still showing the same >> message, then it sounds like you're still trying to send a HashMap using a >> StringSerde. >> >> Cheers, >> Chris >> >> On 8/7/14 12:05 PM, "Telles Nobrega" <[email protected]> wrote: >> >> >Still showing the same message >> > >> > >> >On Thu, Aug 7, 2014 at 3:56 PM, Chris Riccomini < >> >[email protected]> wrote: >> > >> >> Hey Telles, >> >> >> >> The code you've posted in Produce.java shows: >> >> >> >> KeyedMessage<String, String> data = new >> KeyedMessage<String, >> >> String>("consumptions", String.valueOf(key),String.valueOf(value)); >> >> >> >> >> >> Which suggests that you are sending a string for both key and value. If >> >> you have a Samza task consuming from this topic, you should set: >> >> >> >> systems.system-name.samza.key.serde=string >> >> >> >> systems.system-name.samza.msg.serde=string >> >> >> >> >> >> Cheers, >> >> Chris >> >> >> >> On 8/7/14 11:52 AM, "Telles Nobrega" <[email protected]> wrote: >> >> >> >> >Hum, that sounds like a perfect reason for it. >> >> > >> >> >I'm writing to the topic with this code >> >> > >> >> > >> >> >> >> >> https://github.com/tellesnobrega/kafka-producer/blob/master/src/Produce.j >> >>a >> >> >va >> >> > >> >> >My problem is that I need to send numbers as key and value to the >> kafka >> >> >topic so i can read it in samza. >> >> > >> >> >What is the best way to de/serialize this? >> >> > >> >> > >> >> >On Thu, Aug 7, 2014 at 3:43 PM, Yan Fang <[email protected]> >> wrote: >> >> > >> >> >> Hi Telles, >> >> >> >> >> >> One of the possible reasons is that, in your process method, you are >> >> >>trying >> >> >> to send a HashMap, not a String, in the collection.send. Could you >> >> >>check it >> >> >> ? >> >> >> >> >> >> Thanks, >> >> >> >> >> >> Fang, Yan >> >> >> [email protected] >> >> >> +1 (206) 849-4108 >> >> >> >> >> >> >> >> >> On Thu, Aug 7, 2014 at 11:25 AM, Telles Nobrega >> >> >><[email protected]> >> >> >> wrote: >> >> >> >> >> >> > I changed my properties a little to look like this: link >> >> >> > < >> >> >> > >> >> >> >> >> >> >> >> >> >> >> http://mail-archives.apache.org/mod_mbox/samza-dev/201311.mbox/%3CEA1B8C3 >> >> >>[email protected]%3E >> >> >> > > >> >> >> > >> >> >> > here it goes: >> >> >> > >> >> >> > # Job >> >> >> > job.factory.class=org.apache.samza.job.yarn.YarnJobFactory >> >> >> > job.name=consumptions >> >> >> > >> >> >> > # YARN >> >> >> > >> >> >> > >> >> >> >> >> >>yarn.package.path=file://${basedir}/target/${project.artifactId}-${ >> >> pom.ve >> >> >>rsion}-dist.tar.gz >> >> >> > >> >> >> > # Task >> >> >> > task.class=alarm.ConsumptionProducer >> >> >> > task.inputs=kafka.consumptions >> >> >> > >> >> >> > >> >> >> >> >> >> >> >>>>task.checkpoint.factory=org.apache.samza.checkpoint.kafka.KafkaCheckpoi >> >>>>nt >> >> >>ManagerFactory >> >> >> > task.checkpoint.system=kafka >> >> >> > # Normally, this would be 3, but we have only one broker. >> >> >> > task.checkpoint.replication.factor=1 >> >> >> > >> >> >> > # Metrics >> >> >> > metrics.reporters=snapshot,jmx >> >> >> > >> >> >> > >> >> >> >> >> >> >> >>>>metrics.reporter.snapshot.class=org.apache.samza.metrics.reporter.Metri >> >>>>cs >> >> >>SnapshotReporterFactory >> >> >> > metrics.reporter.snapshot.stream=kafka.metrics >> >> >> > >> >> >> > >> >> >> >> >> >> >> >>>>metrics.reporter.jmx.class=org.apache.samza.metrics.reporter.JmxReporte >> >>>>rF >> >> >>actory >> >> >> > >> >> >> > # Serializers >> >> >> > >> >> >> > >> >> >> >> >> >> >> >>>>serializers.registry.string.class=org.apache.samza.serializers.StringSe >> >>>>rd >> >> >>eFactory >> >> >> > >> >> >> > >> >> >> >> >> >> >> >>>>serializers.registry.metrics.class=org.apache.samza.serializers.Metrics >> >>>>Sn >> >> >>apshotSerdeFactory >> >> >> > >> >> >> > # Kafka System >> >> >> > >> >> >> > >> >> >> >> >> >> >> >>>>systems.kafka.samza.factory=org.apache.samza.system.kafka.KafkaSystemFa >> >>>>ct >> >> >>ory >> >> >> > systems.kafka.samza.msg.serde=string >> >> >> > systems.kafka.consumer.zookeeper.connect=localhost:2181/ >> >> >> > systems.kafka.consumer.auto.offset.reset=largest >> >> >> > systems.kafka.producer.metadata.broker.list=localhost:9092 >> >> >> > systems.kafka.producer.producer.type=sync >> >> >> > # Normally, we'd set this much higher, but we want things to look >> >> >>snappy >> >> >> in >> >> >> > the demo. >> >> >> > systems.kafka.producer.batch.num.messages=1 >> >> >> > ystems.kafka.streams.metrics.samza.msg.serde=metrics >> >> >> > >> >> >> > But I'm getting this output, >> >> >> > >> >> >> > 0Caught exception in process loop. >> >> >> > java.lang.ClassCastException: java.util.HashMap cannot be cast to >> >> >> > java.lang.String >> >> >> > at >> >> >> >> >>org.apache.samza.serializers.StringSerde.toBytes(StringSerde.scala:33) >> >> >> > at >> >> >> >> >>org.apache.samza.serializers.SerdeManager.toBytes(SerdeManager.scala:69) >> >> >> > at >> >> >> >> >>org.apache.samza.system.SystemProducers.send(SystemProducers.scala:65) >> >> >> > at >> >> >> > >> >> >> > >> >> >> >> >> >> >> >>>>org.apache.samza.container.TaskInstance$$anonfun$send$2.apply(TaskInsta >> >>>>nc >> >> >>e.scala:170) >> >> >> > at >> >> >> > >> >> >> > >> >> >> >> >> >> >> >>>>org.apache.samza.container.TaskInstance$$anonfun$send$2.apply(TaskInsta >> >>>>nc >> >> >>e.scala:170) >> >> >> > at >> >> >> > >> >> >> > >> >> >> >> >> >> >> >>>>scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.sc >> >>>>al >> >> >>a:59) >> >> >> > at >> >>scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47) >> >> >> > at >> >> >>org.apache.samza.container.TaskInstance.send(TaskInstance.scala:170) >> >> >> > at >> >> >> > >> >> >> >> >> >> >> >>>>org.apache.samza.container.RunLoop$$anonfun$send$2.apply(RunLoop.scala: >> >>>>11 >> >> >>6) >> >> >> > at >> >> >> > >> >> >> >> >> >> >> >>>>org.apache.samza.container.RunLoop$$anonfun$send$2.apply(RunLoop.scala: >> >>>>11 >> >> >>6) >> >> >> > at scala.collection.Iterator$class.foreach(Iterator.scala:727) >> >> >> > at scala.collection.AbstractIterator.foreach(Iterator.scala:1157) >> >> >> > at >> >> >> > >> >> >> >> >>>>scala.collection.MapLike$DefaultValuesIterable.foreach(MapLike.scala:20 >> >>>>6) >> >> >> > at org.apache.samza.container.RunLoop.send(RunLoop.scala:116) >> >> >> > at org.apache.samza.container.RunLoop.run(RunLoop.scala:59) >> >> >> > at >> >> >> >> >>org.apache.samza.container.SamzaContainer.run(SamzaContainer.scala:504) >> >> >> > at >> >> >> > >> >> >> >> >>>>org.apache.samza.container.SamzaContainer$.main(SamzaContainer.scala:81 >> >>>>) >> >> >> > at >> >> >>org.apache.samza.container.SamzaContainer.main(SamzaContainer.scala) >> >> >> > >> >> >> > >> >> >> > >> >> >> > On Thu, Aug 7, 2014 at 3:12 PM, Telles Nobrega >> >> >><[email protected]> >> >> >> > wrote: >> >> >> > >> >> >> > > Hi Chris, I really appreciate the time you are taking to help me >> >> >>out. >> >> >> > > >> >> >> > > This is job.properties file >> >> >> > > >> >> >> > > # Job >> >> >> > > job.factory.class=org.apache.samza.job.yarn.YarnJobFactory >> >> >> > > job.name=consumptions >> >> >> > > >> >> >> > > # YARN >> >> >> > > >> >> >> > > >> >> >> > >> >> >> >> >> >>yarn.package.path=file://${basedir}/target/${project.artifactId}-${ >> >> pom.ve >> >> >>rsion}-dist.tar.gz >> >> >> > > >> >> >> > > # Task >> >> >> > > task.class=alarm.ConsumptionProducer >> >> >> > > task.inputs=kafka.consumptions >> >> >> > > >> >> >> > > >> >> >> > >> >> >> >> >> >> >> >>>>task.checkpoint.factory=org.apache.samza.checkpoint.kafka.KafkaCheckpoi >> >>>>nt >> >> >>ManagerFactory >> >> >> > > task.checkpoint.system=kafka >> >> >> > > # Normally, this would be 3, but we have only one broker. >> >> >> > > task.checkpoint.replication.factor=1 >> >> >> > > >> >> >> > > # Serializers >> >> >> > > >> >> >> > > >> >> >> > >> >> >> >> >> >> >> >>>>serializers.registry.serde.class=org.apache.samza.serializers.StringSer >> >>>>de >> >> >>Factory >> >> >> > > >> >> >> > > # Kafka System >> >> >> > > >> >> >> > > >> >> >> > >> >> >> >> >> >> >> >>>>systems.kafka.samza.factory=org.apache.samza.system.kafka.KafkaSystemFa >> >>>>ct >> >> >>ory >> >> >> > > *systems.kafka.samza.msg.serde=json* >> >> >> > > systems.kafka.consumer.zookeeper.connect=localhost:2181/ >> >> >> > > systems.kafka.consumer.auto.offset.reset=largest >> >> >> > > systems.kafka.producer.metadata.broker.list=localhost:9092 >> >> >> > > systems.kafka.producer.producer.type=sync >> >> >> > > # Normally, we'd set this much higher, but we want things to >> look >> >> >> snappy >> >> >> > > in the demo. >> >> >> > > systems.kafka.producer.batch.num.messages=1 >> >> >> > > >> >> >> > > *systems.kafka.streams.consumptions.key.serde=string* >> >> >> > > *systems.kafka.streams.consumptions.msg.serde=string* >> >> >> > > >> >> >> > > Does this look right? >> >> >> > > I'm running a local cluster, I want to have it running nicely >> >> >>before I >> >> >> > can >> >> >> > > distribute it. >> >> >> > > >> >> >> > > >> >> >> > > >> >> >> > > On Thu, Aug 7, 2014 at 3:08 PM, Chris Riccomini < >> >> >> > > [email protected]> wrote: >> >> >> > > >> >> >> > >> Hey Telles, >> >> >> > >> >> >> >> > >> Sure. In your job.properties file, define the serde: >> >> >> > >> >> >> >> > >> # Serializers >> >> >> > >> >> >> >> > >> >> >> >> > >> >> >> >> >> >> >> >>>>serializers.registry.serde.class=org.apache.samza.serializers.StringSer >> >>>>de >> >> >>Fa >> >> >> > >> ctory >> >> >> > >> >> >> >> > >> >> >> >> > >> Then define the serde for your system: >> >> >> > >> >> >> >> > >> systems.kafka.samza.msg.serde=string >> >> >> > >> >> >> >> > >> >> >> >> > >> Cheers, >> >> >> > >> Chris >> >> >> > >> >> >> >> > >> On 8/7/14 10:54 AM, "Telles Nobrega" <[email protected]> >> >> >>wrote: >> >> >> > >> >> >> >> > >> >Can you give and example on how to use string serde, i'm >> >>getting >> >> >>an >> >> >> > error >> >> >> > >> >when trying to set to string >> >> >> > >> > >> >> >> > >> >:53:26:804Got system producers: Set(kafka) >> >> >> > >> >17:53:26:809Got serdes: Set(string) >> >> >> > >> >17:53:29:206Container container_1407433587782_0001_01_000017 >> >> >>failed >> >> >> > with >> >> >> > >> >exit code 1 - Exception from container-launch: >> >> >> > >> > >> >> >> > >> > >> >> >> > >> > >> >> >> > >> >On Thu, Aug 7, 2014 at 2:41 PM, Telles Nobrega < >> >> >> > [email protected]> >> >> >> > >> >wrote: >> >> >> > >> > >> >> >> > >> >> Thanks. >> >> >> > >> >> >> >> >> > >> >> >> >> >> > >> >> On Thu, Aug 7, 2014 at 1:54 PM, Chris Riccomini < >> >> >> > >> >> [email protected]> wrote: >> >> >> > >> >> >> >> >> > >> >>> Hey Telles, >> >> >> > >> >>> >> >> >> > >> >>> This is definitely a serde error. It sounds like your >> >>message >> >> >>is >> >> >> not >> >> >> > >> >>> properly formatted as a JSON blob. >> >> >> > >> >>> >> >> >> > >> >>> If you are trying to just use a string as the message (vs. >> a >> >> >>well >> >> >> > >> >>> formatted JSON blob), then you should use the StringSerde. >> >> >> > >> >>> >> >> >> > >> >>> Cheers, >> >> >> > >> >>> Chris >> >> >> > >> >>> >> >> >> > >> >>> On 8/7/14 8:05 AM, "Telles Nobrega" >> >><[email protected]> >> >> >> > wrote: >> >> >> > >> >>> >> >> >> > >> >>> >Hi, I'm running a simple samza topology that reads from a >> >> >>kafka >> >> >> > >> topic >> >> >> > >> >>> >that >> >> >> > >> >>> >only has two Strings >> >> >> > >> >>> >xx:xx:xx:xxxx;xx >> >> >> > >> >>> >And its throwing an error >> >> >> > >> >>> > >> >> >> > >> >>> >Caught exception in process loop. >> >> >> > >> >>> >org.codehaus.jackson.JsonParseException: Unexpected >> >>character >> >> >> ('F' >> >> >> > >> >>>(code >> >> >> > >> >>> >70)): expected a valid value (number, String, array, >> >>object, >> >> >> > 'true', >> >> >> > >> >>> >'false' or 'null') >> >> >> > >> >>> > at [Source: [B@56dfb465; line: 1, column: 2] >> >> >> > >> >>> >at >> >> >> > >> >> >> >> >> >>>>>org.codehaus.jackson.JsonParser._constructError(JsonParser.java:1291) >> >> >> > >> >>> > at >> >> >> > >> >>> >> >> >> > >> >>> >> >> >> > >> >> >> >> > >> >> >> >> > >> >> >> >> >> >> >> >>>>>>>>org.codehaus.jackson.impl.JsonParserMinimalBase._reportError(JsonPa >> >>>>>>>>rs >> >> >>>>>>er >> >> >> > >> >>>>Min >> >> >> > >> >>> >imalBase.java:385) >> >> >> > >> >>> >at >> >> >> > >> >>> >> >> >> > >> >>> >> >> >> > >> >> >> >> > >> >> >> >> > >> >> >> >> >> >> >> >>>>>>>>org.codehaus.jackson.impl.JsonParserMinimalBase._reportUnexpectedCh >> >>>>>>>>ar >> >> >>>>>>(J >> >> >> > >> >>>>son >> >> >> > >> >>> >ParserMinimalBase.java:306) >> >> >> > >> >>> > at >> >> >> > >> >>> >> >> >> > >> >>> >> >> >> > >> >> >> >> > >> >> >> >> > >> >> >> >> >> >> >> >>>>>>>>org.codehaus.jackson.impl.Utf8StreamParser._handleUnexpectedValue(U >> >>>>>>>>tf >> >> >>>>>>8S >> >> >> > >> >>>>tre >> >> >> > >> >>> >amParser.java:1581) >> >> >> > >> >>> >at >> >> >> > >> >>> >> >> >> > >> >>> >> >> >> > >> >> >> >> > >> >> >> >> > >> >> >> >> >> >> >> >>>>>>>>org.codehaus.jackson.impl.Utf8StreamParser._nextTokenNotInObject(Ut >> >>>>>>>>f8 >> >> >>>>>>St >> >> >> > >> >>>>rea >> >> >> > >> >>> >mParser.java:436) >> >> >> > >> >>> > at >> >> >> > >> >>> >> >> >> > >> >>> >> >> >> > >> >> >> >> > >> >> >> >> > >> >> >> >> >> >> >> >>>>>>>>org.codehaus.jackson.impl.Utf8StreamParser.nextToken(Utf8StreamPars >> >>>>>>>>er >> >> >>>>>>.j >> >> >> > >> >>>>ava >> >> >> > >> >>> >:322) >> >> >> > >> >>> >at >> >> >> > >> >>> >> >> >> > >> >>> >> >> >> > >> >> >> >> > >> >> >> >> > >> >> >> >> >> >> >> >>>>>>>>org.codehaus.jackson.map.ObjectMapper._initForReading(ObjectMapper. >> >>>>>>>>ja >> >> >>>>>>va >> >> >> > >> >>>>:24 >> >> >> > >> >>> >32) >> >> >> > >> >>> > at >> >> >> > >> >>> >> >> >> > >> >>> >> >> >> > >> >> >> >> > >> >> >> >> > >> >> >> >> >> >> >> >>>>>>>>org.codehaus.jackson.map.ObjectMapper._readMapAndClose(ObjectMapper >> >>>>>>>>.j >> >> >>>>>>av >> >> >> > >> >>>>a:2 >> >> >> > >> >>> >389) >> >> >> > >> >>> >at >> >> >> > >> >>> >> >> >> > >> >> >> >> >> >>org.codehaus.jackson.map.ObjectMapper.readValue(ObjectMapper.java:1667) >> >> >> > >> >>> > at >> >> >> > >> >> >> >> >> >>>>>org.apache.samza.serializers.JsonSerde.fromBytes(JsonSerde.scala:33) >> >> >> > >> >>> >at >> >> >> > >> >>> >> >> >> > >> >>> >> >> >> > >> >> >> >> > >> >> >> >> > >> >> >> >> >> >> >> >>>>>>>>org.apache.samza.serializers.SerdeManager.fromBytes(SerdeManager.sc >> >>>>>>>>al >> >> >>>>>>a: >> >> >> > >> >>>>115 >> >> >> > >> >>> >) >> >> >> > >> >>> > at >> >> >> > >> >>> >> >> >> > >> >>> >> >> >> > >> >> >> >> > >> >> >> >> > >> >> >> >> >> >> >> >>>>>>>>org.apache.samza.system.SystemConsumers$$anonfun$org$apache$samza$s >> >>>>>>>>ys >> >> >>>>>>te >> >> >> > >> >>>>m$S >> >> >> > >> >>> >ystemConsumers$$poll$5.apply(SystemConsumers.scala:245) >> >> >> > >> >>> > at >> >> >> > >> >>> >> >> >> > >> >>> >> >> >> > >> >> >> >> > >> >> >> >> > >> >> >> >> >> >> >> >>>>>>>>org.apache.samza.system.SystemConsumers$$anonfun$org$apache$samza$s >> >>>>>>>>ys >> >> >>>>>>te >> >> >> > >> >>>>m$S >> >> >> > >> >>> >ystemConsumers$$poll$5.apply(SystemConsumers.scala:242) >> >> >> > >> >>> >at >> >>scala.collection.Iterator$class.foreach(Iterator.scala:727) >> >> >> > >> >>> > at >> >> >> scala.collection.AbstractIterator.foreach(Iterator.scala:1157) >> >> >> > >> >>> >at >> >> >> > scala.collection.IterableLike$class.foreach(IterableLike.scala:72) >> >> >> > >> >>> > at >> >> >>scala.collection.AbstractIterable.foreach(Iterable.scala:54) >> >> >> > >> >>> >at org.apache.samza.system.SystemConsumers.org >> >> >> > >> >>> >> >> >> > >> >>>$apache$samza$system$SystemConsumers$$poll(SystemConsumers.scala:242) >> >> >> > >> >>> > at >> >> >> > >> >>> >> >> >> > >> >>> >> >> >> > >> >> >> >> > >> >> >> >> > >> >> >> >> >> >> >> >>>>>>>>org.apache.samza.system.SystemConsumers$$anon$1$$anonfun$call$2.app >> >>>>>>>>ly >> >> >>>>>>(S >> >> >> > >> >>>>yst >> >> >> > >> >>> >emConsumers.scala:180) >> >> >> > >> >>> >at >> >> >> > >> >>> >> >> >> > >> >>> >> >> >> > >> >> >> >> > >> >> >> >> > >> >> >> >> >> >> >> >>>>>>>>org.apache.samza.system.SystemConsumers$$anon$1$$anonfun$call$2.app >> >>>>>>>>ly >> >> >>>>>>(S >> >> >> > >> >>>>yst >> >> >> > >> >>> >emConsumers.scala:180) >> >> >> > >> >>> > at >> >> >> > >> >>> >> >> >> > >> >>> >> >> >> > >> >> >> >> > >> >> >> >> > >> >> >> >> >> >> >> >>>>>>>>scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLi >> >>>>>>>>ke >> >> >>>>>>.s >> >> >> > >> >>>>cal >> >> >> > >> >>> >a:244) >> >> >> > >> >>> >at >> >> >> > >> >>> >> >> >> > >> >>> >> >> >> > >> >> >> >> > >> >> >> >> > >> >> >> >> >> >> >> >>>>>>>>scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLi >> >>>>>>>>ke >> >> >>>>>>.s >> >> >> > >> >>>>cal >> >> >> > >> >>> >a:244) >> >> >> > >> >>> > at >> >> >>scala.collection.Iterator$class.foreach(Iterator.scala:727) >> >> >> > >> >>> >at >> >> >>scala.collection.AbstractIterator.foreach(Iterator.scala:1157) >> >> >> > >> >>> > at >> >> >> > scala.collection.MapLike$DefaultKeySet.foreach(MapLike.scala:174) >> >> >> > >> >>> >at >> >> >> > >> >> >> >> >> >>>>>scala.collection.TraversableLike$class.map(TraversableLike.scala:244) >> >> >> > >> >>> > at >> >> >> > >> >>> >> >> >> > >> >>> >> >> >> > >> >> >> >> > >> >> >> >> > >> >> >> >> >> >> >> >>>>>>>>scala.collection.AbstractSet.scala$collection$SetLike$$super$map(Se >> >>>>>>>>t. >> >> >>>>>>sc >> >> >> > >> >>>>ala >> >> >> > >> >>> >:47) >> >> >> > >> >>> >at scala.collection.SetLike$class.map(SetLike.scala:93) >> >> >> > >> >>> > at scala.collection.AbstractSet.map(Set.scala:47) >> >> >> > >> >>> >at >> >> >> > >> >>> >> >> >> > >> >>> >> >> >> > >> >> >> >> > >> >> >> >> > >> >> >> >> >> >> >> >>>>>>>>org.apache.samza.system.SystemConsumers$$anon$1.call(SystemConsumer >> >>>>>>>>s. >> >> >>>>>>sc >> >> >> > >> >>>>ala >> >> >> > >> >>> >:180) >> >> >> > >> >>> > at >> >> >> > >> >>> >> >> >> > >> >> >> >> > >> >> >> >> > >> >> >> >> >> >> >> >>>>>>>>org.apache.samza.util.DoublingBackOff.maybeCall(DoublingBackOff.sca >> >>>>>>>>la >> >> >>>>>>:4 >> >> >> > >> >>>>4) >> >> >> > >> >>> >at >> >> >> > >> >>> >> >> >> > >> >> >> >> > >> >> >> >> > >> >> >> >> >> >> >> >>>>>>>>org.apache.samza.system.SystemConsumers.choose(SystemConsumers.scal >> >>>>>>>>a: >> >> >>>>>>20 >> >> >> > >> >>>>8) >> >> >> > >> >>> > at >> >> >>org.apache.samza.container.RunLoop.process(RunLoop.scala:73) >> >> >> > >> >>> >at >> org.apache.samza.container.RunLoop.run(RunLoop.scala:57) >> >> >> > >> >>> > at >> >> >> > >> >>> >> >> >> > >> >> >> >> > >> >> >> >> > >> >> >> >> >> >> >> >>>>>>>>org.apache.samza.container.SamzaContainer.run(SamzaContainer.scala: >> >>>>>>>>50 >> >> >>>>>>4) >> >> >> > >> >>> >at >> >> >> > >> >>> >> >> >> > >> >> >> >> > >> >> >> >> > >> >> >> >> >> >> >> >>>>>>>>org.apache.samza.container.SamzaContainer$.main(SamzaContainer.scal >> >>>>>>>>a: >> >> >>>>>>81 >> >> >> > >> >>>>) >> >> >> > >> >>> > at >> >> >> > >> >> >> >> >> >>>>>org.apache.samza.container.SamzaContainer.main(SamzaContainer.scala) >> >> >> > >> >>> > >> >> >> > >> >>> > >> >> >> > >> >>> > >> >> >> > >> >>> >Has anyone experienced this error before? >> >> >> > >> >>> > >> >> >> > >> >>> >-- >> >> >> > >> >>> >------------------------------------------ >> >> >> > >> >>> >Telles Mota Vidal Nobrega >> >> >> > >> >>> >M.sc. Candidate at UFCG >> >> >> > >> >>> >B.sc. in Computer Science at UFCG >> >> >> > >> >>> >Software Engineer at OpenStack Project - HP/LSD-UFCG >> >> >> > >> >>> >> >> >> > >> >>> >> >> >> > >> >> >> >> >> > >> >> >> >> >> > >> >> -- >> >> >> > >> >> ------------------------------------------ >> >> >> > >> >> Telles Mota Vidal Nobrega >> >> >> > >> >> M.sc. Candidate at UFCG >> >> >> > >> >> B.sc. in Computer Science at UFCG >> >> >> > >> >> Software Engineer at OpenStack Project - HP/LSD-UFCG >> >> >> > >> >> >> >> >> > >> > >> >> >> > >> > >> >> >> > >> > >> >> >> > >> >-- >> >> >> > >> >------------------------------------------ >> >> >> > >> >Telles Mota Vidal Nobrega >> >> >> > >> >M.sc. Candidate at UFCG >> >> >> > >> >B.sc. in Computer Science at UFCG >> >> >> > >> >Software Engineer at OpenStack Project - HP/LSD-UFCG >> >> >> > >> >> >> >> > >> >> >> >> > > >> >> >> > > >> >> >> > > -- >> >> >> > > ------------------------------------------ >> >> >> > > Telles Mota Vidal Nobrega >> >> >> > > M.sc. Candidate at UFCG >> >> >> > > B.sc. in Computer Science at UFCG >> >> >> > > Software Engineer at OpenStack Project - HP/LSD-UFCG >> >> >> > > >> >> >> > >> >> >> > >> >> >> > >> >> >> > -- >> >> >> > ------------------------------------------ >> >> >> > Telles Mota Vidal Nobrega >> >> >> > M.sc. Candidate at UFCG >> >> >> > B.sc. in Computer Science at UFCG >> >> >> > Software Engineer at OpenStack Project - HP/LSD-UFCG >> >> >> > >> >> >> >> >> > >> >> > >> >> > >> >> >-- >> >> >------------------------------------------ >> >> >Telles Mota Vidal Nobrega >> >> >M.sc. Candidate at UFCG >> >> >B.sc. in Computer Science at UFCG >> >> >Software Engineer at OpenStack Project - HP/LSD-UFCG >> >> >> >> >> > >> > >> >-- >> >------------------------------------------ >> >Telles Mota Vidal Nobrega >> >M.sc. Candidate at UFCG >> >B.sc. in Computer Science at UFCG >> >Software Engineer at OpenStack Project - HP/LSD-UFCG >> >> > > > -- > ------------------------------------------ > Telles Mota Vidal Nobrega > M.sc. Candidate at UFCG > B.sc. in Computer Science at UFCG > Software Engineer at OpenStack Project - HP/LSD-UFCG > -- ------------------------------------------ Telles Mota Vidal Nobrega M.sc. Candidate at UFCG B.sc. in Computer Science at UFCG Software Engineer at OpenStack Project - HP/LSD-UFCG
