I see, thanks for the quick response and help.
On 07 Aug 2014, at 16:38, Chris Riccomini <[email protected]> wrote: > Hey Telles, > > Great! :) > > We don't usually hit this issue because we use a uniform serialization for > both inputs and outputs. In your case, because input is String and output > is JSON, you have to set things up accordingly. > > Cheers, > Chris > > On 8/7/14 12:36 PM, "Telles Nobrega" <[email protected]> wrote: > >> It worked. >> On 07 Aug 2014, at 16:20, Chris Riccomini >> <[email protected]> wrote: >> >>> Hey Telles, >>> >>> Yea, as Yan suggested, you're sending a map back to the Kafka system: >>> >>> >>> Map<String, Object> outgoingMap = Event.toMap(event); >>> collector.send(new OutgoingMessageEnvelope(OUTPUT_STREAM, outgoingMap)); >>> >>> >>> Samza is going to take the object (outgoingMap) and try to encode it >>> using >>> StringSerde. It won't be able to do this, since the object you've given >>> it >>> is a map, not a string. >>> >>> The fix for this is to configure the default serde as >>> StringSerdeFactory, >>> which you've done, and to configure the output stream ("values") using >>> the >>> JsonSerdeFactory. This can be done with: >>> >>> systems.kafka.streams.values.samza.msg.serde=json >>> >>> >>> And then defining the json serde, if you haven't already done so: >>> >>> >>> serializers.registry.json.class=org.apache.samza.serializers.JsonSerdeFac >>> to >>> ry >>> >>> >>> Cheers, >>> Chris >>> >>> >>> >>> On 8/7/14 12:12 PM, "Telles Nobrega" <[email protected]> wrote: >>> >>>> Here it goes: http://pastebin.com/hhswEnyW >>>> >>>> >>>> On Thu, Aug 7, 2014 at 4:10 PM, Chris Riccomini < >>>> [email protected]> wrote: >>>> >>>>> Hey Telles, >>>>> >>>>> Can you paste the code for your StreamTask? If it's still showing the >>>>> same >>>>> message, then it sounds like you're still trying to send a HashMap >>>>> using a >>>>> StringSerde. >>>>> >>>>> Cheers, >>>>> Chris >>>>> >>>>> On 8/7/14 12:05 PM, "Telles Nobrega" <[email protected]> wrote: >>>>> >>>>>> Still showing the same message >>>>>> >>>>>> >>>>>> On Thu, Aug 7, 2014 at 3:56 PM, Chris Riccomini < >>>>>> [email protected]> wrote: >>>>>> >>>>>>> Hey Telles, >>>>>>> >>>>>>> The code you've posted in Produce.java shows: >>>>>>> >>>>>>> KeyedMessage<String, String> data = new >>>>> KeyedMessage<String, >>>>>>> String>("consumptions", String.valueOf(key),String.valueOf(value)); >>>>>>> >>>>>>> >>>>>>> Which suggests that you are sending a string for both key and value. >>>>> If >>>>>>> you have a Samza task consuming from this topic, you should set: >>>>>>> >>>>>>> systems.system-name.samza.key.serde=string >>>>>>> >>>>>>> systems.system-name.samza.msg.serde=string >>>>>>> >>>>>>> >>>>>>> Cheers, >>>>>>> Chris >>>>>>> >>>>>>> On 8/7/14 11:52 AM, "Telles Nobrega" <[email protected]> >>>>>>> wrote: >>>>>>> >>>>>>>> Hum, that sounds like a perfect reason for it. >>>>>>>> >>>>>>>> I'm writing to the topic with this code >>>>>>>> >>>>>>>> >>>>>>> >>>>>>> >>>>> >>>>> >>>>> https://github.com/tellesnobrega/kafka-producer/blob/master/src/Produce >>>>> .j >>>>>>> a >>>>>>>> va >>>>>>>> >>>>>>>> My problem is that I need to send numbers as key and value to the >>>>> kafka >>>>>>>> topic so i can read it in samza. >>>>>>>> >>>>>>>> What is the best way to de/serialize this? >>>>>>>> >>>>>>>> >>>>>>>> On Thu, Aug 7, 2014 at 3:43 PM, Yan Fang <[email protected]> >>>>> wrote: >>>>>>>> >>>>>>>>> Hi Telles, >>>>>>>>> >>>>>>>>> One of the possible reasons is that, in your process method, you >>>>> are >>>>>>>>> trying >>>>>>>>> to send a HashMap, not a String, in the collection.send. Could you >>>>>>>>> check it >>>>>>>>> ? >>>>>>>>> >>>>>>>>> Thanks, >>>>>>>>> >>>>>>>>> Fang, Yan >>>>>>>>> [email protected] >>>>>>>>> +1 (206) 849-4108 >>>>>>>>> >>>>>>>>> >>>>>>>>> On Thu, Aug 7, 2014 at 11:25 AM, Telles Nobrega >>>>>>>>> <[email protected]> >>>>>>>>> wrote: >>>>>>>>> >>>>>>>>>> I changed my properties a little to look like this: link >>>>>>>>>> < >>>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>> >>>>>>> >>>>> >>>>> >>>>> http://mail-archives.apache.org/mod_mbox/samza-dev/201311.mbox/%3CEA1B8 >>>>> C3 >>>>>>>>> [email protected]%3E >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> here it goes: >>>>>>>>>> >>>>>>>>>> # Job >>>>>>>>>> job.factory.class=org.apache.samza.job.yarn.YarnJobFactory >>>>>>>>>> job.name=consumptions >>>>>>>>>> >>>>>>>>>> # YARN >>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> yarn.package.path=file://${basedir}/target/${project.artifactId}-${ >>>>>>> pom.ve >>>>>>>>> rsion}-dist.tar.gz >>>>>>>>>> >>>>>>>>>> # Task >>>>>>>>>> task.class=alarm.ConsumptionProducer >>>>>>>>>> task.inputs=kafka.consumptions >>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>> >>>>> >>>>>>>>> >>>>>>>>> task.checkpoint.factory=org.apache.samza.checkpoint.kafka.KafkaChec >>>>>>>>> kp >>>>>>>>> oi >>>>>>>>> nt >>>>>>>>> ManagerFactory >>>>>>>>>> task.checkpoint.system=kafka >>>>>>>>>> # Normally, this would be 3, but we have only one broker. >>>>>>>>>> task.checkpoint.replication.factor=1 >>>>>>>>>> >>>>>>>>>> # Metrics >>>>>>>>>> metrics.reporters=snapshot,jmx >>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>> >>>>> >>>>>>>>> >>>>>>>>> metrics.reporter.snapshot.class=org.apache.samza.metrics.reporter.M >>>>>>>>> et >>>>>>>>> ri >>>>>>>>> cs >>>>>>>>> SnapshotReporterFactory >>>>>>>>>> metrics.reporter.snapshot.stream=kafka.metrics >>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>> >>>>> >>>>>>>>> >>>>>>>>> metrics.reporter.jmx.class=org.apache.samza.metrics.reporter.JmxRep >>>>>>>>> or >>>>>>>>> te >>>>>>>>> rF >>>>>>>>> actory >>>>>>>>>> >>>>>>>>>> # Serializers >>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>> >>>>> >>>>>>>>> >>>>>>>>> serializers.registry.string.class=org.apache.samza.serializers.Stri >>>>>>>>> ng >>>>>>>>> Se >>>>>>>>> rd >>>>>>>>> eFactory >>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>> >>>>> >>>>>>>>> >>>>>>>>> serializers.registry.metrics.class=org.apache.samza.serializers.Met >>>>>>>>> ri >>>>>>>>> cs >>>>>>>>> Sn >>>>>>>>> apshotSerdeFactory >>>>>>>>>> >>>>>>>>>> # Kafka System >>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>> >>>>> >>>>>>>>> >>>>>>>>> systems.kafka.samza.factory=org.apache.samza.system.kafka.KafkaSyst >>>>>>>>> em >>>>>>>>> Fa >>>>>>>>> ct >>>>>>>>> ory >>>>>>>>>> systems.kafka.samza.msg.serde=string >>>>>>>>>> systems.kafka.consumer.zookeeper.connect=localhost:2181/ >>>>>>>>>> systems.kafka.consumer.auto.offset.reset=largest >>>>>>>>>> systems.kafka.producer.metadata.broker.list=localhost:9092 >>>>>>>>>> systems.kafka.producer.producer.type=sync >>>>>>>>>> # Normally, we'd set this much higher, but we want things to >>>>> look >>>>>>>>> snappy >>>>>>>>> in >>>>>>>>>> the demo. >>>>>>>>>> systems.kafka.producer.batch.num.messages=1 >>>>>>>>>> ystems.kafka.streams.metrics.samza.msg.serde=metrics >>>>>>>>>> >>>>>>>>>> But I'm getting this output, >>>>>>>>>> >>>>>>>>>> 0Caught exception in process loop. >>>>>>>>>> java.lang.ClassCastException: java.util.HashMap cannot be cast >>>>> to >>>>>>>>>> java.lang.String >>>>>>>>>> at >>>>>>>>> >>>>>>> >>>>>>> org.apache.samza.serializers.StringSerde.toBytes(StringSerde.scala:33 >>>>>>> ) >>>>>>>>>> at >>>>>>>>> >>>>> >>>>>>> >>>>>>> org.apache.samza.serializers.SerdeManager.toBytes(SerdeManager.scala: >>>>>>> 69 >>>>>>> ) >>>>>>>>>> at >>>>>>>>> >>>>>>> >>>>>>> org.apache.samza.system.SystemProducers.send(SystemProducers.scala:65 >>>>>>> ) >>>>>>>>>> at >>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>> >>>>> >>>>>>>>> >>>>>>>>> org.apache.samza.container.TaskInstance$$anonfun$send$2.apply(TaskI >>>>>>>>> ns >>>>>>>>> ta >>>>>>>>> nc >>>>>>>>> e.scala:170) >>>>>>>>>> at >>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>> >>>>> >>>>>>>>> >>>>>>>>> org.apache.samza.container.TaskInstance$$anonfun$send$2.apply(TaskI >>>>>>>>> ns >>>>>>>>> ta >>>>>>>>> nc >>>>>>>>> e.scala:170) >>>>>>>>>> at >>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>> >>>>> >>>>>>>>> >>>>>>>>> scala.collection.mutable.ResizableArray$class.foreach(ResizableArra >>>>>>>>> y. >>>>>>>>> sc >>>>>>>>> al >>>>>>>>> a:59) >>>>>>>>>> at >>>>>>> scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47) >>>>>>>>>> at >>>>>>> >>>>>>> org.apache.samza.container.TaskInstance.send(TaskInstance.scala:170) >>>>>>>>>> at >>>>>>>>>> >>>>>>>>> >>>>>>> >>>>> >>>>>>>>> >>>>>>>>> org.apache.samza.container.RunLoop$$anonfun$send$2.apply(RunLoop.sc >>>>>>>>> al >>>>>>>>> a: >>>>>>>>> 11 >>>>>>>>> 6) >>>>>>>>>> at >>>>>>>>>> >>>>>>>>> >>>>>>> >>>>> >>>>>>>>> >>>>>>>>> org.apache.samza.container.RunLoop$$anonfun$send$2.apply(RunLoop.sc >>>>>>>>> al >>>>>>>>> a: >>>>>>>>> 11 >>>>>>>>> 6) >>>>>>>>>> at scala.collection.Iterator$class.foreach(Iterator.scala:727) >>>>>>>>>> at >>>>> scala.collection.AbstractIterator.foreach(Iterator.scala:1157) >>>>>>>>>> at >>>>>>>>>> >>>>>>> >>>>> >>>>>>>>> >>>>>>>>> scala.collection.MapLike$DefaultValuesIterable.foreach(MapLike.scal >>>>>>>>> a: >>>>>>>>> 20 >>>>>>>>> 6) >>>>>>>>>> at org.apache.samza.container.RunLoop.send(RunLoop.scala:116) >>>>>>>>>> at org.apache.samza.container.RunLoop.run(RunLoop.scala:59) >>>>>>>>>> at >>>>>>>>> >>>>> >>>>>>> >>>>>>> org.apache.samza.container.SamzaContainer.run(SamzaContainer.scala:50 >>>>>>> 4) >>>>>>>>>> at >>>>>>>>>> >>>>>>> >>>>> >>>>>>>>> >>>>>>>>> org.apache.samza.container.SamzaContainer$.main(SamzaContainer.scal >>>>>>>>> a: >>>>>>>>> 81 >>>>>>>>> ) >>>>>>>>>> at >>>>>>> >>>>>>> org.apache.samza.container.SamzaContainer.main(SamzaContainer.scala) >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> On Thu, Aug 7, 2014 at 3:12 PM, Telles Nobrega >>>>>>>>> <[email protected]> >>>>>>>>>> wrote: >>>>>>>>>> >>>>>>>>>>> Hi Chris, I really appreciate the time you are taking to help >>>>> me >>>>>>>>> out. >>>>>>>>>>> >>>>>>>>>>> This is job.properties file >>>>>>>>>>> >>>>>>>>>>> # Job >>>>>>>>>>> job.factory.class=org.apache.samza.job.yarn.YarnJobFactory >>>>>>>>>>> job.name=consumptions >>>>>>>>>>> >>>>>>>>>>> # YARN >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> yarn.package.path=file://${basedir}/target/${project.artifactId}-${ >>>>>>> pom.ve >>>>>>>>> rsion}-dist.tar.gz >>>>>>>>>>> >>>>>>>>>>> # Task >>>>>>>>>>> task.class=alarm.ConsumptionProducer >>>>>>>>>>> task.inputs=kafka.consumptions >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>> >>>>> >>>>>>>>> >>>>>>>>> task.checkpoint.factory=org.apache.samza.checkpoint.kafka.KafkaChec >>>>>>>>> kp >>>>>>>>> oi >>>>>>>>> nt >>>>>>>>> ManagerFactory >>>>>>>>>>> task.checkpoint.system=kafka >>>>>>>>>>> # Normally, this would be 3, but we have only one broker. >>>>>>>>>>> task.checkpoint.replication.factor=1 >>>>>>>>>>> >>>>>>>>>>> # Serializers >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>> >>>>> >>>>>>>>> >>>>>>>>> serializers.registry.serde.class=org.apache.samza.serializers.Strin >>>>>>>>> gS >>>>>>>>> er >>>>>>>>> de >>>>>>>>> Factory >>>>>>>>>>> >>>>>>>>>>> # Kafka System >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>> >>>>> >>>>>>>>> >>>>>>>>> systems.kafka.samza.factory=org.apache.samza.system.kafka.KafkaSyst >>>>>>>>> em >>>>>>>>> Fa >>>>>>>>> ct >>>>>>>>> ory >>>>>>>>>>> *systems.kafka.samza.msg.serde=json* >>>>>>>>>>> systems.kafka.consumer.zookeeper.connect=localhost:2181/ >>>>>>>>>>> systems.kafka.consumer.auto.offset.reset=largest >>>>>>>>>>> systems.kafka.producer.metadata.broker.list=localhost:9092 >>>>>>>>>>> systems.kafka.producer.producer.type=sync >>>>>>>>>>> # Normally, we'd set this much higher, but we want things to >>>>> look >>>>>>>>> snappy >>>>>>>>>>> in the demo. >>>>>>>>>>> systems.kafka.producer.batch.num.messages=1 >>>>>>>>>>> >>>>>>>>>>> *systems.kafka.streams.consumptions.key.serde=string* >>>>>>>>>>> *systems.kafka.streams.consumptions.msg.serde=string* >>>>>>>>>>> >>>>>>>>>>> Does this look right? >>>>>>>>>>> I'm running a local cluster, I want to have it running nicely >>>>>>>>> before I >>>>>>>>>> can >>>>>>>>>>> distribute it. >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> On Thu, Aug 7, 2014 at 3:08 PM, Chris Riccomini < >>>>>>>>>>> [email protected]> wrote: >>>>>>>>>>> >>>>>>>>>>>> Hey Telles, >>>>>>>>>>>> >>>>>>>>>>>> Sure. In your job.properties file, define the serde: >>>>>>>>>>>> >>>>>>>>>>>> # Serializers >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>> >>>>> >>>>>>>>> >>>>>>>>> serializers.registry.serde.class=org.apache.samza.serializers.Strin >>>>>>>>> gS >>>>>>>>> er >>>>>>>>> de >>>>>>>>> Fa >>>>>>>>>>>> ctory >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> Then define the serde for your system: >>>>>>>>>>>> >>>>>>>>>>>> systems.kafka.samza.msg.serde=string >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> Cheers, >>>>>>>>>>>> Chris >>>>>>>>>>>> >>>>>>>>>>>> On 8/7/14 10:54 AM, "Telles Nobrega" >>>>> <[email protected]> >>>>>>>>> wrote: >>>>>>>>>>>> >>>>>>>>>>>>> Can you give and example on how to use string serde, i'm >>>>>>> getting >>>>>>>>> an >>>>>>>>>> error >>>>>>>>>>>>> when trying to set to string >>>>>>>>>>>>> >>>>>>>>>>>>> :53:26:804Got system producers: Set(kafka) >>>>>>>>>>>>> 17:53:26:809Got serdes: Set(string) >>>>>>>>>>>>> 17:53:29:206Container container_1407433587782_0001_01_000017 >>>>>>>>> failed >>>>>>>>>> with >>>>>>>>>>>>> exit code 1 - Exception from container-launch: >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> On Thu, Aug 7, 2014 at 2:41 PM, Telles Nobrega < >>>>>>>>>> [email protected]> >>>>>>>>>>>>> wrote: >>>>>>>>>>>>> >>>>>>>>>>>>>> Thanks. >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> On Thu, Aug 7, 2014 at 1:54 PM, Chris Riccomini < >>>>>>>>>>>>>> [email protected]> wrote: >>>>>>>>>>>>>> >>>>>>>>>>>>>>> Hey Telles, >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> This is definitely a serde error. It sounds like your >>>>>>> message >>>>>>>>> is >>>>>>>>> not >>>>>>>>>>>>>>> properly formatted as a JSON blob. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> If you are trying to just use a string as the message >>>>> (vs. a >>>>>>>>> well >>>>>>>>>>>>>>> formatted JSON blob), then you should use the >>>>> StringSerde. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Cheers, >>>>>>>>>>>>>>> Chris >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> On 8/7/14 8:05 AM, "Telles Nobrega" >>>>>>> <[email protected]> >>>>>>>>>> wrote: >>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Hi, I'm running a simple samza topology that reads from >>>>> a >>>>>>>>> kafka >>>>>>>>>>>> topic >>>>>>>>>>>>>>>> that >>>>>>>>>>>>>>>> only has two Strings >>>>>>>>>>>>>>>> xx:xx:xx:xxxx;xx >>>>>>>>>>>>>>>> And its throwing an error >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Caught exception in process loop. >>>>>>>>>>>>>>>> org.codehaus.jackson.JsonParseException: Unexpected >>>>>>> character >>>>>>>>> ('F' >>>>>>>>>>>>>>> (code >>>>>>>>>>>>>>>> 70)): expected a valid value (number, String, array, >>>>>>> object, >>>>>>>>>> 'true', >>>>>>>>>>>>>>>> 'false' or 'null') >>>>>>>>>>>>>>>> at [Source: [B@56dfb465; line: 1, column: 2] >>>>>>>>>>>>>>>> at >>>>>>>>>>>> >>>>>>>>> >>>>> >>>>>>>>>> >>>>>>>>>> org.codehaus.jackson.JsonParser._constructError(JsonParser.java:12 >>>>>>>>>> 91 >>>>>>>>>> ) >>>>>>>>>>>>>>>> at >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>> >>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> org.codehaus.jackson.impl.JsonParserMinimalBase._reportError(Js >>>>>>>>>>>>> on >>>>>>>>>>>>> Pa >>>>>>>>>>>>> rs >>>>>>>>>>>>> er >>>>>>>>>>>>>>>> Min >>>>>>>>>>>>>>>> imalBase.java:385) >>>>>>>>>>>>>>>> at >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>> >>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> org.codehaus.jackson.impl.JsonParserMinimalBase._reportUnexpect >>>>>>>>>>>>> ed >>>>>>>>>>>>> Ch >>>>>>>>>>>>> ar >>>>>>>>>>>>> (J >>>>>>>>>>>>>>>> son >>>>>>>>>>>>>>>> ParserMinimalBase.java:306) >>>>>>>>>>>>>>>> at >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>> >>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> org.codehaus.jackson.impl.Utf8StreamParser._handleUnexpectedVal >>>>>>>>>>>>> ue >>>>>>>>>>>>> (U >>>>>>>>>>>>> tf >>>>>>>>>>>>> 8S >>>>>>>>>>>>>>>> tre >>>>>>>>>>>>>>>> amParser.java:1581) >>>>>>>>>>>>>>>> at >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>> >>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> org.codehaus.jackson.impl.Utf8StreamParser._nextTokenNotInObjec >>>>>>>>>>>>> t( >>>>>>>>>>>>> Ut >>>>>>>>>>>>> f8 >>>>>>>>>>>>> St >>>>>>>>>>>>>>>> rea >>>>>>>>>>>>>>>> mParser.java:436) >>>>>>>>>>>>>>>> at >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>> >>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> org.codehaus.jackson.impl.Utf8StreamParser.nextToken(Utf8Stream >>>>>>>>>>>>> Pa >>>>>>>>>>>>> rs >>>>>>>>>>>>> er >>>>>>>>>>>>> .j >>>>>>>>>>>>>>>> ava >>>>>>>>>>>>>>>> :322) >>>>>>>>>>>>>>>> at >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>> >>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> org.codehaus.jackson.map.ObjectMapper._initForReading(ObjectMap >>>>>>>>>>>>> pe >>>>>>>>>>>>> r. >>>>>>>>>>>>> ja >>>>>>>>>>>>> va >>>>>>>>>>>>>>>> :24 >>>>>>>>>>>>>>>> 32) >>>>>>>>>>>>>>>> at >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>> >>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> org.codehaus.jackson.map.ObjectMapper._readMapAndClose(ObjectMa >>>>>>>>>>>>> pp >>>>>>>>>>>>> er >>>>>>>>>>>>> .j >>>>>>>>>>>>> av >>>>>>>>>>>>>>>> a:2 >>>>>>>>>>>>>>>> 389) >>>>>>>>>>>>>>>> at >>>>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>> >>>>> >>>>>>> >>>>>>> org.codehaus.jackson.map.ObjectMapper.readValue(ObjectMapper.java:166 >>>>>>> 7) >>>>>>>>>>>>>>>> at >>>>>>>>>>>> >>>>>>>>> >>>>> >>>>>>>>>> >>>>>>>>>> org.apache.samza.serializers.JsonSerde.fromBytes(JsonSerde.scala:3 >>>>>>>>>> 3) >>>>>>>>>>>>>>>> at >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>> >>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> org.apache.samza.serializers.SerdeManager.fromBytes(SerdeManage >>>>>>>>>>>>> r. >>>>>>>>>>>>> sc >>>>>>>>>>>>> al >>>>>>>>>>>>> a: >>>>>>>>>>>>>>>> 115 >>>>>>>>>>>>>>>> ) >>>>>>>>>>>>>>>> at >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>> >>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> org.apache.samza.system.SystemConsumers$$anonfun$org$apache$sam >>>>>>>>>>>>> za >>>>>>>>>>>>> $s >>>>>>>>>>>>> ys >>>>>>>>>>>>> te >>>>>>>>>>>>>>>> m$S >>>>>>>>>>>>>>>> ystemConsumers$$poll$5.apply(SystemConsumers.scala:245) >>>>>>>>>>>>>>>> at >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>> >>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> org.apache.samza.system.SystemConsumers$$anonfun$org$apache$sam >>>>>>>>>>>>> za >>>>>>>>>>>>> $s >>>>>>>>>>>>> ys >>>>>>>>>>>>> te >>>>>>>>>>>>>>>> m$S >>>>>>>>>>>>>>>> ystemConsumers$$poll$5.apply(SystemConsumers.scala:242) >>>>>>>>>>>>>>>> at >>>>>>> scala.collection.Iterator$class.foreach(Iterator.scala:727) >>>>>>>>>>>>>>>> at >>>>>>>>> scala.collection.AbstractIterator.foreach(Iterator.scala:1157) >>>>>>>>>>>>>>>> at >>>>>>>>>> >>>>> scala.collection.IterableLike$class.foreach(IterableLike.scala:72) >>>>>>>>>>>>>>>> at >>>>>>>>> scala.collection.AbstractIterable.foreach(Iterable.scala:54) >>>>>>>>>>>>>>>> at org.apache.samza.system.SystemConsumers.org >>>>>>>>>>>>>>> >>>>>>>>>> >>>>>>>> >>>>>>>> $apache$samza$system$SystemConsumers$$poll(SystemConsumers.scala:242 >>>>>>>> ) >>>>>>>>>>>>>>>> at >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>> >>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> org.apache.samza.system.SystemConsumers$$anon$1$$anonfun$call$2 >>>>>>>>>>>>> .a >>>>>>>>>>>>> pp >>>>>>>>>>>>> ly >>>>>>>>>>>>> (S >>>>>>>>>>>>>>>> yst >>>>>>>>>>>>>>>> emConsumers.scala:180) >>>>>>>>>>>>>>>> at >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>> >>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> org.apache.samza.system.SystemConsumers$$anon$1$$anonfun$call$2 >>>>>>>>>>>>> .a >>>>>>>>>>>>> pp >>>>>>>>>>>>> ly >>>>>>>>>>>>> (S >>>>>>>>>>>>>>>> yst >>>>>>>>>>>>>>>> emConsumers.scala:180) >>>>>>>>>>>>>>>> at >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>> >>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> scala.collection.TraversableLike$$anonfun$map$1.apply(Traversab >>>>>>>>>>>>> le >>>>>>>>>>>>> Li >>>>>>>>>>>>> ke >>>>>>>>>>>>> .s >>>>>>>>>>>>>>>> cal >>>>>>>>>>>>>>>> a:244) >>>>>>>>>>>>>>>> at >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>> >>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> scala.collection.TraversableLike$$anonfun$map$1.apply(Traversab >>>>>>>>>>>>> le >>>>>>>>>>>>> Li >>>>>>>>>>>>> ke >>>>>>>>>>>>> .s >>>>>>>>>>>>>>>> cal >>>>>>>>>>>>>>>> a:244) >>>>>>>>>>>>>>>> at >>>>>>>>> scala.collection.Iterator$class.foreach(Iterator.scala:727) >>>>>>>>>>>>>>>> at >>>>>>>>> scala.collection.AbstractIterator.foreach(Iterator.scala:1157) >>>>>>>>>>>>>>>> at >>>>>>>>>> >>>>> scala.collection.MapLike$DefaultKeySet.foreach(MapLike.scala:174) >>>>>>>>>>>>>>>> at >>>>>>>>>>>> >>>>>>>>> >>>>> >>>>>>>>>> >>>>>>>>>> scala.collection.TraversableLike$class.map(TraversableLike.scala:2 >>>>>>>>>> 44 >>>>>>>>>> ) >>>>>>>>>>>>>>>> at >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>> >>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> scala.collection.AbstractSet.scala$collection$SetLike$$super$ma >>>>>>>>>>>>> p( >>>>>>>>>>>>> Se >>>>>>>>>>>>> t. >>>>>>>>>>>>> sc >>>>>>>>>>>>>>>> ala >>>>>>>>>>>>>>>> :47) >>>>>>>>>>>>>>>> at scala.collection.SetLike$class.map(SetLike.scala:93) >>>>>>>>>>>>>>>> at scala.collection.AbstractSet.map(Set.scala:47) >>>>>>>>>>>>>>>> at >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>> >>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> org.apache.samza.system.SystemConsumers$$anon$1.call(SystemCons >>>>>>>>>>>>> um >>>>>>>>>>>>> er >>>>>>>>>>>>> s. >>>>>>>>>>>>> sc >>>>>>>>>>>>>>>> ala >>>>>>>>>>>>>>>> :180) >>>>>>>>>>>>>>>> at >>>>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>> >>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> org.apache.samza.util.DoublingBackOff.maybeCall(DoublingBackOff >>>>>>>>>>>>> .s >>>>>>>>>>>>> ca >>>>>>>>>>>>> la >>>>>>>>>>>>> :4 >>>>>>>>>>>>>>>> 4) >>>>>>>>>>>>>>>> at >>>>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>> >>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> org.apache.samza.system.SystemConsumers.choose(SystemConsumers. >>>>>>>>>>>>> sc >>>>>>>>>>>>> al >>>>>>>>>>>>> a: >>>>>>>>>>>>> 20 >>>>>>>>>>>>>>>> 8) >>>>>>>>>>>>>>>> at >>>>>>>>> org.apache.samza.container.RunLoop.process(RunLoop.scala:73) >>>>>>>>>>>>>>>> at >>>>> org.apache.samza.container.RunLoop.run(RunLoop.scala:57) >>>>>>>>>>>>>>>> at >>>>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>> >>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> org.apache.samza.container.SamzaContainer.run(SamzaContainer.sc >>>>>>>>>>>>> al >>>>>>>>>>>>> a: >>>>>>>>>>>>> 50 >>>>>>>>>>>>> 4) >>>>>>>>>>>>>>>> at >>>>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>> >>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> org.apache.samza.container.SamzaContainer$.main(SamzaContainer. >>>>>>>>>>>>> sc >>>>>>>>>>>>> al >>>>>>>>>>>>> a: >>>>>>>>>>>>> 81 >>>>>>>>>>>>>>>> ) >>>>>>>>>>>>>>>> at >>>>>>>>>>>> >>>>>>>>> >>>>> >>>>>>>>>> >>>>>>>>>> org.apache.samza.container.SamzaContainer.main(SamzaContainer.scal >>>>>>>>>> a) >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Has anyone experienced this error before? >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>> ------------------------------------------ >>>>>>>>>>>>>>>> Telles Mota Vidal Nobrega >>>>>>>>>>>>>>>> M.sc. Candidate at UFCG >>>>>>>>>>>>>>>> B.sc. in Computer Science at UFCG >>>>>>>>>>>>>>>> Software Engineer at OpenStack Project - HP/LSD-UFCG >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> -- >>>>>>>>>>>>>> ------------------------------------------ >>>>>>>>>>>>>> Telles Mota Vidal Nobrega >>>>>>>>>>>>>> M.sc. Candidate at UFCG >>>>>>>>>>>>>> B.sc. in Computer Science at UFCG >>>>>>>>>>>>>> Software Engineer at OpenStack Project - HP/LSD-UFCG >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> -- >>>>>>>>>>>>> ------------------------------------------ >>>>>>>>>>>>> Telles Mota Vidal Nobrega >>>>>>>>>>>>> M.sc. Candidate at UFCG >>>>>>>>>>>>> B.sc. in Computer Science at UFCG >>>>>>>>>>>>> Software Engineer at OpenStack Project - HP/LSD-UFCG >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> -- >>>>>>>>>>> ------------------------------------------ >>>>>>>>>>> Telles Mota Vidal Nobrega >>>>>>>>>>> M.sc. Candidate at UFCG >>>>>>>>>>> B.sc. in Computer Science at UFCG >>>>>>>>>>> Software Engineer at OpenStack Project - HP/LSD-UFCG >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> -- >>>>>>>>>> ------------------------------------------ >>>>>>>>>> Telles Mota Vidal Nobrega >>>>>>>>>> M.sc. Candidate at UFCG >>>>>>>>>> B.sc. in Computer Science at UFCG >>>>>>>>>> Software Engineer at OpenStack Project - HP/LSD-UFCG >>>>>>>>>> >>>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> -- >>>>>>>> ------------------------------------------ >>>>>>>> Telles Mota Vidal Nobrega >>>>>>>> M.sc. Candidate at UFCG >>>>>>>> B.sc. in Computer Science at UFCG >>>>>>>> Software Engineer at OpenStack Project - HP/LSD-UFCG >>>>>>> >>>>>>> >>>>>> >>>>>> >>>>>> -- >>>>>> ------------------------------------------ >>>>>> Telles Mota Vidal Nobrega >>>>>> M.sc. Candidate at UFCG >>>>>> B.sc. in Computer Science at UFCG >>>>>> Software Engineer at OpenStack Project - HP/LSD-UFCG >>>>> >>>>> >>>> >>>> >>>> -- >>>> ------------------------------------------ >>>> Telles Mota Vidal Nobrega >>>> M.sc. Candidate at UFCG >>>> B.sc. in Computer Science at UFCG >>>> Software Engineer at OpenStack Project - HP/LSD-UFCG >>> >> >
