Here it goes: http://pastebin.com/hhswEnyW


On Thu, Aug 7, 2014 at 4:10 PM, Chris Riccomini <
[email protected]> wrote:

> Hey Telles,
>
> Can you paste the code for your StreamTask? If it's still showing the same
> message, then it sounds like you're still trying to send a HashMap using a
> StringSerde.
>
> Cheers,
> Chris
>
> On 8/7/14 12:05 PM, "Telles Nobrega" <[email protected]> wrote:
>
> >Still showing the same message
> >
> >
> >On Thu, Aug 7, 2014 at 3:56 PM, Chris Riccomini <
> >[email protected]> wrote:
> >
> >> Hey Telles,
> >>
> >> The code you've posted in Produce.java shows:
> >>
> >>             KeyedMessage<String, String> data = new KeyedMessage<String,
> >> String>("consumptions", String.valueOf(key),String.valueOf(value));
> >>
> >>
> >> Which suggests that you are sending a string for both key and value. If
> >> you have a Samza task consuming from this topic, you should set:
> >>
> >> systems.system-name.samza.key.serde=string
> >>
> >> systems.system-name.samza.msg.serde=string
> >>
> >>
> >> Cheers,
> >> Chris
> >>
> >> On 8/7/14 11:52 AM, "Telles Nobrega" <[email protected]> wrote:
> >>
> >> >Hum, that sounds like a perfect reason for it.
> >> >
> >> >I'm writing to the topic with this code
> >> >
> >> >
> >>
> >>
> https://github.com/tellesnobrega/kafka-producer/blob/master/src/Produce.j
> >>a
> >> >va
> >> >
> >> >My problem is that I need to send numbers as key and value to the kafka
> >> >topic so i can read it in samza.
> >> >
> >> >What is the best way to de/serialize this?
> >> >
> >> >
> >> >On Thu, Aug 7, 2014 at 3:43 PM, Yan Fang <[email protected]> wrote:
> >> >
> >> >> Hi Telles,
> >> >>
> >> >> One of the possible reasons is that, in your process method, you are
> >> >>trying
> >> >> to send a HashMap, not a String, in the collection.send. Could you
> >> >>check it
> >> >> ?
> >> >>
> >> >> Thanks,
> >> >>
> >> >> Fang, Yan
> >> >> [email protected]
> >> >> +1 (206) 849-4108
> >> >>
> >> >>
> >> >> On Thu, Aug 7, 2014 at 11:25 AM, Telles Nobrega
> >> >><[email protected]>
> >> >> wrote:
> >> >>
> >> >> > I changed my properties a little to look like this: link
> >> >> > <
> >> >> >
> >> >>
> >> >>
> >>
> >>
> http://mail-archives.apache.org/mod_mbox/samza-dev/201311.mbox/%3CEA1B8C3
> >> >>[email protected]%3E
> >> >> > >
> >> >> >
> >> >> > here it goes:
> >> >> >
> >> >> > # Job
> >> >> > job.factory.class=org.apache.samza.job.yarn.YarnJobFactory
> >> >> > job.name=consumptions
> >> >> >
> >> >> > # YARN
> >> >> >
> >> >> >
> >> >>
> >> >>yarn.package.path=file://${basedir}/target/${project.artifactId}-${
> >> pom.ve
> >> >>rsion}-dist.tar.gz
> >> >> >
> >> >> > # Task
> >> >> > task.class=alarm.ConsumptionProducer
> >> >> > task.inputs=kafka.consumptions
> >> >> >
> >> >> >
> >> >>
> >>
> >>>>task.checkpoint.factory=org.apache.samza.checkpoint.kafka.KafkaCheckpoi
> >>>>nt
> >> >>ManagerFactory
> >> >> > task.checkpoint.system=kafka
> >> >> > # Normally, this would be 3, but we have only one broker.
> >> >> > task.checkpoint.replication.factor=1
> >> >> >
> >> >> > # Metrics
> >> >> > metrics.reporters=snapshot,jmx
> >> >> >
> >> >> >
> >> >>
> >>
> >>>>metrics.reporter.snapshot.class=org.apache.samza.metrics.reporter.Metri
> >>>>cs
> >> >>SnapshotReporterFactory
> >> >> > metrics.reporter.snapshot.stream=kafka.metrics
> >> >> >
> >> >> >
> >> >>
> >>
> >>>>metrics.reporter.jmx.class=org.apache.samza.metrics.reporter.JmxReporte
> >>>>rF
> >> >>actory
> >> >> >
> >> >> > # Serializers
> >> >> >
> >> >> >
> >> >>
> >>
> >>>>serializers.registry.string.class=org.apache.samza.serializers.StringSe
> >>>>rd
> >> >>eFactory
> >> >> >
> >> >> >
> >> >>
> >>
> >>>>serializers.registry.metrics.class=org.apache.samza.serializers.Metrics
> >>>>Sn
> >> >>apshotSerdeFactory
> >> >> >
> >> >> > # Kafka System
> >> >> >
> >> >> >
> >> >>
> >>
> >>>>systems.kafka.samza.factory=org.apache.samza.system.kafka.KafkaSystemFa
> >>>>ct
> >> >>ory
> >> >> > systems.kafka.samza.msg.serde=string
> >> >> > systems.kafka.consumer.zookeeper.connect=localhost:2181/
> >> >> > systems.kafka.consumer.auto.offset.reset=largest
> >> >> > systems.kafka.producer.metadata.broker.list=localhost:9092
> >> >> > systems.kafka.producer.producer.type=sync
> >> >> > # Normally, we'd set this much higher, but we want things to look
> >> >>snappy
> >> >> in
> >> >> > the demo.
> >> >> > systems.kafka.producer.batch.num.messages=1
> >> >> > ystems.kafka.streams.metrics.samza.msg.serde=metrics
> >> >> >
> >> >> > But I'm getting this output,
> >> >> >
> >> >> > 0Caught exception in process loop.
> >> >> > java.lang.ClassCastException: java.util.HashMap cannot be cast to
> >> >> > java.lang.String
> >> >> >  at
> >> >>
> >>org.apache.samza.serializers.StringSerde.toBytes(StringSerde.scala:33)
> >> >> > at
> >> >>
> >>org.apache.samza.serializers.SerdeManager.toBytes(SerdeManager.scala:69)
> >> >> >  at
> >> >>
> >>org.apache.samza.system.SystemProducers.send(SystemProducers.scala:65)
> >> >> > at
> >> >> >
> >> >> >
> >> >>
> >>
> >>>>org.apache.samza.container.TaskInstance$$anonfun$send$2.apply(TaskInsta
> >>>>nc
> >> >>e.scala:170)
> >> >> >  at
> >> >> >
> >> >> >
> >> >>
> >>
> >>>>org.apache.samza.container.TaskInstance$$anonfun$send$2.apply(TaskInsta
> >>>>nc
> >> >>e.scala:170)
> >> >> > at
> >> >> >
> >> >> >
> >> >>
> >>
> >>>>scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.sc
> >>>>al
> >> >>a:59)
> >> >> >  at
> >>scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47)
> >> >> > at
> >> >>org.apache.samza.container.TaskInstance.send(TaskInstance.scala:170)
> >> >> >  at
> >> >> >
> >> >>
> >>
> >>>>org.apache.samza.container.RunLoop$$anonfun$send$2.apply(RunLoop.scala:
> >>>>11
> >> >>6)
> >> >> > at
> >> >> >
> >> >>
> >>
> >>>>org.apache.samza.container.RunLoop$$anonfun$send$2.apply(RunLoop.scala:
> >>>>11
> >> >>6)
> >> >> >  at scala.collection.Iterator$class.foreach(Iterator.scala:727)
> >> >> > at scala.collection.AbstractIterator.foreach(Iterator.scala:1157)
> >> >> >  at
> >> >> >
> >>
> >>>>scala.collection.MapLike$DefaultValuesIterable.foreach(MapLike.scala:20
> >>>>6)
> >> >> > at org.apache.samza.container.RunLoop.send(RunLoop.scala:116)
> >> >> >  at org.apache.samza.container.RunLoop.run(RunLoop.scala:59)
> >> >> > at
> >> >>
> >>org.apache.samza.container.SamzaContainer.run(SamzaContainer.scala:504)
> >> >> >  at
> >> >> >
> >>
> >>>>org.apache.samza.container.SamzaContainer$.main(SamzaContainer.scala:81
> >>>>)
> >> >> > at
> >> >>org.apache.samza.container.SamzaContainer.main(SamzaContainer.scala)
> >> >> >
> >> >> >
> >> >> >
> >> >> > On Thu, Aug 7, 2014 at 3:12 PM, Telles Nobrega
> >> >><[email protected]>
> >> >> > wrote:
> >> >> >
> >> >> > > Hi Chris, I really appreciate the time you are taking to help me
> >> >>out.
> >> >> > >
> >> >> > > This is job.properties file
> >> >> > >
> >> >> > > # Job
> >> >> > > job.factory.class=org.apache.samza.job.yarn.YarnJobFactory
> >> >> > > job.name=consumptions
> >> >> > >
> >> >> > > # YARN
> >> >> > >
> >> >> > >
> >> >> >
> >> >>
> >> >>yarn.package.path=file://${basedir}/target/${project.artifactId}-${
> >> pom.ve
> >> >>rsion}-dist.tar.gz
> >> >> > >
> >> >> > > # Task
> >> >> > > task.class=alarm.ConsumptionProducer
> >> >> > > task.inputs=kafka.consumptions
> >> >> > >
> >> >> > >
> >> >> >
> >> >>
> >>
> >>>>task.checkpoint.factory=org.apache.samza.checkpoint.kafka.KafkaCheckpoi
> >>>>nt
> >> >>ManagerFactory
> >> >> > > task.checkpoint.system=kafka
> >> >> > > # Normally, this would be 3, but we have only one broker.
> >> >> > > task.checkpoint.replication.factor=1
> >> >> > >
> >> >> > > # Serializers
> >> >> > >
> >> >> > >
> >> >> >
> >> >>
> >>
> >>>>serializers.registry.serde.class=org.apache.samza.serializers.StringSer
> >>>>de
> >> >>Factory
> >> >> > >
> >> >> > > # Kafka System
> >> >> > >
> >> >> > >
> >> >> >
> >> >>
> >>
> >>>>systems.kafka.samza.factory=org.apache.samza.system.kafka.KafkaSystemFa
> >>>>ct
> >> >>ory
> >> >> > > *systems.kafka.samza.msg.serde=json*
> >> >> > > systems.kafka.consumer.zookeeper.connect=localhost:2181/
> >> >> > > systems.kafka.consumer.auto.offset.reset=largest
> >> >> > > systems.kafka.producer.metadata.broker.list=localhost:9092
> >> >> > > systems.kafka.producer.producer.type=sync
> >> >> > > # Normally, we'd set this much higher, but we want things to look
> >> >> snappy
> >> >> > > in the demo.
> >> >> > > systems.kafka.producer.batch.num.messages=1
> >> >> > >
> >> >> > > *systems.kafka.streams.consumptions.key.serde=string*
> >> >> > > *systems.kafka.streams.consumptions.msg.serde=string*
> >> >> > >
> >> >> > > Does this look right?
> >> >> > > I'm running a local cluster, I want to have it running nicely
> >> >>before I
> >> >> > can
> >> >> > > distribute it.
> >> >> > >
> >> >> > >
> >> >> > >
> >> >> > > On Thu, Aug 7, 2014 at 3:08 PM, Chris Riccomini <
> >> >> > > [email protected]> wrote:
> >> >> > >
> >> >> > >> Hey Telles,
> >> >> > >>
> >> >> > >> Sure. In your job.properties file, define the serde:
> >> >> > >>
> >> >> > >> # Serializers
> >> >> > >>
> >> >> > >>
> >> >> >
> >> >>
> >>
> >>>>serializers.registry.serde.class=org.apache.samza.serializers.StringSer
> >>>>de
> >> >>Fa
> >> >> > >> ctory
> >> >> > >>
> >> >> > >>
> >> >> > >> Then define the serde for your system:
> >> >> > >>
> >> >> > >> systems.kafka.samza.msg.serde=string
> >> >> > >>
> >> >> > >>
> >> >> > >> Cheers,
> >> >> > >> Chris
> >> >> > >>
> >> >> > >> On 8/7/14 10:54 AM, "Telles Nobrega" <[email protected]>
> >> >>wrote:
> >> >> > >>
> >> >> > >> >Can you give and example on how to use string serde, i'm
> >>getting
> >> >>an
> >> >> > error
> >> >> > >> >when trying to set to string
> >> >> > >> >
> >> >> > >> >:53:26:804Got system producers: Set(kafka)
> >> >> > >> >17:53:26:809Got serdes: Set(string)
> >> >> > >> >17:53:29:206Container container_1407433587782_0001_01_000017
> >> >>failed
> >> >> > with
> >> >> > >> >exit code 1 - Exception from container-launch:
> >> >> > >> >
> >> >> > >> >
> >> >> > >> >
> >> >> > >> >On Thu, Aug 7, 2014 at 2:41 PM, Telles Nobrega <
> >> >> > [email protected]>
> >> >> > >> >wrote:
> >> >> > >> >
> >> >> > >> >> Thanks.
> >> >> > >> >>
> >> >> > >> >>
> >> >> > >> >> On Thu, Aug 7, 2014 at 1:54 PM, Chris Riccomini <
> >> >> > >> >> [email protected]> wrote:
> >> >> > >> >>
> >> >> > >> >>> Hey Telles,
> >> >> > >> >>>
> >> >> > >> >>> This is definitely a serde error. It sounds like your
> >>message
> >> >>is
> >> >> not
> >> >> > >> >>> properly formatted as a JSON blob.
> >> >> > >> >>>
> >> >> > >> >>> If you are trying to just use a string as the message (vs. a
> >> >>well
> >> >> > >> >>> formatted JSON blob), then you should use the StringSerde.
> >> >> > >> >>>
> >> >> > >> >>> Cheers,
> >> >> > >> >>> Chris
> >> >> > >> >>>
> >> >> > >> >>> On 8/7/14 8:05 AM, "Telles Nobrega"
> >><[email protected]>
> >> >> > wrote:
> >> >> > >> >>>
> >> >> > >> >>> >Hi, I'm running a simple samza topology that reads from  a
> >> >>kafka
> >> >> > >> topic
> >> >> > >> >>> >that
> >> >> > >> >>> >only has two Strings
> >> >> > >> >>> >xx:xx:xx:xxxx;xx
> >> >> > >> >>> >And its throwing an error
> >> >> > >> >>> >
> >> >> > >> >>> >Caught exception in process loop.
> >> >> > >> >>> >org.codehaus.jackson.JsonParseException: Unexpected
> >>character
> >> >> ('F'
> >> >> > >> >>>(code
> >> >> > >> >>> >70)): expected a valid value (number, String, array,
> >>object,
> >> >> > 'true',
> >> >> > >> >>> >'false' or 'null')
> >> >> > >> >>> > at [Source: [B@56dfb465; line: 1, column: 2]
> >> >> > >> >>> >at
> >> >> > >>
> >> >>
> >>>>>org.codehaus.jackson.JsonParser._constructError(JsonParser.java:1291)
> >> >> > >> >>> > at
> >> >> > >> >>>
> >> >> > >> >>>
> >> >> > >>
> >> >> > >>
> >> >> >
> >> >>
> >>
> >>>>>>>>org.codehaus.jackson.impl.JsonParserMinimalBase._reportError(JsonPa
> >>>>>>>>rs
> >> >>>>>>er
> >> >> > >> >>>>Min
> >> >> > >> >>> >imalBase.java:385)
> >> >> > >> >>> >at
> >> >> > >> >>>
> >> >> > >> >>>
> >> >> > >>
> >> >> > >>
> >> >> >
> >> >>
> >>
> >>>>>>>>org.codehaus.jackson.impl.JsonParserMinimalBase._reportUnexpectedCh
> >>>>>>>>ar
> >> >>>>>>(J
> >> >> > >> >>>>son
> >> >> > >> >>> >ParserMinimalBase.java:306)
> >> >> > >> >>> > at
> >> >> > >> >>>
> >> >> > >> >>>
> >> >> > >>
> >> >> > >>
> >> >> >
> >> >>
> >>
> >>>>>>>>org.codehaus.jackson.impl.Utf8StreamParser._handleUnexpectedValue(U
> >>>>>>>>tf
> >> >>>>>>8S
> >> >> > >> >>>>tre
> >> >> > >> >>> >amParser.java:1581)
> >> >> > >> >>> >at
> >> >> > >> >>>
> >> >> > >> >>>
> >> >> > >>
> >> >> > >>
> >> >> >
> >> >>
> >>
> >>>>>>>>org.codehaus.jackson.impl.Utf8StreamParser._nextTokenNotInObject(Ut
> >>>>>>>>f8
> >> >>>>>>St
> >> >> > >> >>>>rea
> >> >> > >> >>> >mParser.java:436)
> >> >> > >> >>> > at
> >> >> > >> >>>
> >> >> > >> >>>
> >> >> > >>
> >> >> > >>
> >> >> >
> >> >>
> >>
> >>>>>>>>org.codehaus.jackson.impl.Utf8StreamParser.nextToken(Utf8StreamPars
> >>>>>>>>er
> >> >>>>>>.j
> >> >> > >> >>>>ava
> >> >> > >> >>> >:322)
> >> >> > >> >>> >at
> >> >> > >> >>>
> >> >> > >> >>>
> >> >> > >>
> >> >> > >>
> >> >> >
> >> >>
> >>
> >>>>>>>>org.codehaus.jackson.map.ObjectMapper._initForReading(ObjectMapper.
> >>>>>>>>ja
> >> >>>>>>va
> >> >> > >> >>>>:24
> >> >> > >> >>> >32)
> >> >> > >> >>> > at
> >> >> > >> >>>
> >> >> > >> >>>
> >> >> > >>
> >> >> > >>
> >> >> >
> >> >>
> >>
> >>>>>>>>org.codehaus.jackson.map.ObjectMapper._readMapAndClose(ObjectMapper
> >>>>>>>>.j
> >> >>>>>>av
> >> >> > >> >>>>a:2
> >> >> > >> >>> >389)
> >> >> > >> >>> >at
> >> >> > >> >>>
> >> >> > >>
> >> >>
> >>org.codehaus.jackson.map.ObjectMapper.readValue(ObjectMapper.java:1667)
> >> >> > >> >>> > at
> >> >> > >>
> >> >>
> >>>>>org.apache.samza.serializers.JsonSerde.fromBytes(JsonSerde.scala:33)
> >> >> > >> >>> >at
> >> >> > >> >>>
> >> >> > >> >>>
> >> >> > >>
> >> >> > >>
> >> >> >
> >> >>
> >>
> >>>>>>>>org.apache.samza.serializers.SerdeManager.fromBytes(SerdeManager.sc
> >>>>>>>>al
> >> >>>>>>a:
> >> >> > >> >>>>115
> >> >> > >> >>> >)
> >> >> > >> >>> > at
> >> >> > >> >>>
> >> >> > >> >>>
> >> >> > >>
> >> >> > >>
> >> >> >
> >> >>
> >>
> >>>>>>>>org.apache.samza.system.SystemConsumers$$anonfun$org$apache$samza$s
> >>>>>>>>ys
> >> >>>>>>te
> >> >> > >> >>>>m$S
> >> >> > >> >>> >ystemConsumers$$poll$5.apply(SystemConsumers.scala:245)
> >> >> > >> >>> > at
> >> >> > >> >>>
> >> >> > >> >>>
> >> >> > >>
> >> >> > >>
> >> >> >
> >> >>
> >>
> >>>>>>>>org.apache.samza.system.SystemConsumers$$anonfun$org$apache$samza$s
> >>>>>>>>ys
> >> >>>>>>te
> >> >> > >> >>>>m$S
> >> >> > >> >>> >ystemConsumers$$poll$5.apply(SystemConsumers.scala:242)
> >> >> > >> >>> >at
> >>scala.collection.Iterator$class.foreach(Iterator.scala:727)
> >> >> > >> >>> > at
> >> >> scala.collection.AbstractIterator.foreach(Iterator.scala:1157)
> >> >> > >> >>> >at
> >> >> > scala.collection.IterableLike$class.foreach(IterableLike.scala:72)
> >> >> > >> >>> > at
> >> >>scala.collection.AbstractIterable.foreach(Iterable.scala:54)
> >> >> > >> >>> >at org.apache.samza.system.SystemConsumers.org
> >> >> > >> >>>
> >> >> >
> >>>$apache$samza$system$SystemConsumers$$poll(SystemConsumers.scala:242)
> >> >> > >> >>> > at
> >> >> > >> >>>
> >> >> > >> >>>
> >> >> > >>
> >> >> > >>
> >> >> >
> >> >>
> >>
> >>>>>>>>org.apache.samza.system.SystemConsumers$$anon$1$$anonfun$call$2.app
> >>>>>>>>ly
> >> >>>>>>(S
> >> >> > >> >>>>yst
> >> >> > >> >>> >emConsumers.scala:180)
> >> >> > >> >>> >at
> >> >> > >> >>>
> >> >> > >> >>>
> >> >> > >>
> >> >> > >>
> >> >> >
> >> >>
> >>
> >>>>>>>>org.apache.samza.system.SystemConsumers$$anon$1$$anonfun$call$2.app
> >>>>>>>>ly
> >> >>>>>>(S
> >> >> > >> >>>>yst
> >> >> > >> >>> >emConsumers.scala:180)
> >> >> > >> >>> > at
> >> >> > >> >>>
> >> >> > >> >>>
> >> >> > >>
> >> >> > >>
> >> >> >
> >> >>
> >>
> >>>>>>>>scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLi
> >>>>>>>>ke
> >> >>>>>>.s
> >> >> > >> >>>>cal
> >> >> > >> >>> >a:244)
> >> >> > >> >>> >at
> >> >> > >> >>>
> >> >> > >> >>>
> >> >> > >>
> >> >> > >>
> >> >> >
> >> >>
> >>
> >>>>>>>>scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLi
> >>>>>>>>ke
> >> >>>>>>.s
> >> >> > >> >>>>cal
> >> >> > >> >>> >a:244)
> >> >> > >> >>> > at
> >> >>scala.collection.Iterator$class.foreach(Iterator.scala:727)
> >> >> > >> >>> >at
> >> >>scala.collection.AbstractIterator.foreach(Iterator.scala:1157)
> >> >> > >> >>> > at
> >> >> > scala.collection.MapLike$DefaultKeySet.foreach(MapLike.scala:174)
> >> >> > >> >>> >at
> >> >> > >>
> >> >>
> >>>>>scala.collection.TraversableLike$class.map(TraversableLike.scala:244)
> >> >> > >> >>> > at
> >> >> > >> >>>
> >> >> > >> >>>
> >> >> > >>
> >> >> > >>
> >> >> >
> >> >>
> >>
> >>>>>>>>scala.collection.AbstractSet.scala$collection$SetLike$$super$map(Se
> >>>>>>>>t.
> >> >>>>>>sc
> >> >> > >> >>>>ala
> >> >> > >> >>> >:47)
> >> >> > >> >>> >at scala.collection.SetLike$class.map(SetLike.scala:93)
> >> >> > >> >>> > at scala.collection.AbstractSet.map(Set.scala:47)
> >> >> > >> >>> >at
> >> >> > >> >>>
> >> >> > >> >>>
> >> >> > >>
> >> >> > >>
> >> >> >
> >> >>
> >>
> >>>>>>>>org.apache.samza.system.SystemConsumers$$anon$1.call(SystemConsumer
> >>>>>>>>s.
> >> >>>>>>sc
> >> >> > >> >>>>ala
> >> >> > >> >>> >:180)
> >> >> > >> >>> > at
> >> >> > >> >>>
> >> >> > >>
> >> >> > >>
> >> >> >
> >> >>
> >>
> >>>>>>>>org.apache.samza.util.DoublingBackOff.maybeCall(DoublingBackOff.sca
> >>>>>>>>la
> >> >>>>>>:4
> >> >> > >> >>>>4)
> >> >> > >> >>> >at
> >> >> > >> >>>
> >> >> > >>
> >> >> > >>
> >> >> >
> >> >>
> >>
> >>>>>>>>org.apache.samza.system.SystemConsumers.choose(SystemConsumers.scal
> >>>>>>>>a:
> >> >>>>>>20
> >> >> > >> >>>>8)
> >> >> > >> >>> > at
> >> >>org.apache.samza.container.RunLoop.process(RunLoop.scala:73)
> >> >> > >> >>> >at org.apache.samza.container.RunLoop.run(RunLoop.scala:57)
> >> >> > >> >>> > at
> >> >> > >> >>>
> >> >> > >>
> >> >> > >>
> >> >> >
> >> >>
> >>
> >>>>>>>>org.apache.samza.container.SamzaContainer.run(SamzaContainer.scala:
> >>>>>>>>50
> >> >>>>>>4)
> >> >> > >> >>> >at
> >> >> > >> >>>
> >> >> > >>
> >> >> > >>
> >> >> >
> >> >>
> >>
> >>>>>>>>org.apache.samza.container.SamzaContainer$.main(SamzaContainer.scal
> >>>>>>>>a:
> >> >>>>>>81
> >> >> > >> >>>>)
> >> >> > >> >>> > at
> >> >> > >>
> >> >>
> >>>>>org.apache.samza.container.SamzaContainer.main(SamzaContainer.scala)
> >> >> > >> >>> >
> >> >> > >> >>> >
> >> >> > >> >>> >
> >> >> > >> >>> >​Has anyone experienced this error before?​
> >> >> > >> >>> >
> >> >> > >> >>> >--
> >> >> > >> >>> >------------------------------------------
> >> >> > >> >>> >Telles Mota Vidal Nobrega
> >> >> > >> >>> >M.sc. Candidate at UFCG
> >> >> > >> >>> >B.sc. in Computer Science at UFCG
> >> >> > >> >>> >Software Engineer at OpenStack Project - HP/LSD-UFCG
> >> >> > >> >>>
> >> >> > >> >>>
> >> >> > >> >>
> >> >> > >> >>
> >> >> > >> >> --
> >> >> > >> >> ------------------------------------------
> >> >> > >> >> Telles Mota Vidal Nobrega
> >> >> > >> >> M.sc. Candidate at UFCG
> >> >> > >> >> B.sc. in Computer Science at UFCG
> >> >> > >> >> Software Engineer at OpenStack Project - HP/LSD-UFCG
> >> >> > >> >>
> >> >> > >> >
> >> >> > >> >
> >> >> > >> >
> >> >> > >> >--
> >> >> > >> >------------------------------------------
> >> >> > >> >Telles Mota Vidal Nobrega
> >> >> > >> >M.sc. Candidate at UFCG
> >> >> > >> >B.sc. in Computer Science at UFCG
> >> >> > >> >Software Engineer at OpenStack Project - HP/LSD-UFCG
> >> >> > >>
> >> >> > >>
> >> >> > >
> >> >> > >
> >> >> > > --
> >> >> > > ------------------------------------------
> >> >> > > Telles Mota Vidal Nobrega
> >> >> > > M.sc. Candidate at UFCG
> >> >> > > B.sc. in Computer Science at UFCG
> >> >> > > Software Engineer at OpenStack Project - HP/LSD-UFCG
> >> >> > >
> >> >> >
> >> >> >
> >> >> >
> >> >> > --
> >> >> > ------------------------------------------
> >> >> > Telles Mota Vidal Nobrega
> >> >> > M.sc. Candidate at UFCG
> >> >> > B.sc. in Computer Science at UFCG
> >> >> > Software Engineer at OpenStack Project - HP/LSD-UFCG
> >> >> >
> >> >>
> >> >
> >> >
> >> >
> >> >--
> >> >------------------------------------------
> >> >Telles Mota Vidal Nobrega
> >> >M.sc. Candidate at UFCG
> >> >B.sc. in Computer Science at UFCG
> >> >Software Engineer at OpenStack Project - HP/LSD-UFCG
> >>
> >>
> >
> >
> >--
> >------------------------------------------
> >Telles Mota Vidal Nobrega
> >M.sc. Candidate at UFCG
> >B.sc. in Computer Science at UFCG
> >Software Engineer at OpenStack Project - HP/LSD-UFCG
>
>


-- 
------------------------------------------
Telles Mota Vidal Nobrega
M.sc. Candidate at UFCG
B.sc. in Computer Science at UFCG
Software Engineer at OpenStack Project - HP/LSD-UFCG

Reply via email to