RE: StackOverflow in Spark

2016-06-13 Thread Michel Hubert

I’ve found my problem.

I’ve got a DAG with two consecutive “updateStateByKey” functions .
When I only process (map/foreachRDD/JavaEsSpark) the state of the last 
“updateStateByKey” function, I get an stackoverflow after a while (too long 
linage).

But when I also do some processing (foreachRDD/rdd.take) on the first 
“updatestatebykey”, then there is no problem.

Does this make sense? Probably the “long linage” problem.

But why should I have such a “linage problem” when Sparks claims to be a 
“abstract/high level” architecture? Why should I be worried about “long 
linage”? Its seems a contradiction with the abstract/high level (functional 
programming) approach when I have to know/consider how Spark doest it.



Van: Rishabh Wadhawan [mailto:rishabh...@gmail.com]
Verzonden: donderdag 2 juni 2016 06:06
Aan: Yash Sharma <yash...@gmail.com>
CC: Ted Yu <yuzhih...@gmail.com>; Matthew Young <taige...@gmail.com>; Michel 
Hubert <mich...@phact.nl>; user@spark.apache.org
Onderwerp: Re: StackOverflow in Spark

Stackoverflow is generated when DAG is too log as there are many 
transformations in lot of iterations. Please use checkpointing to store the DAG 
and break the linage to get away from this stack overflow error. Look into 
checkpoint fuction.
Thanks
Hope it helps. Let me know if you need anymore help.
On Jun 1, 2016, at 8:18 PM, Yash Sharma 
<yash...@gmail.com<mailto:yash...@gmail.com>> wrote:

Not sure if its related, But I got a similar stack overflow error some time 
back while reading files and converting them to parquet.



Stack trace-
16/06/02 02:23:54 INFO YarnAllocator: Driver requested a total number of 32769 
executor(s).
16/06/02 02:23:54 INFO ExecutorAllocationManager: Requesting 16384 new 
executors because tasks are backlogged (new desired total will be 32769)
16/06/02 02:23:54 INFO YarnAllocator: Will request 24576 executor containers, 
each with 5 cores and 22528 MB memory including 2048 MB overhead
16/06/02 02:23:55 WARN ApplicationMaster: Reporter thread fails 2 time(s) in a 
row.
at scala.collection.mutable.MapBuilder.$plus$eq(MapBuilder.scala:28)
at scala.collection.mutable.MapBuilder.$plus$eq(MapBuilder.scala:24)
at 
scala.collection.TraversableLike$$anonfun$filter$1.apply(TraversableLike.scala:264)
at 
scala.collection.MapLike$MappedValues$$anonfun$foreach$3.apply(MapLike.scala:245)
at 
scala.collection.MapLike$MappedValues$$anonfun$foreach$3.apply(MapLike.scala:245)
at 
scala.collection.TraversableLike$WithFilter$$anonfun$foreach$1.apply(TraversableLike.scala:772)
at 
scala.collection.MapLike$MappedValues$$anonfun$foreach$3.apply(MapLike.scala:245)
at 
scala.collection.MapLike$MappedValues$$anonfun$foreach$3.apply(MapLike.scala:245)
java.lang.StackOverflowError


On Thu, Jun 2, 2016 at 12:58 PM, Ted Yu 
<yuzhih...@gmail.com<mailto:yuzhih...@gmail.com>> wrote:
Looking at Michel's stack trace, it seems to be different issue.

On Jun 1, 2016, at 7:45 PM, Matthew Young 
<taige...@gmail.com<mailto:taige...@gmail.com>> wrote:
Hi,

It's related to the one fixed bug in Spark, jira ticket 
SPARK-6847<https://issues.apache.org/jira/browse/SPARK-6847>

Matthew Yang

On Wed, May 25, 2016 at 7:48 PM, Michel Hubert 
<mich...@phact.nl<mailto:mich...@phact.nl>> wrote:

Hi,


I have an Spark application which generates StackOverflowError exceptions after 
30+ min.

Anyone any ideas?






16/05/25 10:48:51 WARN scheduler.TaskSetManager: Lost task 0.0 in stage 55449.0 
(TID 5584, host81440-cld.opentsp.com<http://host81440-cld.opentsp.com/>): 
java.lang.StackOverflowError
·at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1382)
·at java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990)
·at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915)
·at java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798)
·at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
·at java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990)
·at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915)
·at java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798)
·at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
·at java.io.ObjectInputStream.readObject(ObjectInputStream.java:370)
·at scala.collection.immutable.$colon$colon.readObject(List.scala:362)
·at sun.reflect.GeneratedMethodAccessor2.invoke(Unknown Source)
·at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
·at java.lang.reflect.Method.invoke(Method.java:606)
·at java.io.ObjectStreamClass.invokeReadObject(ObjectStreamClass.java:1017)
·at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1893)
·at java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798)
·at java.io.ObjectInputStream.readObject0(ObjectInputStre

submitMissingTasks - serialize throws StackOverflow exception

2016-05-27 Thread Michel Hubert
Hi,

My Spark application throws stackoverflow exceptions after a while.
The DAGScheduler function submitMissingTasks tries to serialize a Tuple 
(MapPartitionsRDD, EsSpark..saveToEs) which is handled with a recursive 
algorithm.
The recursive algorithm is too deep and results in a stackoverflow exception.

Should I just try to increase the heap size? Or will it just happen later in 
time?

How can I fix this?

With kind regards,
michel



[cid:image001.png@01D1B7F5.8E8C3980]


StackOverflow in Spark

2016-05-25 Thread Michel Hubert

Hi,


I have an Spark application which generates StackOverflowError exceptions after 
30+ min.

Anyone any ideas?

Seems like problems with deserialization of checkpoint data?





16/05/25 10:48:51 WARN scheduler.TaskSetManager: Lost task 0.0 in stage 55449.0 
(TID 5584, host81440-cld.opentsp.com): java.lang.StackOverflowError
*at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1382)
*at java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990)
*at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915)
*at java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798)
*at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
*at java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990)
*at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915)
*at java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798)
*at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
*at java.io.ObjectInputStream.readObject(ObjectInputStream.java:370)
*at scala.collection.immutable.$colon$colon.readObject(List.scala:362)
*at sun.reflect.GeneratedMethodAccessor2.invoke(Unknown Source)
*at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
*at java.lang.reflect.Method.invoke(Method.java:606)
*at java.io.ObjectStreamClass.invokeReadObject(ObjectStreamClass.java:1017)
*at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1893)
*at java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798)
*at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
*at java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990)
*at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915)
*at java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798)
*at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
*at java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990)
*at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915)
*at java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798)
*at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
*at java.io.ObjectInputStream.readObject(ObjectInputStream.java:370)
*at scala.collection.immutable.$colon$colon.readObject(List.scala:362)
at sun.reflect.GeneratedMethodAccessor2.invoke(Unknown Source)




Driver stacktrace:
*at 
org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1431)
*at 
org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1419)
*at 
org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1418)
*at 
scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
*at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47)
*at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1418)
*at 
org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:799)
*at 
org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:799)
*at scala.Option.foreach(Option.scala:236)
*at 
org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:799)
*at 
org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:1640)
*at 
org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1599)
*at 
org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1588)
*at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:48)
*at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:620)
*at org.apache.spark.SparkContext.runJob(SparkContext.scala:1843)
*at org.apache.spark.SparkContext.runJob(SparkContext.scala:1856)
*at org.apache.spark.SparkContext.runJob(SparkContext.scala:1933)
*at org.elasticsearch.spark.rdd.EsSpark$.saveToEs(EsSpark.scala:67)
*at org.elasticsearch.spark.rdd.EsSpark$.saveToEs(EsSpark.scala:54)
*at org.elasticsearch.spark.rdd.EsSpark$.saveJsonToEs(EsSpark.scala:90)
*at 
org.elasticsearch.spark.rdd.api.java.JavaEsSpark$.saveJsonToEs(JavaEsSpark.scala:62)
at 
org.elasticsearch.spark.rdd.api.java.JavaEsSpark.saveJsonToEs(JavaEsSpark.scala)



Kafka 0.9 and spark-streaming-kafka_2.10

2016-05-09 Thread Michel Hubert
Hi,

I'm thinking of upgdrading our kafka cluster to 0.9.

Will this be a problem for the Spark Streaming + Kafka Direct Approach 
Integration using artifact spark-streaming-kafka_2.10 (1.6.1)?


groupId = org.apache.spark
artifactId = spark-streaming-kafka_2.10
version = 1.6.1


Because the documentation states:
Kafka: Spark Streaming 1.6.1 is compatible with Kafka 0.8.2.1.


Thanks.


RE: run-example streaming.KafkaWordCount fails on CDH 5.7.0

2016-05-04 Thread Michel Hubert
We're running Kafka 0.8.2.2
Is that the problem, why?

-Oorspronkelijk bericht-
Van: Sean Owen [mailto:so...@cloudera.com] 
Verzonden: woensdag 4 mei 2016 10:41
Aan: Michel Hubert <mich...@phact.nl>
CC: user@spark.apache.org
Onderwerp: Re: run-example streaming.KafkaWordCount fails on CDH 5.7.0

Please try the CDH forums; this is the Spark list:
http://community.cloudera.com/t5/Advanced-Analytics-Apache-Spark/bd-p/Spark

Before you even do that, I can tell you to double check you're running Kafka 
0.9.

On Wed, May 4, 2016 at 9:29 AM, Michel Hubert <mich...@phact.nl> wrote:
>
>
> Hi,
>
>
>
> After an upgrade to CDH 5.7.0 we have troubles with the Kafka to Spark 
> Streaming.
>
>
>
> The example jar doesn’t work:
>
>
>
> /opt/cloudera/parcels/CDH/lib/spark/bin/run-example 
> streaming.KafkaWordCount ….
>
>
>
> Attached is a log file.
>
>
>
> 16/05/04 10:06:23 WARN consumer.ConsumerFetcherThread:
> [ConsumerFetcherThread-wordcount1_host81436-cld.domain.com-14623491821
> 72-13b9f2e7-0-2], Error in fetch 
> kafka.consumer.ConsumerFetcherThread$FetchRequest@47fb7e6d.
> Possible cause: java.nio.BufferUnderflowException
>
> 16/05/04 10:06:24 WARN consumer.ConsumerFetcherThread:
> [ConsumerFetcherThread-wordcount1_host81436-cld.domain.com-14623491821
> 72-13b9f2e7-0-1], Error in fetch 
> kafka.consumer.ConsumerFetcherThread$FetchRequest@73b9a762.
> Possible cause: java.lang.IllegalArgumentException
>
>
>
> We have no problem running this from my development pc, only in 
> product op CDH 5.7.0 environment.
>
>
>
>
>
> Any ideas?
>
>
>
> Thanks,
>
> Michel
>
>
>
>
>
> -
> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org For 
> additional commands, e-mail: user-h...@spark.apache.org



RE: Kafka exception in Apache Spark

2016-04-26 Thread Michel Hubert
This is production.

Van: Mich Talebzadeh [mailto:mich.talebza...@gmail.com]
Verzonden: dinsdag 26 april 2016 12:01
Aan: Michel Hubert <mich...@phact.nl>
CC: user@spark.apache.org
Onderwerp: Re: Kafka exception in Apache Spark

Hi Michael,

Is this production or test?


Dr Mich Talebzadeh



LinkedIn  
https://www.linkedin.com/profile/view?id=AAEWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw



http://talebzadehmich.wordpress.com<http://talebzadehmich.wordpress.com/>



On 26 April 2016 at 09:07, Michel Hubert 
<mich...@phact.nl<mailto:mich...@phact.nl>> wrote:
Hi,


I use a Kafka direct stream approach.
My Spark application was running ok.
This morning we upgraded to CDH 5.7.0
And when I re-started my Spark application I get exceptions.

It seems a problem with the direct stream approach.
Any ideas how to fix this?



User class threw exception: org.apache.spark.SparkException: Job aborted due to 
stage failure: Task 3 in stage 0.0 failed 4 times, most recent failure: Lost 
task 3.3 in stage 0.0 (TID 26, 
bfravicsvr81439-cld.opentsp.com<http://bfravicsvr81439-cld.opentsp.com>): 
java.lang.IllegalArgumentException
at java.nio.Buffer.limit(Buffer.java:267)
at kafka.api.FetchResponsePartitionData$.readFrom(FetchResponse.scala:38)
at kafka.api.TopicData$$anonfun$1.apply(FetchResponse.scala:100)
at kafka.api.TopicData$$anonfun$1.apply(FetchResponse.scala:98)
at 
scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:244)
at 
scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:244)
at scala.collection.immutable.Range.foreach(Range.scala:141)
at scala.collection.TraversableLike$class.map(TraversableLike.scala:244)
at scala.collection.AbstractTraversable.map(Traversable.scala:105)
at kafka.api.TopicData$.readFrom(FetchResponse.scala:98)
at kafka.api.FetchResponse$$anonfun$4.apply(FetchResponse.scala:170)
at kafka.api.FetchResponse$$anonfun$4.apply(FetchResponse.scala:169)
at 
scala.collection.TraversableLike$$anonfun$flatMap$1.apply(TraversableLike.scala:251)
at 
scala.collection.TraversableLike$$anonfun$flatMap$1.apply(TraversableLike.scala:251)
at scala.collection.immutable.Range.foreach(Range.scala:141)
at scala.collection.TraversableLike$class.flatMap(TraversableLike.scala:251)
at scala.collection.AbstractTraversable.flatMap(Traversable.scala:105)
at kafka.api.FetchResponse$.readFrom(FetchResponse.scala:169)
at kafka.consumer.SimpleConsumer.fetch(SimpleConsumer.scala:135)
at 
org.apache.spark.streaming.kafka.KafkaRDD$KafkaRDDIterator.fetchBatch(KafkaRDD.scala:192)
at 
org.apache.spark.streaming.kafka.KafkaRDD$KafkaRDDIterator.getNext(KafkaRDD.scala:208)
at org.apache.spark.util.NextIterator.hasNext(NextIterator.scala:73)
at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:327)
at scala.collection.Iterator$$anon$14.hasNext(Iterator.scala:388)
at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:327)
at 
org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.write(BypassMergeSortShuffleWriter.java:126)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
at org.apache.spark.scheduler.Task.run(Task.scala:89)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)




Kafka exception in Apache Spark

2016-04-26 Thread Michel Hubert
Hi,


I use a Kafka direct stream approach.
My Spark application was running ok.
This morning we upgraded to CDH 5.7.0
And when I re-started my Spark application I get exceptions.

It seems a problem with the direct stream approach.
Any ideas how to fix this?



User class threw exception: org.apache.spark.SparkException: Job aborted due to 
stage failure: Task 3 in stage 0.0 failed 4 times, most recent failure: Lost 
task 3.3 in stage 0.0 (TID 26, bfravicsvr81439-cld.opentsp.com): 
java.lang.IllegalArgumentException
at java.nio.Buffer.limit(Buffer.java:267)
at kafka.api.FetchResponsePartitionData$.readFrom(FetchResponse.scala:38)
at kafka.api.TopicData$$anonfun$1.apply(FetchResponse.scala:100)
at kafka.api.TopicData$$anonfun$1.apply(FetchResponse.scala:98)
at 
scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:244)
at 
scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:244)
at scala.collection.immutable.Range.foreach(Range.scala:141)
at scala.collection.TraversableLike$class.map(TraversableLike.scala:244)
at scala.collection.AbstractTraversable.map(Traversable.scala:105)
at kafka.api.TopicData$.readFrom(FetchResponse.scala:98)
at kafka.api.FetchResponse$$anonfun$4.apply(FetchResponse.scala:170)
at kafka.api.FetchResponse$$anonfun$4.apply(FetchResponse.scala:169)
at 
scala.collection.TraversableLike$$anonfun$flatMap$1.apply(TraversableLike.scala:251)
at 
scala.collection.TraversableLike$$anonfun$flatMap$1.apply(TraversableLike.scala:251)
at scala.collection.immutable.Range.foreach(Range.scala:141)
at scala.collection.TraversableLike$class.flatMap(TraversableLike.scala:251)
at scala.collection.AbstractTraversable.flatMap(Traversable.scala:105)
at kafka.api.FetchResponse$.readFrom(FetchResponse.scala:169)
at kafka.consumer.SimpleConsumer.fetch(SimpleConsumer.scala:135)
at 
org.apache.spark.streaming.kafka.KafkaRDD$KafkaRDDIterator.fetchBatch(KafkaRDD.scala:192)
at 
org.apache.spark.streaming.kafka.KafkaRDD$KafkaRDDIterator.getNext(KafkaRDD.scala:208)
at org.apache.spark.util.NextIterator.hasNext(NextIterator.scala:73)
at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:327)
at scala.collection.Iterator$$anon$14.hasNext(Iterator.scala:388)
at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:327)
at 
org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.write(BypassMergeSortShuffleWriter.java:126)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
at org.apache.spark.scheduler.Task.run(Task.scala:89)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)



RE: apache spark errors

2016-03-24 Thread Michel Hubert
No.

But I may be on to something.
I use Jedis to send data to Redis.

I used a ThreadLocal construct:

private static final ThreadLocal jedis = new ThreadLocal(){
@Override
protected Jedis initialValue()
{
return new Jedis("10.101.41.19",6379);
}
};

and then


.foreachRDD(new VoidFunction<JavaRDD>() {
public void call(JavaRDD rdd) throws Exception {

for (TopData t: rdd.take(top)) {
jedis …
}

May this resulted in a memory leak?

Van: Ted Yu [mailto:yuzhih...@gmail.com]
Verzonden: donderdag 24 maart 2016 15:15
Aan: Michel Hubert <mich...@phact.nl>
CC: user@spark.apache.org
Onderwerp: Re: apache spark errors

Do you have history server enabled ?

Posting your code snippet would help us understand your use case (and reproduce 
the leak).

Thanks

On Thu, Mar 24, 2016 at 6:40 AM, Michel Hubert 
<mich...@phact.nl<mailto:mich...@phact.nl>> wrote:

 
org.apache.spark
spark-core_2.10
1.6.1


org.apache.spark
spark-streaming_2.10
1.6.1


org.apache.spark
spark-streaming-kafka_2.10
1.6.1



org.elasticsearch
elasticsearch
2.2.0



org.apache.kafka
   kafka_2.10
0.8.2.2




org.elasticsearch
elasticsearch-spark_2.10
2.2.0


redis.clients
jedis
2.8.0
jar
compile




How can I look at those tasks?

Van: Ted Yu [mailto:yuzhih...@gmail.com<mailto:yuzhih...@gmail.com>]
Verzonden: donderdag 24 maart 2016 14:33
Aan: Michel Hubert <mich...@phact.nl<mailto:mich...@phact.nl>>
CC: user@spark.apache.org<mailto:user@spark.apache.org>
Onderwerp: Re: apache spark errors

Which release of Spark are you using ?

Have you looked the tasks whose Ids were printed to see if there was more clue ?

Thanks

On Thu, Mar 24, 2016 at 6:12 AM, Michel Hubert 
<mich...@phact.nl<mailto:mich...@phact.nl>> wrote:
HI,

I constantly get these errors:

0[Executor task launch worker-15] ERROR org.apache.spark.executor.Executor  
- Managed memory leak detected; size = 6564500 bytes, TID = 38969
310002 [Executor task launch worker-12] ERROR 
org.apache.spark.executor.Executor  - Managed memory leak detected; size = 
5523550 bytes, TID = 43270
318445 [Executor task launch worker-12] ERROR 
org.apache.spark.executor.Executor  - Managed memory leak detected; size = 
6879566 bytes, TID = 43408
388893 [Executor task launch worker-12] ERROR 
org.apache.spark.executor.Executor  - Managed memory leak detected; size = 
5572546 bytes, TID = 44382
418186 [Executor task launch worker-13] ERROR 
org.apache.spark.executor.Executor  - Managed memory leak detected; size = 
5289222 bytes, TID = 44795
488421 [Executor task launch worker-4] ERROR org.apache.spark.executor.Executor 
 - Managed memory leak detected; size = 8738142 bytes, TID = 45769
619276 [Executor task launch worker-4] ERROR org.apache.spark.executor.Executor 
 - Managed memory leak detected; size = 5759312 bytes, TID = 47571
632275 [Executor task launch worker-12] ERROR 
org.apache.spark.executor.Executor  - Managed memory leak detected; size = 
5602240 bytes, TID = 47709
644989 [Executor task launch worker-13] ERROR 
org.apache.spark.executor.Executor  - Managed memory leak detected; size = 
5326260 bytes, TID = 47863
720701 [Executor task launch worker-12] ERROR 
org.apache.spark.executor.Executor  - Managed memory leak detected; size = 
5399578 bytes, TID = 48959
1147961 [Executor task launch worker-16] ERROR 
org.apache.spark.executor.Executor  - Managed memory leak detected; size = 
5251872 bytes, TID = 54922


How can I fix this?

With kind regard,

Michel




RE: apache spark errors

2016-03-24 Thread Michel Hubert

 
org.apache.spark
spark-core_2.10
1.6.1


org.apache.spark
spark-streaming_2.10
1.6.1


org.apache.spark
spark-streaming-kafka_2.10
1.6.1



org.elasticsearch
elasticsearch
2.2.0



org.apache.kafka
   kafka_2.10
0.8.2.2




org.elasticsearch
elasticsearch-spark_2.10
2.2.0


redis.clients
jedis
2.8.0
jar
compile




How can I look at those tasks?

Van: Ted Yu [mailto:yuzhih...@gmail.com]
Verzonden: donderdag 24 maart 2016 14:33
Aan: Michel Hubert <mich...@phact.nl>
CC: user@spark.apache.org
Onderwerp: Re: apache spark errors

Which release of Spark are you using ?

Have you looked the tasks whose Ids were printed to see if there was more clue ?

Thanks

On Thu, Mar 24, 2016 at 6:12 AM, Michel Hubert 
<mich...@phact.nl<mailto:mich...@phact.nl>> wrote:
HI,

I constantly get these errors:

0[Executor task launch worker-15] ERROR org.apache.spark.executor.Executor  
- Managed memory leak detected; size = 6564500 bytes, TID = 38969
310002 [Executor task launch worker-12] ERROR 
org.apache.spark.executor.Executor  - Managed memory leak detected; size = 
5523550 bytes, TID = 43270
318445 [Executor task launch worker-12] ERROR 
org.apache.spark.executor.Executor  - Managed memory leak detected; size = 
6879566 bytes, TID = 43408
388893 [Executor task launch worker-12] ERROR 
org.apache.spark.executor.Executor  - Managed memory leak detected; size = 
5572546 bytes, TID = 44382
418186 [Executor task launch worker-13] ERROR 
org.apache.spark.executor.Executor  - Managed memory leak detected; size = 
5289222 bytes, TID = 44795
488421 [Executor task launch worker-4] ERROR org.apache.spark.executor.Executor 
 - Managed memory leak detected; size = 8738142 bytes, TID = 45769
619276 [Executor task launch worker-4] ERROR org.apache.spark.executor.Executor 
 - Managed memory leak detected; size = 5759312 bytes, TID = 47571
632275 [Executor task launch worker-12] ERROR 
org.apache.spark.executor.Executor  - Managed memory leak detected; size = 
5602240 bytes, TID = 47709
644989 [Executor task launch worker-13] ERROR 
org.apache.spark.executor.Executor  - Managed memory leak detected; size = 
5326260 bytes, TID = 47863
720701 [Executor task launch worker-12] ERROR 
org.apache.spark.executor.Executor  - Managed memory leak detected; size = 
5399578 bytes, TID = 48959
1147961 [Executor task launch worker-16] ERROR 
org.apache.spark.executor.Executor  - Managed memory leak detected; size = 
5251872 bytes, TID = 54922


How can I fix this?

With kind regard,

Michel



apache spark errors

2016-03-24 Thread Michel Hubert
HI,

I constantly get these errors:

0[Executor task launch worker-15] ERROR org.apache.spark.executor.Executor  
- Managed memory leak detected; size = 6564500 bytes, TID = 38969
310002 [Executor task launch worker-12] ERROR 
org.apache.spark.executor.Executor  - Managed memory leak detected; size = 
5523550 bytes, TID = 43270
318445 [Executor task launch worker-12] ERROR 
org.apache.spark.executor.Executor  - Managed memory leak detected; size = 
6879566 bytes, TID = 43408
388893 [Executor task launch worker-12] ERROR 
org.apache.spark.executor.Executor  - Managed memory leak detected; size = 
5572546 bytes, TID = 44382
418186 [Executor task launch worker-13] ERROR 
org.apache.spark.executor.Executor  - Managed memory leak detected; size = 
5289222 bytes, TID = 44795
488421 [Executor task launch worker-4] ERROR org.apache.spark.executor.Executor 
 - Managed memory leak detected; size = 8738142 bytes, TID = 45769
619276 [Executor task launch worker-4] ERROR org.apache.spark.executor.Executor 
 - Managed memory leak detected; size = 5759312 bytes, TID = 47571
632275 [Executor task launch worker-12] ERROR 
org.apache.spark.executor.Executor  - Managed memory leak detected; size = 
5602240 bytes, TID = 47709
644989 [Executor task launch worker-13] ERROR 
org.apache.spark.executor.Executor  - Managed memory leak detected; size = 
5326260 bytes, TID = 47863
720701 [Executor task launch worker-12] ERROR 
org.apache.spark.executor.Executor  - Managed memory leak detected; size = 
5399578 bytes, TID = 48959
1147961 [Executor task launch worker-16] ERROR 
org.apache.spark.executor.Executor  - Managed memory leak detected; size = 
5251872 bytes, TID = 54922


How can I fix this?

With kind regard,

Michel


Spark 1.6.0 op CDH 5.6.0

2016-03-22 Thread Michel Hubert
Hi,


I'm trying to run a Spark 1.6.0 application on a CDH 5.6.0 cluster.

How do I submit the uber-jar so it's totally self-reliant?

With kind regards,
Mitchel


spark-submit --class TEST --master yarn-cluster ./uber-TEST-1.0-SNAPSHOT.jar


Spark 1.6.1
Version: Cloudera Express 5.6.0

16/03/22 09:16:33 INFO yarn.ApplicationMaster: Unregistering ApplicationMaster 
with FAILED (diag message: User class threw exception: 
java.lang.NoSuchMethodError: 
org.apache.spark.streaming.api.java.JavaPairDStream.mapWithState(Lorg/apache/spark/streaming/StateSpec;)Lorg/apache/spark/streaming/api/java/JavaMapWithStateDStream;)





updateStateByKey schedule time

2015-07-15 Thread Michel Hubert
Hi,


I want to implement a time-out mechanism in de updateStateByKey(...) routine.

But is there a way the retrieve the time of the start of the batch 
corresponding to the call to my updateStateByKey routines?

Suppose the streaming has build up some delay then a System.currentTimeMillis() 
will not be the time of the time the batch was scheduled.

I want to retrieve the job/task schedule time of the batch for which my 
updateStateByKey(..) routine is called.

Is this possible?

With kind regards,
Michel Hubert




spark streaming performance

2015-07-09 Thread Michel Hubert

Hi,

I've developed a POC Spark Streaming application.
But it seems to perform better on my development machine  than on our cluster.
I submit it to yarn on our cloudera cluster.

But my first question is more detailed:

In de application UI (:4040) I see in the streaming section that the batch 
processing took 6 sec.
Then when I look at the stages I indeed see a stage with duration 5s.

For example:
1678

map at LogonAnalysis.scala:215+details

2015/07/09 09:17:00

5 s

50/50

173.5 KB


But when I look into the details of state 1678 it tells me the duration was 14 
ms and the aggregated metrics by executor has 1.0s as Task Time.
What is responsible for the gap between 14 ms, 1s and 5 sec?


Details for Stage 1678
* Total task time across all tasks: 0.8 s
* Shuffle write: 173.5 KB / 2031
 Show additional metrics
Summary Metrics for 50 Completed Tasks
Metric

Min

25th percentile

Median

75th percentile

Max

Duration

14 ms

14 ms

15 ms

15 ms

24 ms

GC Time

0 ms

0 ms

0 ms

0 ms

0 ms

Shuffle Write Size / Records

2.6 KB / 28

3.1 KB / 35

3.5 KB / 42

3.9 KB / 46

4.4 KB / 53

Aggregated Metrics by Executor
Executor ID

Address

Task Time

Total Tasks

Failed Tasks

Succeeded Tasks

Shuffle Write Size / Records

2

:44231

1.0 s

50

0

50

173.5 KB / 2031








RE: Breaking lineage and reducing stages in Spark Streaming

2015-07-09 Thread Michel Hubert
Hi,

I was just wondering how you generated to second image with the charts.
What product?

From: Anand Nalya [mailto:anand.na...@gmail.com]
Sent: donderdag 9 juli 2015 11:48
To: spark users
Subject: Breaking lineage and reducing stages in Spark Streaming

Hi,

I've an application in which an rdd is being updated with tuples coming from 
RDDs in a DStream with following pattern.

dstream.foreachRDD(rdd = {
  myRDD = myRDD.union(rdd.filter(myfilter)).reduceByKey(_+_)
})

I'm using cache() and checkpointin to cache results. Over the time, the lineage 
of myRDD keeps increasing and stages in each batch of dstream keeps increasing, 
even though all the earlier stages are skipped. When the number of stages grow 
big enough, the overall delay due to scheduling delay starts increasing. The 
processing time for each batch is still fixed.

Following figures illustrate the problem:

Job execution: https://i.imgur.com/GVHeXH3.png?1
[Image removed by sender.]
Delays: https://i.imgur.com/1DZHydw.png?1
[Image removed by sender.]
Is there some pattern that I can use to avoid this?

Regards,
Anand