using Spark Kmeans. Any
solutions from the experts. Would be really helpful.
val kMeans = new KMeans().setK(reductionCount).setMaxIter(30)
val kMeansModel = kMeans.fit(df)
Error is occured when calling kmeans.fit
Exception in thread "main" java.lang.OutOfMemoryError: Java
;main" java.lang.OutOfMemoryError: Java heap space
at
org.apache.spark.mllib.linalg.SparseVector.toArray(Vectors.scala:760)
at
org.apache.spark.mllib.clustering.VectorWithNorm.toDense(KMeans.scala:614)
at
org.apache.spark.mllib.clustering.KMeans$$anonfun$initKMeansParall
nt
val tmp = graph.triangleCount().vertices.filter{ case (vid, count) => count >
0 }
var numberOfTriangles = tmp.map(a => a._2).sum()
Even though it manages to compute the number of triplets, I can’t compute the
number of triangles. Every time I get an exception OOM - Java Heap Space on
s
k.util.Utils -
Suppressing exception in finally: Java heap space
java.lang.OutOfMemoryError: Java heap space
at java.nio.HeapByteBuffer.(HeapByteBuffer.java:57)
~[na:1.8.0_162]
at java.nio.ByteBuffer.allocate(ByteBuffer.java:335) ~[na:1.8.0_162]
at
org.apache.spark.broadcast.TorrentBroadc
row object increased from 20kb to about 4mb. Now the
same code runs into java heap space issues and application is shut down
with an out of memory exception.
Seems dataframe sort operations cannot handle large objects. I took an
heap dump and saw an large array-of-array. I would expect such object
is being created is huge. Could it be due to
that?
Thanks!
Vinay
18/02/17 00:59:02 ERROR Utils: throw uncaught fatal error in thread
SparkListenerBus
java.lang.OutOfMemoryError: Java heap space
at java.util.Arrays.copyOfRange(Arrays.java:3664)
at java.lang.String.(String.java:207
ang.OutOfMemoryError: Java heap space
at java.util.Arrays.copyOf(Arrays.java:3230)
at java.io.ByteArrayOutputStream.grow(ByteArrayOutputStream.java:113)
at
java.io.ByteArrayOutputStream.ensureCapacity(ByteArrayOutputStream.java:93)
at java.io.ByteArr
May be you can repartition?
2017-09-04 9:25 GMT+08:00 KhajaAsmath Mohammed :
> Hi,
>
> I am getting java.lang.OutOfMemoryError: Java heap space error whenever I
> ran the spark sql job.
>
> I came to conclusion issue is because of reading number of files from
> spark
Hi,
I am getting java.lang.OutOfMemoryError: Java heap space error whenever I
ran the spark sql job.
I came to conclusion issue is because of reading number of files from spark.
I am reading 37 partitions and each partition has around 2000 files with
filesize more than 128 MB 37*2000 files
10:09:26 INFO BlockManagerInfo: Removed taskresult_362 on
ip-...-45.dev:40963 in memory (size: 5.2 MB, free: 8.9 GB)
17/04/24 10:09:26 INFO TaskSetManager: Finished task 125.0 in stage 1.0
(TID 359) in 4383 ms on ip-...-45.dev (125/234)
#
# java.lang.OutOfMemoryError: Java heap space
Hi,
I have 1 master and 4 slave node. Input data size is 14GB.
Slave Node config : 32GB Ram,16 core
I am trying to train word embedding model using spark. It is going out of
memory. To train 14GB of data how much memory do i require?.
I have givem 20gb per executor but below shows it is using
Hi,
Need a help to figure out and solve heap space problem.
I have query which contains 15+ table and when i trying to print out the
result(Just 23 rows) it throws heap space error.
Following command i tried in standalone mode:
(My mac machine having 8 core and 15GB ram)
spark.conf().set("spark
ion(master =
>>>>>>>> "spark://ip-172-31-6-116:7077",sparkConfig=list(spark.execut
>>>>>>>> or.memory="10g",spark.app.name="Testing",spark.driver.memory
>>>>>>>> ="14g&
>>>>>> LinkedIn *
>>>>>> https://www.linkedin.com/profile/view?id=AAEWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw
>>>>>> <https://www.linkedin.com/profile/view?id=AAEWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw>*
>>>&
claimer:* Use it at your own risk. Any and all responsibility for
>>>>> any loss, damage or destruction of data or any other property which may
>>>>> arise from relying on this email's technical content is explicitly
>>>>> disclaimed. The author wi
, damage or destruction.
>>>>
>>>>
>>>>
>>>> On 24 October 2016 at 12:19, sankarmittapally <
>>>> sankar.mittapa...@creditvidya.com> wrote:
>>>>
>>>>> Hi,
>>>>>
>>>>> I have a thr
ng to
>>>> analyzing
>>>> the data of 200MB and running out of memory every time. This is the
>>>> command
>>>> I am using
>>>>
>>>> Driver Memory = 10G
>>>> Executor memory=10G
>>>>
>>>> sc
=
>>> "spark://ip-172-31-6-116:7077",sparkConfig=list(spark.execut
>>> or.memory="10g",spark.app.name="Testing",spark.driver.memory
>>> ="14g",spark.executor.extraJavaOption="-Xms2g
>>>
ver.
>> memory="14g",spark.executor.extraJavaOption="-Xms2g
>> -Xmx5g -XX:MaxPermSize=1024M",spark.driver.extraJavaOption="-Xms2g -Xmx5g
>> -XX:MaxPermSize=1024M",spark.cores.max="2"))
>>
>>
>> [D 16:43:51.437 Notebo
ms2g
> -Xmx5g -XX:MaxPermSize=1024M",spark.driver.extraJavaOption="-Xms2g -Xmx5g
> -XX:MaxPermSize=1024M",spark.cores.max="2"))
>
>
> [D 16:43:51.437 NotebookApp] 200 GET
> /api/contents?type=directory&_=1477289197671 (123.176.38.226) 7.96ms
> Exception in
ot;))
[D 16:43:51.437 NotebookApp] 200 GET
/api/contents?type=directory&_=1477289197671 (123.176.38.226) 7.96ms
Exception in thread "broadcast-exchange-0" java.lang.OutOfMemoryError: Java
heap space
am concerned that
this will reduce concurrency
Thanks
Andy
From: Ted Yu
Date: Friday, July 22, 2016 at 2:54 PM
To: Andrew Davidson
Cc: "user @spark"
Subject: Re: Exception in thread "dispatcher-event-loop-1"
java.lang.OutOfMemoryError: Java heap space
> How much
TaskSetManager: Stage 146 contains a task of very
> large size (145 KB). The maximum recommended task size is 100 KB.
>
> 16/07/22 18:39:47 WARN HeartbeatReceiver: Removing executor 2 with no
> recent heartbeats: 153037 ms exceeds timeout 12 ms
>
> Exception in th
java.lang.OutOfMemoryError:
Java heap space
at java.util.jar.Manifest$FastInputStream.(Manifest.java:332)
at java.util.jar.Manifest$FastInputStream.(Manifest.java:327)
at java.util.jar.Manifest.read(Manifest.java:195)
at java.util.jar.Manifest.(Manifest.java:69)
at java.util.jar.JarFile.getManif
>
>>>>> Hi,
>>>>>
>>>>> I have a Java memory issue with Spark. The same application working on my
>>>>> 8GB Mac crashes on my 72GB Ubuntu server...
>>>>>
>>>>> I have changed things in the conf file,
gt;> 8GB Mac crashes on my 72GB Ubuntu server...
>>>>
>>>> I have changed things in the conf file, but it looks like Spark does not
>>>> care, so I wonder if my issues are with the driver or executor.
>>>>
>>>> I set:
>>&g
20g
>>> spark.executor.memory 20g
>>> And, whatever I do, the crash is always at the same spot in the app, which
>>> makes me think that it is a driver problem.
>>>
>>> The exception I get is:
>>>
>>> 16/07/13 20
ich
>> makes me think that it is a driver problem.
>>
>> The exception I get is:
>>
>> 16/07/13 20:36:30 WARN TaskSetManager: Lost task 0.0 in stage 7.0 (TID 208,
>> micha.nc.rr.com): java.lang.OutOfMemoryError: Java heap space
>> at java.nio.HeapCharBuffer
tion I get is:
>
> 16/07/13 20:36:30 WARN TaskSetManager: Lost task 0.0 in stage 7.0 (TID 208,
> micha.nc.rr.com): java.lang.OutOfMemoryError: Java heap space
> at java.nio.HeapCharBuffer.(HeapCharBuffer.java:57)
> at java.nio.CharBuffer.allocate(CharBuffer.java:335)
>
): java.lang.OutOfMemoryError: Java heap space
at java.nio.HeapCharBuffer.(HeapCharBuffer.java:57)
at java.nio.CharBuffer.allocate(CharBuffer.java:335)
at java.nio.charset.CharsetDecoder.decode(CharsetDecoder.java:810)
at org.apache.hadoop.io.Text.decode(Text.java:412)
at
Thanks Ted for the input. I was able to get it working with pyspark shell
but the same job submitted via 'spark-submit' using client or cluster
deploy mode ends up with these errors:
~
java.lang.OutOfMemoryError: Java heap space
at java.lang.Object.clone(Native
;
>> On Sat, Apr 9, 2016 at 7:51 PM, Buntu Dev wrote:
>> > I'm running this motif pattern against 1.5M vertices (5.5mb) and 10M
>> (60mb)
>> > edges:
>> >
>> > tgraph.find("(a)-[]->(b); (c)-[]->(b); (c)-[]->(d)")
>>
is motif pattern against 1.5M vertices (5.5mb) and 10M
> (60mb)
> > edges:
> >
> > tgraph.find("(a)-[]->(b); (c)-[]->(b); (c)-[]->(d)")
> >
> > I keep running into Java heap space errors:
> >
> > ~
>
s (5.5mb) and 10M (60mb)
> edges:
>
> tgraph.find("(a)-[]->(b); (c)-[]->(b); (c)-[]->(d)")
>
> I keep running into Java heap space errors:
>
> ~
>
> ERROR actor.ActorSystemImpl: Uncaught fatal error from thread
> [sparkDriver-ak
I'm running this motif pattern against 1.5M vertices (5.5mb) and 10M (60mb)
edges:
tgraph.find("(a)-[]->(b); (c)-[]->(b); (c)-[]->(d)")
I keep running into Java heap space errors:
~
ERROR actor.ActorSystemImpl: Uncaught fatal error from thread
[sparkDriver-akka.act
down
ActorSystem [sparkDriver]
java.lang.OutOfMemoryError: Java heap space
at
com.google.protobuf.AbstractMessageLite.toByteArray(AbstractMessageLite.java:62)
at
akka.remote.transport.AkkaPduProtobufCodec$.constructMessage(AkkaPduCodec.scala:138)
at akka.remote.EndpointWriter.writeSend(Endpoint.scal
y the new
>>> version of Spark.
>>>
>>>
>>>
>>> So I want to know any new setup I should set in Spark 1.5 to make it
>>> work?
>>>
>>>
>>>
>>> Regards,
>>>
>>>
>>>
>>> S
gt;> *From:* Shuai Zheng [mailto:szheng.c...@gmail.com]
>> *Sent:* Wednesday, November 04, 2015 3:22 PM
>> *To:* user@spark.apache.org
>> *Subject:* [Spark 1.5]: Exception in thread "broadcast-hash-join-2"
>> java.lang.OutOfMemoryError: Java heap space
>>
>>
>>
&
oin-2"
java.lang.OutOfMemoryError: Java heap space
Hi All,
I have a program which actually run a bit complex business (join) in spark.
And I have below exception:
I running on Spark 1.5, and with parameter:
spark-submit --deploy-mode client --executor-cores=24 --driver-memory=2G
;spark.sql.autoBroadcastJoinThreshold",
"104857600");
This is running on AWS c3*8xlarge instance. I am not sure what kind of
parameter I should set if I have below OutOfMemoryError exception.
#
# java.lang.OutOfMemoryError: Java heap space
# -XX:OnOutOfMemoryError="kill -9
y what you want, to consolidate all the
> productionName and product catagory together, without even consider removing
> duplication.
>
> But both query still should push similar records count per partition, but
> with much of different volume size of data.
>
> Yong
>
> S
ving
> duplication.
>
> But both query still should push similar records count per partition, but
> with much of different volume size of data.
>
> Yong
>
> Subject: Re: Java Heap Space Error
> From: yu...@useinsider.com
> Date: Thu, 24 Sep 2015 18:56:51 +0300
&
r, without even consider removing duplication.
But both query still should push similar records count per partition, but with
much of different volume size of data.
Yong
Subject: Re: Java Heap Space Error
From: yu...@useinsider.com
Date: Thu, 24 Sep 2015 18:56:51 +0300
CC: jingyu.zh...@news.com.au; u
e is huge and the others are very small.
>
>
> ——
> So how can i balance this shuffle read size between partitions?
>
>
> On 24 Sep 2015, at 03:35, Zhang, Jingyu <mailto:jingyu.zh...@news.com.au>> wrote:
>
> Is you sql works if do not runs a regex on
simply depends on the Hash partitioner of "userid".
Can you show us the query after you add "regex" and "concatenation"?
Yong
Subject: Re: Java Heap Space Error
From: yu...@useinsider.com
Date: Thu, 24 Sep 2015 15:34:48 +0300
CC: user@spark.apache.org
To: jingyu.zh...@
the stuff without String operations?
>
> On 24 September 2015 at 10:11, java8964 <mailto:java8...@hotmail.com>> wrote:
> Try to increase partitions count, that will make each partition has less data.
>
> Yong
>
> Subject: Re: Java Heap Space Error
> From: yu...@useinsi
t;
> Yong
>
> --
> Subject: Re: Java Heap Space Error
> From: yu...@useinsider.com
> Date: Thu, 24 Sep 2015 00:32:47 +0300
> CC: user@spark.apache.org
> To: java8...@hotmail.com
>
>
> Yes, it’s possible. I use S3 as data source. My external tables has
> partition
Try to increase partitions count, that will make each partition has less data.
Yong
Subject: Re: Java Heap Space Error
From: yu...@useinsider.com
Date: Thu, 24 Sep 2015 00:32:47 +0300
CC: user@spark.apache.org
To: java8...@hotmail.com
Yes, it’s possible. I use S3 as data source. My external
query, and this is my guess of what happening), then it simple means that one
> partition having way more data than the rest of partitions.
>
> Yong
>
> From: yu...@useinsider.com
> Subject: Java Heap Space Error
> Date: Wed, 23 Sep 2015 23:07:17 +0300
> To: user@spark.
, which led to the dreaded Java heap space error. See the
stack trace at the end of this message.
When this happens, I see 10's of executors in "EXITED" state, a couple in
"LOADING" and one in "RUNNING". All of them are retrying the same task all
over again, and ke
une 11, 2015 8:43 AM
Subject: spark-sql from CLI --->EXCEPTION: java.lang.OutOfMemoryError: Java
heap space
hey guys
Using Hive and Impala daily intensively.Want to transition to spark-sql in CLI
mode
Currently in my sandbox I am using the Spark (standalone mode) in the CDH
distribution (sta
y questions on all CDH groups,
>> Spark, Hive
>>
>> best regards
>>
>> sanjay
>>
>>
>>
>> From: Josh Rosen
>> To: Sanjay Subramanian
>> Cc: "user@spark.apache.org"
>> Sent: Friday, June 12, 2015 7:15 AM
>&g
t`,
>`aers_demo_v5`.`age`,
>`aers_demo_v5`.`age_cod`,
>`aers_demo_v5`.`gndr_cod`,
>`aers_demo_v5`.`year`,
>`aers_demo_v5`.`quarter`
> FROM
> `aers`.`aers_demo_v5`
> UNION ALL
> SELECT
>`aers_demo_v6`.`primaryid` AS `ISR`,
:-) to my questions on all CDH groups, Spark, Hive
best regards
sanjay
From: Josh Rosen
To: Sanjay Subramanian
Cc: "user@spark.apache.org"
Sent: Friday, June 12, 2015 7:15 AM
Subject: Re: spark-sql from CLI --->EXCEPTION: java.lang.OutOfMemoryError:
Java heap space
` AS `ISR`,
>`aers_demo_v6`.`event_dt`,
>`aers_demo_v6`.`age`,
>`aers_demo_v6`.`age_cod`,
>`aers_demo_v6`.`sex` AS `GNDR_COD`,
>`aers_demo_v6`.`year`,
> `aers_demo_v6`.`quarter`
> FROM
> `aers`.`aers_demo_v6`) `aers_demo_view`
>
>
>
>
ELECT
>`aers_demo_v6`.`primaryid` AS `ISR`,
>`aers_demo_v6`.`event_dt`,
>`aers_demo_v6`.`age`,
>`aers_demo_v6`.`age_cod`,
>`aers_demo_v6`.`sex` AS `GNDR_COD`,
>`aers_demo_v6`.`year`,
> `aers_demo_v6`.`quarter`
> FROM
> `aers`.`aers_demo_v6`)
hrown by a
user handler while handling an exception event ([id: 0x01b99855,
/10.0.0.19:58117 => /10.0.0.19:52016] EXCEPTION: java.lang.OutOfMemoryError:
Java heap space)java.lang.OutOfMemoryError: Java heap space at
org.jboss.netty.buffer.HeapChannelBuffer.(HeapChannelBuffer.java:4
Try increasing your driver memory.
Thanks
Best Regards
On Thu, Apr 16, 2015 at 6:09 PM, sarath wrote:
> Hi,
>
> I'm trying to train an SVM on KDD2010 dataset (available from libsvm). But
> I'm getting "java.lang.OutOfMemoryError: Java heap space" error. The
Hi,
I'm trying to train an SVM on KDD2010 dataset (available from libsvm). But
I'm getting "java.lang.OutOfMemoryError: Java heap space" error. The dataset
is really sparse and have around 8 million data points and 20 million
features. I'm using a cluster of 8 nodes (each
ecutor, it
>> will lower the memory requirement, with running in a slower speed.
>>
>> Yong
>>
>> --
>> Date: Wed, 8 Apr 2015 04:57:22 +0800
>> Subject: Re: 'Java heap space' error occured when query 4G data file from
oncurrency of your executor, it
> will lower the memory requirement, with running in a slower speed.
>
> Yong
>
> --
> Date: Wed, 8 Apr 2015 04:57:22 +0800
> Subject: Re: 'Java heap space' error occured when query 4G data file from
&
ower the
cores for executor by set "-Dspark.deploy.defaultCores=". When you have not
enough memory, reduce the concurrency of your executor, it will lower the
memory requirement, with running in a slower speed.
Yong
Date: Wed, 8 Apr 2015 04:57:22 +0800
Subject: Re: 'Java heap space' error
:11:03 INFO BlockManagerInfo: Added taskresult_29 on disk on
> cloud2:49451 (size: 163.7 MB)
> 15/04/07 18:11:03 INFO BlockManagerInfo: Added taskresult_29 on disk on
> cloud2:49451 (size: 163.7 MB)
> 15/04/07 18:11:03 INFO TaskSetManager: Starting task 30.0 in stage 1.0
> (TID 32, cloud2
)
15/04/07 18:11:03 ERROR Utils: Uncaught exception in thread
task-result-getter-0
java.lang.OutOfMemoryError: Java heap space
at
org.apache.spark.scheduler.DirectTaskResult$$anonfun$readExternal$1.apply$mcV$sp(TaskResult.scala:61)
at org.apache.spark.util.Utils$.tryOrIOException(Utils.scala:985)
at
Hi all,
sometimes you can see "OutOfMemoryException: Java heap space" of executor in
Spark. There many ideas about how to work arounds.
My question is: how does executor execute tasks from the point of view of
memory usage and parallelism?
Picture in my mind is:
Executor is JVM instan
eeSpace(61480) called with cur
> Mem=270794224, maxMem=311387750
> java.lang.OutOfMemoryError: Java heap space
> at java.io.BufferedOutputStream.(Unknown Source)
> at
> org.apache.spark.api.python.PythonRDD$$anon$2.run(PythonRDD.scala:62)
> 14/07/09 01:46:15
pache-spark-user-list.1001560.n3.nabble.com/KMeans-with-large-clusters-Java-Heap-Space-tp21432p22153.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.
-
To unsubscribe, e-mail: user-unsubscr...@spar
e has
> ideas on this? thanks so much!
>
> (I think the memory is sufficient, spark.executor.memory 30GB )
>
>
> 15/02/09 00:37:12 ERROR Executor: Exception in task 162.0 in stage 719.0 (TID
> 7653)
> java.lang.OutOfMemoryError: Java heap space
> a
nt, spark.executor.memory 30GB )
>>
>>
>> 15/02/09 00:37:12 ERROR Executor: Exception in task 162.0 in stage 719.0
>> (TID 7653)
>> java.lang.OutOfMemoryError: Java heap space
>> at
>> com.esotericsoftware.kryo.util.IdentityObjectIntMap.resize
>
>> On Mon, Feb 9, 2015 at 3:30 PM, Yifan LI > <mailto:iamyifa...@gmail.com>> wrote:
>> Hi,
>>
>> I just found the following errors during computation(graphx), anyone has
>> ideas on this? thanks so much!
>>
>> (I think th
>> I just found the following errors during computation(graphx), anyone has
>> ideas on this? thanks so much!
>>
>> (I think the memory is sufficient, spark.executor.memory 30GB )
>>
>>
>> 15/02/09 0
t; 15/02/09 00:37:12 ERROR Executor: Exception in task 162.0 in stage 719.0 (TID
> 7653)
> java.lang.OutOfMemoryError: Java heap space
> at
> com.esotericsoftware.kryo.util.IdentityObjectIntMap.resize(IdentityObjectIntMap.java:410)
> at
> com.esotericsoft
>
> (I think the memory is sufficient, spark.executor.memory 30GB )
>
>
> 15/02/09 00:37:12 ERROR Executor: Exception in task 162.0 in stage 719.0 (TID
> 7653)
> java.lang.OutOfMemoryError: Java heap space
> at
> com.esotericsoftware.kryo.util.IdentityObjectIntM
heap space
at
com.esotericsoftware.kryo.util.IdentityObjectIntMap.resize(IdentityObjectIntMap.java:410)
at
com.esotericsoftware.kryo.util.IdentityObjectIntMap.put(IdentityObjectIntMap.java:113)
at
com.esotericsoftware.kryo.util.MapReferenceResolver.addWrittenObject
Vector], k: Int): KMeansModel = {
KMeans.train(raw, k, embeddingNames = List(LOW_DIMENSIONAL_RI)
}
--
View this message in context:
http://apache-spark-user-list.1001560.n3.nabble.com/KMeans-with-large-clusters-Java-Heap-Space-tp21432p21437.html
Sent from the Apache Spark User List mailin
getting Java Heap
Space error. Working with 3 nodes cluster with each 8 GB memory and 2 cores.
Played with different configuration, but no luck...
what am I missing any suggestions?
here is my code
val sparkConf = new SparkConf().setMaster("spark://master:7077")
.setAppName("Spa
Hi Kane-
http://spark.apache.org/docs/latest/tuning.html has excellent information that
may be helpful. In particular increasing the number of tasks may help, as well
as confirming that you don’t have more data than you're expecting landing on a
key.
Also, if you are using spark < 1.2.0, set
I'm trying to process a large dataset, mapping/filtering works ok, but
as long as I try to reduceByKey, I get out of memory errors:
http://pastebin.com/70M5d0Bn
Any ideas how I can fix that?
Thanks.
-
To unsubscribe, e-mail: us
t;) //default is 0.2
> conf.set("spark.storage.memoryFraction","0.3")//default is 0.6
> I have to set rank value under 40, otherwise occure this problem.
>
>
>
> --
> View this message in context:
> http://apache-spark-user-list.1001560.n3.nabble.com/MLLib-AL
ction","0.65") //default is 0.2
conf.set("spark.storage.memoryFraction","0.3")//default is 0.6
I have to set rank value under 40, otherwise occure this problem.
--
View this message in context:
http://apache-spark-user-list.1001560.n3.nabble.com/MLLib-ALS-ja
memory. I tried with 100 executors.Can some one
> please point me in the right direction ?Thanks,Jay
--
View this message in context:
http://apache-spark-user-list.1001560.n3.nabble.com/MLLib-ALS-java-lang-OutOfMemoryError-Java-heap-space-tp20584p20714.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.
you can try to decrease the rank value.
--
View this message in context:
http://apache-spark-user-list.1001560.n3.nabble.com/MLLib-ALS-java-lang-OutOfMemoryError-Java-heap-space-tp20584p20711.html
Sent from the Apache Spark User List mailing list archive at Nabble.com
How many working nodes do these 100 executors locate at?
--
View this message in context:
http://apache-spark-user-list.1001560.n3.nabble.com/MLLib-ALS-java-lang-OutOfMemoryError-Java-heap-space-tp20584p20610.html
Sent from the Apache Spark User List mailing list archive at Nabble.com
Hi all,
I am running into an out of memory error while running ALS using MLLIB on a
reasonably small data set consisting of around 6 Million ratings.
The stack trace is below:
java.lang.OutOfMemoryError: Java heap space
at org.jblas.DoubleMatrix.(DoubleMatrix.java:323)
at
113:34602], 1 messages pending
14/10/20 22:38:41 INFO ConnectionManager: Accepted connection from
[cse-hadoop-113/192.168.0.113]
Exception in thread "pool-5-thread-3" java.lang.OutOfMemoryError: Java heap
space
at java.nio.HeapByteBuffer.(HeapByteBuffer.java:5
--conf spark.driver.memory=2g \
> > target/scala-2.10/my-job_2.10-1.0.jar
> >
> >
> > I get the following error :
> >
> > Exception in thread "stdin writer for List(patch_matching_similarity)"
> > java.lang.OutOfMemoryError: Java heap space
>
\
> --conf spark.executor.memory=4g \
> --conf spark.driver.memory=2g \
> target/scala-2.10/my-job_2.10-1.0.jar
>
>
> I get the following error :
>
> Exception in thread "stdin writer for List(patch_matching_similarity)
alue.jobs.MyJob \*
* --master local[4] \*
* --conf spark.executor.memory=4g \*
* --conf spark.driver.memory=2g \*
* target/scala-2.10/my-job_2.10-1.0.jar*
I get the following error :
*Exception in thread "stdin writer for List(patch_matching_similarity)"
java.la
(rectangle) associated
with image id. My goal is to draw these primitives on the corresponding
image. So my first attempt is to join images and primitives by image ids
and then do the drawing.
But, when I do
*primitives.join(images) *
I got the following error :
*java.lang.OutOfMemoryError: Java heap s
hat looks like the same thing. I'll follow the Jira ticket
> for updates.
>
>
>
> --
> View this message in context:
> http://apache-spark-user-list.1001560.n3.nabble.com/Yarn-Driver-OOME-Java-heap-space-when-executors-request-map-output-locations-tp13827p13829.html
> Sent
Thanks Marcelo, that looks like the same thing. I'll follow the Jira ticket
for updates.
--
View this message in context:
http://apache-spark-user-list.1001560.n3.nabble.com/Yarn-Driver-OOME-Java-heap-space-when-executors-request-map-output-locations-tp13827p13829.html
Sent from the A
anyone else run into this? Maybe I'm misunderstanding the underlying
> cause. I don't have a copy of the stack trace handy but can recreate it if
> necessary. It was somewhere in the for HeapByteBuffer. Any advice
> would be helpful.
>
>
>
> --
> View this message i
into this? Maybe I'm misunderstanding the underlying
cause. I don't have a copy of the stack trace handy but can recreate it if
necessary. It was somewhere in the for HeapByteBuffer. Any advice
would be helpful.
--
View this message in context:
http://apache-spark-user-list.1001560.n3.nabble.com/Ya
153-141-230.ec2.internal
> :53906
> 14/08/21 19:29:32 ERROR actor.ActorSystemImpl: Uncaught fatal error from
> thread [spark-akka.actor.default-dispatcher-20] shutting down ActorSystem
> [spark]
> java.lang.OutOfMemoryError: Java heap space
> at
> com.go
1 19:29:32 ERROR actor.ActorSystemImpl: Uncaught fatal error from
thread [spark-akka.actor.default-dispatcher-20] shutting down ActorSystem
[spark]
java.lang.OutOfMemoryError: Java heap space
at
com.google.protobuf_spark.AbstractMessageLite.toByteArray(AbstractMessageLite.java:62)
g 2 non-empty blocks out of 2 blocks
>> 14/07/31 09:48:09 INFO BlockFetcherIterator$BasicBlockFetcherIterator:
>> Getting 2 non-empty blocks out of 2 blocks
>> 14/07/31 09:48:09 INFO BlockFetcherIterator$BasicBlockFetcherIterator:
>> Started 0 remote fetch
t; 14/07/31 09:48:09 INFO BlockFetcherIterator$BasicBlockFetcherIterator:
> Started 0 remote fetches in 1 ms
> 14/07/31 09:48:17 ERROR Executor: Exception in task ID 5
> java.lang.OutOfMemoryError: Java heap space
> at java.util.Arrays.copyOf(Arrays.java:2271)
>
fetches in 1
ms14/07/31 09:48:09 INFO BlockFetcherIterator$BasicBlockFetcherIterator:
Started 0 remote fetches in 1 ms14/07/31 09:48:17 ERROR Executor: Exception in
task ID 5java.lang.OutOfMemoryError: Java heap space at
java.util.Arrays.copyOf(Arrays.java:2271)at
java.io.ByteArrayOu
, init =
626, finish = 162
Exception in thread "stdin writer for python" 14/07/09 01:46:14 INFO
MemoryStore: ensureFreeSpace(61480) called with cur
Mem=270794224, maxMem=311387750
java.lang.OutOfMemoryError: Java heap space
at java.io.BufferedOutputStream.(Unknown Source)
loop. java.lang.OutOfMemoryError: Java heap space
Also, the machine on which the driver program runs constantly uses about
7~8% of 100Mbps network connection.
Is the driver program involved in the reduceByKey() somehow?
BTW, currently an accumulator is used, but the network usage does not drop
even when accumulat
1 - 100 of 121 matches
Mail list logo