Re: Exception in task 0.0 in stage 13.0 (TID 13) java.lang.OutOfMemoryError: Java heap space

2016-02-16 Thread Dmitriy Lyubimov
Caused by: java.lang.OutOfMemoryError: GC overhead limit exceeded > ……. > ….. > ... > .. > . > > > Thanks for advance > > El 2/2/2016, a las 7:48, Pat Ferrel mailto:p...@occamsmachete.com >> escribió: > > You probably need to increase your driver memor

Re: Exception in task 0.0 in stage 13.0 (TID 13) java.lang.OutOfMemoryError: Java heap space

2016-02-16 Thread Dmitriy Lyubimov
"main" org.apache.spark.SparkException: Job aborted >> due to stage failure: Task 0 in stage 12.0 failed 1 times, most recent >> failure: Lost task 0.0 in stage 12.0 (TID 24, localhost): >> java.lang.OutOfMemoryError: GC overhead limit exceeded >> ……. >> …..

Re: Exception in task 0.0 in stage 13.0 (TID 13) java.lang.OutOfMemoryError: Java heap space

2016-02-16 Thread Dmitriy Lyubimov
;> escribió: > > You probably need to increase your driver memory and 8g will not work. 16g > is probably the smallest stand alone machine that will work since the > driver and executors run on it. > > On Feb 1, 2016, at 1:24 AM, jg...@konodrac.com <mailto:jg...@konodrac.com

Re: Exception in task 0.0 in stage 13.0 (TID 13) java.lang.OutOfMemoryError: Java heap space

2016-02-13 Thread Pat Ferrel
ad limit exceeded >>>> ……. >>>> ….. >>>> .. >>>> . >>>> >>>> Driver stacktrace: >>>> Caused by: java.lang.OutOfMemoryError: GC overhead limit exceeded >>>> ……. >>>> ….. >>&g

Re: Exception in task 0.0 in stage 13.0 (TID 13) java.lang.OutOfMemoryError: Java heap space

2016-02-13 Thread Angelo Leto
gt;>> .. >>> . >>> >>> >>> Thanks for advance >>> >>>> El 2/2/2016, a las 7:48, Pat Ferrel >>> <mailto:p...@occamsmachete.com>> escribió: >>>> >>>> You probably need to increase your dr

Re: Exception in task 0.0 in stage 13.0 (TID 13) java.lang.OutOfMemoryError: Java heap space

2016-02-13 Thread Angelo Leto
chete.com>> escribió: >>> >>> You probably need to increase your driver memory and 8g will not work. 16g >>> is probably the smallest stand alone machine that will work since the >>> driver and executors run on it. >>> >>>&

Re: Exception in task 0.0 in stage 13.0 (TID 13) java.lang.OutOfMemoryError: Java heap space

2016-02-12 Thread Pat Ferrel
t stand alone machine that will work since the driver >> and executors run on it. >> >>> On Feb 1, 2016, at 1:24 AM, jg...@konodrac.com <mailto:jg...@konodrac.com> >>> wrote: >>> >>> Hello everybody, >>> >>> We are exp

Re: Exception in task 0.0 in stage 13.0 (TID 13) java.lang.OutOfMemoryError: Java heap space

2016-02-10 Thread Jaume Galí
t; >> wrote: >> >> Hello everybody, >> >> We are experimenting problems when we use "mahout spark-rowsimilarity” >> operation. We have an input matrix with 100k rows and 100 items and process >> throws an exception about “Exception in task 0.0 in

Re: Exception in task 0.0 in stage 13.0 (TID 13) java.lang.OutOfMemoryError: Java heap space

2016-02-01 Thread Pat Ferrel
g problems when we use "mahout spark-rowsimilarity” > operation. We have an input matrix with 100k rows and 100 items and process > throws an exception about “Exception in task 0.0 in stage 13.0 (TID 13) > java.lang.OutOfMemoryError: Java heap space” and we try to increase JAVA HEAP

Exception in task 0.0 in stage 13.0 (TID 13) java.lang.OutOfMemoryError: Java heap space

2016-02-01 Thread jgali
Hello everybody, We are experimenting problems when we use "mahout spark-rowsimilarity” operation. We have an input matrix with 100k rows and 100 items and process throws an exception about “Exception in task 0.0 in stage 13.0 (TID 13) java.lang.OutOfMemoryError: Java heap space” and we t

Re: spark-rowsimilarity java.lang.OutOfMemoryError: Java heap space

2015-05-19 Thread Pat Ferrel
cause this and is fixed by increasing > you client JVM heap space but try the above first. > > BTW setting the executor memory twice, is not necessary. > > > On May 13, 2015, at 2:21 AM, Xavier Rampino > wrote: > > Hello, > > I've tried spark-rowsimilarity

Re: spark-rowsimilarity java.lang.OutOfMemoryError: Java heap space

2015-05-18 Thread Xavier Rampino
org/apache/mahout/drivers/TextDelimitedReaderWriter.scala#L157 > > remove the line that says: interactions.collect() > > See this Jira https://issues.apache.org/jira/browse/MAHOUT-1707 > > There is one other thing that can cause this and is fixed by increasing > you client JVM heap

Re: spark-rowsimilarity java.lang.OutOfMemoryError: Java heap space

2015-05-13 Thread Pat Ferrel
the line that says: interactions.collect() See this Jira https://issues.apache.org/jira/browse/MAHOUT-1707 There is one other thing that can cause this and is fixed by increasing you client JVM heap space but try the above first. BTW setting the executor memory twice, is not necessary. On May 13

spark-rowsimilarity java.lang.OutOfMemoryError: Java heap space

2015-05-13 Thread Xavier Rampino
Hello, I've tried spark-rowsimilarity with out-of-the-box setup (downloaded mahout distribution and spark, and set up the PATH), and I stumble upon a Java Heap space error. My input file is ~100MB. It seems the various parameters I tried to give won't change this. I do : ~/mahout-di

Re: java heap space error - Naive Bayes

2015-02-23 Thread chirag lakhani
This is what I am running from command line mahout trainnb -i /trainingMahout/ -o /outputNaiveBayes/ -el -li /labelIndex -a .5 -ow On Mon, Feb 23, 2015 at 1:17 PM, chirag lakhani wrote: > I am trying to train a Naive Bayes model in Mahout and I keep getting a > java heap space error. I

java heap space error - Naive Bayes

2015-02-23 Thread chirag lakhani
I am trying to train a Naive Bayes model in Mahout and I keep getting a java heap space error. It is a strange thing because I am using a hashing vectorizer where I am fitting 1-gram and 2-gram tokens into a vector of size 2^20. My cluster consists of 7 nodes with 16 gb of ram and each time I

Java heap space using TrainNaiveBayesJob

2014-08-23 Thread Sergi -
lts in the following exception: Exception in thread "main" java.lang.OutOfMemoryError: Java heap space at org.apache.mahout.math.map.OpenIntDoubleHashMap.rehash(OpenIntDoubleHashMap.java:491) at org.apache.mahout.math.map.OpenIntDoubleHashMap.put(OpenIntDoubleHashMap.

Re: Error running Mahout NB Classifier in Train step - java.lang.OutOfMemoryError: Java heap space

2014-07-11 Thread Alex Ott
Hi, > > I am getting following error while training a model. All Mapreduce runs > fine without error on HDP Ambari1.5(as per logs). But, I assume, while > creating models, it gives following error. Please help > > Exception in thread "main" java.lang.OutOfMemoryErro

Error running Mahout NB Classifier in Train step - java.lang.OutOfMemoryError: Java heap space

2014-07-10 Thread prabhat
Hi, I am getting following error while training a model. All Mapreduce runs fine without error on HDP Ambari1.5(as per logs). But, I assume, while creating models, it gives following error. Please help Exception in thread "main" java.lang.OutOfMemoryError: Java heap space

Re: Heap space

2014-03-11 Thread Mahmood Naderan
14 4:00 PM, Mahmood Naderan >  wrote: > > Excuse me, I added the -Xmx option and restarted the > hadoop services using > sbin/stop-all.sh && sbin/start-all.sh > > however still I get heap size error. How can I find the correct and needed > heap size? > > &

Re: Heap space

2014-03-11 Thread Andrew Musselman
l.sh && sbin/start-all.sh > > however still I get heap size error. How can I find the correct and needed > heap size? > > > Regards, > Mahmood > > > > On Sunday, March 9, 2014 1:37 PM, Mahmood Naderan > wrote: > > OK I found that I have to add this

Re: Heap space

2014-03-11 Thread Mahmood Naderan
e: Hello, I ran this command     ./bin/mahout wikipediaXMLSplitter -d examples/temp/enwiki-latest-pages-articles.xml -o wikipedia/chunks -c 64 but got this error Exception in thread "main" java.lang.OutOfMemoryError: Java heap space There are many web pages regarding this and the so

Re: Heap space

2014-03-11 Thread Mahmood Naderan
ch 9, 2014 11:39 AM, Mahmood Naderan wrote: Hello, I ran this command     ./bin/mahout wikipediaXMLSplitter -d examples/temp/enwiki-latest-pages-articles.xml -o wikipedia/chunks -c 64 but got this error Exception in thread "main" java.lang.OutOfMemoryError: Java heap space The

Re: Heap space

2014-03-10 Thread Mahmood Naderan
Regards, Mahmood On Sunday, March 9, 2014 11:39 AM, Mahmood Naderan wrote: Hello, I ran this command     ./bin/mahout wikipediaXMLSplitter -d examples/temp/enwiki-latest-pages-articles.xml -o wikipedia/chunks -c 64 but got this error Exception in thread "main" java.lang.OutOfMemor

Re: Heap space

2014-03-10 Thread Mahmood Naderan
latest-pages-articles.xml -o wikipedia/chunks -c 64 but got this error Exception in thread "main" java.lang.OutOfMemoryError: Java heap space There are many web pages regarding this and the solution is to add "-Xmx 2048M" for example. My question is, that option should be p

Re: Heap space

2014-03-10 Thread Suneel Marthi
h 9, 2014 11:39 AM, Mahmood Naderan wrote: Hello, I ran this command     ./bin/mahout wikipediaXMLSplitter -d examples/temp/enwiki-latest-pages-articles.xml -o wikipedia/chunks -c 64 but got this error Exception in thread "main" java.lang.OutOfMemoryError: Java heap space The

Re: Heap space

2014-03-09 Thread Mahmood Naderan
opts -Xmx2048m   Regards, Mahmood On Sunday, March 9, 2014 11:39 AM, Mahmood Naderan wrote: Hello, I ran this command     ./bin/mahout wikipediaXMLSplitter -d examples/temp/enwiki-latest-pages-articles.xml -o wikipedia/chunks -c 64 but got this error Exception in thread "m

Re: Heap space

2014-03-09 Thread Suneel Marthi
/chunks -c 64 but got this error Exception in thread "main" java.lang.OutOfMemoryError: Java heap space There are many web pages regarding this and the solution is to add "-Xmx 2048M" for example. My question is, that option should be passed to java command and not Mahout. As  result

RE: Heap space

2014-03-09 Thread Jason Xin
, 2014 4:26 PM To: user@mahout.apache.org Subject: Re: Heap space I usually do try and error. Start with some very large value and do a binary search :) --sebastian On 03/09/2014 01:30 PM, Mahmood Naderan wrote: > Excuse me, I added the -Xmx option and restarted the hadoop services > usin

Re: Heap space

2014-03-09 Thread Sebastian Schelter
11:39 AM, Mahmood Naderan wrote: Hello, I ran this command ./bin/mahout wikipediaXMLSplitter -d examples/temp/enwiki-latest-pages-articles.xml -o wikipedia/chunks -c 64 but got this error Exception in thread "main" java.lang.OutOfMemoryError: Java heap space There are many web pa

Re: Heap space

2014-03-09 Thread Mahmood Naderan
ikipedia/chunks -c 64 but got this error Exception in thread "main" java.lang.OutOfMemoryError: Java heap space There are many web pages regarding this and the solution is to add "-Xmx 2048M" for example. My question is, that option should be passed to java command

Re: Heap space

2014-03-09 Thread Mahmood Naderan
-articles.xml -o wikipedia/chunks -c 64 but got this error Exception in thread "main" java.lang.OutOfMemoryError: Java heap space There are many web pages regarding this and the solution is to add "-Xmx 2048M" for example. My question is, that option should be passed to java com

Heap space

2014-03-09 Thread Mahmood Naderan
Hello, I ran this command     ./bin/mahout wikipediaXMLSplitter -d examples/temp/enwiki-latest-pages-articles.xml -o wikipedia/chunks -c 64 but got this error Exception in thread "main" java.lang.OutOfMemoryError: Java heap space There are many web pages regarding this and the s

Re: OutOfMemoryError: Java Heap Space in DocumentProcessor.tokenizeDocuments

2014-02-24 Thread Justin Kay
limit exceeded" caused by running out of heap space. (I've tried running it with the -XX:-UseGCOverheadLimit option and it just gives me the same Java heap error.) I've also tried running it with Hadoop 1.2.1 and Mahout 0.8 and had the same problem. On Sat, Feb 22, 2014 at 12:22 PM, Jo

Re: OutOfMemoryError: Java Heap Space in DocumentProcessor.tokenizeDocuments

2014-02-22 Thread Johannes Schulte
g has seemed to work. My Java heap settings are similar and > configured to "-Xms512m -Xmx1536m" when running the project. The data I'm > using is 100,000 sequence files totally ~250mb. It doesn't fail on a data > set of 63 sequence files ~2mb. Here is an example stack trace: &

OutOfMemoryError: Java Heap Space in DocumentProcessor.tokenizeDocuments

2014-02-20 Thread Justin Kay
equence files totally ~250mb. It doesn't fail on a data set of 63 sequence files ~2mb. Here is an example stack trace: Exception in thread "Thread-18" java.lang.OutOfMemoryError: Java heap space at sun.util.resources.TimeZoneNames.getContents(TimeZoneNames.java:205) at sun.util.resources.Op

Running seq2sparse on Amazon EMR / Error: Java heap space

2013-08-21 Thread Necati Demir
a:597)* *at org.apache.hadoop.util.RunJar.main(RunJar.java:187)* with lots of *2013-08-21 17:59:50,995 INFO org.apache.hadoop.mapred.JobClient (main): Task Id : attempt_201308211756_0001_m_00_2, Status : FAILED* *Error: Java heap space* Note: Minimum size of a chunk file is 110 MB.

Re: mahout clusterdump : java.lang.OutOfMemoryError: Java heap space error

2012-10-12 Thread Rajesh Nikam
> > > mahout clusterdump -dt sequencefile -i > > clean-kmeans-clusters/clusters-1-final/part-r-0 -n 20 -b 100 -o > > cdump-clean.txt -p clean-kmeans-clusters/clusteredPoints/ > > > > Exception in thread "main" java.lang.OutOfMemoryError: Java he

Re: mahout clusterdump : java.lang.OutOfMemoryError: Java heap space error

2012-10-12 Thread paritosh ranjan
; Java OutOfMemory error. > > mahout clusterdump -dt sequencefile -i > clean-kmeans-clusters/clusters-1-final/part-r-0 -n 20 -b 100 -o > cdump-clean.txt -p clean-kmeans-clusters/clusteredPoints/ > > Exception in thread "main" java.lang.OutOfMemoryError: Java heap space >

mahout clusterdump : java.lang.OutOfMemoryError: Java heap space error

2012-10-12 Thread Rajesh Nikam
nal/part-r-0 -n 20 -b 100 -o cdump-clean.txt -p clean-kmeans-clusters/clusteredPoints/ Exception in thread "main" java.lang.OutOfMemoryError: Java heap space at org.apache.mahout.math.DenseVector.(DenseVector.java:44) at org.apache.mahout.math.DenseVector.(DenseVector

Re: Heap Space Problem while running in cluster in map reduce

2012-10-04 Thread paritosh ranjan
found it is throwing > Java heap space > at > > org.apache.mahout.math.map.OpenIntDoubleHashMap.rehash(OpenIntDoubleHashMap.java:434) > at > > org.apache.mahout.math.map.OpenIntDoubleHashMap.put(OpenIntDoubleHash

Re: KMeans job fails during 2nd iteration. Java Heap space

2012-08-09 Thread Ted Dunning
t; >>>> >>>> Thank you! >>>> >>>> Pavel >>>> >>>> >>>> 08.08.12 16:15 пользователь "Jeff Eastman" >>>> написал: >>>> >>>>> Consider that each cluster retains 4

Re: KMeans job fails during 2nd iteration. Java Heap space

2012-08-09 Thread Ted Dunning
>>> Pavel >>> >>> >>> 08.08.12 16:15 пользователь "Jeff Eastman" >>> написал: >>> >>>> Consider that each cluster retains 4 vectors in memory in each mapper >>>> and reducer, and that these vectors tend to become

Re: KMeans job fails during 2nd iteration. Java Heap space

2012-08-09 Thread Paritosh Ranjan
in each mapper and reducer, and that these vectors tend to become more dense (through addition of multiple sparse components) as iterations proceed. With 1000 clusters and 200k terms in your dictionary this will cause the heap space to be consumed rapidly as you have noted. Some times you can w

Re: KMeans job fails during 2nd iteration. Java Heap space

2012-08-09 Thread Abramov Pavel
er that each cluster retains 4 vectors in memory in each mapper >>>and reducer, and that these vectors tend to become more dense (through >>>addition of multiple sparse components) as iterations proceed. With 1000 >>>clusters and 200k terms in your dictionary this will cause th

Re: KMeans job fails during 2nd iteration. Java Heap space

2012-08-08 Thread Lance Norskog
reducer, and that these vectors tend to become more dense (through >>addition of multiple sparse components) as iterations proceed. With 1000 >>clusters and 200k terms in your dictionary this will cause the heap >>space to be consumed rapidly as you have noted. Some times you can work >

Re: KMeans job fails during 2nd iteration. Java Heap space

2012-08-08 Thread Abramov Pavel
that these vectors tend to become more dense (through >addition of multiple sparse components) as iterations proceed. With 1000 >clusters and 200k terms in your dictionary this will cause the heap >space to be consumed rapidly as you have noted. Some times you can work >around this problem

Re: KMeans job fails during 2nd iteration. Java Heap space

2012-08-08 Thread Jeff Eastman
Consider that each cluster retains 4 vectors in memory in each mapper and reducer, and that these vectors tend to become more dense (through addition of multiple sparse components) as iterations proceed. With 1000 clusters and 200k terms in your dictionary this will cause the heap space to be

Re: KMeans job fails during 2nd iteration. Java Heap space

2012-08-08 Thread Paritosh Ranjan
A stacktrace of error would have helped in finding the exact error. However, number of clusters can create Heap Space problems ( If the vector dimension is also high ). Either try to reduce the number of initial clusters ( In my opinion, the best way to know about initial clusters is Canopy

KMeans job fails during 2nd iteration. Java Heap space

2012-08-07 Thread Abramov Pavel
Hello, I am trying to run KMeans example on 15 000 000 documents (seq2sparse output). There are 1 000 clusters, 200 000 terms dictionary and 3-10 terms document size (titles). seq2sparse produces 200 files 80 MB each. My job failed with Java heap space Error. 1st iteration passes while 2nd

Re: ERROR: OutOfMemoryError: Java heap space

2012-08-01 Thread Lance Norskog
-Xmx2048m >> > map heap size for child task >> > >> > >> > This ups the task heap to 2G. >> > >> > >> > On 7/26/12 7:12 PM, Lance Norskog wrote: >> > >> >> Increase the memory size or split the file! >&

Re: ERROR: OutOfMemoryError: Java heap space

2012-08-01 Thread pricila rr
On Thu, Jul 26, 2012 at 5:37 AM, pricila rr > wrote: > >> > >>> I'm trying to transform a file .txt of 1gb for seqfile and the error > >>> occurs: OutOfMemoryError: Java heap space > >>> How to solve? > >>> I am using Hadoop

Re: ERROR: OutOfMemoryError: Java heap space

2012-07-28 Thread Anandha L Ranganathan
#x27;m trying to transform a file .txt of 1gb for seqfile and the error >>> occurs: OutOfMemoryError: Java heap space >>> How to solve? >>> I am using Hadoop and Mahout. >>> >>> $MAHOUT_HOME/bin/mahout seqdirectory --input '/home/usuario/Área de >&g

Re: ERROR: OutOfMemoryError: Java heap space

2012-07-28 Thread Pat Ferrel
icila rr wrote: I'm trying to transform a file .txt of 1gb for seqfile and the error occurs: OutOfMemoryError: Java heap space How to solve? I am using Hadoop and Mahout. $MAHOUT_HOME/bin/mahout seqdirectory --input '/home/usuario/Área de Trabalho/Dados/base1.txt' --output &#

Re: ERROR: OutOfMemoryError: Java heap space

2012-07-26 Thread Lance Norskog
Increase the memory size or split the file! On Thu, Jul 26, 2012 at 5:37 AM, pricila rr wrote: > I'm trying to transform a file .txt of 1gb for seqfile and the error > occurs: OutOfMemoryError: Java heap space > How to solve? > I am using Hadoop and Mahout. > >

Re: Error: Java heap space on mahout cvb command

2012-05-26 Thread Jake Mannix
e to check next week. > > So, I guess the key is the jack up the map task heap space to support a > dense matrix? So per your O(num topics * num terms) below, I guess "k - > #topics" could also have been a culprit, in particular when I had k=200. > The total heapsize will need t

Re: Error: Java heap space on mahout cvb command

2012-05-25 Thread Andy Schlaikjer
On Fri, May 25, 2012 at 4:25 PM, DAN HELM wrote: > Out of curiosity, if one were to cluster 1 million documents, what would > be a reasonable k? I guess it depends to the nature of the data (domain) > and application but it would seem if k is too small then the clusters would > be way too fat an

Re: Error: Java heap space on mahout cvb command

2012-05-25 Thread DAN HELM
Hi Andy,   I ran this at work so don't have the data and log now but somehow I seem to recall log output (after the rowid step) saying there were around 90K terms/columns in the resulting matrix...but I would have to check next week.   So, I guess the key is the jack up the map task heap

Re: Error: Java heap space on mahout cvb command

2012-05-25 Thread Andy Schlaikjer
on. Some of the documents are > larger than the reuters files but most are not particularly large files. > > When attempting to cluster that collection, I get a “Java heap space” > error at start of first iteration of the “mahout cvb” run. I wanted to run > for 4 iterations and gen

Error: Java heap space on mahout cvb command

2012-05-25 Thread DAN HELM
than the reuters files but most are not particularly large files.  When attempting to cluster that collection, I get a “Java heap space” error at start of first iteration of the “mahout cvb” run.  I wanted to run for 4 iterations and generate 200 clusters.   The command I ran was: mahout cvb

Re: Heap Space Issues with Complementary Naive Bayes

2012-05-04 Thread Lance Norskog
th 24GB of RAM >>>>> and 24 cores. When I get to the "Computing TF-IDF" stage, the whole >>>>> process fails with the following exception. I have 14000 documents and >>>>> 2 classes. The lexicon consists of 2705284 trigrams which I created >>

Re: Heap Space Issues with Complementary Naive Bayes

2012-05-04 Thread Ryan Rosario
rams which I created >>>> myself. I then set the ng parameter equal to 1 since I already >>>> tokenized the words myself. >>>> >>>> The system at max has only been using 4-5GB total, and I have set >>>> MAHOUT_OPTIONS=-Xmx4g, MAHOUT_HEAP

Re: Heap Space Issues with Complementary Naive Bayes

2012-05-04 Thread Lance Norskog
since I already >>> tokenized the words myself. >>> >>> The system at max has only been using 4-5GB total, and I have set >>> MAHOUT_OPTIONS=-Xmx4g, MAHOUT_HEAPSIZE=24000, >>> mapred.map.child.java.opts=-Xmx24g just to see if I could get Mahout >>

Re: Heap Space Issues with Complementary Naive Bayes

2012-05-04 Thread Ryan Rosario
okenized the words myself. >> >> The system at max has only been using 4-5GB total, and I have set >> MAHOUT_OPTIONS=-Xmx4g, MAHOUT_HEAPSIZE=24000, >> mapred.map.child.java.opts=-Xmx24g just to see if I could get Mahout >> to acknowledge the increase in heap space

Re: Heap Space Issues with Complementary Naive Bayes

2012-05-04 Thread Lance Norskog
4000, > mapred.map.child.java.opts=-Xmx24g just to see if I could get Mahout > to acknowledge the increase in heap space, but this does not seem to > be helping at all. > > What else can I try to get past this problem? The system has plenty of RAM. > > Thanks, > Ryan > >

Heap Space Issues with Complementary Naive Bayes

2012-05-04 Thread Ryan Rosario
4g just to see if I could get Mahout to acknowledge the increase in heap space, but this does not seem to be helping at all. What else can I try to get past this problem? The system has plenty of RAM. Thanks, Ryan ./bin/mahout trainclassifier -i /user/ryan/pageclass-train -o pageclass-out -type cb

RE: Canopy Job failed processing, Error: Java heap space

2012-03-14 Thread WangRamon
Yes, it's the reason b), I'm using the Cosin measure, so i put the wrong value to t1 and t2, thanks Pranjan. > Date: Thu, 15 Mar 2012 07:48:55 +0530 > From: pran...@xebia.com > To: user@mahout.apache.org > Subject: Re: Canopy Job failed processing, Error: Java heap space &

Re: Canopy Job failed processing, Error: Java heap space

2012-03-14 Thread Jeff Eastman
09:51:40,822 INFO org.apache.hadoop.mapred.Merger: Down to the >> last merge-pass, with 1 segments left of total size: 136745350 bytes >> 2012-03-15 10:03:25,273 FATAL org.apache.hadoop.mapred.TaskTracker: Error >> running child : java.lang.OutOfMemoryError: Java heap space >

Re: Canopy Job failed processing, Error: Java heap space

2012-03-14 Thread Paritosh Ranjan
che.hadoop.mapred.TaskTracker: Error > running child : java.lang.OutOfMemoryError: Java heap space > at > org.apache.mahout.math.map.OpenIntDoubleHashMap.rehash(OpenIntDoubleHashMap.java:434) > at > org.apache.mahout.math.map.OpenIntDoubleHa

RE: Canopy Job failed processing, Error: Java heap space

2012-03-14 Thread WangRamon
segments left of total size: 136745350 bytes 2012-03-15 10:03:25,273 FATAL org.apache.hadoop.mapred.TaskTracker: Error running child : java.lang.OutOfMemoryError: Java heap space at org.apache.mahout.math.map.OpenIntDoubleHashMap.rehash(OpenIntDoubleHashMap.java:434) at

Canopy Job failed processing, Error: Java heap space

2012-03-14 Thread WangRamon
Hi All I'm using the Canopy driver to find the cluster center points, the mapred.child.java.opts parameter for Hadoop is set to 1024M, I'm processing 11000 records, I was supprised to got the Java heap space error during cluster, did i miss something? Thanks. BTW, i did succee

Re: PFPGrowth : Error: Java heap space

2011-05-25 Thread Robin Anil
FPGrowth doesnt use Hadoop. Its in-memory implementation of the algo. You need to increase the java heapspace of your client. (if the data can fit the memory) Robin On Wed, May 25, 2011 at 1:44 PM, Sean Owen wrote: > I don't think setting this on the command line does anything. It affects > t

Re: PFPGrowth : Error: Java heap space

2011-05-25 Thread Sean Owen
I don't think setting this on the command line does anything. It affects the client JVM not the cluster. Set this in find/makers-site.conf On May 25, 2011 2:42 AM, "Vipul Pandey" wrote: > hi! > > I had been running PFPGrowth embedded in my code so far in that my main > class launches PFPGrowth aft

PFPGrowth : Error: Java heap space

2011-05-24 Thread Vipul Pandey
hi! I had been running PFPGrowth embedded in my code so far in that my main class launches PFPGrowth after accepting the config params from command line. I used to set -Dmapred.child.java.opts="-server -Xmx2G -Xms2G" for my jobs and the config used to be passed along to PFPGrowth as well. I just

Re: FW: OutOfMemoryError: Java heap space

2011-04-01 Thread Lance Norskog
ohini > > -Original Message- > From: moving...@gmail.com [mailto:moving...@gmail.com] On Behalf Of Kate > Ericson > Sent: Tuesday, March 29, 2011 8:22 PM > To: user@mahout.apache.org > Subject: Re: FW: OutOfMemoryError: Java heap space > > Yes, that should help.

RE: FW: OutOfMemoryError: Java heap space

2011-04-01 Thread Sengupta, Sohini IN BLR SISL
lto:moving...@gmail.com] On Behalf Of Kate Ericson Sent: Tuesday, March 29, 2011 8:22 PM To: user@mahout.apache.org Subject: Re: FW: OutOfMemoryError: Java heap space Yes, that should help. While you are executing using Hadoop, and that now has more memory, you're still trying to run a Mahout pr

Re: FW: OutOfMemoryError: Java heap space

2011-03-29 Thread Kate Ericson
r@mahout.apache.org > Subject: Re: FW: OutOfMemoryError: Java heap space > > Did you go through the mahout shell script in $MAHOUT_HOME/bin/ ? > For the stack trace posted in the first mail in this chain, heap space need > to be increased in mahout shell script and not in hadoop-env

RE: FW: OutOfMemoryError: Java heap space

2011-03-29 Thread Sengupta, Sohini IN BLR SISL
space Did you go through the mahout shell script in $MAHOUT_HOME/bin/ ? For the stack trace posted in the first mail in this chain, heap space need to be increased in mahout shell script and not in hadoop-env.sh Thanks, MIS. On Tue, Mar 29, 2011 at 1:56 PM, Sengupta, Sohini IN BLR SISL

Re: FW: OutOfMemoryError: Java heap space

2011-03-29 Thread MIS
Did you go through the mahout shell script in $MAHOUT_HOME/bin/ ? For the stack trace posted in the first mail in this chain, heap space need to be increased in mahout shell script and not in hadoop-env.sh Thanks, MIS. On Tue, Mar 29, 2011 at 1:56 PM, Sengupta, Sohini IN BLR SISL < sohini.se

RE: FW: OutOfMemoryError: Java heap space

2011-03-29 Thread Sengupta, Sohini IN BLR SISL
ct: Re: FW: OutOfMemoryError: Java heap space try increasing MAHOUT_HEAPSIZE which by default is set to 1000 MB. check-out the mahout.sh script file for further reference Thanks, MIS. On Tue, Mar 29, 2011 at 9:21 AM, Sengupta, Sohini IN BLR SISL mailto:sohini.sengu...@siemens.com>> wrote: Would

Re: FW: OutOfMemoryError: Java heap space

2011-03-29 Thread MIS
following. > Thanks a lot in advance, > Sohini > > > From: Sengupta, Sohini IN BLR SISL > Sent: Monday, March 28, 2011 3:36 PM > To: user@mahout.apache.org > Cc: Sengupta, Sohini IN BLR SISL > Subject: OutOfMemoryError: Java heap space > >

FW: OutOfMemoryError: Java heap space

2011-03-28 Thread Sengupta, Sohini IN BLR SISL
Would anyone have any pointers regarding following. Thanks a lot in advance, Sohini From: Sengupta, Sohini IN BLR SISL Sent: Monday, March 28, 2011 3:36 PM To: user@mahout.apache.org Cc: Sengupta, Sohini IN BLR SISL Subject: OutOfMemoryError: Java heap space Hi

OutOfMemoryError: Java heap space

2011-03-28 Thread Sengupta, Sohini IN BLR SISL
records=2524062 Exception in thread "main" java.lang.OutOfMemoryError: Java heap space at org.apache.mahout.math.map.OpenIntDoubleHashMap.rehash(OpenIntDoubleHashMap.java:434) at org.apache.mahout.math.map.OpenIntDoubleHashMap.put(OpenIntDoubleHashMap.java:387)

RE: java heap space exception using org.apache.mahout.fpm.pfpgrowth.FPGrowthDriver

2011-02-08 Thread praveen.peddi
something like 512m or higher and try again. Praveen From: ext Yang Sun [soushare@gmail.com] Sent: Monday, February 07, 2011 2:55 PM To: user@mahout.apache.org Subject: java heap space exception using org.apache.mahout.fpm.pfpgrowth.FPGrowthDriver Hi

java heap space exception using org.apache.mahout.fpm.pfpgrowth.FPGrowthDriver

2011-02-07 Thread Yang Sun
Hi, I'm trying to use parallel FPGrowth on a text based data set with about 14K documents. But when I run mahout, I got the following exception: FATAL org.apache.hadoop.mapred.TaskTracker: Error running child : java.lang.OutOfMemoryError: Java heap space

Re: Naive Bayes testclassifier Java heap space exception

2010-12-08 Thread Frank Wang
Hi David, Thanks for your reply. I just check, the docs are 39MB and models are 301MB. I'm running on a single node with pseudo cluster setup. I think giving the hadoop worker 1GB of memory should be more than enough. Am I missing something here? On Wed, Dec 8, 2010 at 3:24 AM, David Hagar wr

Re: Naive Bayes testclassifier Java heap space exception

2010-12-08 Thread David Hagar
Hi Frank -- One major caveat to the below: I've hacked the 0.4 distribution of Mahout quite a bit to get Naive Bayes running smoothly on Amazon's s3 and elastic mapreduce services, thus my experience may not be typical and the memory problems I ran into might well be of my own making. That said I

Naive Bayes testclassifier Java heap space exception

2010-12-08 Thread Frank Wang
Hi, I was trying out Naive Bayes with a setup similar to 20NewsGroup setup. There are 5 categories, each category with 150 articles, and each article is about 50~150kb in size. Training was successful: $MAHOUT_HOME/bin/mahout trainclassifier -i news-input -o news-model -type bayes -ng 3 -s

RE: Java heap space error on PFPGrowth

2010-11-11 Thread praveen.peddi
tatic.void@gmail.com] Sent: Thursday, November 11, 2010 12:10 PM To: user@mahout.apache.org Subject: Re: Java heap space error on PFPGrowth That did it. Thanks. What do you have set for your HADOOP_HEAPSIZE in hadoop-env.sh? On 11/11/10 8:28 AM, praveen.pe...@nokia.com wrote: > Hi Mark, > I g

Re: Java heap space error on PFPGrowth

2010-11-11 Thread Mark
...@hadoop.apache.org; user@mahout.apache.org Subject: Java heap space error on PFPGrowth I am trying to run PFPGrowth but I keep receiving this Java heap space error at the end of the first step/beginning of second step. I am using the following parameters: -method mapreduce -regex [\\t] -s 5 -g 55000

RE: Java heap space error on PFPGrowth

2010-11-11 Thread praveen.peddi
issue. Praveen -Original Message- From: ext Mark [mailto:static.void@gmail.com] Sent: Thursday, November 11, 2010 11:24 AM To: common-u...@hadoop.apache.org; user@mahout.apache.org Subject: Java heap space error on PFPGrowth I am trying to run PFPGrowth but I keep receiving this Java

Java heap space error on PFPGrowth

2010-11-11 Thread Mark
I am trying to run PFPGrowth but I keep receiving this Java heap space error at the end of the first step/beginning of second step. I am using the following parameters: -method mapreduce -regex [\\t] -s 5 -g 55000 Output: .. 10/11/11 08:12:56 INFO mapred.JobClient: map 100% reduce

Re: Error: Java heap space when running FPGrowth

2010-08-05 Thread tanweiguo 00124421
597) at org.apache.hadoop.util.RunJar.main(RunJar.java:156) -邮件原件- 发件人: Ankur C. Goel [mailto:gan...@yahoo-inc.com] 发送时间: 2010年8月4日 18:02 收件人: user@mahout.apache.org 主题: Re: Error: Java heap space when running FPGrowth Hi tanweiguo, Which version of hadoop are you us

Re: Error: Java heap space when running FPGrowth

2010-08-04 Thread Ankur C. Goel
d.JobClient: Task Id : attempt_201007271506_0025_m_00_0, Status : FAILED Error: java.lang.OutOfMemoryError: Java heap space at org.apache.mahout.fpm.pfpgrowth.TransactionTree.resize(TransactionTree.java: 446) at org.apache.mahout.fpm.pfpgrowth.TransactionTree.createNode(Trans

Error: Java heap space when running FPGrowth

2010-08-04 Thread tanweiguo
: map 0% reduce 0% 10/08/04 15:24:31 INFO mapred.JobClient: Task Id : attempt_201007271506_0025_m_00_0, Status : FAILED Error: java.lang.OutOfMemoryError: Java heap space at org.apache.mahout.fpm.pfpgrowth.TransactionTree.resize(TransactionTree.java: 446) at

Error: Java heap space when running FPGrowth

2010-08-04 Thread tanweiguo
: map 0% reduce 0% 10/08/04 15:24:31 INFO mapred.JobClient: Task Id : attempt_201007271506_0025_m_00_0, Status : FAILED Error: java.lang.OutOfMemoryError: Java heap space at org.apache.mahout.fpm.pfpgrowth.TransactionTree.resize(TransactionTree.java: 446) at

Re: mahout seqdirectory: Java heap space error

2010-06-17 Thread rmx
After following a suggestion I splitted the file into several parts and it worked out. -- View this message in context: http://lucene.472066.n3.nabble.com/mahout-seqdirectory-Java-heap-space-error-tp902299p902886.html Sent from the Mahout User List mailing list archive at Nabble.com.

mahout seqdirectory: Java heap space error

2010-06-17 Thread rmx
Hi! I am trying to convert a dataset from text to SeqFormat. The file is inside txtData directory and is 532 MB. I am using a virtual machine with ubuntu with 1500MB of RAM assigned. I got a heap space error when I execute the following command: mahout seqdirectory -i ./txtData/ -o ./seqData