the OOM, something like:
log2= self.sqlContext.jsonFile(path)
log2.count()
...
out.count()
...
On Thu, Mar 26, 2015 at 10:34 AM, Eduardo Cusa
eduardo.c...@usmediaconsulting.com wrote:
the last try was without log2.cache() and still getting out of memory
I using the following conf, maybe
, Mar 26, 2015 at 10:02 AM, Eduardo Cusa
eduardo.c...@usmediaconsulting.com wrote:
I running on ec2 :
1 Master : 4 CPU 15 GB RAM (2 GB swap)
2 Slaves 4 CPU 15 GB RAM
the uncompressed dataset size is 15 GB
On Thu, Mar 26, 2015 at 10:41 AM, Eduardo Cusa
eduardo.c
I running on ec2 :
1 Master : 4 CPU 15 GB RAM (2 GB swap)
2 Slaves 4 CPU 15 GB RAM
the uncompressed dataset size is 15 GB
On Thu, Mar 26, 2015 at 10:41 AM, Eduardo Cusa
eduardo.c...@usmediaconsulting.com wrote:
Hi Davies, I upgrade to 1.3.0 and still getting Out of Memory.
I ran
a taste for the new DataFrame API.
On Wed, Mar 25, 2015 at 11:49 AM, Eduardo Cusa
eduardo.c...@usmediaconsulting.com wrote:
Hi Davies, I running 1.1.0.
Now I'm following this thread that recommend use batchsize parameter = 1
http://apache-spark-user-list.1001560.n3.nabble.com/pySpark
Hi Guys, I running the following function with spark-submmit and de SO is
killing my process :
def getRdd(self,date,provider):
path='s3n://'+AWS_BUCKET+'/'+date+'/*.log.gz'
log2= self.sqlContext.jsonFile(path)
log2.registerTempTable('log_test')
log2.cache()
Liu dav...@databricks.com wrote:
What's the version of Spark you are running?
There is a bug in SQL Python API [1], it's fixed in 1.2.1 and 1.3,
[1] https://issues.apache.org/jira/browse/SPARK-6055
On Wed, Mar 25, 2015 at 10:33 AM, Eduardo Cusa
eduardo.c...@usmediaconsulting.com wrote:
Hi
the build
file https://github.com/knoldus/Play-Spark-Scala/blob/master/build.sbt
of your play application it seems that it uses Spark 1.0.1.
Thanks
Best Regards
On Fri, Jan 9, 2015 at 7:17 PM, Eduardo Cusa
eduardo.c...@usmediaconsulting.com wrote:
Hi guys, I running the following example
Hi guys, I running the following example :
https://github.com/knoldus/Play-Spark-Scala in the same machine as the
spark master, and the spark cluster was lauched with ec2 script.
I'm stuck with this errors, any idea how to fix it?
Regards
Eduardo
call the play app prints the following
and restart the process.
Does this always happen, or was it just once?
Nick
On Thu, Dec 18, 2014 at 9:42 AM, Eduardo Cusa
eduardo.c...@usmediaconsulting.com wrote:
Hi guys.
I run the folling command to lauch a new cluster :
./spark-ec2 -k test -i test.pem -s 1 --vpc-id vpc-X
Hi guys.
I run the folling command to lauch a new cluster :
./spark-ec2 -k test -i test.pem -s 1 --vpc-id vpc-X --subnet-id
subnet-X launch vpc_spark
The instances started ok but the command never end. With the following
output:
Setting up security groups...
Searching for existing
Hi guys.
I run the folling command to lauch a new cluster :
./spark-ec2 -k test -i test.pem -s 1 --vpc-id vpc-X --subnet-id
subnet-X launch vpc_spark
The instances started ok but the command never end. With the following
output:
Setting up security groups...
Searching for existing
HI guys, I starting to working with spark from java and when i run the
folliwing code :
SparkConf conf = new SparkConf().setMaster(spark://10.0.2.20:7077
).setAppName(SparkTest);
JavaSparkContext sc = new JavaSparkContext(conf);
I recived the following error and the java process exit ends:
12 matches
Mail list logo