Hi All,
I am running 3 executors in my spark streaming application with 3
cores per executors. I have written my custom receiver for receiving network
data.
In my current configuration I am launching 3 receivers , one receiver per
executor.
In the run if 2 of my executor dies, I am left
Yes that is my understanding of how it should work.
But in my case when I call collect first time, it reads the data from files
on the disk.
Subsequent collect queries are not reading data files ( Verified from the
logs.)
On spark ui I see only shuffle read and no shuffle write.
--
View this
Hi,
I have two RDDs A and B which are created from reading file from HDFS.
I have a third RDD C which is created by taking join of A and B. All three
RDDs (A, B and C ) are not cached.
Now if I perform any action on C (let say collect), action is served without
reading any data from the disk.
Hi All,
Is it possible to map and filter a javardd in a single operation?
Thanks
--
View this message in context:
http://apache-spark-user-list.1001560.n3.nabble.com/Map-with-filter-on-JavaRdd-tp8401.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.
Thanks Mayur for clarification..
--
View this message in context:
http://apache-spark-user-list.1001560.n3.nabble.com/Map-with-filter-on-JavaRdd-tp8401p8410.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.