a JavaRDD is just a wrapper around a normal RDD defined in scala, which is
stored in the "rdd" field.  You can access everything that way.  The
JavaRDD wrappers just provide some interfaces that are a bit easier to work
with in Java.

If this is at all convincing, here's me demonstrating it inside the
spark-shell (yes its scala, but I'm using the java api)

scala> val jsc = new JavaSparkContext(sc)
> jsc: org.apache.spark.api.java.JavaSparkContext =
> org.apache.spark.api.java.JavaSparkContext@7d365529
>


scala> val data = jsc.parallelize(java.util.Arrays.asList(Array("a", "b",
> "c")))
> data: org.apache.spark.api.java.JavaRDD[Array[String]] =
> ParallelCollectionRDD[0] at parallelize at <console>:15
>


scala> data.rdd.partitioner
> res0: Option[org.apache.spark.Partitioner] = None


On Tue, Feb 17, 2015 at 3:44 PM, Darin McBeath <ddmcbe...@yahoo.com.invalid>
wrote:

> In an 'early release' of the Learning Spark book, there is the following
> reference:
>
> In Scala and Java, you can determine how an RDD is partitioned using its
> partitioner property (or partitioner() method in Java)
>
> However, I don't see the mentioned 'partitioner()' method in Spark 1.2 or
> a way of getting this information.
>
> I'm curious if anyone has any suggestions for how I might go about finding
> how an RDD is partitioned in a Java program.
>
> Thanks.
>
> Darin.
>
> ---------------------------------------------------------------------
> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
> For additional commands, e-mail: user-h...@spark.apache.org
>
>

Reply via email to