[09/51] [partial] spark-website git commit: Add 2.1.2 docs
http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6155a89/site/docs/2.1.2/api/java/org/apache/spark/api/java/JavaRDD.html -- diff --git a/site/docs/2.1.2/api/java/org/apache/spark/api/java/JavaRDD.html b/site/docs/2.1.2/api/java/org/apache/spark/api/java/JavaRDD.html new file mode 100644 index 000..a7004a5 --- /dev/null +++ b/site/docs/2.1.2/api/java/org/apache/spark/api/java/JavaRDD.html @@ -0,0 +1,1854 @@ +http://www.w3.org/TR/html4/loose.dtd";> + + + + +JavaRDD (Spark 2.1.2 JavaDoc) + + + + + + + +JavaScript is disabled on your browser. + + + + + + + + +Overview +Package +Class +Tree +Deprecated +Index +Help + + + + +Prev Class +Next Class + + +Frames +No Frames + + +All Classes + + + + + + + +Summary: +Nested | +Field | +Constr | +Method + + +Detail: +Field | +Constr | +Method + + + + + + + + +org.apache.spark.api.java +Class JavaRDD+ + + +Object + + +org.apache.spark.api.java.JavaRDD + + + + + + + +All Implemented Interfaces: +java.io.Serializable, JavaRDDLike > + + + +public class JavaRDD +extends Object +See Also:Serialized Form + + + + + + + + + + + +Constructor Summary + +Constructors + +Constructor and Description + + +JavaRDD(RDD rdd, + scala.reflect.ClassTag classTag) + + + + + + + + + +Method Summary + +Methods + +Modifier and Type +Method and Description + + +static U +aggregate(U zeroValue, + Function2 seqOp, + Function2 combOp) + + +JavaRDD +cache() +Persist this RDD with the default storage level (MEMORY_ONLY). + + + +static JavaPairRDD +cartesian(JavaRDDLike other) + + +static void +checkpoint() + + +scala.reflect.ClassTag +classTag() + + +JavaRDD +coalesce(int numPartitions) +Return a new RDD that is reduced into numPartitions partitions. + + + +JavaRDD +coalesce(int numPartitions, +boolean shuffle) +Return a new RDD that is reduced into numPartitions partitions. + + + +static java.util.List +collect() + + +static JavaFutureAction > +collectAsync() + + +static java.util.List [] +collectPartitions(int[] partitionIds) + + +static SparkContext +context() + + +static long +count() + + +static PartialResult +countApprox(long timeout) + + +static PartialResult +countApprox(long timeout, + double confidence) + + +static long +countApproxDistinct(double relativeSD) + + +static JavaFutureAction +countAsync() + + +static java.util.Map +countByValue() + + +static PartialResult > +countByValueApprox(long timeout) + + +static PartialResult > +countByValueApprox(long timeout, + double confidence) + + +JavaRDD +distinct() +Return a new RDD containing the distinct elements in this RDD. + + + +JavaRDD +distinct(int numPartitions) +Return a new RDD containing the distinct elements in this RDD. + + + +JavaRDD +filter(Function f) +Return a new RDD containing only the elements that satisfy a predicate. + + + +static T +first() + + +static JavaRDD +flatMap(FlatMapFunction f) + + +static JavaDoubleRDD +flatMapToDouble(DoubleFlatMapFunction f) + + +static JavaPairRDD +flatMapToPair(PairFlatMapFunction f) + + +static T +fold(T zeroValue, +Function2 f) + + +static void +foreach(VoidFunction f) + + +static JavaFutureAction +foreachAsync(VoidFunction f) + + +static void +foreachPartition(VoidFunction > f) + + +static JavaFutureAction +foreachPartitionAsync(VoidFunction > f) + + +static JavaRDD +fromRDD(RDD rdd, + scala.reflect.ClassTag evidence$1) + + +static Optional +getCheckpointFile() + + +static int +getNumPartitions() + + +static StorageLevel +getStorageLevel() + + +static JavaRDD > +glom() + + +static JavaPairRDD> +groupBy(Function f) + + +static JavaPairRDD> +groupBy(Function f, + int numPartitions) + + +static int +id() + + +JavaRDD +intersection(JavaRDD other) +Return the intersection of this RDD and another one. + + + +static boolean +isCheckpointed() + + +static boolean +isEmpty() + + +static java.util.Iterator +iterator(Partition split, +TaskContext taskContext) + + +static JavaPairRDD +keyBy(Function f) + + +static JavaRDD +map(Function f) + + +stat
[09/51] [partial] spark-website git commit: Add 2.1.2 docs
http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/java/org/apache/spark/api/java/JavaRDD.html -- diff --git a/site/docs/2.1.2/api/java/org/apache/spark/api/java/JavaRDD.html b/site/docs/2.1.2/api/java/org/apache/spark/api/java/JavaRDD.html new file mode 100644 index 000..a7004a5 --- /dev/null +++ b/site/docs/2.1.2/api/java/org/apache/spark/api/java/JavaRDD.html @@ -0,0 +1,1854 @@ +http://www.w3.org/TR/html4/loose.dtd";> + + + + +JavaRDD (Spark 2.1.2 JavaDoc) + + + + + + + +JavaScript is disabled on your browser. + + + + + + + + +Overview +Package +Class +Tree +Deprecated +Index +Help + + + + +Prev Class +Next Class + + +Frames +No Frames + + +All Classes + + + + + + + +Summary: +Nested | +Field | +Constr | +Method + + +Detail: +Field | +Constr | +Method + + + + + + + + +org.apache.spark.api.java +Class JavaRDD+ + + +Object + + +org.apache.spark.api.java.JavaRDD + + + + + + + +All Implemented Interfaces: +java.io.Serializable, JavaRDDLike > + + + +public class JavaRDD +extends Object +See Also:Serialized Form + + + + + + + + + + + +Constructor Summary + +Constructors + +Constructor and Description + + +JavaRDD(RDD rdd, + scala.reflect.ClassTag classTag) + + + + + + + + + +Method Summary + +Methods + +Modifier and Type +Method and Description + + +static U +aggregate(U zeroValue, + Function2 seqOp, + Function2 combOp) + + +JavaRDD +cache() +Persist this RDD with the default storage level (MEMORY_ONLY). + + + +static JavaPairRDD +cartesian(JavaRDDLike other) + + +static void +checkpoint() + + +scala.reflect.ClassTag +classTag() + + +JavaRDD +coalesce(int numPartitions) +Return a new RDD that is reduced into numPartitions partitions. + + + +JavaRDD +coalesce(int numPartitions, +boolean shuffle) +Return a new RDD that is reduced into numPartitions partitions. + + + +static java.util.List +collect() + + +static JavaFutureAction > +collectAsync() + + +static java.util.List [] +collectPartitions(int[] partitionIds) + + +static SparkContext +context() + + +static long +count() + + +static PartialResult +countApprox(long timeout) + + +static PartialResult +countApprox(long timeout, + double confidence) + + +static long +countApproxDistinct(double relativeSD) + + +static JavaFutureAction +countAsync() + + +static java.util.Map +countByValue() + + +static PartialResult > +countByValueApprox(long timeout) + + +static PartialResult > +countByValueApprox(long timeout, + double confidence) + + +JavaRDD +distinct() +Return a new RDD containing the distinct elements in this RDD. + + + +JavaRDD +distinct(int numPartitions) +Return a new RDD containing the distinct elements in this RDD. + + + +JavaRDD +filter(Function f) +Return a new RDD containing only the elements that satisfy a predicate. + + + +static T +first() + + +static JavaRDD +flatMap(FlatMapFunction f) + + +static JavaDoubleRDD +flatMapToDouble(DoubleFlatMapFunction f) + + +static JavaPairRDD +flatMapToPair(PairFlatMapFunction f) + + +static T +fold(T zeroValue, +Function2 f) + + +static void +foreach(VoidFunction f) + + +static JavaFutureAction +foreachAsync(VoidFunction f) + + +static void +foreachPartition(VoidFunction > f) + + +static JavaFutureAction +foreachPartitionAsync(VoidFunction > f) + + +static JavaRDD +fromRDD(RDD rdd, + scala.reflect.ClassTag evidence$1) + + +static Optional +getCheckpointFile() + + +static int +getNumPartitions() + + +static StorageLevel +getStorageLevel() + + +static JavaRDD > +glom() + + +static JavaPairRDD> +groupBy(Function f) + + +static JavaPairRDD> +groupBy(Function f, + int numPartitions) + + +static int +id() + + +JavaRDD +intersection(JavaRDD other) +Return the intersection of this RDD and another one. + + + +static boolean +isCheckpointed() + + +static boolean +isEmpty() + + +static java.util.Iterator +iterator(Partition split, +TaskContext taskContext) + + +static JavaPairRDD +keyBy(Function f) + + +static JavaRDD +map(Function f) + + +stat
[09/51] [partial] spark-website git commit: Add 2.1.2 docs
http://git-wip-us.apache.org/repos/asf/spark-website/blob/0b563c84/site/docs/2.1.2/api/java/org/apache/spark/api/java/JavaRDD.html -- diff --git a/site/docs/2.1.2/api/java/org/apache/spark/api/java/JavaRDD.html b/site/docs/2.1.2/api/java/org/apache/spark/api/java/JavaRDD.html new file mode 100644 index 000..a7004a5 --- /dev/null +++ b/site/docs/2.1.2/api/java/org/apache/spark/api/java/JavaRDD.html @@ -0,0 +1,1854 @@ +http://www.w3.org/TR/html4/loose.dtd";> + + + + +JavaRDD (Spark 2.1.2 JavaDoc) + + + + + + + +JavaScript is disabled on your browser. + + + + + + + + +Overview +Package +Class +Tree +Deprecated +Index +Help + + + + +Prev Class +Next Class + + +Frames +No Frames + + +All Classes + + + + + + + +Summary: +Nested | +Field | +Constr | +Method + + +Detail: +Field | +Constr | +Method + + + + + + + + +org.apache.spark.api.java +Class JavaRDD+ + + +Object + + +org.apache.spark.api.java.JavaRDD + + + + + + + +All Implemented Interfaces: +java.io.Serializable, JavaRDDLike > + + + +public class JavaRDD +extends Object +See Also:Serialized Form + + + + + + + + + + + +Constructor Summary + +Constructors + +Constructor and Description + + +JavaRDD(RDD rdd, + scala.reflect.ClassTag classTag) + + + + + + + + + +Method Summary + +Methods + +Modifier and Type +Method and Description + + +static U +aggregate(U zeroValue, + Function2 seqOp, + Function2 combOp) + + +JavaRDD +cache() +Persist this RDD with the default storage level (MEMORY_ONLY). + + + +static JavaPairRDD +cartesian(JavaRDDLike other) + + +static void +checkpoint() + + +scala.reflect.ClassTag +classTag() + + +JavaRDD +coalesce(int numPartitions) +Return a new RDD that is reduced into numPartitions partitions. + + + +JavaRDD +coalesce(int numPartitions, +boolean shuffle) +Return a new RDD that is reduced into numPartitions partitions. + + + +static java.util.List +collect() + + +static JavaFutureAction > +collectAsync() + + +static java.util.List [] +collectPartitions(int[] partitionIds) + + +static SparkContext +context() + + +static long +count() + + +static PartialResult +countApprox(long timeout) + + +static PartialResult +countApprox(long timeout, + double confidence) + + +static long +countApproxDistinct(double relativeSD) + + +static JavaFutureAction +countAsync() + + +static java.util.Map +countByValue() + + +static PartialResult > +countByValueApprox(long timeout) + + +static PartialResult > +countByValueApprox(long timeout, + double confidence) + + +JavaRDD +distinct() +Return a new RDD containing the distinct elements in this RDD. + + + +JavaRDD +distinct(int numPartitions) +Return a new RDD containing the distinct elements in this RDD. + + + +JavaRDD +filter(Function f) +Return a new RDD containing only the elements that satisfy a predicate. + + + +static T +first() + + +static JavaRDD +flatMap(FlatMapFunction f) + + +static JavaDoubleRDD +flatMapToDouble(DoubleFlatMapFunction f) + + +static JavaPairRDD +flatMapToPair(PairFlatMapFunction f) + + +static T +fold(T zeroValue, +Function2 f) + + +static void +foreach(VoidFunction f) + + +static JavaFutureAction +foreachAsync(VoidFunction f) + + +static void +foreachPartition(VoidFunction > f) + + +static JavaFutureAction +foreachPartitionAsync(VoidFunction > f) + + +static JavaRDD +fromRDD(RDD rdd, + scala.reflect.ClassTag evidence$1) + + +static Optional +getCheckpointFile() + + +static int +getNumPartitions() + + +static StorageLevel +getStorageLevel() + + +static JavaRDD > +glom() + + +static JavaPairRDD> +groupBy(Function f) + + +static JavaPairRDD> +groupBy(Function f, + int numPartitions) + + +static int +id() + + +JavaRDD +intersection(JavaRDD other) +Return the intersection of this RDD and another one. + + + +static boolean +isCheckpointed() + + +static boolean +isEmpty() + + +static java.util.Iterator +iterator(Partition split, +TaskContext taskContext) + + +static JavaPairRDD +keyBy(Function f) + + +static JavaRDD +map(Function f) + + +stat