[09/51] [partial] spark-website git commit: Add 2.1.2 docs

2017-10-17 Thread holden
http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6155a89/site/docs/2.1.2/api/java/org/apache/spark/api/java/JavaRDD.html
--
diff --git a/site/docs/2.1.2/api/java/org/apache/spark/api/java/JavaRDD.html 
b/site/docs/2.1.2/api/java/org/apache/spark/api/java/JavaRDD.html
new file mode 100644
index 000..a7004a5
--- /dev/null
+++ b/site/docs/2.1.2/api/java/org/apache/spark/api/java/JavaRDD.html
@@ -0,0 +1,1854 @@
+http://www.w3.org/TR/html4/loose.dtd";>
+
+
+
+
+JavaRDD (Spark 2.1.2 JavaDoc)
+
+
+
+
+
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+
+
+
+Overview
+Package
+Class
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev 
Class
+Next 
Class
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+Summary: 
+Nested | 
+Field | 
+Constr | 
+Method
+
+
+Detail: 
+Field | 
+Constr | 
+Method
+
+
+
+
+
+
+
+
+org.apache.spark.api.java
+Class JavaRDD
+
+
+
+Object
+
+
+org.apache.spark.api.java.JavaRDD
+
+
+
+
+
+
+
+All Implemented Interfaces:
+java.io.Serializable, JavaRDDLike>
+
+
+
+public class JavaRDD
+extends Object
+See Also:Serialized
 Form
+
+
+
+
+
+
+
+
+
+
+
+Constructor Summary
+
+Constructors 
+
+Constructor and Description
+
+
+JavaRDD(RDD rdd,
+   scala.reflect.ClassTag classTag) 
+
+
+
+
+
+
+
+
+
+Method Summary
+
+Methods 
+
+Modifier and Type
+Method and Description
+
+
+static  U
+aggregate(U zeroValue,
+ Function2 seqOp,
+ Function2 combOp) 
+
+
+JavaRDD
+cache()
+Persist this RDD with the default storage level 
(MEMORY_ONLY).
+
+
+
+static  JavaPairRDD
+cartesian(JavaRDDLike other) 
+
+
+static void
+checkpoint() 
+
+
+scala.reflect.ClassTag
+classTag() 
+
+
+JavaRDD
+coalesce(int numPartitions)
+Return a new RDD that is reduced into 
numPartitions partitions.
+
+
+
+JavaRDD
+coalesce(int numPartitions,
+boolean shuffle)
+Return a new RDD that is reduced into 
numPartitions partitions.
+
+
+
+static java.util.List
+collect() 
+
+
+static JavaFutureAction>
+collectAsync() 
+
+
+static java.util.List[]
+collectPartitions(int[] partitionIds) 
+
+
+static SparkContext
+context() 
+
+
+static long
+count() 
+
+
+static PartialResult
+countApprox(long timeout) 
+
+
+static PartialResult
+countApprox(long timeout,
+   double confidence) 
+
+
+static long
+countApproxDistinct(double relativeSD) 
+
+
+static JavaFutureAction
+countAsync() 
+
+
+static java.util.Map
+countByValue() 
+
+
+static PartialResult>
+countByValueApprox(long timeout) 
+
+
+static PartialResult>
+countByValueApprox(long timeout,
+  double confidence) 
+
+
+JavaRDD
+distinct()
+Return a new RDD containing the distinct elements in this 
RDD.
+
+
+
+JavaRDD
+distinct(int numPartitions)
+Return a new RDD containing the distinct elements in this 
RDD.
+
+
+
+JavaRDD
+filter(Function f)
+Return a new RDD containing only the elements that satisfy 
a predicate.
+
+
+
+static T
+first() 
+
+
+static  JavaRDD
+flatMap(FlatMapFunction f) 
+
+
+static JavaDoubleRDD
+flatMapToDouble(DoubleFlatMapFunction f) 
+
+
+static  JavaPairRDD
+flatMapToPair(PairFlatMapFunction f) 
+
+
+static T
+fold(T zeroValue,
+Function2 f) 
+
+
+static void
+foreach(VoidFunction f) 
+
+
+static JavaFutureAction
+foreachAsync(VoidFunction f) 
+
+
+static void
+foreachPartition(VoidFunction> f) 
+
+
+static JavaFutureAction
+foreachPartitionAsync(VoidFunction> f) 
+
+
+static  JavaRDD
+fromRDD(RDD rdd,
+   scala.reflect.ClassTag evidence$1) 
+
+
+static Optional
+getCheckpointFile() 
+
+
+static int
+getNumPartitions() 
+
+
+static StorageLevel
+getStorageLevel() 
+
+
+static JavaRDD>
+glom() 
+
+
+static  JavaPairRDD>
+groupBy(Function f) 
+
+
+static  JavaPairRDD>
+groupBy(Function f,
+   int numPartitions) 
+
+
+static int
+id() 
+
+
+JavaRDD
+intersection(JavaRDD other)
+Return the intersection of this RDD and another one.
+
+
+
+static boolean
+isCheckpointed() 
+
+
+static boolean
+isEmpty() 
+
+
+static java.util.Iterator
+iterator(Partition split,
+TaskContext taskContext) 
+
+
+static  JavaPairRDD
+keyBy(Function f) 
+
+
+static  JavaRDD
+map(Function f) 
+
+
+stat

[09/51] [partial] spark-website git commit: Add 2.1.2 docs

http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/java/org/apache/spark/api/java/JavaRDD.html
--
diff --git a/site/docs/2.1.2/api/java/org/apache/spark/api/java/JavaRDD.html 
b/site/docs/2.1.2/api/java/org/apache/spark/api/java/JavaRDD.html
new file mode 100644
index 000..a7004a5
--- /dev/null
+++ b/site/docs/2.1.2/api/java/org/apache/spark/api/java/JavaRDD.html
@@ -0,0 +1,1854 @@
+http://www.w3.org/TR/html4/loose.dtd";>
+
+
+
+
+JavaRDD (Spark 2.1.2 JavaDoc)
+
+
+
+
+
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+
+
+
+Overview
+Package
+Class
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev 
Class
+Next 
Class
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+Summary: 
+Nested | 
+Field | 
+Constr | 
+Method
+
+
+Detail: 
+Field | 
+Constr | 
+Method
+
+
+
+
+
+
+
+
+org.apache.spark.api.java
+Class JavaRDD
+
+
+
+Object
+
+
+org.apache.spark.api.java.JavaRDD
+
+
+
+
+
+
+
+All Implemented Interfaces:
+java.io.Serializable, JavaRDDLike>
+
+
+
+public class JavaRDD
+extends Object
+See Also:Serialized
 Form
+
+
+
+
+
+
+
+
+
+
+
+Constructor Summary
+
+Constructors 
+
+Constructor and Description
+
+
+JavaRDD(RDD rdd,
+   scala.reflect.ClassTag classTag) 
+
+
+
+
+
+
+
+
+
+Method Summary
+
+Methods 
+
+Modifier and Type
+Method and Description
+
+
+static  U
+aggregate(U zeroValue,
+ Function2 seqOp,
+ Function2 combOp) 
+
+
+JavaRDD
+cache()
+Persist this RDD with the default storage level 
(MEMORY_ONLY).
+
+
+
+static  JavaPairRDD
+cartesian(JavaRDDLike other) 
+
+
+static void
+checkpoint() 
+
+
+scala.reflect.ClassTag
+classTag() 
+
+
+JavaRDD
+coalesce(int numPartitions)
+Return a new RDD that is reduced into 
numPartitions partitions.
+
+
+
+JavaRDD
+coalesce(int numPartitions,
+boolean shuffle)
+Return a new RDD that is reduced into 
numPartitions partitions.
+
+
+
+static java.util.List
+collect() 
+
+
+static JavaFutureAction>
+collectAsync() 
+
+
+static java.util.List[]
+collectPartitions(int[] partitionIds) 
+
+
+static SparkContext
+context() 
+
+
+static long
+count() 
+
+
+static PartialResult
+countApprox(long timeout) 
+
+
+static PartialResult
+countApprox(long timeout,
+   double confidence) 
+
+
+static long
+countApproxDistinct(double relativeSD) 
+
+
+static JavaFutureAction
+countAsync() 
+
+
+static java.util.Map
+countByValue() 
+
+
+static PartialResult>
+countByValueApprox(long timeout) 
+
+
+static PartialResult>
+countByValueApprox(long timeout,
+  double confidence) 
+
+
+JavaRDD
+distinct()
+Return a new RDD containing the distinct elements in this 
RDD.
+
+
+
+JavaRDD
+distinct(int numPartitions)
+Return a new RDD containing the distinct elements in this 
RDD.
+
+
+
+JavaRDD
+filter(Function f)
+Return a new RDD containing only the elements that satisfy 
a predicate.
+
+
+
+static T
+first() 
+
+
+static  JavaRDD
+flatMap(FlatMapFunction f) 
+
+
+static JavaDoubleRDD
+flatMapToDouble(DoubleFlatMapFunction f) 
+
+
+static  JavaPairRDD
+flatMapToPair(PairFlatMapFunction f) 
+
+
+static T
+fold(T zeroValue,
+Function2 f) 
+
+
+static void
+foreach(VoidFunction f) 
+
+
+static JavaFutureAction
+foreachAsync(VoidFunction f) 
+
+
+static void
+foreachPartition(VoidFunction> f) 
+
+
+static JavaFutureAction
+foreachPartitionAsync(VoidFunction> f) 
+
+
+static  JavaRDD
+fromRDD(RDD rdd,
+   scala.reflect.ClassTag evidence$1) 
+
+
+static Optional
+getCheckpointFile() 
+
+
+static int
+getNumPartitions() 
+
+
+static StorageLevel
+getStorageLevel() 
+
+
+static JavaRDD>
+glom() 
+
+
+static  JavaPairRDD>
+groupBy(Function f) 
+
+
+static  JavaPairRDD>
+groupBy(Function f,
+   int numPartitions) 
+
+
+static int
+id() 
+
+
+JavaRDD
+intersection(JavaRDD other)
+Return the intersection of this RDD and another one.
+
+
+
+static boolean
+isCheckpointed() 
+
+
+static boolean
+isEmpty() 
+
+
+static java.util.Iterator
+iterator(Partition split,
+TaskContext taskContext) 
+
+
+static  JavaPairRDD
+keyBy(Function f) 
+
+
+static  JavaRDD
+map(Function f) 
+
+
+stat

[09/51] [partial] spark-website git commit: Add 2.1.2 docs

http://git-wip-us.apache.org/repos/asf/spark-website/blob/0b563c84/site/docs/2.1.2/api/java/org/apache/spark/api/java/JavaRDD.html
--
diff --git a/site/docs/2.1.2/api/java/org/apache/spark/api/java/JavaRDD.html 
b/site/docs/2.1.2/api/java/org/apache/spark/api/java/JavaRDD.html
new file mode 100644
index 000..a7004a5
--- /dev/null
+++ b/site/docs/2.1.2/api/java/org/apache/spark/api/java/JavaRDD.html
@@ -0,0 +1,1854 @@
+http://www.w3.org/TR/html4/loose.dtd";>
+
+
+
+
+JavaRDD (Spark 2.1.2 JavaDoc)
+
+
+
+
+
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+
+
+
+Overview
+Package
+Class
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev 
Class
+Next 
Class
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+Summary: 
+Nested | 
+Field | 
+Constr | 
+Method
+
+
+Detail: 
+Field | 
+Constr | 
+Method
+
+
+
+
+
+
+
+
+org.apache.spark.api.java
+Class JavaRDD
+
+
+
+Object
+
+
+org.apache.spark.api.java.JavaRDD
+
+
+
+
+
+
+
+All Implemented Interfaces:
+java.io.Serializable, JavaRDDLike>
+
+
+
+public class JavaRDD
+extends Object
+See Also:Serialized
 Form
+
+
+
+
+
+
+
+
+
+
+
+Constructor Summary
+
+Constructors 
+
+Constructor and Description
+
+
+JavaRDD(RDD rdd,
+   scala.reflect.ClassTag classTag) 
+
+
+
+
+
+
+
+
+
+Method Summary
+
+Methods 
+
+Modifier and Type
+Method and Description
+
+
+static  U
+aggregate(U zeroValue,
+ Function2 seqOp,
+ Function2 combOp) 
+
+
+JavaRDD
+cache()
+Persist this RDD with the default storage level 
(MEMORY_ONLY).
+
+
+
+static  JavaPairRDD
+cartesian(JavaRDDLike other) 
+
+
+static void
+checkpoint() 
+
+
+scala.reflect.ClassTag
+classTag() 
+
+
+JavaRDD
+coalesce(int numPartitions)
+Return a new RDD that is reduced into 
numPartitions partitions.
+
+
+
+JavaRDD
+coalesce(int numPartitions,
+boolean shuffle)
+Return a new RDD that is reduced into 
numPartitions partitions.
+
+
+
+static java.util.List
+collect() 
+
+
+static JavaFutureAction>
+collectAsync() 
+
+
+static java.util.List[]
+collectPartitions(int[] partitionIds) 
+
+
+static SparkContext
+context() 
+
+
+static long
+count() 
+
+
+static PartialResult
+countApprox(long timeout) 
+
+
+static PartialResult
+countApprox(long timeout,
+   double confidence) 
+
+
+static long
+countApproxDistinct(double relativeSD) 
+
+
+static JavaFutureAction
+countAsync() 
+
+
+static java.util.Map
+countByValue() 
+
+
+static PartialResult>
+countByValueApprox(long timeout) 
+
+
+static PartialResult>
+countByValueApprox(long timeout,
+  double confidence) 
+
+
+JavaRDD
+distinct()
+Return a new RDD containing the distinct elements in this 
RDD.
+
+
+
+JavaRDD
+distinct(int numPartitions)
+Return a new RDD containing the distinct elements in this 
RDD.
+
+
+
+JavaRDD
+filter(Function f)
+Return a new RDD containing only the elements that satisfy 
a predicate.
+
+
+
+static T
+first() 
+
+
+static  JavaRDD
+flatMap(FlatMapFunction f) 
+
+
+static JavaDoubleRDD
+flatMapToDouble(DoubleFlatMapFunction f) 
+
+
+static  JavaPairRDD
+flatMapToPair(PairFlatMapFunction f) 
+
+
+static T
+fold(T zeroValue,
+Function2 f) 
+
+
+static void
+foreach(VoidFunction f) 
+
+
+static JavaFutureAction
+foreachAsync(VoidFunction f) 
+
+
+static void
+foreachPartition(VoidFunction> f) 
+
+
+static JavaFutureAction
+foreachPartitionAsync(VoidFunction> f) 
+
+
+static  JavaRDD
+fromRDD(RDD rdd,
+   scala.reflect.ClassTag evidence$1) 
+
+
+static Optional
+getCheckpointFile() 
+
+
+static int
+getNumPartitions() 
+
+
+static StorageLevel
+getStorageLevel() 
+
+
+static JavaRDD>
+glom() 
+
+
+static  JavaPairRDD>
+groupBy(Function f) 
+
+
+static  JavaPairRDD>
+groupBy(Function f,
+   int numPartitions) 
+
+
+static int
+id() 
+
+
+JavaRDD
+intersection(JavaRDD other)
+Return the intersection of this RDD and another one.
+
+
+
+static boolean
+isCheckpointed() 
+
+
+static boolean
+isEmpty() 
+
+
+static java.util.Iterator
+iterator(Partition split,
+TaskContext taskContext) 
+
+
+static  JavaPairRDD
+keyBy(Function f) 
+
+
+static  JavaRDD
+map(Function f) 
+
+
+stat