[20/51] [partial] spark-website git commit: Add 2.1.2 docs

2017-10-17 Thread holden
http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6155a89/site/docs/2.1.2/api/java/org/apache/spark/SparkConf.html
--
diff --git a/site/docs/2.1.2/api/java/org/apache/spark/SparkConf.html 
b/site/docs/2.1.2/api/java/org/apache/spark/SparkConf.html
new file mode 100644
index 000..4c33803
--- /dev/null
+++ b/site/docs/2.1.2/api/java/org/apache/spark/SparkConf.html
@@ -0,0 +1,1147 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+
+
+SparkConf (Spark 2.1.2 JavaDoc)
+
+
+
+
+
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+
+
+
+Overview
+Package
+Class
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev Class
+Next Class
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+
+
+Detail:
+Field|
+Constr|
+Method
+
+
+
+
+
+
+
+
+org.apache.spark
+Class SparkConf
+
+
+
+Object
+
+
+org.apache.spark.SparkConf
+
+
+
+
+
+
+
+All Implemented Interfaces:
+java.io.Serializable, Cloneable
+
+
+
+public class SparkConf
+extends Object
+implements scala.Cloneable, scala.Serializable
+Configuration for a Spark application. Used to set various 
Spark parameters as key-value pairs.
+ 
+ Most of the time, you would create a SparkConf object with new 
SparkConf(), which will load
+ values from any spark.* Java system properties set in your 
application as well. In this case,
+ parameters you set directly on the SparkConf object take 
priority over system properties.
+ 
+ For unit tests, you can also call new SparkConf(false) to skip 
loading external settings and
+ get the same configuration no matter what the system properties are.
+ 
+ All setter methods in this class support chaining. For example, you can write
+ new SparkConf().setMaster("local").setAppName("My app").
+ 
+ param:  loadDefaults whether to also load values from Java system properties
+ 
+See Also:Serialized 
FormNote:
+  Once a SparkConf object is passed to Spark, it is cloned and can no 
longer be modified
+ by the user. Spark does not support modifying the configuration at 
runtime.
+
+
+
+
+
+
+
+
+
+
+
+Constructor Summary
+
+Constructors
+
+Constructor and Description
+
+
+SparkConf()
+Create a SparkConf that loads defaults from system 
properties and the classpath
+
+
+
+SparkConf(booleanloadDefaults)
+
+
+
+
+
+
+
+
+
+Method Summary
+
+Methods
+
+Modifier and Type
+Method and Description
+
+
+SparkConf
+clone()
+Copy this object
+
+
+
+boolean
+contains(Stringkey)
+Does the configuration contain a given parameter?
+
+
+
+String
+get(Stringkey)
+Get a parameter; throws a NoSuchElementException if it's 
not set
+
+
+
+String
+get(Stringkey,
+   StringdefaultValue)
+Get a parameter, falling back to a default if not set
+
+
+
+scala.Tuple2String,String[]
+getAll()
+Get all parameters as a list of pairs
+
+
+
+scala.Tuple2String,String[]
+getAllWithPrefix(Stringprefix)
+Get all parameters that start with prefix
+
+
+
+String
+getAppId()
+Returns the Spark application id, valid in the Driver after 
TaskScheduler registration and
+ from the start in the Executor.
+
+
+
+scala.collection.immutable.MapObject,String
+getAvroSchema()
+Gets all the avro schemas in the configuration used in the 
generic Avro record serializer
+
+
+
+boolean
+getBoolean(Stringkey,
+  booleandefaultValue)
+Get a parameter as a boolean, falling back to a default if 
not set
+
+
+
+static scala.OptionString
+getDeprecatedConfig(Stringkey,
+   SparkConfconf)
+Looks for available deprecated keys for the given config 
option, and return the first
+ value available.
+
+
+
+double
+getDouble(Stringkey,
+ doubledefaultValue)
+Get a parameter as a double, falling back to a default if 
not set
+
+
+
+scala.collection.Seqscala.Tuple2String,String
+getExecutorEnv()
+Get all executor environment variables set on this 
SparkConf
+
+
+
+int
+getInt(Stringkey,
+  intdefaultValue)
+Get a parameter as an integer, falling back to a default if 
not set
+
+
+
+long
+getLong(Stringkey,
+   longdefaultValue)
+Get a parameter as a long, falling back to a default if not 
set
+
+
+
+scala.OptionString
+getOption(Stringkey)
+Get a parameter as an Option
+
+
+
+long
+getSizeAsBytes(Stringkey)
+Get a size parameter as bytes; throws a 
NoSuchElementException if it's not set.
+
+
+
+long
+getSizeAsBytes(Stringkey,
+  longdefaultValue)
+Get a size parameter as bytes, falling back to a default if 
not set.
+
+
+
+long
+getSizeAsBytes(Stringkey,
+  StringdefaultValue)
+Get a size parameter as bytes, falling back to a 

[09/51] [partial] spark-website git commit: Add 2.1.2 docs

2017-10-17 Thread holden
http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6155a89/site/docs/2.1.2/api/java/org/apache/spark/api/java/JavaRDD.html
--
diff --git a/site/docs/2.1.2/api/java/org/apache/spark/api/java/JavaRDD.html 
b/site/docs/2.1.2/api/java/org/apache/spark/api/java/JavaRDD.html
new file mode 100644
index 000..a7004a5
--- /dev/null
+++ b/site/docs/2.1.2/api/java/org/apache/spark/api/java/JavaRDD.html
@@ -0,0 +1,1854 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+
+
+JavaRDD (Spark 2.1.2 JavaDoc)
+
+
+
+
+
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+
+
+
+Overview
+Package
+Class
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev 
Class
+Next 
Class
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+
+
+Detail:
+Field|
+Constr|
+Method
+
+
+
+
+
+
+
+
+org.apache.spark.api.java
+Class JavaRDDT
+
+
+
+Object
+
+
+org.apache.spark.api.java.JavaRDDT
+
+
+
+
+
+
+
+All Implemented Interfaces:
+java.io.Serializable, JavaRDDLikeT,JavaRDDT
+
+
+
+public class JavaRDDT
+extends Object
+See Also:Serialized
 Form
+
+
+
+
+
+
+
+
+
+
+
+Constructor Summary
+
+Constructors
+
+Constructor and Description
+
+
+JavaRDD(RDDTrdd,
+   scala.reflect.ClassTagTclassTag)
+
+
+
+
+
+
+
+
+
+Method Summary
+
+Methods
+
+Modifier and Type
+Method and Description
+
+
+static UU
+aggregate(UzeroValue,
+ Function2U,T,UseqOp,
+ Function2U,U,UcombOp)
+
+
+JavaRDDT
+cache()
+Persist this RDD with the default storage level 
(MEMORY_ONLY).
+
+
+
+static UJavaPairRDDT,U
+cartesian(JavaRDDLikeU,?other)
+
+
+static void
+checkpoint()
+
+
+scala.reflect.ClassTagT
+classTag()
+
+
+JavaRDDT
+coalesce(intnumPartitions)
+Return a new RDD that is reduced into 
numPartitions partitions.
+
+
+
+JavaRDDT
+coalesce(intnumPartitions,
+booleanshuffle)
+Return a new RDD that is reduced into 
numPartitions partitions.
+
+
+
+static java.util.ListT
+collect()
+
+
+static JavaFutureActionjava.util.ListT
+collectAsync()
+
+
+static java.util.ListT[]
+collectPartitions(int[]partitionIds)
+
+
+static SparkContext
+context()
+
+
+static long
+count()
+
+
+static PartialResultBoundedDouble
+countApprox(longtimeout)
+
+
+static PartialResultBoundedDouble
+countApprox(longtimeout,
+   doubleconfidence)
+
+
+static long
+countApproxDistinct(doublerelativeSD)
+
+
+static JavaFutureActionLong
+countAsync()
+
+
+static java.util.MapT,Long
+countByValue()
+
+
+static PartialResultjava.util.MapT,BoundedDouble
+countByValueApprox(longtimeout)
+
+
+static PartialResultjava.util.MapT,BoundedDouble
+countByValueApprox(longtimeout,
+  doubleconfidence)
+
+
+JavaRDDT
+distinct()
+Return a new RDD containing the distinct elements in this 
RDD.
+
+
+
+JavaRDDT
+distinct(intnumPartitions)
+Return a new RDD containing the distinct elements in this 
RDD.
+
+
+
+JavaRDDT
+filter(FunctionT,Booleanf)
+Return a new RDD containing only the elements that satisfy 
a predicate.
+
+
+
+static T
+first()
+
+
+static UJavaRDDU
+flatMap(FlatMapFunctionT,Uf)
+
+
+static JavaDoubleRDD
+flatMapToDouble(DoubleFlatMapFunctionTf)
+
+
+static K2,V2JavaPairRDDK2,V2
+flatMapToPair(PairFlatMapFunctionT,K2,V2f)
+
+
+static T
+fold(TzeroValue,
+Function2T,T,Tf)
+
+
+static void
+foreach(VoidFunctionTf)
+
+
+static JavaFutureActionVoid
+foreachAsync(VoidFunctionTf)
+
+
+static void
+foreachPartition(VoidFunctionjava.util.IteratorTf)
+
+
+static JavaFutureActionVoid
+foreachPartitionAsync(VoidFunctionjava.util.IteratorTf)
+
+
+static TJavaRDDT
+fromRDD(RDDTrdd,
+   scala.reflect.ClassTagTevidence$1)
+
+
+static OptionalString
+getCheckpointFile()
+
+
+static int
+getNumPartitions()
+
+
+static StorageLevel
+getStorageLevel()
+
+
+static JavaRDDjava.util.ListT
+glom()
+
+
+static UJavaPairRDDU,IterableT
+groupBy(FunctionT,Uf)
+
+
+static UJavaPairRDDU,IterableT
+groupBy(FunctionT,Uf,
+   intnumPartitions)
+
+
+static int
+id()
+
+
+JavaRDDT
+intersection(JavaRDDTother)
+Return the intersection of this RDD and another one.
+
+
+
+static boolean
+isCheckpointed()
+
+
+static boolean
+isEmpty()
+
+
+static java.util.IteratorT
+iterator(Partitionsplit,
+TaskContexttaskContext)
+
+
+static UJavaPairRDDU,T
+keyBy(FunctionT,Uf)
+
+
+static RJavaRDDR
+map(FunctionT,Rf)
+
+
+static UJavaRDDU
+mapPartitions(FlatMapFunctionjava.util.IteratorT,Uf)
+
+
+static UJavaRDDU
+mapPartitions(FlatMapFunctionjava.util.IteratorT,Uf,
+ booleanpreservesPartitioning)
+
+
+static JavaDoubleRDD

[45/51] [partial] spark-website git commit: Add 2.1.2 docs

2017-10-17 Thread holden
http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6155a89/site/docs/2.1.2/api/R/date_format.html
--
diff --git a/site/docs/2.1.2/api/R/date_format.html 
b/site/docs/2.1.2/api/R/date_format.html
new file mode 100644
index 000..7f9ad3c
--- /dev/null
+++ b/site/docs/2.1.2/api/R/date_format.html
@@ -0,0 +1,87 @@
+http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd;>http://www.w3.org/1999/xhtml;>R: date_format
+
+
+
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css;>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js";>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js";>
+hljs.initHighlightingOnLoad();
+
+
+date_format {SparkR}R 
Documentation
+
+date_format
+
+Description
+
+Converts a date/timestamp/string to a value of string in the format 
specified by the date
+format given by the second argument.
+
+
+
+Usage
+
+
+date_format(y, x)
+
+## S4 method for signature 'Column,character'
+date_format(y, x)
+
+
+
+Arguments
+
+
+y
+
+Column to compute on.
+
+x
+
+date format specification.
+
+
+
+
+Details
+
+A pattern could be for instance 
+dd.MM. and could return a string like '18.03.1993'. All
+pattern letters of java.text.SimpleDateFormat can be used.
+
+Note: Use when ever possible specialized functions like year. 
These benefit from a
+specialized implementation.
+
+
+
+Note
+
+date_format since 1.5.0
+
+
+
+See Also
+
+Other datetime_funcs: add_months,
+date_add, date_sub,
+datediff, dayofmonth,
+dayofyear, from_unixtime,
+from_utc_timestamp, 
hour,
+last_day, minute,
+months_between, month,
+next_day, quarter,
+second, to_date,
+to_utc_timestamp,
+unix_timestamp, weekofyear,
+window, year
+
+
+
+Examples
+
+## Not run: date_format(df$t, MM/dd/yyy)
+
+
+
+[Package SparkR version 2.1.2 
Index]
+

http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6155a89/site/docs/2.1.2/api/R/date_sub.html
--
diff --git a/site/docs/2.1.2/api/R/date_sub.html 
b/site/docs/2.1.2/api/R/date_sub.html
new file mode 100644
index 000..89d6661
--- /dev/null
+++ b/site/docs/2.1.2/api/R/date_sub.html
@@ -0,0 +1,75 @@
+http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd;>http://www.w3.org/1999/xhtml;>R: date_sub
+
+
+
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css;>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js";>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js";>
+hljs.initHighlightingOnLoad();
+
+
+date_sub 
{SparkR}R Documentation
+
+date_sub
+
+Description
+
+Returns the date that is x days before
+
+
+
+Usage
+
+
+date_sub(y, x)
+
+## S4 method for signature 'Column,numeric'
+date_sub(y, x)
+
+
+
+Arguments
+
+
+y
+
+Column to compute on
+
+x
+
+Number of days to substract
+
+
+
+
+Note
+
+date_sub since 1.5.0
+
+
+
+See Also
+
+Other datetime_funcs: add_months,
+date_add, date_format,
+datediff, dayofmonth,
+dayofyear, from_unixtime,
+from_utc_timestamp, 
hour,
+last_day, minute,
+months_between, month,
+next_day, quarter,
+second, to_date,
+to_utc_timestamp,
+unix_timestamp, weekofyear,
+window, year
+
+
+
+Examples
+
+## Not run: date_sub(df$d, 1)
+
+
+
+[Package SparkR version 2.1.2 
Index]
+

http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6155a89/site/docs/2.1.2/api/R/datediff.html
--
diff --git a/site/docs/2.1.2/api/R/datediff.html 
b/site/docs/2.1.2/api/R/datediff.html
new file mode 100644
index 000..78ff84c
--- /dev/null
+++ b/site/docs/2.1.2/api/R/datediff.html
@@ -0,0 +1,75 @@
+http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd;>http://www.w3.org/1999/xhtml;>R: datediff
+
+
+
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css;>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js";>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js";>
+hljs.initHighlightingOnLoad();
+
+
+datediff 
{SparkR}R Documentation
+
+datediff
+
+Description
+
+Returns the number of days from start to end.
+
+
+
+Usage
+
+
+datediff(y, x)
+
+## S4 method for signature 'Column'
+datediff(y, x)
+
+
+
+Arguments
+
+
+y
+
+end Column to use.
+
+x
+
+start Column to use.
+
+
+
+
+Note
+
+datediff since 1.5.0
+
+
+
+See Also
+
+Other datetime_funcs: add_months,
+date_add, date_format,
+date_sub, dayofmonth,
+dayofyear, from_unixtime,
+from_utc_timestamp, 
hour,
+last_day, minute,
+months_between, month,
+next_day, quarter,
+second, to_date,
+to_utc_timestamp,
+unix_timestamp, weekofyear,
+window, year
+
+
+
+Examples
+
+## Not run: datediff(df$c, x)
+
+
+
+[Package SparkR version 2.1.2 
Index]
+

http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6155a89/site/docs/2.1.2/api/R/dayofmonth.html

[23/51] [partial] spark-website git commit: Add 2.1.2 docs

2017-10-17 Thread holden
http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6155a89/site/docs/2.1.2/api/java/org/apache/spark/InternalAccumulator.html
--
diff --git a/site/docs/2.1.2/api/java/org/apache/spark/InternalAccumulator.html 
b/site/docs/2.1.2/api/java/org/apache/spark/InternalAccumulator.html
new file mode 100644
index 000..b236e74
--- /dev/null
+++ b/site/docs/2.1.2/api/java/org/apache/spark/InternalAccumulator.html
@@ -0,0 +1,499 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+
+
+InternalAccumulator (Spark 2.1.2 JavaDoc)
+
+
+
+
+
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+
+
+
+Overview
+Package
+Class
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev Class
+Next 
Class
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+
+
+Detail:
+Field|
+Constr|
+Method
+
+
+
+
+
+
+
+
+org.apache.spark
+Class 
InternalAccumulator
+
+
+
+Object
+
+
+org.apache.spark.InternalAccumulator
+
+
+
+
+
+
+
+
+public class InternalAccumulator
+extends Object
+A collection of fields and methods concerned with internal 
accumulators that represent
+ task level metrics.
+
+
+
+
+
+
+
+
+
+
+
+Nested Class Summary
+
+Nested Classes
+
+Modifier and Type
+Class and Description
+
+
+static class
+InternalAccumulator.input$
+
+
+static class
+InternalAccumulator.output$
+
+
+static class
+InternalAccumulator.shuffleRead$
+
+
+static class
+InternalAccumulator.shuffleWrite$
+
+
+
+
+
+
+
+
+
+Constructor Summary
+
+Constructors
+
+Constructor and Description
+
+
+InternalAccumulator()
+
+
+
+
+
+
+
+
+
+Method Summary
+
+Methods
+
+Modifier and Type
+Method and Description
+
+
+static String
+DISK_BYTES_SPILLED()
+
+
+static String
+EXECUTOR_CPU_TIME()
+
+
+static String
+EXECUTOR_DESERIALIZE_CPU_TIME()
+
+
+static String
+EXECUTOR_DESERIALIZE_TIME()
+
+
+static String
+EXECUTOR_RUN_TIME()
+
+
+static String
+INPUT_METRICS_PREFIX()
+
+
+static String
+JVM_GC_TIME()
+
+
+static String
+MEMORY_BYTES_SPILLED()
+
+
+static String
+METRICS_PREFIX()
+
+
+static String
+OUTPUT_METRICS_PREFIX()
+
+
+static String
+PEAK_EXECUTION_MEMORY()
+
+
+static String
+RESULT_SERIALIZATION_TIME()
+
+
+static String
+RESULT_SIZE()
+
+
+static String
+SHUFFLE_READ_METRICS_PREFIX()
+
+
+static String
+SHUFFLE_WRITE_METRICS_PREFIX()
+
+
+static String
+TEST_ACCUM()
+
+
+static String
+UPDATED_BLOCK_STATUSES()
+
+
+
+
+
+
+Methods inherited from classObject
+equals, getClass, hashCode, notify, notifyAll, toString, wait, wait, 
wait
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+Constructor Detail
+
+
+
+
+
+InternalAccumulator
+publicInternalAccumulator()
+
+
+
+
+
+
+
+
+
+Method Detail
+
+
+
+
+
+METRICS_PREFIX
+public staticStringMETRICS_PREFIX()
+
+
+
+
+
+
+
+SHUFFLE_READ_METRICS_PREFIX
+public staticStringSHUFFLE_READ_METRICS_PREFIX()
+
+
+
+
+
+
+
+SHUFFLE_WRITE_METRICS_PREFIX
+public staticStringSHUFFLE_WRITE_METRICS_PREFIX()
+
+
+
+
+
+
+
+OUTPUT_METRICS_PREFIX
+public staticStringOUTPUT_METRICS_PREFIX()
+
+
+
+
+
+
+
+INPUT_METRICS_PREFIX
+public staticStringINPUT_METRICS_PREFIX()
+
+
+
+
+
+
+
+EXECUTOR_DESERIALIZE_TIME
+public staticStringEXECUTOR_DESERIALIZE_TIME()
+
+
+
+
+
+
+
+EXECUTOR_DESERIALIZE_CPU_TIME
+public staticStringEXECUTOR_DESERIALIZE_CPU_TIME()
+
+
+
+
+
+
+
+EXECUTOR_RUN_TIME
+public staticStringEXECUTOR_RUN_TIME()
+
+
+
+
+
+
+
+EXECUTOR_CPU_TIME
+public staticStringEXECUTOR_CPU_TIME()
+
+
+
+
+
+
+
+RESULT_SIZE
+public staticStringRESULT_SIZE()
+
+
+
+
+
+
+
+JVM_GC_TIME
+public staticStringJVM_GC_TIME()
+
+
+
+
+
+
+
+RESULT_SERIALIZATION_TIME
+public staticStringRESULT_SERIALIZATION_TIME()
+
+
+
+
+
+
+
+MEMORY_BYTES_SPILLED
+public staticStringMEMORY_BYTES_SPILLED()
+
+
+
+
+
+
+
+DISK_BYTES_SPILLED
+public staticStringDISK_BYTES_SPILLED()
+
+
+
+
+
+
+
+PEAK_EXECUTION_MEMORY
+public staticStringPEAK_EXECUTION_MEMORY()
+
+
+
+
+
+
+
+UPDATED_BLOCK_STATUSES
+public staticStringUPDATED_BLOCK_STATUSES()
+
+
+
+
+
+
+
+TEST_ACCUM
+public staticStringTEST_ACCUM()
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+Overview
+Package
+Class
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev Class
+Next 
Class
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+
+
+Detail:
+Field|
+Constr|
+Method
+
+
+
+
+
+
+
+


[24/51] [partial] spark-website git commit: Add 2.1.2 docs

2017-10-17 Thread holden
http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6155a89/site/docs/2.1.2/api/java/org/apache/spark/ExecutorRemoved.html
--
diff --git a/site/docs/2.1.2/api/java/org/apache/spark/ExecutorRemoved.html 
b/site/docs/2.1.2/api/java/org/apache/spark/ExecutorRemoved.html
new file mode 100644
index 000..60dcc93
--- /dev/null
+++ b/site/docs/2.1.2/api/java/org/apache/spark/ExecutorRemoved.html
@@ -0,0 +1,356 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+
+
+ExecutorRemoved (Spark 2.1.2 JavaDoc)
+
+
+
+
+
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+
+
+
+Overview
+Package
+Class
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev Class
+Next Class
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+
+
+Detail:
+Field|
+Constr|
+Method
+
+
+
+
+
+
+
+
+org.apache.spark
+Class ExecutorRemoved
+
+
+
+Object
+
+
+org.apache.spark.ExecutorRemoved
+
+
+
+
+
+
+
+All Implemented Interfaces:
+java.io.Serializable, scala.Equals, scala.Product
+
+
+
+public class ExecutorRemoved
+extends Object
+implements scala.Product, scala.Serializable
+See Also:Serialized
 Form
+
+
+
+
+
+
+
+
+
+
+
+Constructor Summary
+
+Constructors
+
+Constructor and Description
+
+
+ExecutorRemoved(StringexecutorId)
+
+
+
+
+
+
+
+
+
+Method Summary
+
+Methods
+
+Modifier and Type
+Method and Description
+
+
+abstract static boolean
+canEqual(Objectthat)
+
+
+abstract static boolean
+equals(Objectthat)
+
+
+String
+executorId()
+
+
+abstract static int
+productArity()
+
+
+abstract static Object
+productElement(intn)
+
+
+static 
scala.collection.IteratorObject
+productIterator()
+
+
+static String
+productPrefix()
+
+
+
+
+
+
+Methods inherited from classObject
+equals, getClass, hashCode, notify, notifyAll, toString, wait, wait, 
wait
+
+
+
+
+
+Methods inherited from interfacescala.Product
+productArity, productElement, productIterator, productPrefix
+
+
+
+
+
+Methods inherited from interfacescala.Equals
+canEqual, equals
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+Constructor Detail
+
+
+
+
+
+ExecutorRemoved
+publicExecutorRemoved(StringexecutorId)
+
+
+
+
+
+
+
+
+
+Method Detail
+
+
+
+
+
+canEqual
+public abstract staticbooleancanEqual(Objectthat)
+
+
+
+
+
+
+
+equals
+public abstract staticbooleanequals(Objectthat)
+
+
+
+
+
+
+
+productElement
+public abstract staticObjectproductElement(intn)
+
+
+
+
+
+
+
+productArity
+public abstract staticintproductArity()
+
+
+
+
+
+
+
+productIterator
+public 
staticscala.collection.IteratorObjectproductIterator()
+
+
+
+
+
+
+
+productPrefix
+public staticStringproductPrefix()
+
+
+
+
+
+
+
+executorId
+publicStringexecutorId()
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+Overview
+Package
+Class
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev Class
+Next Class
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+
+
+Detail:
+Field|
+Constr|
+Method
+
+
+
+
+
+
+
+

http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6155a89/site/docs/2.1.2/api/java/org/apache/spark/ExpireDeadHosts.html
--
diff --git a/site/docs/2.1.2/api/java/org/apache/spark/ExpireDeadHosts.html 
b/site/docs/2.1.2/api/java/org/apache/spark/ExpireDeadHosts.html
new file mode 100644
index 000..4edc799
--- /dev/null
+++ b/site/docs/2.1.2/api/java/org/apache/spark/ExpireDeadHosts.html
@@ -0,0 +1,323 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+
+
+ExpireDeadHosts (Spark 2.1.2 JavaDoc)
+
+
+
+
+
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+
+
+
+Overview
+Package
+Class
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev Class
+Next Class
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+
+
+Detail:
+Field|
+Constr|
+Method
+
+
+
+
+
+
+
+
+org.apache.spark
+Class ExpireDeadHosts
+
+
+
+Object
+
+

[01/51] [partial] spark-website git commit: Add 2.1.2 docs

2017-10-17 Thread holden
Repository: spark-website
Updated Branches:
  refs/heads/asf-site 0490125a8 -> a6155a89d


http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6155a89/site/docs/2.1.2/api/java/org/apache/spark/api/r/RRDD.html
--
diff --git a/site/docs/2.1.2/api/java/org/apache/spark/api/r/RRDD.html 
b/site/docs/2.1.2/api/java/org/apache/spark/api/r/RRDD.html
new file mode 100644
index 000..19fea40
--- /dev/null
+++ b/site/docs/2.1.2/api/java/org/apache/spark/api/r/RRDD.html
@@ -0,0 +1,2158 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+
+
+RRDD (Spark 2.1.2 JavaDoc)
+
+
+
+
+
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+
+
+
+Overview
+Package
+Class
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev 
Class
+Next 
Class
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+
+
+Detail:
+Field|
+Constr|
+Method
+
+
+
+
+
+
+
+
+org.apache.spark.api.r
+Class RRDDT
+
+
+
+Object
+
+
+org.apache.spark.rdd.RDDU
+
+
+org.apache.spark.api.r.BaseRRDDT,byte[]
+
+
+org.apache.spark.api.r.RRDDT
+
+
+
+
+
+
+
+
+
+
+
+All Implemented Interfaces:
+java.io.Serializable
+
+
+
+public class RRDDT
+extends BaseRRDDT,byte[]
+An RDD that stores serialized R objects as 
Array[Byte].
+See Also:Serialized
 Form
+
+
+
+
+
+
+
+
+
+
+
+Constructor Summary
+
+Constructors
+
+Constructor and Description
+
+
+RRDD(RDDTparent,
+byte[]func,
+Stringdeserializer,
+Stringserializer,
+byte[]packageNames,
+Object[]broadcastVars,
+scala.reflect.ClassTagTevidence$4)
+
+
+
+
+
+
+
+
+
+Method Summary
+
+Methods
+
+Modifier and Type
+Method and Description
+
+
+static RDDT
+$plus$plus(RDDTother)
+
+
+static UU
+aggregate(UzeroValue,
+ scala.Function2U,T,UseqOp,
+ scala.Function2U,U,UcombOp,
+ scala.reflect.ClassTagUevidence$30)
+
+
+JavaRDDbyte[]
+asJavaRDD()
+
+
+static RDDT
+cache()
+
+
+static URDDscala.Tuple2T,U
+cartesian(RDDUother,
+ scala.reflect.ClassTagUevidence$5)
+
+
+static void
+checkpoint()
+
+
+static RDDT
+coalesce(intnumPartitions,
+booleanshuffle,
+scala.OptionPartitionCoalescerpartitionCoalescer,
+scala.math.OrderingTord)
+
+
+static boolean
+coalesce$default$2()
+
+
+static scala.OptionPartitionCoalescer
+coalesce$default$3()
+
+
+static scala.math.OrderingT
+coalesce$default$4(intnumPartitions,
+  booleanshuffle,
+  scala.OptionPartitionCoalescerpartitionCoalescer)
+
+
+static Object
+collect()
+
+
+static URDDU
+collect(scala.PartialFunctionT,Uf,
+   scala.reflect.ClassTagUevidence$29)
+
+
+static 
scala.collection.IteratorU
+compute(Partitionpartition,
+   TaskContextcontext)
+
+
+static SparkContext
+context()
+
+
+static long
+count()
+
+
+static PartialResultBoundedDouble
+countApprox(longtimeout,
+   doubleconfidence)
+
+
+static double
+countApprox$default$2()
+
+
+static long
+countApproxDistinct(doublerelativeSD)
+
+
+static long
+countApproxDistinct(intp,
+   intsp)
+
+
+static double
+countApproxDistinct$default$1()
+
+
+static 
scala.collection.MapT,Object
+countByValue(scala.math.OrderingTord)
+
+
+static scala.math.OrderingT
+countByValue$default$1()
+
+
+static PartialResultscala.collection.MapT,BoundedDouble
+countByValueApprox(longtimeout,
+  doubleconfidence,
+  scala.math.OrderingTord)
+
+
+static double
+countByValueApprox$default$2()
+
+
+static scala.math.OrderingT
+countByValueApprox$default$3(longtimeout,
+doubleconfidence)
+
+
+static JavaRDDbyte[]
+createRDDFromArray(JavaSparkContextjsc,
+  byte[][]arr)
+Create an RRDD given a sequence of byte arrays.
+
+
+
+static JavaRDDbyte[]
+createRDDFromFile(JavaSparkContextjsc,
+ StringfileName,
+ intparallelism)
+Create an RRDD given a temporary file name.
+
+
+
+static JavaSparkContext
+createSparkContext(Stringmaster,
+  StringappName,
+  StringsparkHome,
+  String[]jars,
+  java.util.MapObject,ObjectsparkEnvirMap,
+  
java.util.MapObject,ObjectsparkExecutorEnvMap)
+
+
+static scala.collection.SeqDependency?
+dependencies()
+
+
+static RDDT
+distinct()
+
+
+static RDDT
+distinct(intnumPartitions,
+scala.math.OrderingTord)
+
+
+static scala.math.OrderingT
+distinct$default$2(intnumPartitions)
+
+
+static RDDT
+filter(scala.Function1T,Objectf)
+
+
+static T
+first()
+
+
+static URDDU

[46/51] [partial] spark-website git commit: Add 2.1.2 docs

2017-10-17 Thread holden
http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6155a89/site/docs/2.1.2/api/R/columns.html
--
diff --git a/site/docs/2.1.2/api/R/columns.html 
b/site/docs/2.1.2/api/R/columns.html
new file mode 100644
index 000..3e7242a
--- /dev/null
+++ b/site/docs/2.1.2/api/R/columns.html
@@ -0,0 +1,137 @@
+http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd;>http://www.w3.org/1999/xhtml;>R: Column Names of 
SparkDataFrame
+
+
+
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css;>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js";>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js";>
+hljs.initHighlightingOnLoad();
+
+
+colnames 
{SparkR}R Documentation
+
+Column Names of SparkDataFrame
+
+Description
+
+Return all column names as a list.
+
+
+
+Usage
+
+
+colnames(x, do.NULL = TRUE, prefix = "col")
+
+colnames(x) - value
+
+columns(x)
+
+## S4 method for signature 'SparkDataFrame'
+columns(x)
+
+## S4 method for signature 'SparkDataFrame'
+names(x)
+
+## S4 replacement method for signature 'SparkDataFrame'
+names(x) - value
+
+## S4 method for signature 'SparkDataFrame'
+colnames(x)
+
+## S4 replacement method for signature 'SparkDataFrame'
+colnames(x) - value
+
+
+
+Arguments
+
+
+x
+
+a SparkDataFrame.
+
+do.NULL
+
+currently not used.
+
+prefix
+
+currently not used.
+
+value
+
+a character vector. Must have the same length as the number
+of columns in the SparkDataFrame.
+
+
+
+
+Note
+
+columns since 1.4.0
+
+names since 1.5.0
+
+names- since 1.5.0
+
+colnames since 1.6.0
+
+colnames- since 1.6.0
+
+
+
+See Also
+
+Other SparkDataFrame functions: SparkDataFrame-class,
+agg, arrange,
+as.data.frame, attach,
+cache, coalesce,
+collect, coltypes,
+createOrReplaceTempView,
+crossJoin, dapplyCollect,
+dapply, describe,
+dim, distinct,
+dropDuplicates, dropna,
+drop, dtypes,
+except, explain,
+filter, first,
+gapplyCollect, gapply,
+getNumPartitions, group_by,
+head, histogram,
+insertInto, intersect,
+isLocal, join,
+limit, merge,
+mutate, ncol,
+nrow, persist,
+printSchema, randomSplit,
+rbind, registerTempTable,
+rename, repartition,
+sample, saveAsTable,
+schema, selectExpr,
+select, showDF,
+show, storageLevel,
+str, subset,
+take, union,
+unpersist, withColumn,
+with, write.df,
+write.jdbc, write.json,
+write.orc, write.parquet,
+write.text
+
+
+
+Examples
+
+## Not run: 
+##D sparkR.session()
+##D path - path/to/file.json
+##D df - read.json(path)
+##D columns(df)
+##D colnames(df)
+## End(Not run)
+
+
+
+[Package SparkR version 2.1.2 
Index]
+

http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6155a89/site/docs/2.1.2/api/R/concat.html
--
diff --git a/site/docs/2.1.2/api/R/concat.html 
b/site/docs/2.1.2/api/R/concat.html
new file mode 100644
index 000..6d226c8
--- /dev/null
+++ b/site/docs/2.1.2/api/R/concat.html
@@ -0,0 +1,77 @@
+http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd;>http://www.w3.org/1999/xhtml;>R: concat
+
+
+
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css;>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js";>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js";>
+hljs.initHighlightingOnLoad();
+
+
+concat 
{SparkR}R Documentation
+
+concat
+
+Description
+
+Concatenates multiple input string columns together into a single string 
column.
+
+
+
+Usage
+
+
+concat(x, ...)
+
+## S4 method for signature 'Column'
+concat(x, ...)
+
+
+
+Arguments
+
+
+x
+
+Column to compute on
+
+...
+
+other columns
+
+
+
+
+Note
+
+concat since 1.5.0
+
+
+
+See Also
+
+Other string_funcs: ascii,
+base64, concat_ws,
+decode, encode,
+format_number, format_string,
+initcap, instr,
+length, levenshtein,
+locate, lower,
+lpad, ltrim,
+regexp_extract,
+regexp_replace, reverse,
+rpad, rtrim,
+soundex, substring_index,
+translate, trim,
+unbase64, upper
+
+
+
+Examples
+
+## Not run: concat(df$strings, df$strings2)
+
+
+
+[Package SparkR version 2.1.2 
Index]
+

http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6155a89/site/docs/2.1.2/api/R/concat_ws.html
--
diff --git a/site/docs/2.1.2/api/R/concat_ws.html 
b/site/docs/2.1.2/api/R/concat_ws.html
new file mode 100644
index 000..e30c492
--- /dev/null
+++ b/site/docs/2.1.2/api/R/concat_ws.html
@@ -0,0 +1,82 @@
+http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd;>http://www.w3.org/1999/xhtml;>R: concat_ws
+
+
+
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css;>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js";>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js";>
+hljs.initHighlightingOnLoad();
+
+
+concat_ws 
{SparkR}R Documentation
+
+concat_ws
+

spark git commit: [SPARK-22271][SQL] mean overflows and returns null for some decimal variables

2017-10-17 Thread lixiao
Repository: spark
Updated Branches:
  refs/heads/branch-2.2 71d1cb6a4 -> c42309143


[SPARK-22271][SQL] mean overflows and returns null for some decimal variables

## What changes were proposed in this pull request?

In Average.scala, it has
```
  override lazy val evaluateExpression = child.dataType match {
case DecimalType.Fixed(p, s) =>
  // increase the precision and scale to prevent precision loss
  val dt = DecimalType.bounded(p + 14, s + 4)
  Cast(Cast(sum, dt) / Cast(count, dt), resultType)
case _ =>
  Cast(sum, resultType) / Cast(count, resultType)
  }

  def setChild (newchild: Expression) = {
child = newchild
  }

```
It is possible that  Cast(count, dt), resultType) will make the precision of 
the decimal number bigger than 38, and this causes over flow.  Since count is 
an integer and doesn't need a scale, I will cast it using 
DecimalType.bounded(38,0)
## How was this patch tested?
In DataFrameSuite, I will add a test case.

Please review http://spark.apache.org/contributing.html before opening a pull 
request.

Author: Huaxin Gao 

Closes #19496 from huaxingao/spark-22271.

(cherry picked from commit 28f9f3f22511e9f2f900764d9bd5b90d2773)
Signed-off-by: gatorsmile 

# Conflicts:
#   sql/core/src/test/scala/org/apache/spark/sql/DataFrameSuite.scala


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/c4230914
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/c4230914
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/c4230914

Branch: refs/heads/branch-2.2
Commit: c423091432d55fa88c2799181b9d3575213530c8
Parents: 71d1cb6
Author: Huaxin Gao 
Authored: Tue Oct 17 12:50:41 2017 -0700
Committer: gatorsmile 
Committed: Tue Oct 17 12:53:14 2017 -0700

--
 .../spark/sql/catalyst/expressions/aggregate/Average.scala  | 3 ++-
 .../test/scala/org/apache/spark/sql/DataFrameSuite.scala| 9 +
 2 files changed, 11 insertions(+), 1 deletion(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/c4230914/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/aggregate/Average.scala
--
diff --git 
a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/aggregate/Average.scala
 
b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/aggregate/Average.scala
index c423e17..708bdbf 100644
--- 
a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/aggregate/Average.scala
+++ 
b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/aggregate/Average.scala
@@ -80,7 +80,8 @@ case class Average(child: Expression) extends 
DeclarativeAggregate with Implicit
 case DecimalType.Fixed(p, s) =>
   // increase the precision and scale to prevent precision loss
   val dt = DecimalType.bounded(p + 14, s + 4)
-  Cast(Cast(sum, dt) / Cast(count, dt), resultType)
+  Cast(Cast(sum, dt) / Cast(count, 
DecimalType.bounded(DecimalType.MAX_PRECISION, 0)),
+resultType)
 case _ =>
   Cast(sum, resultType) / Cast(count, resultType)
   }

http://git-wip-us.apache.org/repos/asf/spark/blob/c4230914/sql/core/src/test/scala/org/apache/spark/sql/DataFrameSuite.scala
--
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/DataFrameSuite.scala 
b/sql/core/src/test/scala/org/apache/spark/sql/DataFrameSuite.scala
index 1f29cb7..5d02239 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/DataFrameSuite.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/DataFrameSuite.scala
@@ -1772,4 +1772,13 @@ class DataFrameSuite extends QueryTest with 
SharedSQLContext {
   checkAnswer(spark.table("t4"), Row(0, 0))
 }
   }
+
+  test("SPARK-22271: mean overflows and returns null for some decimal 
variables") {
+val d = 0.034567890
+val df = Seq(d, d, d, d, d, d, d, d, d, d).toDF("DecimalCol")
+val result = df.select('DecimalCol cast DecimalType(38, 33))
+  .select(col("DecimalCol")).describe()
+val mean = result.select("DecimalCol").where($"summary" === "mean")
+assert(mean.collect().toSet === 
Set(Row("0.034567890")))
+  }
 }


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[29/51] [partial] spark-website git commit: Add 2.1.2 docs

2017-10-17 Thread holden
http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/java/lib/jquery.js
--
diff --git a/site/docs/2.1.2/api/java/lib/jquery.js 
b/site/docs/2.1.2/api/java/lib/jquery.js
new file mode 100644
index 000..bc3fbc8
--- /dev/null
+++ b/site/docs/2.1.2/api/java/lib/jquery.js
@@ -0,0 +1,2 @@
+/*! jQuery v1.8.2 jquery.com | jquery.org/license */
+(function(a,b){function G(a){var b=F[a]={};return 
p.each(a.split(s),function(a,c){b[c]=!0}),b}function 
J(a,c,d){if(d===b&===1){var 
e="data-"+c.replace(I,"-$1").toLowerCase();d=a.getAttribute(e);if(typeof 
d=="string"){try{d=d==="true"?!0:d==="false"?!1:d==="null"?null:+d+""===d?+d:H.test(d)?p.parseJSON(d):d}catch(f){}p.data(a,c,d)}else
 d=b}return d}function K(a){var b;for(b in 
a){if(b==="data"&(a[b]))continue;if(b!=="toJSON")return!1}return!0}function
 ba(){return!1}function bb(){return!0}function 
bh(a){return!a||!a.parentNode||a.parentNode.nodeType===11}function bi(a,b){do 
a=a[b];while(a&!==1);return a}function 
bj(a,b,c){b=b||0;if(p.isFunction(b))return p.grep(a,function(a,d){var 
e=!!b.call(a,d,a);return e===c});if(b.nodeType)return 
p.grep(a,function(a,d){return a===b===c});if(typeof b=="string"){var 
d=p.grep(a,function(a){return a.nodeType===1});if(be.test(b))return 
p.filter(b,d,!c);b=p.filter(b,d)}return p.grep(a,function(a,d){return p.inArray(
 a,b)>=0===c})}function bk(a){var 
b=bl.split("|"),c=a.createDocumentFragment();if(c.createElement)while(b.length)c.createElement(b.pop());return
 c}function bC(a,b){return 
a.getElementsByTagName(b)[0]||a.appendChild(a.ownerDocument.createElement(b))}function
 bD(a,b){if(b.nodeType!==1||!p.hasData(a))return;var 
c,d,e,f=p._data(a),g=p._data(b,f),h=f.events;if(h){delete 
g.handle,g.events={};for(c in 
h)for(d=0,e=h[c].length;d").appendTo(e.body),c=b.css("display");b.remove();if(c==="none"||c===""){bI=e.body.appendChild(bI||p.extend(e.createElement("iframe"),{frameBorder:0,width:0,height:0}));if(!bJ||!bI.
 
createElement)bJ=(bI.contentWindow||bI.contentDocument).document,bJ.write(""),bJ.close();b=bJ.body.appendChild(bJ.createElement(a)),c=bH(b,"display"),e.body.removeChild(bI)}return
 bS[a]=c,c}function ci(a,b,c,d){var 
e;if(p.isArray(b))p.each(b,function(b,e){c||ce.test(a)?d(a,e):ci(a+"["+(typeof 
e=="object"?b:"")+"]",e,c,d)});else if(!c&(b)==="object")for(e in 
b)ci(a+"["+e+"]",b[e],c,d);else d(a,b)}function cz(a){return 
function(b,c){typeof b!="string"&&(c=b,b="*");var 

[22/51] [partial] spark-website git commit: Add 2.1.2 docs

2017-10-17 Thread holden
http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/java/org/apache/spark/JobExecutionStatus.html
--
diff --git a/site/docs/2.1.2/api/java/org/apache/spark/JobExecutionStatus.html 
b/site/docs/2.1.2/api/java/org/apache/spark/JobExecutionStatus.html
new file mode 100644
index 000..be22a87
--- /dev/null
+++ b/site/docs/2.1.2/api/java/org/apache/spark/JobExecutionStatus.html
@@ -0,0 +1,358 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+
+
+JobExecutionStatus (Spark 2.1.2 JavaDoc)
+
+
+
+
+
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+
+
+
+Overview
+Package
+Class
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev 
Class
+Next Class
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+Summary:
+Nested|
+Enum Constants|
+Field|
+Method
+
+
+Detail:
+Enum Constants|
+Field|
+Method
+
+
+
+
+
+
+
+
+org.apache.spark
+Enum JobExecutionStatus
+
+
+
+Object
+
+
+EnumJobExecutionStatus
+
+
+org.apache.spark.JobExecutionStatus
+
+
+
+
+
+
+
+
+
+All Implemented Interfaces:
+java.io.Serializable, ComparableJobExecutionStatus
+
+
+
+public enum JobExecutionStatus
+extends EnumJobExecutionStatus
+
+
+
+
+
+
+
+
+
+
+
+Enum Constant Summary
+
+Enum Constants
+
+Enum Constant and Description
+
+
+FAILED
+
+
+RUNNING
+
+
+SUCCEEDED
+
+
+UNKNOWN
+
+
+
+
+
+
+
+
+
+Method Summary
+
+Methods
+
+Modifier and Type
+Method and Description
+
+
+static JobExecutionStatus
+fromString(Stringstr)
+
+
+static JobExecutionStatus
+valueOf(Stringname)
+Returns the enum constant of this type with the specified 
name.
+
+
+
+static JobExecutionStatus[]
+values()
+Returns an array containing the constants of this enum 
type, in
+the order they are declared.
+
+
+
+
+
+
+
+Methods inherited from classEnum
+compareTo, equals, getDeclaringClass, hashCode, name, ordinal, toString, 
valueOf
+
+
+
+
+
+Methods inherited from classObject
+getClass, notify, notifyAll, wait, wait, wait
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+Enum Constant Detail
+
+
+
+
+
+RUNNING
+public static finalJobExecutionStatus RUNNING
+
+
+
+
+
+
+
+SUCCEEDED
+public static finalJobExecutionStatus SUCCEEDED
+
+
+
+
+
+
+
+FAILED
+public static finalJobExecutionStatus FAILED
+
+
+
+
+
+
+
+UNKNOWN
+public static finalJobExecutionStatus UNKNOWN
+
+
+
+
+
+
+
+
+
+Method Detail
+
+
+
+
+
+values
+public staticJobExecutionStatus[]values()
+Returns an array containing the constants of this enum 
type, in
+the order they are declared.  This method may be used to iterate
+over the constants as follows:
+
+for (JobExecutionStatus c : JobExecutionStatus.values())
+   System.out.println(c);
+
+Returns:an array containing the 
constants of this enum type, in the order they are declared
+
+
+
+
+
+
+
+valueOf
+public staticJobExecutionStatusvalueOf(Stringname)
+Returns the enum constant of this type with the specified 
name.
+The string must match exactly an identifier used to declare an
+enum constant in this type.  (Extraneous whitespace characters are 
+not permitted.)
+Parameters:name - 
the name of the enum constant to be returned.
+Returns:the enum constant with the 
specified name
+Throws:
+IllegalArgumentException - if this enum type has no constant 
with the specified name
+NullPointerException - if the argument is null
+
+
+
+
+
+
+
+fromString
+public staticJobExecutionStatusfromString(Stringstr)
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+Overview
+Package
+Class
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev 
Class
+Next Class
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+Summary:
+Nested|
+Enum Constants|
+Field|
+Method
+
+
+Detail:
+Enum Constants|
+Field|
+Method
+
+
+
+
+
+
+
+

http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/java/org/apache/spark/JobSubmitter.html
--
diff --git a/site/docs/2.1.2/api/java/org/apache/spark/JobSubmitter.html 
b/site/docs/2.1.2/api/java/org/apache/spark/JobSubmitter.html
new file mode 100644
index 000..84e3160
--- /dev/null
+++ b/site/docs/2.1.2/api/java/org/apache/spark/JobSubmitter.html
@@ -0,0 +1,225 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+
+
+JobSubmitter (Spark 2.1.2 JavaDoc)
+
+
+
+
+
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+
+
+
+Overview
+Package
+Class
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev Class
+Next Class
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+
+
+Detail:
+Field|
+Constr|
+Method
+
+
+
+
+
+
+
+
+org.apache.spark.api.java.function
+Interface 
CoGroupFunctionK,V1,V2,R
+
+
+
+
+
+
+All Superinterfaces:
+java.io.Serializable
+
+
+
+public interface CoGroupFunctionK,V1,V2,R
+extends java.io.Serializable
+A function that returns zero or more output records from 
each grouping key and its values from 2
+ Datasets.
+
+
+
+
+
+
+
+
+
+
+
+Method Summary
+
+Methods
+
+Modifier and Type
+Method and Description
+
+
+java.util.IteratorR
+call(Kkey,
+java.util.IteratorV1left,
+java.util.IteratorV2right)
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+Method Detail
+
+
+
+
+
+
+
+call
+java.util.IteratorRcall(Kkey,
+ java.util.IteratorV1left,
+ java.util.IteratorV2right)
+   throws Exception
+Throws:
+Exception
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+Overview
+Package
+Class
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev Class
+Next Class
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+
+
+Detail:
+Field|
+Constr|
+Method
+
+
+
+
+
+
+
+

http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/java/org/apache/spark/api/java/function/DoubleFlatMapFunction.html
--
diff --git 
a/site/docs/2.1.2/api/java/org/apache/spark/api/java/function/DoubleFlatMapFunction.html
 
b/site/docs/2.1.2/api/java/org/apache/spark/api/java/function/DoubleFlatMapFunction.html
new file mode 100644
index 000..628dfa8
--- /dev/null
+++ 
b/site/docs/2.1.2/api/java/org/apache/spark/api/java/function/DoubleFlatMapFunction.html
@@ -0,0 +1,219 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+
+
+DoubleFlatMapFunction (Spark 2.1.2 JavaDoc)
+
+
+
+
+
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+
+
+
+Overview
+Package
+Class
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev Class
+Next Class
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+
+
+Detail:
+Field|
+Constr|
+Method
+
+
+
+
+
+
+
+
+org.apache.spark.api.java.function
+Interface 
DoubleFlatMapFunctionT
+
+
+
+
+
+
+All Superinterfaces:
+java.io.Serializable
+
+
+
+public interface DoubleFlatMapFunctionT
+extends java.io.Serializable
+A function that returns zero or more records of type Double 
from each input record.
+
+
+
+
+
+
+
+
+
+
+
+Method Summary
+
+Methods
+
+Modifier and Type
+Method and Description
+
+
+java.util.IteratorDouble
+call(Tt)
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+Method Detail
+
+
+
+
+
+
+
+call
+java.util.IteratorDoublecall(Tt)
+throws Exception
+Throws:
+Exception
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+Overview
+Package
+Class
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev Class
+Next Class
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+
+
+
+Overview
+Package
+Class
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev Class
+Next Class
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+
+
+Detail:
+Field|
+Constr|
+Method
+
+
+
+
+
+
+
+
+org.apache.spark
+Class ExecutorRemoved
+
+
+
+Object
+
+
+org.apache.spark.ExecutorRemoved
+
+
+
+
+
+
+
+All Implemented Interfaces:
+java.io.Serializable, scala.Equals, scala.Product
+
+
+
+public class ExecutorRemoved
+extends Object
+implements scala.Product, scala.Serializable
+See Also:Serialized
 Form
+
+
+
+
+
+
+
+
+
+
+
+Constructor Summary
+
+Constructors
+
+Constructor and Description
+
+
+ExecutorRemoved(StringexecutorId)
+
+
+
+
+
+
+
+
+
+Method Summary
+
+Methods
+
+Modifier and Type
+Method and Description
+
+
+abstract static boolean
+canEqual(Objectthat)
+
+
+abstract static boolean
+equals(Objectthat)
+
+
+String
+executorId()
+
+
+abstract static int
+productArity()
+
+
+abstract static Object
+productElement(intn)
+
+
+static 
scala.collection.IteratorObject
+productIterator()
+
+
+static String
+productPrefix()
+
+
+
+
+
+
+Methods inherited from classObject
+equals, getClass, hashCode, notify, notifyAll, toString, wait, wait, 
wait
+
+
+
+
+
+Methods inherited from interfacescala.Product
+productArity, productElement, productIterator, productPrefix
+
+
+
+
+
+Methods inherited from interfacescala.Equals
+canEqual, equals
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+Constructor Detail
+
+
+
+
+
+ExecutorRemoved
+publicExecutorRemoved(StringexecutorId)
+
+
+
+
+
+
+
+
+
+Method Detail
+
+
+
+
+
+canEqual
+public abstract staticbooleancanEqual(Objectthat)
+
+
+
+
+
+
+
+equals
+public abstract staticbooleanequals(Objectthat)
+
+
+
+
+
+
+
+productElement
+public abstract staticObjectproductElement(intn)
+
+
+
+
+
+
+
+productArity
+public abstract staticintproductArity()
+
+
+
+
+
+
+
+productIterator
+public 
staticscala.collection.IteratorObjectproductIterator()
+
+
+
+
+
+
+
+productPrefix
+public staticStringproductPrefix()
+
+
+
+
+
+
+
+executorId
+publicStringexecutorId()
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+Overview
+Package
+Class
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev Class
+Next Class
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+
+
+Detail:
+Field|
+Constr|
+Method
+
+
+
+
+
+
+
+

http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/java/org/apache/spark/ExpireDeadHosts.html
--
diff --git a/site/docs/2.1.2/api/java/org/apache/spark/ExpireDeadHosts.html 
b/site/docs/2.1.2/api/java/org/apache/spark/ExpireDeadHosts.html
new file mode 100644
index 000..4edc799
--- /dev/null
+++ b/site/docs/2.1.2/api/java/org/apache/spark/ExpireDeadHosts.html
@@ -0,0 +1,323 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+
+
+ExpireDeadHosts (Spark 2.1.2 JavaDoc)
+
+
+
+
+
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+
+
+
+Overview
+Package
+Class
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev Class
+Next Class
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+
+
+Detail:
+Field|
+Constr|
+Method
+
+
+
+
+
+
+
+
+org.apache.spark
+Class ExpireDeadHosts
+
+
+
+Object
+
+

[19/51] [partial] spark-website git commit: Add 2.1.2 docs

2017-10-17 Thread holden
http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/java/org/apache/spark/SparkContext.html
--
diff --git a/site/docs/2.1.2/api/java/org/apache/spark/SparkContext.html 
b/site/docs/2.1.2/api/java/org/apache/spark/SparkContext.html
new file mode 100644
index 000..38c0b01
--- /dev/null
+++ b/site/docs/2.1.2/api/java/org/apache/spark/SparkContext.html
@@ -0,0 +1,2527 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+
+
+SparkContext (Spark 2.1.2 JavaDoc)
+
+
+
+
+
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+
+
+
+Overview
+Package
+Class
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev Class
+Next Class
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+
+
+Detail:
+Field|
+Constr|
+Method
+
+
+
+
+
+
+
+
+org.apache.spark
+Class SparkContext
+
+
+
+Object
+
+
+org.apache.spark.SparkContext
+
+
+
+
+
+
+
+
+public class SparkContext
+extends Object
+Main entry point for Spark functionality. A SparkContext 
represents the connection to a Spark
+ cluster, and can be used to create RDDs, accumulators and broadcast variables 
on that cluster.
+ 
+ Only one SparkContext may be active per JVM.  You must stop() 
the active SparkContext before
+ creating a new one.  This limitation may eventually be removed; see 
SPARK-2243 for more details.
+ 
+ param:  config a Spark Config object describing the application 
configuration. Any settings in
+   this config overrides the default configs as well as system 
properties.
+
+
+
+
+
+
+
+
+
+
+
+Constructor Summary
+
+Constructors
+
+Constructor and Description
+
+
+SparkContext()
+Create a SparkContext that loads settings from system 
properties (for instance, when
+ launching with ./bin/spark-submit).
+
+
+
+SparkContext(SparkConfconfig)
+
+
+SparkContext(Stringmaster,
+StringappName,
+SparkConfconf)
+Alternative constructor that allows setting common Spark 
properties directly
+
+
+
+SparkContext(Stringmaster,
+StringappName,
+StringsparkHome,
+scala.collection.SeqStringjars,
+scala.collection.MapString,Stringenvironment)
+Alternative constructor that allows setting common Spark 
properties directly
+
+
+
+
+
+
+
+
+
+
+Method Summary
+
+Methods
+
+Modifier and Type
+Method and Description
+
+
+R,TAccumulableR,T
+accumulable(RinitialValue,
+   AccumulableParamR,Tparam)
+Deprecated.
+use AccumulatorV2. Since 2.0.0.
+
+
+
+
+R,TAccumulableR,T
+accumulable(RinitialValue,
+   Stringname,
+   AccumulableParamR,Tparam)
+Deprecated.
+use AccumulatorV2. Since 2.0.0.
+
+
+
+
+R,TAccumulableR,T
+accumulableCollection(RinitialValue,
+ 
scala.Function1R,scala.collection.generic.GrowableTevidence$9,
+ scala.reflect.ClassTagRevidence$10)
+Deprecated.
+use AccumulatorV2. Since 2.0.0.
+
+
+
+
+TAccumulatorT
+accumulator(TinitialValue,
+   AccumulatorParamTparam)
+Deprecated.
+use AccumulatorV2. Since 2.0.0.
+
+
+
+
+TAccumulatorT
+accumulator(TinitialValue,
+   Stringname,
+   AccumulatorParamTparam)
+Deprecated.
+use AccumulatorV2. Since 2.0.0.
+
+
+
+
+void
+addFile(Stringpath)
+Add a file to be downloaded with this Spark job on every 
node.
+
+
+
+void
+addFile(Stringpath,
+   booleanrecursive)
+Add a file to be downloaded with this Spark job on every 
node.
+
+
+
+void
+addJar(Stringpath)
+Adds a JAR dependency for all tasks to be executed on this 
SparkContext in the future.
+
+
+
+void
+addSparkListener(org.apache.spark.scheduler.SparkListenerInterfacelistener)
+:: DeveloperApi ::
+ Register a listener to receive up-calls from events that happen during 
execution.
+
+
+
+scala.OptionString
+applicationAttemptId()
+
+
+String
+applicationId()
+A unique identifier for the Spark application.
+
+
+
+String
+appName()
+
+
+RDDscala.Tuple2String,PortableDataStream
+binaryFiles(Stringpath,
+   intminPartitions)
+Get an RDD for a Hadoop-readable dataset as 
PortableDataStream for each file
+ (useful for binary data)
+
+
+
+RDDbyte[]
+binaryRecords(Stringpath,
+ intrecordLength,
+ org.apache.hadoop.conf.Configurationconf)
+Load data from a flat binary file, assuming the length of 
each record is constant.
+
+
+
+TBroadcastT
+broadcast(Tvalue,
+ scala.reflect.ClassTagTevidence$11)
+Broadcast a read-only variable to the cluster, returning a
+ Broadcast object for reading it in 
distributed functions.
+
+
+
+void

[48/51] [partial] spark-website git commit: Add 2.1.2 docs

2017-10-17 Thread holden
http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/R/00frame_toc.html
--
diff --git a/site/docs/2.1.2/api/R/00frame_toc.html 
b/site/docs/2.1.2/api/R/00frame_toc.html
new file mode 100644
index 000..1c62835
--- /dev/null
+++ b/site/docs/2.1.2/api/R/00frame_toc.html
@@ -0,0 +1,406 @@
+
+
+
+
+
+R Documentation of SparkR
+
+
+window.onload = function() {
+  var imgs = document.getElementsByTagName('img'), i, img;
+  for (i = 0; i < imgs.length; i++) {
+img = imgs[i];
+// center an image if it is the only element of its parent
+if (img.parentElement.childElementCount === 1)
+  img.parentElement.style.textAlign = 'center';
+  }
+};
+
+
+
+
+
+
+
+* {
+   font-family: "Trebuchet MS", "Lucida Grande", "Lucida Sans Unicode", 
"Lucida Sans", Arial, sans-serif;
+   font-size: 14px;
+}
+body {
+  padding: 0 5px; 
+  margin: 0 auto; 
+  width: 80%;
+  max-width: 60em; /* 960px */
+}
+
+h1, h2, h3, h4, h5, h6 {
+   color: #666;
+}
+h1, h2 {
+   text-align: center;
+}
+h1 {
+   font-size: x-large;
+}
+h2, h3 {
+   font-size: large;
+}
+h4, h6 {
+   font-style: italic;
+}
+h3 {
+   border-left: solid 5px #ddd;
+   padding-left: 5px;
+   font-variant: small-caps;
+}
+
+p img {
+   display: block;
+   margin: auto;
+}
+
+span, code, pre {
+   font-family: Monaco, "Lucida Console", "Courier New", Courier, 
monospace;
+}
+span.acronym {}
+span.env {
+   font-style: italic;
+}
+span.file {}
+span.option {}
+span.pkg {
+   font-weight: bold;
+}
+span.samp{}
+
+dt, p code {
+   background-color: #F7F7F7;
+}
+
+
+
+
+
+
+
+
+SparkR
+
+
+AFTSurvivalRegressionModel-class
+ALSModel-class
+GBTClassificationModel-class
+GBTRegressionModel-class
+GaussianMixtureModel-class
+GeneralizedLinearRegressionModel-class
+GroupedData
+IsotonicRegressionModel-class
+KMeansModel-class
+KSTest-class
+LDAModel-class
+LogisticRegressionModel-class
+MultilayerPerceptronClassificationModel-class
+NaiveBayesModel-class
+RandomForestClassificationModel-class
+RandomForestRegressionModel-class
+SparkDataFrame
+WindowSpec
+abs
+acos
+add_months
+alias
+approxCountDistinct
+approxQuantile
+arrange
+array_contains
+as.data.frame
+ascii
+asin
+atan
+atan2
+attach
+avg
+base64
+between
+bin
+bitwiseNOT
+bround
+cache
+cacheTable
+cancelJobGroup
+cast
+cbrt
+ceil
+clearCache
+clearJobGroup
+coalesce
+collect
+coltypes
+column
+columnfunctions
+columns
+concat
+concat_ws
+conv
+corr
+cos
+cosh
+count
+countDistinct
+cov
+covar_pop
+crc32
+createDataFrame
+createExternalTable
+createOrReplaceTempView
+crossJoin
+crosstab
+cume_dist
+dapply
+dapplyCollect
+date_add
+date_format
+date_sub
+datediff
+dayofmonth
+dayofyear
+decode
+dense_rank
+dim
+distinct
+drop
+dropDuplicates
+dropTempTable-deprecated
+dropTempView
+dtypes
+encode
+endsWith
+except
+exp
+explain
+explode
+expm1
+expr
+factorial
+filter
+first
+fitted
+floor
+format_number
+format_string
+freqItems
+from_unixtime
+fromutctimestamp
+gapply
+gapplyCollect
+generateAliasesForIntersectedCols
+getNumPartitions
+glm
+greatest
+groupBy
+hash
+hashCode
+head
+hex
+histogram
+hour
+hypot
+ifelse
+initcap
+insertInto
+install.spark
+instr
+intersect
+is.nan
+isLocal
+join
+kurtosis
+lag
+last
+last_day
+lead
+least
+length
+levenshtein
+limit
+lit
+locate
+log
+log10
+log1p
+log2
+lower
+lpad
+ltrim
+match
+max
+md5
+mean
+merge
+min
+minute
+monotonicallyincreasingid
+month
+months_between
+mutate
+nafunctions
+nanvl
+ncol
+negate
+next_day
+nrow
+ntile
+orderBy
+otherwise
+over
+partitionBy
+percent_rank
+persist
+pivot
+pmod
+posexplode
+predict
+print.jobj
+print.structField
+print.structType
+printSchema
+quarter
+rand
+randn
+randomSplit
+rangeBetween
+rank
+rbind
+read.df
+read.jdbc
+read.json
+read.ml
+read.orc
+read.parquet
+read.text
+regexp_extract
+regexp_replace
+registerTempTable-deprecated
+rename
+repartition
+reverse
+rint
+round
+row_number
+rowsBetween
+rpad
+rtrim
+sample
+sampleBy
+saveAsTable
+schema
+sd
+second
+select
+selectExpr
+setJobGroup
+setLogLevel
+sha1
+sha2
+shiftLeft
+shiftRight
+shiftRightUnsigned
+show
+showDF
+sign
+sin
+sinh
+size
+skewness
+sort_array
+soundex
+spark.addFile
+spark.als
+spark.gaussianMixture
+spark.gbt
+spark.getSparkFiles
+spark.getSparkFilesRootDirectory
+spark.glm
+spark.isoreg
+spark.kmeans
+spark.kstest
+spark.lapply
+spark.lda
+spark.logit
+spark.mlp
+spark.naiveBayes
+spark.randomForest
+spark.survreg
+sparkR.callJMethod
+sparkR.callJStatic
+sparkR.conf
+sparkR.init-deprecated
+sparkR.newJObject
+sparkR.session
+sparkR.session.stop
+sparkR.uiWebUrl
+sparkR.version
+sparkRHive.init-deprecated
+sparkRSQL.init-deprecated
+sparkpartitionid
+sql
+sqrt
+startsWith
+stddev_pop
+stddev_samp
+storageLevel
+str
+struct
+structField
+structType
+subset
+substr
+substring_index
+sum
+sumDistinct
+summarize
+summary
+tableNames
+tableToDF
+tables
+take
+tan
+tanh

[42/51] [partial] spark-website git commit: Add 2.1.2 docs

2017-10-17 Thread holden
http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/R/limit.html
--
diff --git a/site/docs/2.1.2/api/R/limit.html b/site/docs/2.1.2/api/R/limit.html
new file mode 100644
index 000..309b624
--- /dev/null
+++ b/site/docs/2.1.2/api/R/limit.html
@@ -0,0 +1,109 @@
+http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd;>http://www.w3.org/1999/xhtml;>R: Limit
+
+
+
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css;>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js";>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js";>
+hljs.initHighlightingOnLoad();
+
+
+limit 
{SparkR}R Documentation
+
+Limit
+
+Description
+
+Limit the resulting SparkDataFrame to the number of rows specified.
+
+
+
+Usage
+
+
+limit(x, num)
+
+## S4 method for signature 'SparkDataFrame,numeric'
+limit(x, num)
+
+
+
+Arguments
+
+
+x
+
+A SparkDataFrame
+
+num
+
+The number of rows to return
+
+
+
+
+Value
+
+A new SparkDataFrame containing the number of rows specified.
+
+
+
+Note
+
+limit since 1.4.0
+
+
+
+See Also
+
+Other SparkDataFrame functions: SparkDataFrame-class,
+agg, arrange,
+as.data.frame, attach,
+cache, coalesce,
+collect, colnames,
+coltypes,
+createOrReplaceTempView,
+crossJoin, dapplyCollect,
+dapply, describe,
+dim, distinct,
+dropDuplicates, dropna,
+drop, dtypes,
+except, explain,
+filter, first,
+gapplyCollect, gapply,
+getNumPartitions, group_by,
+head, histogram,
+insertInto, intersect,
+isLocal, join,
+merge, mutate,
+ncol, nrow,
+persist, printSchema,
+randomSplit, rbind,
+registerTempTable, rename,
+repartition, sample,
+saveAsTable, schema,
+selectExpr, select,
+showDF, show,
+storageLevel, str,
+subset, take,
+union, unpersist,
+withColumn, with,
+write.df, write.jdbc,
+write.json, write.orc,
+write.parquet, write.text
+
+
+
+Examples
+
+## Not run: 
+##D sparkR.session()
+##D path - path/to/file.json
+##D df - read.json(path)
+##D limitedDF - limit(df, 10)
+## End(Not run)
+
+
+
+[Package SparkR version 2.1.2 
Index]
+

http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/R/lit.html
--
diff --git a/site/docs/2.1.2/api/R/lit.html b/site/docs/2.1.2/api/R/lit.html
new file mode 100644
index 000..e227436
--- /dev/null
+++ b/site/docs/2.1.2/api/R/lit.html
@@ -0,0 +1,72 @@
+http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd;>http://www.w3.org/1999/xhtml;>R: lit
+
+
+
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css;>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js";>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js";>
+hljs.initHighlightingOnLoad();
+
+
+lit 
{SparkR}R Documentation
+
+lit
+
+Description
+
+A new Column is created to represent the literal 
value.
+If the parameter is a Column, it is returned 
unchanged.
+
+
+
+Usage
+
+
+lit(x)
+
+## S4 method for signature 'ANY'
+lit(x)
+
+
+
+Arguments
+
+
+x
+
+a literal value or a Column.
+
+
+
+
+Note
+
+lit since 1.5.0
+
+
+
+See Also
+
+Other normal_funcs: abs,
+bitwiseNOT, coalesce,
+column, expr,
+greatest, ifelse,
+isnan, least,
+nanvl, negate,
+randn, rand,
+struct, when
+
+
+
+Examples
+
+## Not run: 
+##D lit(df$name)
+##D select(df, lit(x))
+##D select(df, lit(2015-01-01))
+## End(Not run)
+
+
+
+[Package SparkR version 2.1.2 
Index]
+

http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/R/locate.html
--
diff --git a/site/docs/2.1.2/api/R/locate.html 
b/site/docs/2.1.2/api/R/locate.html
new file mode 100644
index 000..8240ffd
--- /dev/null
+++ b/site/docs/2.1.2/api/R/locate.html
@@ -0,0 +1,92 @@
+http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd;>http://www.w3.org/1999/xhtml;>R: locate
+
+
+
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css;>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js";>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js";>
+hljs.initHighlightingOnLoad();
+
+
+locate 
{SparkR}R Documentation
+
+locate
+
+Description
+
+Locate the position of the first occurrence of substr.
+
+
+
+Usage
+
+
+locate(substr, str, ...)
+
+## S4 method for signature 'character,Column'
+locate(substr, str, pos = 1)
+
+
+
+Arguments
+
+
+substr
+
+a character string to be matched.
+
+str
+
+a Column where matches are sought for each entry.
+
+...
+
+further arguments to be passed to or from other methods.
+
+pos
+
+start position of search.
+
+
+
+
+Details
+
+Note: The position is not zero based, but 1 based index. Returns 0 if substr
+could not be found in str.
+
+
+
+Note
+
+locate since 1.5.0
+
+
+
+See Also
+
+Other string_funcs: ascii,
+base64, concat_ws,

[08/51] [partial] spark-website git commit: Add 2.1.2 docs

2017-10-17 Thread holden
http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/java/org/apache/spark/api/java/JavaRDDLike.html
--
diff --git 
a/site/docs/2.1.2/api/java/org/apache/spark/api/java/JavaRDDLike.html 
b/site/docs/2.1.2/api/java/org/apache/spark/api/java/JavaRDDLike.html
new file mode 100644
index 000..c249acb
--- /dev/null
+++ b/site/docs/2.1.2/api/java/org/apache/spark/api/java/JavaRDDLike.html
@@ -0,0 +1,1786 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+
+
+JavaRDDLike (Spark 2.1.2 JavaDoc)
+
+
+
+
+
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+
+
+
+Overview
+Package
+Class
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev 
Class
+Next 
Class
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+
+
+Detail:
+Field|
+Constr|
+Method
+
+
+
+
+
+
+
+
+org.apache.spark.api.java
+Interface 
JavaRDDLikeT,This extends JavaRDDLikeT,This
+
+
+
+
+
+
+All Superinterfaces:
+java.io.Serializable
+
+
+All Known Implementing Classes:
+JavaDoubleRDD, JavaHadoopRDD, JavaNewHadoopRDD, JavaPairRDD, JavaRDD
+
+
+
+public interface JavaRDDLikeT,This extends 
JavaRDDLikeT,This
+extends scala.Serializable
+Defines operations common to several Java RDD 
implementations.
+ 
+Note:
+  This trait is not intended to be implemented by user code.
+
+
+
+
+
+
+
+
+
+
+
+Method Summary
+
+Methods
+
+Modifier and Type
+Method and Description
+
+
+UU
+aggregate(UzeroValue,
+ Function2U,T,UseqOp,
+ Function2U,U,UcombOp)
+Aggregate the elements of each partition, and then the 
results for all the partitions, using
+ given combine functions and a neutral "zero value".
+
+
+
+UJavaPairRDDT,U
+cartesian(JavaRDDLikeU,?other)
+Return the Cartesian product of this RDD and another one, 
that is, the RDD of all pairs of
+ elements (a, b) where a is in this and b is in 
other.
+
+
+
+void
+checkpoint()
+Mark this RDD for checkpointing.
+
+
+
+scala.reflect.ClassTagT
+classTag()
+
+
+java.util.ListT
+collect()
+Return an array that contains all of the elements in this 
RDD.
+
+
+
+JavaFutureActionjava.util.ListT
+collectAsync()
+The asynchronous version of collect, which 
returns a future for
+ retrieving an array containing all of the elements in this RDD.
+
+
+
+java.util.ListT[]
+collectPartitions(int[]partitionIds)
+Return an array that contains all of the elements in a 
specific partition of this RDD.
+
+
+
+SparkContext
+context()
+The SparkContext that this RDD was created 
on.
+
+
+
+long
+count()
+Return the number of elements in the RDD.
+
+
+
+PartialResultBoundedDouble
+countApprox(longtimeout)
+Approximate version of count() that returns a potentially 
incomplete result
+ within a timeout, even if not all tasks have finished.
+
+
+
+PartialResultBoundedDouble
+countApprox(longtimeout,
+   doubleconfidence)
+Approximate version of count() that returns a potentially 
incomplete result
+ within a timeout, even if not all tasks have finished.
+
+
+
+long
+countApproxDistinct(doublerelativeSD)
+Return approximate number of distinct elements in the 
RDD.
+
+
+
+JavaFutureActionLong
+countAsync()
+The asynchronous version of count, which 
returns a
+ future for counting the number of elements in this RDD.
+
+
+
+java.util.MapT,Long
+countByValue()
+Return the count of each unique value in this RDD as a map 
of (value, count) pairs.
+
+
+
+PartialResultjava.util.MapT,BoundedDouble
+countByValueApprox(longtimeout)
+Approximate version of countByValue().
+
+
+
+PartialResultjava.util.MapT,BoundedDouble
+countByValueApprox(longtimeout,
+  doubleconfidence)
+Approximate version of countByValue().
+
+
+
+T
+first()
+Return the first element in this RDD.
+
+
+
+UJavaRDDU
+flatMap(FlatMapFunctionT,Uf)
+Return a new RDD by first applying a function to all 
elements of this
+  RDD, and then flattening the results.
+
+
+
+JavaDoubleRDD
+flatMapToDouble(DoubleFlatMapFunctionTf)
+Return a new RDD by first applying a function to all 
elements of this
+  RDD, and then flattening the results.
+
+
+
+K2,V2JavaPairRDDK2,V2
+flatMapToPair(PairFlatMapFunctionT,K2,V2f)
+Return a new RDD by first applying a function to all 
elements of this
+  RDD, and then flattening the results.
+
+
+
+T
+fold(TzeroValue,
+Function2T,T,Tf)
+Aggregate the elements of each partition, and then the 
results for all the partitions, using a
+ given associative function and a neutral "zero value".
+
+
+
+void
+foreach(VoidFunctionTf)
+Applies a function f to all elements of this RDD.
+
+
+

[49/51] [partial] spark-website git commit: Add 2.1.2 docs

2017-10-17 Thread holden
http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/R/00Index.html
--
diff --git a/site/docs/2.1.2/api/R/00Index.html 
b/site/docs/2.1.2/api/R/00Index.html
new file mode 100644
index 000..541a952
--- /dev/null
+++ b/site/docs/2.1.2/api/R/00Index.html
@@ -0,0 +1,1585 @@
+http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd;>
+http://www.w3.org/1999/xhtml;>
+R: R Frontend for Apache Spark
+
+
+
+ R Frontend for Apache Spark
+http://stat.ethz.ch/R-manual/R-devel/doc/html/Rlogo.svg; alt="[R logo]" />
+
+
+
+http://stat.ethz.ch/R-manual/R-devel/doc/html/packages.html;>http://stat.ethz.ch/R-manual/R-devel/doc/html/left.jpg; 
alt="[Up]" />
+http://stat.ethz.ch/R-manual/R-devel/doc/html/index.html;>http://stat.ethz.ch/R-manual/R-devel/doc/html/up.jpg; 
alt="[Top]" />
+Documentation for package SparkR version 2.1.2
+
+DESCRIPTION file.
+
+
+Help Pages
+
+
+
+A
+B
+C
+D
+E
+F
+G
+H
+I
+J
+K
+L
+M
+N
+O
+P
+Q
+R
+S
+T
+U
+V
+W
+Y
+misc
+
+
+
+-- A --
+
+
+abs
+abs
+abs-method
+abs
+acos
+acos
+acos-method
+acos
+add_months
+add_months
+add_months-method
+add_months
+AFTSurvivalRegressionModel-class
+S4 class that represents a AFTSurvivalRegressionModel
+agg
+summarize
+agg-method
+summarize
+alias
+alias
+alias-method
+alias
+ALSModel-class
+S4 class that represents an ALSModel
+approxCountDistinct
+Returns the approximate number of distinct items in a group
+approxCountDistinct-method
+Returns the approximate number of distinct items in a group
+approxQuantile
+Calculates the approximate quantiles of a numerical column of a 
SparkDataFrame
+approxQuantile-method
+Calculates the approximate quantiles of a numerical column of a 
SparkDataFrame
+arrange
+Arrange Rows by Variables
+arrange-method
+Arrange Rows by Variables
+array_contains
+array_contains
+array_contains-method
+array_contains
+as.data.frame
+Download data from a SparkDataFrame into a R data.frame
+as.data.frame-method
+Download data from a SparkDataFrame into a R data.frame
+as.DataFrame
+Create a SparkDataFrame
+as.DataFrame.default
+Create a SparkDataFrame
+asc
+A set of operations working with SparkDataFrame columns
+ascii
+ascii
+ascii-method
+ascii
+asin
+asin
+asin-method
+asin
+atan
+atan
+atan-method
+atan
+atan2
+atan2
+atan2-method
+atan2
+attach
+Attach SparkDataFrame to R search path
+attach-method
+Attach SparkDataFrame to R search path
+avg
+avg
+avg-method
+avg
+
+
+-- B --
+
+
+base64
+base64
+base64-method
+base64
+between
+between
+between-method
+between
+bin
+bin
+bin-method
+bin
+bitwiseNOT
+bitwiseNOT
+bitwiseNOT-method
+bitwiseNOT
+bround
+bround
+bround-method
+bround
+
+
+-- C --
+
+
+cache
+Cache
+cache-method
+Cache
+cacheTable
+Cache Table
+cacheTable.default
+Cache Table
+cancelJobGroup
+Cancel active jobs for the specified group
+cancelJobGroup.default
+Cancel active jobs for the specified group
+cast
+Casts the column to a different data type.
+cast-method
+Casts the column to a different data type.
+cbrt
+cbrt
+cbrt-method
+cbrt
+ceil
+Computes the ceiling of the given value
+ceil-method
+Computes the ceiling of the given value
+ceiling
+Computes the ceiling of the given value
+ceiling-method
+Computes the ceiling of the given value
+clearCache
+Clear Cache
+clearCache.default
+Clear Cache
+clearJobGroup
+Clear current job group ID and its description
+clearJobGroup.default
+Clear current job group ID and its description
+coalesce
+Coalesce
+coalesce-method
+Coalesce
+collect
+Collects all the elements of a SparkDataFrame and coerces them into an R 
data.frame.
+collect-method
+Collects all the elements of a SparkDataFrame and coerces them into an R 
data.frame.
+colnames
+Column Names of SparkDataFrame
+colnames-method
+Column Names of SparkDataFrame
+colnames-
+Column Names of SparkDataFrame
+colnames--method
+Column Names of SparkDataFrame
+coltypes
+coltypes
+coltypes-method
+coltypes
+coltypes-
+coltypes
+coltypes--method
+coltypes
+column
+S4 class that represents a SparkDataFrame column
+Column-class
+S4 class that represents a SparkDataFrame column
+column-method
+S4 class that represents a SparkDataFrame column
+columnfunctions
+A set of operations working with SparkDataFrame columns
+columns
+Column Names of SparkDataFrame
+columns-method
+Column Names of SparkDataFrame
+concat
+concat
+concat-method
+concat
+concat_ws
+concat_ws
+concat_ws-method
+concat_ws
+contains
+A set of operations working with SparkDataFrame columns
+conv
+conv
+conv-method
+conv
+corr
+corr
+corr-method
+corr
+cos
+cos
+cos-method
+cos
+cosh
+cosh
+cosh-method
+cosh
+count
+Count
+count-method
+Count
+count-method
+Returns the number of rows in a SparkDataFrame
+countDistinct
+Count Distinct Values
+countDistinct-method
+Count Distinct Values
+cov
+cov
+cov-method
+cov
+covar_pop
+covar_pop
+covar_pop-method
+covar_pop
+covar_samp
+cov
+covar_samp-method
+cov
+crc32
+crc32
+crc32-method
+crc32
+createDataFrame
+Create a 

[32/51] [partial] spark-website git commit: Add 2.1.2 docs

2017-10-17 Thread holden
http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/java/constant-values.html
--
diff --git a/site/docs/2.1.2/api/java/constant-values.html 
b/site/docs/2.1.2/api/java/constant-values.html
new file mode 100644
index 000..2782c51
--- /dev/null
+++ b/site/docs/2.1.2/api/java/constant-values.html
@@ -0,0 +1,237 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+
+
+Constant Field Values (Spark 2.1.2 JavaDoc)
+
+
+
+
+
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+
+
+
+Overview
+Package
+Class
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev
+Next
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+
+
+
+Constant Field Values
+Contents
+
+org.apache.*
+
+
+
+
+
+org.apache.*
+
+
+
+org.apache.spark.launcher.SparkLauncher
+
+Modifier and Type
+Constant Field
+Value
+
+
+
+
+
+publicstaticfinalString
+CHILD_CONNECTION_TIMEOUT
+"spark.launcher.childConectionTimeout"
+
+
+
+
+publicstaticfinalString
+CHILD_PROCESS_LOGGER_NAME
+"spark.launcher.childProcLoggerName"
+
+
+
+
+publicstaticfinalString
+DEPLOY_MODE
+"spark.submit.deployMode"
+
+
+
+
+publicstaticfinalString
+DRIVER_EXTRA_CLASSPATH
+"spark.driver.extraClassPath"
+
+
+
+
+publicstaticfinalString
+DRIVER_EXTRA_JAVA_OPTIONS
+"spark.driver.extraJavaOptions"
+
+
+
+
+publicstaticfinalString
+DRIVER_EXTRA_LIBRARY_PATH
+"spark.driver.extraLibraryPath"
+
+
+
+
+publicstaticfinalString
+DRIVER_MEMORY
+"spark.driver.memory"
+
+
+
+
+publicstaticfinalString
+EXECUTOR_CORES
+"spark.executor.cores"
+
+
+
+
+publicstaticfinalString
+EXECUTOR_EXTRA_CLASSPATH
+"spark.executor.extraClassPath"
+
+
+
+
+publicstaticfinalString
+EXECUTOR_EXTRA_JAVA_OPTIONS
+"spark.executor.extraJavaOptions"
+
+
+
+
+publicstaticfinalString
+EXECUTOR_EXTRA_LIBRARY_PATH
+"spark.executor.extraLibraryPath"
+
+
+
+
+publicstaticfinalString
+EXECUTOR_MEMORY
+"spark.executor.memory"
+
+
+
+
+publicstaticfinalString
+NO_RESOURCE
+"spark-internal"
+
+
+
+
+publicstaticfinalString
+SPARK_MASTER
+"spark.master"
+
+
+
+
+
+
+
+
+
+
+
+
+
+Overview
+Package
+Class
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev
+Next
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+
+
+
+

http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/java/deprecated-list.html
--
diff --git a/site/docs/2.1.2/api/java/deprecated-list.html 
b/site/docs/2.1.2/api/java/deprecated-list.html
new file mode 100644
index 000..51c3c20
--- /dev/null
+++ b/site/docs/2.1.2/api/java/deprecated-list.html
@@ -0,0 +1,611 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+
+
+Deprecated List (Spark 2.1.2 JavaDoc)
+
+
+
+
+
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+
+
+
+Overview
+Package
+Class
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev
+Next
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+
+
+
+Deprecated API
+Contents
+
+Deprecated Interfaces
+Deprecated Classes
+Deprecated Methods
+Deprecated Constructors
+
+
+
+
+
+
+
+
+Deprecated Interfaces
+
+Interface and Description
+
+
+
+org.apache.spark.AccumulableParam
+use AccumulatorV2. Since 2.0.0.
+
+
+
+org.apache.spark.AccumulatorParam
+use AccumulatorV2. Since 2.0.0.
+
+
+
+
+
+
+
+
+
+
+
+
+Deprecated Classes
+
+Class and Description
+
+
+
+org.apache.spark.Accumulable
+use AccumulatorV2. Since 2.0.0.
+
+
+
+org.apache.spark.Accumulator
+use AccumulatorV2. Since 2.0.0.
+
+
+
+org.apache.spark.AccumulatorParam.DoubleAccumulatorParam$
+use AccumulatorV2. Since 2.0.0.
+
+
+
+org.apache.spark.AccumulatorParam.FloatAccumulatorParam$
+use AccumulatorV2. Since 2.0.0.
+
+
+
+org.apache.spark.AccumulatorParam.IntAccumulatorParam$
+use AccumulatorV2. Since 2.0.0.
+
+
+
+org.apache.spark.AccumulatorParam.LongAccumulatorParam$
+use AccumulatorV2. Since 2.0.0.
+
+
+
+org.apache.spark.AccumulatorParam.StringAccumulatorParam$
+use 

[07/51] [partial] spark-website git commit: Add 2.1.2 docs

2017-10-17 Thread holden
http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/java/org/apache/spark/api/java/JavaSparkContext.html
--
diff --git 
a/site/docs/2.1.2/api/java/org/apache/spark/api/java/JavaSparkContext.html 
b/site/docs/2.1.2/api/java/org/apache/spark/api/java/JavaSparkContext.html
new file mode 100644
index 000..6927b66
--- /dev/null
+++ b/site/docs/2.1.2/api/java/org/apache/spark/api/java/JavaSparkContext.html
@@ -0,0 +1,2088 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+
+
+JavaSparkContext (Spark 2.1.2 JavaDoc)
+
+
+
+
+
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+
+
+
+Overview
+Package
+Class
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev 
Class
+Next 
Class
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+
+
+Detail:
+Field|
+Constr|
+Method
+
+
+
+
+
+
+
+
+org.apache.spark.api.java
+Class JavaSparkContext
+
+
+
+Object
+
+
+org.apache.spark.api.java.JavaSparkContext
+
+
+
+
+
+
+
+All Implemented Interfaces:
+java.io.Closeable, AutoCloseable
+
+
+
+public class JavaSparkContext
+extends Object
+implements java.io.Closeable
+A Java-friendly version of SparkContext that returns
+ JavaRDDs and works with Java 
collections instead of Scala ones.
+ 
+ Only one SparkContext may be active per JVM.  You must stop() 
the active SparkContext before
+ creating a new one.  This limitation may eventually be removed; see 
SPARK-2243 for more details.
+
+
+
+
+
+
+
+
+
+
+
+Constructor Summary
+
+Constructors
+
+Constructor and Description
+
+
+JavaSparkContext()
+Create a JavaSparkContext that loads settings from system 
properties (for instance, when
+ launching with ./bin/spark-submit).
+
+
+
+JavaSparkContext(SparkConfconf)
+
+
+JavaSparkContext(SparkContextsc)
+
+
+JavaSparkContext(Stringmaster,
+StringappName)
+
+
+JavaSparkContext(Stringmaster,
+StringappName,
+SparkConfconf)
+
+
+JavaSparkContext(Stringmaster,
+StringappName,
+StringsparkHome,
+StringjarFile)
+
+
+JavaSparkContext(Stringmaster,
+StringappName,
+StringsparkHome,
+String[]jars)
+
+
+JavaSparkContext(Stringmaster,
+StringappName,
+StringsparkHome,
+String[]jars,
+
java.util.MapString,Stringenvironment)
+
+
+
+
+
+
+
+
+
+Method Summary
+
+Methods
+
+Modifier and Type
+Method and Description
+
+
+T,RAccumulableT,R
+accumulable(TinitialValue,
+   AccumulableParamT,Rparam)
+Deprecated.
+use AccumulatorV2. Since 2.0.0.
+
+
+
+
+T,RAccumulableT,R
+accumulable(TinitialValue,
+   Stringname,
+   AccumulableParamT,Rparam)
+Deprecated.
+use AccumulatorV2. Since 2.0.0.
+
+
+
+
+AccumulatorDouble
+accumulator(doubleinitialValue)
+Deprecated.
+use sc().doubleAccumulator(). Since 2.0.0.
+
+
+
+
+AccumulatorDouble
+accumulator(doubleinitialValue,
+   Stringname)
+Deprecated.
+use sc().doubleAccumulator(String). Since 
2.0.0.
+
+
+
+
+AccumulatorInteger
+accumulator(intinitialValue)
+Deprecated.
+use sc().longAccumulator(). Since 2.0.0.
+
+
+
+
+AccumulatorInteger
+accumulator(intinitialValue,
+   Stringname)
+Deprecated.
+use sc().longAccumulator(String). Since 2.0.0.
+
+
+
+
+TAccumulatorT
+accumulator(TinitialValue,
+   AccumulatorParamTaccumulatorParam)
+Deprecated.
+use AccumulatorV2. Since 2.0.0.
+
+
+
+
+TAccumulatorT
+accumulator(TinitialValue,
+   Stringname,
+   AccumulatorParamTaccumulatorParam)
+Deprecated.
+use AccumulatorV2. Since 2.0.0.
+
+
+
+
+void
+addFile(Stringpath)
+Add a file to be downloaded with this Spark job on every 
node.
+
+
+
+void
+addFile(Stringpath,
+   booleanrecursive)
+Add a file to be downloaded with this Spark job on every 
node.
+
+
+
+void
+addJar(Stringpath)
+Adds a JAR dependency for all tasks to be executed on this 
SparkContext in the future.
+
+
+
+String
+appName()
+
+
+JavaPairRDDString,PortableDataStream
+binaryFiles(Stringpath)
+Read a directory of binary files from HDFS, a local file 
system (available on all nodes),
+ or any Hadoop-supported file system URI as a byte array.
+
+
+
+JavaPairRDDString,PortableDataStream
+binaryFiles(Stringpath,
+   intminPartitions)
+Read a directory of binary files from HDFS, a local file 
system (available on all nodes),
+ or any Hadoop-supported file system URI as a byte array.
+
+
+
+JavaRDDbyte[]
+binaryRecords(Stringpath,
+  

[20/51] [partial] spark-website git commit: Add 2.1.2 docs

2017-10-17 Thread holden
http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/java/org/apache/spark/SparkConf.html
--
diff --git a/site/docs/2.1.2/api/java/org/apache/spark/SparkConf.html 
b/site/docs/2.1.2/api/java/org/apache/spark/SparkConf.html
new file mode 100644
index 000..4c33803
--- /dev/null
+++ b/site/docs/2.1.2/api/java/org/apache/spark/SparkConf.html
@@ -0,0 +1,1147 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+
+
+SparkConf (Spark 2.1.2 JavaDoc)
+
+
+
+
+
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+
+
+
+Overview
+Package
+Class
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev Class
+Next Class
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+
+
+Detail:
+Field|
+Constr|
+Method
+
+
+
+
+
+
+
+
+org.apache.spark
+Class SparkConf
+
+
+
+Object
+
+
+org.apache.spark.SparkConf
+
+
+
+
+
+
+
+All Implemented Interfaces:
+java.io.Serializable, Cloneable
+
+
+
+public class SparkConf
+extends Object
+implements scala.Cloneable, scala.Serializable
+Configuration for a Spark application. Used to set various 
Spark parameters as key-value pairs.
+ 
+ Most of the time, you would create a SparkConf object with new 
SparkConf(), which will load
+ values from any spark.* Java system properties set in your 
application as well. In this case,
+ parameters you set directly on the SparkConf object take 
priority over system properties.
+ 
+ For unit tests, you can also call new SparkConf(false) to skip 
loading external settings and
+ get the same configuration no matter what the system properties are.
+ 
+ All setter methods in this class support chaining. For example, you can write
+ new SparkConf().setMaster("local").setAppName("My app").
+ 
+ param:  loadDefaults whether to also load values from Java system properties
+ 
+See Also:Serialized 
FormNote:
+  Once a SparkConf object is passed to Spark, it is cloned and can no 
longer be modified
+ by the user. Spark does not support modifying the configuration at 
runtime.
+
+
+
+
+
+
+
+
+
+
+
+Constructor Summary
+
+Constructors
+
+Constructor and Description
+
+
+SparkConf()
+Create a SparkConf that loads defaults from system 
properties and the classpath
+
+
+
+SparkConf(booleanloadDefaults)
+
+
+
+
+
+
+
+
+
+Method Summary
+
+Methods
+
+Modifier and Type
+Method and Description
+
+
+SparkConf
+clone()
+Copy this object
+
+
+
+boolean
+contains(Stringkey)
+Does the configuration contain a given parameter?
+
+
+
+String
+get(Stringkey)
+Get a parameter; throws a NoSuchElementException if it's 
not set
+
+
+
+String
+get(Stringkey,
+   StringdefaultValue)
+Get a parameter, falling back to a default if not set
+
+
+
+scala.Tuple2String,String[]
+getAll()
+Get all parameters as a list of pairs
+
+
+
+scala.Tuple2String,String[]
+getAllWithPrefix(Stringprefix)
+Get all parameters that start with prefix
+
+
+
+String
+getAppId()
+Returns the Spark application id, valid in the Driver after 
TaskScheduler registration and
+ from the start in the Executor.
+
+
+
+scala.collection.immutable.MapObject,String
+getAvroSchema()
+Gets all the avro schemas in the configuration used in the 
generic Avro record serializer
+
+
+
+boolean
+getBoolean(Stringkey,
+  booleandefaultValue)
+Get a parameter as a boolean, falling back to a default if 
not set
+
+
+
+static scala.OptionString
+getDeprecatedConfig(Stringkey,
+   SparkConfconf)
+Looks for available deprecated keys for the given config 
option, and return the first
+ value available.
+
+
+
+double
+getDouble(Stringkey,
+ doubledefaultValue)
+Get a parameter as a double, falling back to a default if 
not set
+
+
+
+scala.collection.Seqscala.Tuple2String,String
+getExecutorEnv()
+Get all executor environment variables set on this 
SparkConf
+
+
+
+int
+getInt(Stringkey,
+  intdefaultValue)
+Get a parameter as an integer, falling back to a default if 
not set
+
+
+
+long
+getLong(Stringkey,
+   longdefaultValue)
+Get a parameter as a long, falling back to a default if not 
set
+
+
+
+scala.OptionString
+getOption(Stringkey)
+Get a parameter as an Option
+
+
+
+long
+getSizeAsBytes(Stringkey)
+Get a size parameter as bytes; throws a 
NoSuchElementException if it's not set.
+
+
+
+long
+getSizeAsBytes(Stringkey,
+  longdefaultValue)
+Get a size parameter as bytes, falling back to a default if 
not set.
+
+
+
+long
+getSizeAsBytes(Stringkey,
+  StringdefaultValue)
+Get a size parameter as bytes, falling back to a 

[09/51] [partial] spark-website git commit: Add 2.1.2 docs

2017-10-17 Thread holden
http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/java/org/apache/spark/api/java/JavaRDD.html
--
diff --git a/site/docs/2.1.2/api/java/org/apache/spark/api/java/JavaRDD.html 
b/site/docs/2.1.2/api/java/org/apache/spark/api/java/JavaRDD.html
new file mode 100644
index 000..a7004a5
--- /dev/null
+++ b/site/docs/2.1.2/api/java/org/apache/spark/api/java/JavaRDD.html
@@ -0,0 +1,1854 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+
+
+JavaRDD (Spark 2.1.2 JavaDoc)
+
+
+
+
+
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+
+
+
+Overview
+Package
+Class
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev 
Class
+Next 
Class
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+
+
+Detail:
+Field|
+Constr|
+Method
+
+
+
+
+
+
+
+
+org.apache.spark.api.java
+Class JavaRDDT
+
+
+
+Object
+
+
+org.apache.spark.api.java.JavaRDDT
+
+
+
+
+
+
+
+All Implemented Interfaces:
+java.io.Serializable, JavaRDDLikeT,JavaRDDT
+
+
+
+public class JavaRDDT
+extends Object
+See Also:Serialized
 Form
+
+
+
+
+
+
+
+
+
+
+
+Constructor Summary
+
+Constructors
+
+Constructor and Description
+
+
+JavaRDD(RDDTrdd,
+   scala.reflect.ClassTagTclassTag)
+
+
+
+
+
+
+
+
+
+Method Summary
+
+Methods
+
+Modifier and Type
+Method and Description
+
+
+static UU
+aggregate(UzeroValue,
+ Function2U,T,UseqOp,
+ Function2U,U,UcombOp)
+
+
+JavaRDDT
+cache()
+Persist this RDD with the default storage level 
(MEMORY_ONLY).
+
+
+
+static UJavaPairRDDT,U
+cartesian(JavaRDDLikeU,?other)
+
+
+static void
+checkpoint()
+
+
+scala.reflect.ClassTagT
+classTag()
+
+
+JavaRDDT
+coalesce(intnumPartitions)
+Return a new RDD that is reduced into 
numPartitions partitions.
+
+
+
+JavaRDDT
+coalesce(intnumPartitions,
+booleanshuffle)
+Return a new RDD that is reduced into 
numPartitions partitions.
+
+
+
+static java.util.ListT
+collect()
+
+
+static JavaFutureActionjava.util.ListT
+collectAsync()
+
+
+static java.util.ListT[]
+collectPartitions(int[]partitionIds)
+
+
+static SparkContext
+context()
+
+
+static long
+count()
+
+
+static PartialResultBoundedDouble
+countApprox(longtimeout)
+
+
+static PartialResultBoundedDouble
+countApprox(longtimeout,
+   doubleconfidence)
+
+
+static long
+countApproxDistinct(doublerelativeSD)
+
+
+static JavaFutureActionLong
+countAsync()
+
+
+static java.util.MapT,Long
+countByValue()
+
+
+static PartialResultjava.util.MapT,BoundedDouble
+countByValueApprox(longtimeout)
+
+
+static PartialResultjava.util.MapT,BoundedDouble
+countByValueApprox(longtimeout,
+  doubleconfidence)
+
+
+JavaRDDT
+distinct()
+Return a new RDD containing the distinct elements in this 
RDD.
+
+
+
+JavaRDDT
+distinct(intnumPartitions)
+Return a new RDD containing the distinct elements in this 
RDD.
+
+
+
+JavaRDDT
+filter(FunctionT,Booleanf)
+Return a new RDD containing only the elements that satisfy 
a predicate.
+
+
+
+static T
+first()
+
+
+static UJavaRDDU
+flatMap(FlatMapFunctionT,Uf)
+
+
+static JavaDoubleRDD
+flatMapToDouble(DoubleFlatMapFunctionTf)
+
+
+static K2,V2JavaPairRDDK2,V2
+flatMapToPair(PairFlatMapFunctionT,K2,V2f)
+
+
+static T
+fold(TzeroValue,
+Function2T,T,Tf)
+
+
+static void
+foreach(VoidFunctionTf)
+
+
+static JavaFutureActionVoid
+foreachAsync(VoidFunctionTf)
+
+
+static void
+foreachPartition(VoidFunctionjava.util.IteratorTf)
+
+
+static JavaFutureActionVoid
+foreachPartitionAsync(VoidFunctionjava.util.IteratorTf)
+
+
+static TJavaRDDT
+fromRDD(RDDTrdd,
+   scala.reflect.ClassTagTevidence$1)
+
+
+static OptionalString
+getCheckpointFile()
+
+
+static int
+getNumPartitions()
+
+
+static StorageLevel
+getStorageLevel()
+
+
+static JavaRDDjava.util.ListT
+glom()
+
+
+static UJavaPairRDDU,IterableT
+groupBy(FunctionT,Uf)
+
+
+static UJavaPairRDDU,IterableT
+groupBy(FunctionT,Uf,
+   intnumPartitions)
+
+
+static int
+id()
+
+
+JavaRDDT
+intersection(JavaRDDTother)
+Return the intersection of this RDD and another one.
+
+
+
+static boolean
+isCheckpointed()
+
+
+static boolean
+isEmpty()
+
+
+static java.util.IteratorT
+iterator(Partitionsplit,
+TaskContexttaskContext)
+
+
+static UJavaPairRDDU,T
+keyBy(FunctionT,Uf)
+
+
+static RJavaRDDR
+map(FunctionT,Rf)
+
+
+static UJavaRDDU
+mapPartitions(FlatMapFunctionjava.util.IteratorT,Uf)
+
+
+static UJavaRDDU
+mapPartitions(FlatMapFunctionjava.util.IteratorT,Uf,
+ booleanpreservesPartitioning)
+
+
+static JavaDoubleRDD

[06/51] [partial] spark-website git commit: Add 2.1.2 docs

2017-10-17 Thread holden
http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/java/org/apache/spark/api/java/JavaSparkStatusTracker.html
--
diff --git 
a/site/docs/2.1.2/api/java/org/apache/spark/api/java/JavaSparkStatusTracker.html
 
b/site/docs/2.1.2/api/java/org/apache/spark/api/java/JavaSparkStatusTracker.html
new file mode 100644
index 000..a54f037
--- /dev/null
+++ 
b/site/docs/2.1.2/api/java/org/apache/spark/api/java/JavaSparkStatusTracker.html
@@ -0,0 +1,323 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+
+
+JavaSparkStatusTracker (Spark 2.1.2 JavaDoc)
+
+
+
+
+
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+
+
+
+Overview
+Package
+Class
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev 
Class
+Next 
Class
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+
+
+Detail:
+Field|
+Constr|
+Method
+
+
+
+
+
+
+
+
+org.apache.spark.api.java
+Class 
JavaSparkStatusTracker
+
+
+
+Object
+
+
+org.apache.spark.api.java.JavaSparkStatusTracker
+
+
+
+
+
+
+
+
+public class JavaSparkStatusTracker
+extends Object
+Low-level status reporting APIs for monitoring job and 
stage progress.
+ 
+ These APIs intentionally provide very weak consistency semantics; consumers 
of these APIs should
+ be prepared to handle empty / missing information.  For example, a job's 
stage ids may be known
+ but the status API may not have any information about the details of those 
stages, so
+ getStageInfo could potentially return null for a 
valid stage id.
+ 
+ To limit memory usage, these APIs only provide information on recent jobs / 
stages.  These APIs
+ will provide information for the last spark.ui.retainedStages 
stages and
+ spark.ui.retainedJobs jobs.
+ 
+Note:
+  This class's constructor should be considered private and may be subject 
to change.
+
+
+
+
+
+
+
+
+
+
+
+Method Summary
+
+Methods
+
+Modifier and Type
+Method and Description
+
+
+int[]
+getActiveJobIds()
+Returns an array containing the ids of all active 
jobs.
+
+
+
+int[]
+getActiveStageIds()
+Returns an array containing the ids of all active 
stages.
+
+
+
+int[]
+getJobIdsForGroup(StringjobGroup)
+Return a list of all known jobs in a particular job 
group.
+
+
+
+SparkJobInfo
+getJobInfo(intjobId)
+Returns job information, or null if the job 
info could not be found or was garbage collected.
+
+
+
+SparkStageInfo
+getStageInfo(intstageId)
+Returns stage information, or null if the 
stage info could not be found or was
+ garbage collected.
+
+
+
+
+
+
+
+Methods inherited from classObject
+equals, getClass, hashCode, notify, notifyAll, toString, wait, wait, 
wait
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+Method Detail
+
+
+
+
+
+getJobIdsForGroup
+publicint[]getJobIdsForGroup(StringjobGroup)
+Return a list of all known jobs in a particular job group.  
If jobGroup is null, then
+ returns all known jobs that are not associated with a job group.
+ 
+ The returned list may contain running, failed, and completed jobs, and may 
vary across
+ invocations of this method.  This method does not guarantee the order of the 
elements in
+ its result.
+Parameters:jobGroup 
- (undocumented)
+Returns:(undocumented)
+
+
+
+
+
+
+
+getActiveStageIds
+publicint[]getActiveStageIds()
+Returns an array containing the ids of all active stages.
+ 
+ This method does not guarantee the order of the elements in its result.
+Returns:(undocumented)
+
+
+
+
+
+
+
+getActiveJobIds
+publicint[]getActiveJobIds()
+Returns an array containing the ids of all active jobs.
+ 
+ This method does not guarantee the order of the elements in its result.
+Returns:(undocumented)
+
+
+
+
+
+
+
+getJobInfo
+publicSparkJobInfogetJobInfo(intjobId)
+Returns job information, or null if the job 
info could not be found or was garbage collected.
+Parameters:jobId - 
(undocumented)
+Returns:(undocumented)
+
+
+
+
+
+
+
+getStageInfo
+publicSparkStageInfogetStageInfo(intstageId)
+Returns stage information, or null if the 
stage info could not be found or was
+ garbage collected.
+Parameters:stageId - 
(undocumented)
+Returns:(undocumented)
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+Overview
+Package
+Class
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev 
Class
+Next 
Class
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+

[37/51] [partial] spark-website git commit: Add 2.1.2 docs

2017-10-17 Thread holden
http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/R/sparkR.conf.html
--
diff --git a/site/docs/2.1.2/api/R/sparkR.conf.html 
b/site/docs/2.1.2/api/R/sparkR.conf.html
new file mode 100644
index 000..322fbfd
--- /dev/null
+++ b/site/docs/2.1.2/api/R/sparkR.conf.html
@@ -0,0 +1,68 @@
+http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd;>http://www.w3.org/1999/xhtml;>R: Get Runtime Config from 
the current active SparkSession
+
+
+
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css;>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js";>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js";>
+hljs.initHighlightingOnLoad();
+
+
+sparkR.conf {SparkR}R 
Documentation
+
+Get Runtime Config from the current active SparkSession
+
+Description
+
+Get Runtime Config from the current active SparkSession.
+To change SparkSession Runtime Config, please see 
sparkR.session().
+
+
+
+Usage
+
+
+sparkR.conf(key, defaultValue)
+
+
+
+Arguments
+
+
+key
+
+(optional) The key of the config to get, if omitted, all config is 
returned
+
+defaultValue
+
+(optional) The default value of the config to return if they config is not
+set, if omitted, the call fails if the config key is not set
+
+
+
+
+Value
+
+a list of config values with keys as their names
+
+
+
+Note
+
+sparkR.conf since 2.0.0
+
+
+
+Examples
+
+## Not run: 
+##D sparkR.session()
+##D allConfigs - sparkR.conf()
+##D masterValue - unlist(sparkR.conf(spark.master))
+##D namedConfig - sparkR.conf(spark.executor.memory, 
0g)
+## End(Not run)
+
+
+
+[Package SparkR version 2.1.2 
Index]
+

http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/R/sparkR.init-deprecated.html
--
diff --git a/site/docs/2.1.2/api/R/sparkR.init-deprecated.html 
b/site/docs/2.1.2/api/R/sparkR.init-deprecated.html
new file mode 100644
index 000..d86da83
--- /dev/null
+++ b/site/docs/2.1.2/api/R/sparkR.init-deprecated.html
@@ -0,0 +1,92 @@
+http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd;>http://www.w3.org/1999/xhtml;>R: (Deprecated) Initialize a 
new Spark Context
+
+
+
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css;>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js";>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js";>
+hljs.initHighlightingOnLoad();
+
+
+sparkR.init {SparkR}R 
Documentation
+
+(Deprecated) Initialize a new Spark Context
+
+Description
+
+This function initializes a new SparkContext.
+
+
+
+Usage
+
+
+sparkR.init(master = "", appName = "SparkR",
+  sparkHome = Sys.getenv("SPARK_HOME"), sparkEnvir = list(),
+  sparkExecutorEnv = list(), sparkJars = "", sparkPackages = "")
+
+
+
+Arguments
+
+
+master
+
+The Spark master URL
+
+appName
+
+Application name to register with cluster manager
+
+sparkHome
+
+Spark Home directory
+
+sparkEnvir
+
+Named list of environment variables to set on worker nodes
+
+sparkExecutorEnv
+
+Named list of environment variables to be used when launching executors
+
+sparkJars
+
+Character vector of jar files to pass to the worker nodes
+
+sparkPackages
+
+Character vector of package coordinates
+
+
+
+
+Note
+
+sparkR.init since 1.4.0
+
+
+
+See Also
+
+sparkR.session
+
+
+
+Examples
+
+## Not run: 
+##D sc - sparkR.init(local[2], SparkR, 
/home/spark)
+##D sc - sparkR.init(local[2], SparkR, 
/home/spark,
+##D  list(spark.executor.memory=1g))
+##D sc - sparkR.init(yarn-client, SparkR, 
/home/spark,
+##D  list(spark.executor.memory=4g),
+##D  list(LD_LIBRARY_PATH=/directory of JVM libraries 
(libjvm.so) on workers/),
+##D  c(one.jar, two.jar, 
three.jar),
+##D  c(com.databricks:spark-avro_2.10:2.0.1))
+## End(Not run)
+
+
+
+[Package SparkR version 2.1.2 
Index]
+

http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/R/sparkR.newJObject.html
--
diff --git a/site/docs/2.1.2/api/R/sparkR.newJObject.html 
b/site/docs/2.1.2/api/R/sparkR.newJObject.html
new file mode 100644
index 000..8ecda40
--- /dev/null
+++ b/site/docs/2.1.2/api/R/sparkR.newJObject.html
@@ -0,0 +1,86 @@
+http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd;>http://www.w3.org/1999/xhtml;>R: Create Java Objects
+
+
+
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css;>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js";>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js";>
+hljs.initHighlightingOnLoad();
+
+
+sparkR.newJObject {SparkR}R Documentation
+
+Create Java Objects
+
+Description
+
+Create a new Java object in 

[30/51] [partial] spark-website git commit: Add 2.1.2 docs

2017-10-17 Thread holden
http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/java/index.html
--
diff --git a/site/docs/2.1.2/api/java/index.html 
b/site/docs/2.1.2/api/java/index.html
new file mode 100644
index 000..08d3ceb
--- /dev/null
+++ b/site/docs/2.1.2/api/java/index.html
@@ -0,0 +1,75 @@
+http://www.w3.org/TR/html4/frameset.dtd;>
+
+
+
+
+Spark 2.1.2 JavaDoc
+
+tmpTargetPage = "" + window.location.search;
+if (tmpTargetPage != "" && tmpTargetPage != "undefined")
+tmpTargetPage = tmpTargetPage.substring(1);
+if (tmpTargetPage.indexOf(":") != -1 || (tmpTargetPage != "" && 
!validURL(tmpTargetPage)))
+tmpTargetPage = "undefined";
+targetPage = tmpTargetPage;
+function validURL(url) {
+try {
+url = decodeURIComponent(url);
+}
+catch (error) {
+return false;
+}
+var pos = url.indexOf(".html");
+if (pos == -1 || pos != url.length - 5)
+return false;
+var allowNumber = false;
+var allowSep = false;
+var seenDot = false;
+for (var i = 0; i < url.length - 5; i++) {
+var ch = url.charAt(i);
+if ('a' <= ch && ch <= 'z' ||
+'A' <= ch && ch <= 'Z' ||
+ch == '$' ||
+ch == '_' ||
+ch.charCodeAt(0) > 127) {
+allowNumber = true;
+allowSep = true;
+} else if ('0' <= ch && ch <= '9'
+|| ch == '-') {
+if (!allowNumber)
+ return false;
+} else if (ch == '/' || ch == '.') {
+if (!allowSep)
+return false;
+allowNumber = false;
+allowSep = false;
+if (ch == '.')
+ seenDot = true;
+if (ch == '/' && seenDot)
+ return false;
+} else {
+return false;
+}
+}
+return true;
+}
+function loadFrames() {
+if (targetPage != "" && targetPage != "undefined")
+ top.classFrame.location = top.targetPage;
+}
+
+
+
+
+
+
+
+
+
+
+JavaScript is disabled on your browser.
+
+Frame Alert
+This document is designed to be viewed using the frames feature. If you see 
this message, you are using a non-frame-capable web client. Link to Non-frame version.
+
+
+

http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/java/lib/api-javadocs.js
--
diff --git a/site/docs/2.1.2/api/java/lib/api-javadocs.js 
b/site/docs/2.1.2/api/java/lib/api-javadocs.js
new file mode 100644
index 000..ead13d6
--- /dev/null
+++ b/site/docs/2.1.2/api/java/lib/api-javadocs.js
@@ -0,0 +1,60 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one or more
+ * contributor license agreements.  See the NOTICE file distributed with
+ * this work for additional information regarding copyright ownership.
+ * The ASF licenses this file to You under the Apache License, Version 2.0
+ * (the "License"); you may not use this file except in compliance with
+ * the License.  You may obtain a copy of the License at
+ *
+ *http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+/* Dynamically injected post-processing code for the API docs */
+
+$(document).ready(function() {
+  addBadges(":: AlphaComponent ::", 'Alpha 
Component');
+  addBadges(":: DeveloperApi ::", 'Developer 
API');
+  addBadges(":: Experimental ::", 'Experimental');
+});
+
+function addBadges(tag, html) {
+  var tags = $(".block:contains(" + tag + ")")
+
+  // Remove identifier tags
+  tags.each(function(index) {
+var oldHTML = $(this).html();
+var newHTML = oldHTML.replace(tag, "");
+$(this).html(newHTML);
+  });
+
+  // Add html badge tags
+  tags.each(function(index) {
+if ($(this).parent().is('td.colLast')) {
+  $(this).parent().prepend(html);
+} else if ($(this).parent('li.blockList')
+  .parent('ul.blockList')
+  .parent('div.description')
+  .parent().is('div.contentContainer')) {
+  var contentContainer = $(this).parent('li.blockList')
+.parent('ul.blockList')
+.parent('div.description')
+.parent('div.contentContainer')
+  var header = contentContainer.prev('div.header');
+  if (header.length > 0) {
+

[17/51] [partial] spark-website git commit: Add 2.1.2 docs

2017-10-17 Thread holden
http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/java/org/apache/spark/SparkJobInfo.html
--
diff --git a/site/docs/2.1.2/api/java/org/apache/spark/SparkJobInfo.html 
b/site/docs/2.1.2/api/java/org/apache/spark/SparkJobInfo.html
new file mode 100644
index 000..a311bae
--- /dev/null
+++ b/site/docs/2.1.2/api/java/org/apache/spark/SparkJobInfo.html
@@ -0,0 +1,247 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+
+
+SparkJobInfo (Spark 2.1.2 JavaDoc)
+
+
+
+
+
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+
+
+
+Overview
+Package
+Class
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev 
Class
+Next Class
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+
+
+Detail:
+Field|
+Constr|
+Method
+
+
+
+
+
+
+
+
+org.apache.spark
+Interface SparkJobInfo
+
+
+
+
+
+
+All Superinterfaces:
+java.io.Serializable
+
+
+All Known Implementing Classes:
+SparkJobInfoImpl
+
+
+
+public interface SparkJobInfo
+extends java.io.Serializable
+Exposes information about Spark Jobs.
+
+ This interface is not designed to be implemented outside of Spark.  We may 
add additional methods
+ which may break binary compatibility with outside implementations.
+
+
+
+
+
+
+
+
+
+
+
+Method Summary
+
+Methods
+
+Modifier and Type
+Method and Description
+
+
+int
+jobId()
+
+
+int[]
+stageIds()
+
+
+JobExecutionStatus
+status()
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+Method Detail
+
+
+
+
+
+jobId
+intjobId()
+
+
+
+
+
+
+
+stageIds
+int[]stageIds()
+
+
+
+
+
+
+
+status
+JobExecutionStatusstatus()
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+Overview
+Package
+Class
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev 
Class
+Next Class
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+
+
+Detail:
+Field|
+Constr|
+Method
+
+
+
+
+
+
+
+

http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/java/org/apache/spark/SparkJobInfoImpl.html
--
diff --git a/site/docs/2.1.2/api/java/org/apache/spark/SparkJobInfoImpl.html 
b/site/docs/2.1.2/api/java/org/apache/spark/SparkJobInfoImpl.html
new file mode 100644
index 000..1a94cdf
--- /dev/null
+++ b/site/docs/2.1.2/api/java/org/apache/spark/SparkJobInfoImpl.html
@@ -0,0 +1,306 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+
+
+SparkJobInfoImpl (Spark 2.1.2 JavaDoc)
+
+
+
+
+
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+
+
+
+Overview
+Package
+Class
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev Class
+Next Class
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+
+
+Detail:
+Field|
+Constr|
+Method
+
+
+
+
+
+
+
+
+org.apache.spark
+Class SparkJobInfoImpl
+
+
+
+Object
+
+
+org.apache.spark.SparkJobInfoImpl
+
+
+
+
+
+
+
+All Implemented Interfaces:
+java.io.Serializable, SparkJobInfo
+
+
+
+public class SparkJobInfoImpl
+extends Object
+implements SparkJobInfo
+See Also:Serialized
 Form
+
+
+
+
+
+
+
+
+
+
+
+Constructor Summary
+
+Constructors
+
+Constructor and Description
+
+
+SparkJobInfoImpl(intjobId,
+int[]stageIds,
+JobExecutionStatusstatus)
+
+
+
+
+
+
+
+
+
+Method Summary
+
+Methods
+
+Modifier and Type
+Method and Description
+
+
+int
+jobId()
+
+
+int[]
+stageIds()
+
+
+JobExecutionStatus
+status()
+
+
+
+
+
+
+Methods inherited from classObject
+equals, getClass, hashCode, notify, notifyAll, toString, wait, wait, 
wait
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+Constructor Detail
+
+
+
+
+
+SparkJobInfoImpl
+publicSparkJobInfoImpl(intjobId,
+int[]stageIds,
+JobExecutionStatusstatus)
+
+
+
+
+
+
+
+
+
+Method Detail
+
+
+
+
+
+jobId
+publicintjobId()
+
+Specified by:
+jobIdin
 interfaceSparkJobInfo
+
+
+
+
+
+
+
+
+stageIds
+publicint[]stageIds()
+
+Specified by:

[33/51] [partial] spark-website git commit: Add 2.1.2 docs

2017-10-17 Thread holden
http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/java/allclasses-noframe.html
--
diff --git a/site/docs/2.1.2/api/java/allclasses-noframe.html 
b/site/docs/2.1.2/api/java/allclasses-noframe.html
new file mode 100644
index 000..413f8db
--- /dev/null
+++ b/site/docs/2.1.2/api/java/allclasses-noframe.html
@@ -0,0 +1,1138 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+
+
+All Classes (Spark 2.1.2 JavaDoc)
+
+
+
+
+All Classes
+
+
+AbsoluteError
+Accumulable
+AccumulableInfo
+AccumulableInfo
+AccumulableParam
+Accumulator
+AccumulatorContext
+AccumulatorParam
+AccumulatorParam.DoubleAccumulatorParam$
+AccumulatorParam.FloatAccumulatorParam$
+AccumulatorParam.IntAccumulatorParam$
+AccumulatorParam.LongAccumulatorParam$
+AccumulatorParam.StringAccumulatorParam$
+AccumulatorV2
+AFTAggregator
+AFTCostFun
+AFTSurvivalRegression
+AFTSurvivalRegressionModel
+AggregatedDialect
+AggregatingEdgeContext
+Aggregator
+Aggregator
+Algo
+AllJobsCancelled
+AllReceiverIds
+ALS
+ALS
+ALS.InBlock$
+ALS.Rating
+ALS.Rating$
+ALS.RatingBlock$
+ALSModel
+AnalysisException
+And
+AnyDataType
+ApplicationAttemptInfo
+ApplicationInfo
+ApplicationsListResource
+ApplicationStatus
+ApplyInPlace
+AreaUnderCurve
+ArrayType
+AskPermissionToCommitOutput
+AssociationRules
+AssociationRules.Rule
+AsyncRDDActions
+Attribute
+AttributeGroup
+AttributeKeys
+AttributeType
+BaseRelation
+BaseRRDD
+BatchInfo
+BernoulliCellSampler
+BernoulliSampler
+Binarizer
+BinaryAttribute
+BinaryClassificationEvaluator
+BinaryClassificationMetrics
+BinaryLogisticRegressionSummary
+BinaryLogisticRegressionTrainingSummary
+BinarySample
+BinaryType
+BinomialBounds
+BisectingKMeans
+BisectingKMeans
+BisectingKMeansModel
+BisectingKMeansModel
+BisectingKMeansModel.SaveLoadV1_0$
+BisectingKMeansSummary
+BlacklistTracker
+BLAS
+BLAS
+BlockId
+BlockManagerId
+BlockManagerMessages
+BlockManagerMessages.BlockManagerHeartbeat
+BlockManagerMessages.BlockManagerHeartbeat$
+BlockManagerMessages.GetBlockStatus
+BlockManagerMessages.GetBlockStatus$
+BlockManagerMessages.GetExecutorEndpointRef
+BlockManagerMessages.GetExecutorEndpointRef$
+BlockManagerMessages.GetLocations
+BlockManagerMessages.GetLocations$
+BlockManagerMessages.GetLocationsMultipleBlockIds
+BlockManagerMessages.GetLocationsMultipleBlockIds$
+BlockManagerMessages.GetMatchingBlockIds
+BlockManagerMessages.GetMatchingBlockIds$
+BlockManagerMessages.GetMemoryStatus$
+BlockManagerMessages.GetPeers
+BlockManagerMessages.GetPeers$
+BlockManagerMessages.GetStorageStatus$
+BlockManagerMessages.HasCachedBlocks
+BlockManagerMessages.HasCachedBlocks$
+BlockManagerMessages.RegisterBlockManager
+BlockManagerMessages.RegisterBlockManager$
+BlockManagerMessages.RemoveBlock
+BlockManagerMessages.RemoveBlock$
+BlockManagerMessages.RemoveBroadcast
+BlockManagerMessages.RemoveBroadcast$
+BlockManagerMessages.RemoveExecutor
+BlockManagerMessages.RemoveExecutor$
+BlockManagerMessages.RemoveRdd
+BlockManagerMessages.RemoveRdd$
+BlockManagerMessages.RemoveShuffle
+BlockManagerMessages.RemoveShuffle$
+BlockManagerMessages.StopBlockManagerMaster$
+BlockManagerMessages.ToBlockManagerMaster
+BlockManagerMessages.ToBlockManagerSlave
+BlockManagerMessages.TriggerThreadDump$
+BlockManagerMessages.UpdateBlockInfo
+BlockManagerMessages.UpdateBlockInfo$
+BlockMatrix
+BlockNotFoundException
+BlockReplicationPolicy
+BlockStatus
+BlockUpdatedInfo
+BloomFilter
+BloomFilter.Version
+BooleanParam
+BooleanType
+BoostingStrategy
+BoundedDouble
+BreezeUtil
+Broadcast
+BroadcastBlockId
+Broker
+BucketedRandomProjectionLSH
+BucketedRandomProjectionLSHModel
+Bucketizer
+BufferReleasingInputStream
+BytecodeUtils
+ByteType
+CalendarIntervalType
+Catalog
+CatalystScan
+CategoricalSplit
+CausedBy
+CharType
+CheckpointReader
+CheckpointState
+ChiSqSelector
+ChiSqSelector
+ChiSqSelectorModel
+ChiSqSelectorModel
+ChiSqSelectorModel.SaveLoadV1_0$
+ChiSqTest
+ChiSqTest.Method
+ChiSqTest.Method$
+ChiSqTest.NullHypothesis$
+ChiSqTestResult
+CholeskyDecomposition
+ChunkedByteBufferInputStream
+ClassificationModel
+ClassificationModel
+Classifier
+CleanAccum
+CleanBroadcast
+CleanCheckpoint
+CleanRDD
+CleanShuffle
+CleanupTask
+CleanupTaskWeakReference
+ClosureCleaner
+ClusteringSummary
+CoarseGrainedClusterMessages
+CoarseGrainedClusterMessages.AddWebUIFilter
+CoarseGrainedClusterMessages.AddWebUIFilter$
+CoarseGrainedClusterMessages.GetExecutorLossReason
+CoarseGrainedClusterMessages.GetExecutorLossReason$
+CoarseGrainedClusterMessages.KillExecutors
+CoarseGrainedClusterMessages.KillExecutors$
+CoarseGrainedClusterMessages.KillTask
+CoarseGrainedClusterMessages.KillTask$
+CoarseGrainedClusterMessages.LaunchTask
+CoarseGrainedClusterMessages.LaunchTask$
+CoarseGrainedClusterMessages.RegisterClusterManager
+CoarseGrainedClusterMessages.RegisterClusterManager$
+CoarseGrainedClusterMessages.RegisteredExecutor$
+CoarseGrainedClusterMessages.RegisterExecutor

[41/51] [partial] spark-website git commit: Add 2.1.2 docs

2017-10-17 Thread holden
http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/R/ncol.html
--
diff --git a/site/docs/2.1.2/api/R/ncol.html b/site/docs/2.1.2/api/R/ncol.html
new file mode 100644
index 000..8677389
--- /dev/null
+++ b/site/docs/2.1.2/api/R/ncol.html
@@ -0,0 +1,97 @@
+http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd;>http://www.w3.org/1999/xhtml;>R: Returns the number of 
columns in a SparkDataFrame
+
+
+
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css;>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js";>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js";>
+hljs.initHighlightingOnLoad();
+
+
+ncol 
{SparkR}R Documentation
+
+Returns the number of columns in a SparkDataFrame
+
+Description
+
+Returns the number of columns in a SparkDataFrame
+
+
+
+Usage
+
+
+## S4 method for signature 'SparkDataFrame'
+ncol(x)
+
+
+
+Arguments
+
+
+x
+
+a SparkDataFrame
+
+
+
+
+Note
+
+ncol since 1.5.0
+
+
+
+See Also
+
+Other SparkDataFrame functions: SparkDataFrame-class,
+agg, arrange,
+as.data.frame, attach,
+cache, coalesce,
+collect, colnames,
+coltypes,
+createOrReplaceTempView,
+crossJoin, dapplyCollect,
+dapply, describe,
+dim, distinct,
+dropDuplicates, dropna,
+drop, dtypes,
+except, explain,
+filter, first,
+gapplyCollect, gapply,
+getNumPartitions, group_by,
+head, histogram,
+insertInto, intersect,
+isLocal, join,
+limit, merge,
+mutate, nrow,
+persist, printSchema,
+randomSplit, rbind,
+registerTempTable, rename,
+repartition, sample,
+saveAsTable, schema,
+selectExpr, select,
+showDF, show,
+storageLevel, str,
+subset, take,
+union, unpersist,
+withColumn, with,
+write.df, write.jdbc,
+write.json, write.orc,
+write.parquet, write.text
+
+
+
+Examples
+
+## Not run: 
+##D sparkR.session()
+##D path - path/to/file.json
+##D df - read.json(path)
+##D ncol(df)
+## End(Not run)
+
+
+
+[Package SparkR version 2.1.2 
Index]
+

http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/R/negate.html
--
diff --git a/site/docs/2.1.2/api/R/negate.html 
b/site/docs/2.1.2/api/R/negate.html
new file mode 100644
index 000..cc0e06d
--- /dev/null
+++ b/site/docs/2.1.2/api/R/negate.html
@@ -0,0 +1,67 @@
+http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd;>http://www.w3.org/1999/xhtml;>R: negate
+
+
+
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css;>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js";>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js";>
+hljs.initHighlightingOnLoad();
+
+
+negate 
{SparkR}R Documentation
+
+negate
+
+Description
+
+Unary minus, i.e. negate the expression.
+
+
+
+Usage
+
+
+negate(x)
+
+## S4 method for signature 'Column'
+negate(x)
+
+
+
+Arguments
+
+
+x
+
+Column to compute on.
+
+
+
+
+Note
+
+negate since 1.5.0
+
+
+
+See Also
+
+Other normal_funcs: abs,
+bitwiseNOT, coalesce,
+column, expr,
+greatest, ifelse,
+isnan, least,
+lit, nanvl,
+randn, rand,
+struct, when
+
+
+
+Examples
+
+## Not run: negate(df$c)
+
+
+
+[Package SparkR version 2.1.2 
Index]
+

http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/R/next_day.html
--
diff --git a/site/docs/2.1.2/api/R/next_day.html 
b/site/docs/2.1.2/api/R/next_day.html
new file mode 100644
index 000..736d7e4
--- /dev/null
+++ b/site/docs/2.1.2/api/R/next_day.html
@@ -0,0 +1,89 @@
+http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd;>http://www.w3.org/1999/xhtml;>R: next_day
+
+
+
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css;>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js";>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js";>
+hljs.initHighlightingOnLoad();
+
+
+next_day 
{SparkR}R Documentation
+
+next_day
+
+Description
+
+Given a date column, returns the first date which is later than the value 
of the date column
+that is on the specified day of the week.
+
+
+
+Usage
+
+
+next_day(y, x)
+
+## S4 method for signature 'Column,character'
+next_day(y, x)
+
+
+
+Arguments
+
+
+y
+
+Column to compute on.
+
+x
+
+Day of the week string.
+
+
+
+
+Details
+
+For example, next_day('2015-07-27', "Sunday") returns 
2015-08-02 because that is the first
+Sunday after 2015-07-27.
+
+Day of the week parameter is case insensitive, and accepts first three or 
two characters:
+Mon, Tue, Wed, Thu, 
Fri, Sat, Sun.
+
+
+
+Note
+
+next_day since 1.5.0
+
+
+
+See Also
+
+Other datetime_funcs: add_months,
+date_add, date_format,
+date_sub, datediff,
+dayofmonth, dayofyear,
+from_unixtime,
+from_utc_timestamp, 
hour,
+last_day, minute,
+months_between, month,
+quarter, second,

[46/51] [partial] spark-website git commit: Add 2.1.2 docs

2017-10-17 Thread holden
http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/R/columns.html
--
diff --git a/site/docs/2.1.2/api/R/columns.html 
b/site/docs/2.1.2/api/R/columns.html
new file mode 100644
index 000..3e7242a
--- /dev/null
+++ b/site/docs/2.1.2/api/R/columns.html
@@ -0,0 +1,137 @@
+http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd;>http://www.w3.org/1999/xhtml;>R: Column Names of 
SparkDataFrame
+
+
+
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css;>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js";>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js";>
+hljs.initHighlightingOnLoad();
+
+
+colnames 
{SparkR}R Documentation
+
+Column Names of SparkDataFrame
+
+Description
+
+Return all column names as a list.
+
+
+
+Usage
+
+
+colnames(x, do.NULL = TRUE, prefix = "col")
+
+colnames(x) - value
+
+columns(x)
+
+## S4 method for signature 'SparkDataFrame'
+columns(x)
+
+## S4 method for signature 'SparkDataFrame'
+names(x)
+
+## S4 replacement method for signature 'SparkDataFrame'
+names(x) - value
+
+## S4 method for signature 'SparkDataFrame'
+colnames(x)
+
+## S4 replacement method for signature 'SparkDataFrame'
+colnames(x) - value
+
+
+
+Arguments
+
+
+x
+
+a SparkDataFrame.
+
+do.NULL
+
+currently not used.
+
+prefix
+
+currently not used.
+
+value
+
+a character vector. Must have the same length as the number
+of columns in the SparkDataFrame.
+
+
+
+
+Note
+
+columns since 1.4.0
+
+names since 1.5.0
+
+names- since 1.5.0
+
+colnames since 1.6.0
+
+colnames- since 1.6.0
+
+
+
+See Also
+
+Other SparkDataFrame functions: SparkDataFrame-class,
+agg, arrange,
+as.data.frame, attach,
+cache, coalesce,
+collect, coltypes,
+createOrReplaceTempView,
+crossJoin, dapplyCollect,
+dapply, describe,
+dim, distinct,
+dropDuplicates, dropna,
+drop, dtypes,
+except, explain,
+filter, first,
+gapplyCollect, gapply,
+getNumPartitions, group_by,
+head, histogram,
+insertInto, intersect,
+isLocal, join,
+limit, merge,
+mutate, ncol,
+nrow, persist,
+printSchema, randomSplit,
+rbind, registerTempTable,
+rename, repartition,
+sample, saveAsTable,
+schema, selectExpr,
+select, showDF,
+show, storageLevel,
+str, subset,
+take, union,
+unpersist, withColumn,
+with, write.df,
+write.jdbc, write.json,
+write.orc, write.parquet,
+write.text
+
+
+
+Examples
+
+## Not run: 
+##D sparkR.session()
+##D path - path/to/file.json
+##D df - read.json(path)
+##D columns(df)
+##D colnames(df)
+## End(Not run)
+
+
+
+[Package SparkR version 2.1.2 
Index]
+

http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/R/concat.html
--
diff --git a/site/docs/2.1.2/api/R/concat.html 
b/site/docs/2.1.2/api/R/concat.html
new file mode 100644
index 000..6d226c8
--- /dev/null
+++ b/site/docs/2.1.2/api/R/concat.html
@@ -0,0 +1,77 @@
+http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd;>http://www.w3.org/1999/xhtml;>R: concat
+
+
+
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css;>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js";>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js";>
+hljs.initHighlightingOnLoad();
+
+
+concat 
{SparkR}R Documentation
+
+concat
+
+Description
+
+Concatenates multiple input string columns together into a single string 
column.
+
+
+
+Usage
+
+
+concat(x, ...)
+
+## S4 method for signature 'Column'
+concat(x, ...)
+
+
+
+Arguments
+
+
+x
+
+Column to compute on
+
+...
+
+other columns
+
+
+
+
+Note
+
+concat since 1.5.0
+
+
+
+See Also
+
+Other string_funcs: ascii,
+base64, concat_ws,
+decode, encode,
+format_number, format_string,
+initcap, instr,
+length, levenshtein,
+locate, lower,
+lpad, ltrim,
+regexp_extract,
+regexp_replace, reverse,
+rpad, rtrim,
+soundex, substring_index,
+translate, trim,
+unbase64, upper
+
+
+
+Examples
+
+## Not run: concat(df$strings, df$strings2)
+
+
+
+[Package SparkR version 2.1.2 
Index]
+

http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/R/concat_ws.html
--
diff --git a/site/docs/2.1.2/api/R/concat_ws.html 
b/site/docs/2.1.2/api/R/concat_ws.html
new file mode 100644
index 000..e30c492
--- /dev/null
+++ b/site/docs/2.1.2/api/R/concat_ws.html
@@ -0,0 +1,82 @@
+http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd;>http://www.w3.org/1999/xhtml;>R: concat_ws
+
+
+
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css;>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js";>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js";>
+hljs.initHighlightingOnLoad();
+
+
+concat_ws 
{SparkR}R Documentation
+
+concat_ws
+

[21/51] [partial] spark-website git commit: Add 2.1.2 docs

2017-10-17 Thread holden
http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/java/org/apache/spark/RangePartitioner.html
--
diff --git a/site/docs/2.1.2/api/java/org/apache/spark/RangePartitioner.html 
b/site/docs/2.1.2/api/java/org/apache/spark/RangePartitioner.html
new file mode 100644
index 000..75e4293
--- /dev/null
+++ b/site/docs/2.1.2/api/java/org/apache/spark/RangePartitioner.html
@@ -0,0 +1,394 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+
+
+RangePartitioner (Spark 2.1.2 JavaDoc)
+
+
+
+
+
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+
+
+
+Overview
+Package
+Class
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev Class
+Next Class
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+
+
+Detail:
+Field|
+Constr|
+Method
+
+
+
+
+
+
+
+
+org.apache.spark
+Class 
RangePartitionerK,V
+
+
+
+Object
+
+
+org.apache.spark.Partitioner
+
+
+org.apache.spark.RangePartitionerK,V
+
+
+
+
+
+
+
+
+
+All Implemented Interfaces:
+java.io.Serializable
+
+
+
+public class RangePartitionerK,V
+extends Partitioner
+A Partitioner that partitions 
sortable records by range into roughly
+ equal ranges. The ranges are determined by sampling the content of the RDD 
passed in.
+ 
+See Also:Serialized
 FormNote:
+  The actual number of partitions created by the RangePartitioner might 
not be the same
+ as the partitions parameter, in the case where the number of 
sampled records is less than
+ the value of partitions.
+
+
+
+
+
+
+
+
+
+
+
+Constructor Summary
+
+Constructors
+
+Constructor and Description
+
+
+RangePartitioner(intpartitions,
+RDD? extends scala.Product2K,Vrdd,
+booleanascending,
+scala.math.OrderingKevidence$1,
+scala.reflect.ClassTagKevidence$2)
+
+
+
+
+
+
+
+
+
+Method Summary
+
+Methods
+
+Modifier and Type
+Method and Description
+
+
+static KObject
+determineBounds(scala.collection.mutable.ArrayBufferscala.Tuple2K,Objectcandidates,
+   intpartitions,
+   scala.math.OrderingKevidence$4,
+   scala.reflect.ClassTagKevidence$5)
+Determines the bounds for range partitioning from 
candidates with weights indicating how many
+ items each represents.
+
+
+
+boolean
+equals(Objectother)
+
+
+int
+getPartition(Objectkey)
+
+
+int
+hashCode()
+
+
+int
+numPartitions()
+
+
+static 
Kscala.Tuple2Object,scala.Tuple3Object,Object,Object[]
+sketch(RDDKrdd,
+  intsampleSizePerPartition,
+  scala.reflect.ClassTagKevidence$3)
+Sketches the input RDD via reservoir sampling on each 
partition.
+
+
+
+
+
+
+
+Methods inherited from classorg.apache.spark.Partitioner
+defaultPartitioner
+
+
+
+
+
+Methods inherited from classObject
+getClass, notify, notifyAll, toString, wait, wait, wait
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+Constructor Detail
+
+
+
+
+
+RangePartitioner
+publicRangePartitioner(intpartitions,
+RDD? extends scala.Product2K,Vrdd,
+booleanascending,
+scala.math.OrderingKevidence$1,
+scala.reflect.ClassTagKevidence$2)
+
+
+
+
+
+
+
+
+
+Method Detail
+
+
+
+
+
+sketch
+public 
staticKscala.Tuple2Object,scala.Tuple3Object,Object,Object[]sketch(RDDKrdd,
+   
intsampleSizePerPartition,
+   
scala.reflect.ClassTagKevidence$3)
+Sketches the input RDD via reservoir sampling on each 
partition.
+ 
+Parameters:rdd - the 
input RDD to sketchsampleSizePerPartition - max sample 
size per partitionevidence$3 - (undocumented)
+Returns:(total number of items, an 
array of (partitionId, number of items, sample))
+
+
+
+
+
+
+
+determineBounds
+public 
staticKObjectdetermineBounds(scala.collection.mutable.ArrayBufferscala.Tuple2K,Objectcandidates,
+ intpartitions,
+ scala.math.OrderingKevidence$4,
+ scala.reflect.ClassTagKevidence$5)
+Determines the bounds for range partitioning from 
candidates with weights indicating how many
+ items each represents. Usually this is 1 over the probability used to sample 
this candidate.
+ 
+Parameters:candidates - unordered 
candidates with weightspartitions - number of 
partitionsevidence$4 - 
(undocumented)evidence$5 - (undocumented)
+Returns:selected bounds
+
+
+
+
+
+
+
+numPartitions
+publicintnumPartitions()
+
+Specified by:
+numPartitionsin
 classPartitioner
+
+
+
+
+
+
+
+

[16/51] [partial] spark-website git commit: Add 2.1.2 docs

2017-10-17 Thread holden
http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/java/org/apache/spark/StopMapOutputTracker.html
--
diff --git 
a/site/docs/2.1.2/api/java/org/apache/spark/StopMapOutputTracker.html 
b/site/docs/2.1.2/api/java/org/apache/spark/StopMapOutputTracker.html
new file mode 100644
index 000..ffb5849
--- /dev/null
+++ b/site/docs/2.1.2/api/java/org/apache/spark/StopMapOutputTracker.html
@@ -0,0 +1,323 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+
+
+StopMapOutputTracker (Spark 2.1.2 JavaDoc)
+
+
+
+
+
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+
+
+
+Overview
+Package
+Class
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev Class
+Next Class
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+
+
+Detail:
+Field|
+Constr|
+Method
+
+
+
+
+
+
+
+
+org.apache.spark
+Class 
StopMapOutputTracker
+
+
+
+Object
+
+
+org.apache.spark.StopMapOutputTracker
+
+
+
+
+
+
+
+
+public class StopMapOutputTracker
+extends Object
+
+
+
+
+
+
+
+
+
+
+
+Constructor Summary
+
+Constructors
+
+Constructor and Description
+
+
+StopMapOutputTracker()
+
+
+
+
+
+
+
+
+
+Method Summary
+
+Methods
+
+Modifier and Type
+Method and Description
+
+
+abstract static boolean
+canEqual(Objectthat)
+
+
+abstract static boolean
+equals(Objectthat)
+
+
+abstract static int
+productArity()
+
+
+abstract static Object
+productElement(intn)
+
+
+static 
scala.collection.IteratorObject
+productIterator()
+
+
+static String
+productPrefix()
+
+
+
+
+
+
+Methods inherited from classObject
+equals, getClass, hashCode, notify, notifyAll, toString, wait, wait, 
wait
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+Constructor Detail
+
+
+
+
+
+StopMapOutputTracker
+publicStopMapOutputTracker()
+
+
+
+
+
+
+
+
+
+Method Detail
+
+
+
+
+
+canEqual
+public abstract staticbooleancanEqual(Objectthat)
+
+
+
+
+
+
+
+equals
+public abstract staticbooleanequals(Objectthat)
+
+
+
+
+
+
+
+productElement
+public abstract staticObjectproductElement(intn)
+
+
+
+
+
+
+
+productArity
+public abstract staticintproductArity()
+
+
+
+
+
+
+
+productIterator
+public 
staticscala.collection.IteratorObjectproductIterator()
+
+
+
+
+
+
+
+productPrefix
+public staticStringproductPrefix()
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+Overview
+Package
+Class
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev Class
+Next Class
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+
+
+Detail:
+Field|
+Constr|
+Method
+
+
+
+
+
+
+
+

http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/java/org/apache/spark/Success.html
--
diff --git a/site/docs/2.1.2/api/java/org/apache/spark/Success.html 
b/site/docs/2.1.2/api/java/org/apache/spark/Success.html
new file mode 100644
index 000..830b4d8
--- /dev/null
+++ b/site/docs/2.1.2/api/java/org/apache/spark/Success.html
@@ -0,0 +1,325 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+
+
+Success (Spark 2.1.2 JavaDoc)
+
+
+
+
+
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+
+
+
+Overview
+Package
+Class
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev Class
+Next Class
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+
+
+Detail:
+Field|
+Constr|
+Method
+
+
+
+
+
+
+
+
+org.apache.spark
+Class Success
+
+
+
+Object
+
+
+org.apache.spark.Success
+
+
+
+
+
+
+
+
+public class Success
+extends Object
+:: DeveloperApi ::
+ Task succeeded.
+
+
+
+
+
+
+
+
+
+
+
+Constructor Summary
+
+Constructors
+
+Constructor and Description
+
+
+Success()
+
+
+
+
+
+
+
+
+
+Method Summary
+
+Methods
+
+Modifier and Type
+Method and Description
+
+
+abstract static boolean
+canEqual(Objectthat)
+
+
+abstract static boolean
+equals(Objectthat)
+
+
+abstract static int
+productArity()
+
+
+abstract static 

[11/51] [partial] spark-website git commit: Add 2.1.2 docs

2017-10-17 Thread holden
http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/java/org/apache/spark/api/java/JavaNewHadoopRDD.html
--
diff --git 
a/site/docs/2.1.2/api/java/org/apache/spark/api/java/JavaNewHadoopRDD.html 
b/site/docs/2.1.2/api/java/org/apache/spark/api/java/JavaNewHadoopRDD.html
new file mode 100644
index 000..6aefe3e
--- /dev/null
+++ b/site/docs/2.1.2/api/java/org/apache/spark/api/java/JavaNewHadoopRDD.html
@@ -0,0 +1,325 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+
+
+JavaNewHadoopRDD (Spark 2.1.2 JavaDoc)
+
+
+
+
+
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+
+
+
+Overview
+Package
+Class
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev 
Class
+Next 
Class
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+
+
+Detail:
+Field|
+Constr|
+Method
+
+
+
+
+
+
+
+
+org.apache.spark.api.java
+Class 
JavaNewHadoopRDDK,V
+
+
+
+Object
+
+
+org.apache.spark.api.java.JavaPairRDDK,V
+
+
+org.apache.spark.api.java.JavaNewHadoopRDDK,V
+
+
+
+
+
+
+
+
+
+All Implemented Interfaces:
+java.io.Serializable, JavaRDDLikescala.Tuple2K,V,JavaPairRDDK,V
+
+
+
+public class JavaNewHadoopRDDK,V
+extends JavaPairRDDK,V
+See Also:Serialized
 Form
+
+
+
+
+
+
+
+
+
+
+
+Constructor Summary
+
+Constructors
+
+Constructor and Description
+
+
+JavaNewHadoopRDD(NewHadoopRDDK,Vrdd,
+scala.reflect.ClassTagKkClassTag,
+scala.reflect.ClassTagVvClassTag)
+
+
+
+
+
+
+
+
+
+Method Summary
+
+Methods
+
+Modifier and Type
+Method and Description
+
+
+scala.reflect.ClassTagK
+kClassTag()
+
+
+RJavaRDDR
+mapPartitionsWithInputSplit(Function2org.apache.hadoop.mapreduce.InputSplit,java.util.Iteratorscala.Tuple2K,V,java.util.IteratorRf,
+   booleanpreservesPartitioning)
+Maps over a partition, providing the InputSplit that was 
used as the base of the partition.
+
+
+
+scala.reflect.ClassTagV
+vClassTag()
+
+
+
+
+
+
+Methods inherited from classorg.apache.spark.api.java.JavaPairRDD
+aggregate,
 aggregateByKey,
 aggregateByKey,
 aggregateByKey,
 cache,
 cartesian, checkpoint,
 classTag,
 coalesce,
 coalesce,
 cogroup,
 cogroup,
 cogroup,
 cogroup, cogroup,
 cogroup,
 cogroup,
 cogroup,
 cogroup,
 collect,
 collectAsMap,
 collectAsync,
 collectPartitions,
 combineByKey,
 combineByKey,
  combineByKey,
 combineByKey,
 context,
 count,
 countApprox,
 countApprox,
 countApproxDistinct,
 countApproxDistinctByKey,
 countApproxDistinctByKey,
 countApproxDistinctByKey,
 countAsync,
 countByKey,
 countByKeyApprox,
 countByKeyApprox,
 countByValue,
 countByValueApprox,
 countByValueApprox,
 distinct,
 distinct,
 filter,
 first,
 flatMap,
 flatMapToDouble, flatMapToPair,
 flatMapValues,
 fold,
 foldByKey,
 foldByKey,
 foldByKey,
 foreach,
 foreachAsync,
 foreachPartition,
 foreachPartitionAsync,
 fromJavaRDD,
 fromRDD,
 fullOuterJoin
 , fullOuterJoin,
 fullOuterJoin,
 getCheckpointFile,
 getNumPartitions,
 getStorageLevel,
 glom,
 groupBy,
 gr
 oupBy, groupByKey,
 groupByKey,
 groupByKey,
 groupWith,
 groupWith,
 groupWith,
 id, 
intersection, isCheckpointed,
 isEmpty,
 iterator,
 join,
 join,
 join,
 keyBy,
 keys, leftOuterJoin,
 leftOuterJoin,
 leftOuterJoin,
 lookup,
 map,
 mapPartitions,
 mapPartitions, mapPartitionsToDouble,
 mapPartitionsToDouble,
 mapPartitionsToPair,
 mapPartitionsToPair,
 mapPartitionsWithIndex,
 mapPartitionsWithIndex$default$2, mapToDouble,
 mapToPair,
 mapValues,
 max,
 min,
 name,
 partitionBy,
 
 partitioner, partitions,
 persist,
 pipe,
 pipe,
 pipe,
 pipe,
 pipe,
 rdd, 
reduce, reduceByKey,
 reduceByKey,
 reduceByKey,
 reduceByKeyLocally,
 repartition,
 repartitionAndSortWithinPartitions,
 repartitionAndSortWithinPartitions,
 rightOuterJoin,
 rightOuterJoin,
 rightOuterJoin,
 sample,
 sample,
 sampleByKey,
 sampleByKey,
 sampleByKeyExact,
 sampleByKeyExact,
 saveAsHadoopDataset,
 saveAsHadoopFile,
 saveAsHadoopFile,
 saveAsHadoopFile,
 saveAsNewAPIHadoopDataset,
 saveAsNewAPIHadoopFile,
 saveAsNewAPIHadoopFile,
 saveAsObjectFile,
 saveAsTextFile,
 saveAsTextFile,
 setName,
 sortByKey,
 sortByKey,
 sortByKey,
 sortByKey,
 sortByKey,
 sortByKey,
 subtract, subtract,
 subtract,
 subtractByKey,
 subtractByKey,
 subtractByKey,
 take,
 takeAsync,
 takeOrdered,
 takeOrdered,
 takeSample,
 takeSample,
 toDebugString,
 toLocalIterator,
 top,
 top,
 toRDD,
 treeAggregate,
 treeAggregate,
 treeReduce,
 treeReduce,
 union,
 unpersist,
 unpersist,
 values, 
wrapRDD,
 zip,
 zipPartitions,
 zipWithIndex,
 

[23/51] [partial] spark-website git commit: Add 2.1.2 docs

2017-10-17 Thread holden
http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/java/org/apache/spark/InternalAccumulator.html
--
diff --git a/site/docs/2.1.2/api/java/org/apache/spark/InternalAccumulator.html 
b/site/docs/2.1.2/api/java/org/apache/spark/InternalAccumulator.html
new file mode 100644
index 000..b236e74
--- /dev/null
+++ b/site/docs/2.1.2/api/java/org/apache/spark/InternalAccumulator.html
@@ -0,0 +1,499 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+
+
+InternalAccumulator (Spark 2.1.2 JavaDoc)
+
+
+
+
+
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+
+
+
+Overview
+Package
+Class
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev Class
+Next 
Class
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+
+
+Detail:
+Field|
+Constr|
+Method
+
+
+
+
+
+
+
+
+org.apache.spark
+Class 
InternalAccumulator
+
+
+
+Object
+
+
+org.apache.spark.InternalAccumulator
+
+
+
+
+
+
+
+
+public class InternalAccumulator
+extends Object
+A collection of fields and methods concerned with internal 
accumulators that represent
+ task level metrics.
+
+
+
+
+
+
+
+
+
+
+
+Nested Class Summary
+
+Nested Classes
+
+Modifier and Type
+Class and Description
+
+
+static class
+InternalAccumulator.input$
+
+
+static class
+InternalAccumulator.output$
+
+
+static class
+InternalAccumulator.shuffleRead$
+
+
+static class
+InternalAccumulator.shuffleWrite$
+
+
+
+
+
+
+
+
+
+Constructor Summary
+
+Constructors
+
+Constructor and Description
+
+
+InternalAccumulator()
+
+
+
+
+
+
+
+
+
+Method Summary
+
+Methods
+
+Modifier and Type
+Method and Description
+
+
+static String
+DISK_BYTES_SPILLED()
+
+
+static String
+EXECUTOR_CPU_TIME()
+
+
+static String
+EXECUTOR_DESERIALIZE_CPU_TIME()
+
+
+static String
+EXECUTOR_DESERIALIZE_TIME()
+
+
+static String
+EXECUTOR_RUN_TIME()
+
+
+static String
+INPUT_METRICS_PREFIX()
+
+
+static String
+JVM_GC_TIME()
+
+
+static String
+MEMORY_BYTES_SPILLED()
+
+
+static String
+METRICS_PREFIX()
+
+
+static String
+OUTPUT_METRICS_PREFIX()
+
+
+static String
+PEAK_EXECUTION_MEMORY()
+
+
+static String
+RESULT_SERIALIZATION_TIME()
+
+
+static String
+RESULT_SIZE()
+
+
+static String
+SHUFFLE_READ_METRICS_PREFIX()
+
+
+static String
+SHUFFLE_WRITE_METRICS_PREFIX()
+
+
+static String
+TEST_ACCUM()
+
+
+static String
+UPDATED_BLOCK_STATUSES()
+
+
+
+
+
+
+Methods inherited from classObject
+equals, getClass, hashCode, notify, notifyAll, toString, wait, wait, 
wait
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+Constructor Detail
+
+
+
+
+
+InternalAccumulator
+publicInternalAccumulator()
+
+
+
+
+
+
+
+
+
+Method Detail
+
+
+
+
+
+METRICS_PREFIX
+public staticStringMETRICS_PREFIX()
+
+
+
+
+
+
+
+SHUFFLE_READ_METRICS_PREFIX
+public staticStringSHUFFLE_READ_METRICS_PREFIX()
+
+
+
+
+
+
+
+SHUFFLE_WRITE_METRICS_PREFIX
+public staticStringSHUFFLE_WRITE_METRICS_PREFIX()
+
+
+
+
+
+
+
+OUTPUT_METRICS_PREFIX
+public staticStringOUTPUT_METRICS_PREFIX()
+
+
+
+
+
+
+
+INPUT_METRICS_PREFIX
+public staticStringINPUT_METRICS_PREFIX()
+
+
+
+
+
+
+
+EXECUTOR_DESERIALIZE_TIME
+public staticStringEXECUTOR_DESERIALIZE_TIME()
+
+
+
+
+
+
+
+EXECUTOR_DESERIALIZE_CPU_TIME
+public staticStringEXECUTOR_DESERIALIZE_CPU_TIME()
+
+
+
+
+
+
+
+EXECUTOR_RUN_TIME
+public staticStringEXECUTOR_RUN_TIME()
+
+
+
+
+
+
+
+EXECUTOR_CPU_TIME
+public staticStringEXECUTOR_CPU_TIME()
+
+
+
+
+
+
+
+RESULT_SIZE
+public staticStringRESULT_SIZE()
+
+
+
+
+
+
+
+JVM_GC_TIME
+public staticStringJVM_GC_TIME()
+
+
+
+
+
+
+
+RESULT_SERIALIZATION_TIME
+public staticStringRESULT_SERIALIZATION_TIME()
+
+
+
+
+
+
+
+MEMORY_BYTES_SPILLED
+public staticStringMEMORY_BYTES_SPILLED()
+
+
+
+
+
+
+
+DISK_BYTES_SPILLED
+public staticStringDISK_BYTES_SPILLED()
+
+
+
+
+
+
+
+PEAK_EXECUTION_MEMORY
+public staticStringPEAK_EXECUTION_MEMORY()
+
+
+
+
+
+
+
+UPDATED_BLOCK_STATUSES
+public staticStringUPDATED_BLOCK_STATUSES()
+
+
+
+
+
+
+
+TEST_ACCUM
+public staticStringTEST_ACCUM()
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+Overview
+Package
+Class
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev Class
+Next 
Class
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+
+
+Detail:
+Field|
+Constr|
+Method
+
+
+
+
+
+
+
+


[38/51] [partial] spark-website git commit: Add 2.1.2 docs

2017-10-17 Thread holden
http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/R/spark.gbt.html
--
diff --git a/site/docs/2.1.2/api/R/spark.gbt.html 
b/site/docs/2.1.2/api/R/spark.gbt.html
new file mode 100644
index 000..98b2b03
--- /dev/null
+++ b/site/docs/2.1.2/api/R/spark.gbt.html
@@ -0,0 +1,244 @@
+http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd;>http://www.w3.org/1999/xhtml;>R: Gradient Boosted Tree 
Model for Regression and Classification
+
+
+
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css;>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js";>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js";>
+hljs.initHighlightingOnLoad();
+
+
+spark.gbt 
{SparkR}R Documentation
+
+Gradient Boosted Tree Model for Regression and Classification
+
+Description
+
+spark.gbt fits a Gradient Boosted Tree Regression model or 
Classification model on a
+SparkDataFrame. Users can call summary to get a summary of the 
fitted
+Gradient Boosted Tree model, predict to make predictions on new 
data, and
+write.ml/read.ml to save/load fitted models.
+For more details, see
+http://spark.apache.org/docs/latest/ml-classification-regression.html#gradient-boosted-tree-regression;>
+GBT Regression and
+http://spark.apache.org/docs/latest/ml-classification-regression.html#gradient-boosted-tree-classifier;>
+GBT Classification
+
+
+
+Usage
+
+
+spark.gbt(data, formula, ...)
+
+## S4 method for signature 'SparkDataFrame,formula'
+spark.gbt(data, formula,
+  type = c("regression", "classification"), maxDepth = 5, maxBins = 32,
+  maxIter = 20, stepSize = 0.1, lossType = NULL, seed = NULL,
+  subsamplingRate = 1, minInstancesPerNode = 1, minInfoGain = 0,
+  checkpointInterval = 10, maxMemoryInMB = 256, cacheNodeIds = FALSE)
+
+## S4 method for signature 'GBTRegressionModel'
+predict(object, newData)
+
+## S4 method for signature 'GBTClassificationModel'
+predict(object, newData)
+
+## S4 method for signature 'GBTRegressionModel,character'
+write.ml(object, path,
+  overwrite = FALSE)
+
+## S4 method for signature 'GBTClassificationModel,character'
+write.ml(object, path,
+  overwrite = FALSE)
+
+## S4 method for signature 'GBTRegressionModel'
+summary(object)
+
+## S4 method for signature 'GBTClassificationModel'
+summary(object)
+
+## S3 method for class 'summary.GBTRegressionModel'
+print(x, ...)
+
+## S3 method for class 'summary.GBTClassificationModel'
+print(x, ...)
+
+
+
+Arguments
+
+
+data
+
+a SparkDataFrame for training.
+
+formula
+
+a symbolic description of the model to be fitted. Currently only a few 
formula
+operators are supported, including '~', ':', '+', and '-'.
+
+...
+
+additional arguments passed to the method.
+
+type
+
+type of model, one of regression or classification, 
to fit
+
+maxDepth
+
+Maximum depth of the tree (= 0).
+
+maxBins
+
+Maximum number of bins used for discretizing continuous features and for 
choosing
+how to split on features at each node. More bins give higher granularity. Must 
be
+= 2 and = number of categories in any categorical feature.
+
+maxIter
+
+Param for maximum number of iterations (= 0).
+
+stepSize
+
+Param for Step size to be used for each iteration of optimization.
+
+lossType
+
+Loss function which GBT tries to minimize.
+For classification, must be logistic. For regression, must be one 
of
+squared (L2) and absolute (L1), default is 
squared.
+
+seed
+
+integer seed for random number generation.
+
+subsamplingRate
+
+Fraction of the training data used for learning each decision tree, in
+range (0, 1].
+
+minInstancesPerNode
+
+Minimum number of instances each child must have after split. If a
+split causes the left or right child to have fewer than
+minInstancesPerNode, the split will be discarded as invalid. Should be
+= 1.
+
+minInfoGain
+
+Minimum information gain for a split to be considered at a tree node.
+
+checkpointInterval
+
+Param for set checkpoint interval (= 1) or disable checkpoint (-1).
+
+maxMemoryInMB
+
+Maximum memory in MB allocated to histogram aggregation.
+
+cacheNodeIds
+
+If FALSE, the algorithm will pass trees to executors to match instances with
+nodes. If TRUE, the algorithm will cache node IDs for each instance. Caching
+can speed up training of deeper trees. Users can set how often should the
+cache be checkpointed or disable it by setting checkpointInterval.
+
+object
+
+A fitted Gradient Boosted Tree regression model or classification model.
+
+newData
+
+a SparkDataFrame for testing.
+
+path
+
+The directory where the model is saved.
+
+overwrite
+
+Overwrites or not if the output path already exists. Default is FALSE
+which means throw exception if the output path exists.
+
+x
+
+summary object of Gradient Boosted Tree regression model or classification 
model
+returned by summary.
+
+
+
+
+Value
+
+spark.gbt returns a fitted Gradient Boosted Tree model.
+

[01/51] [partial] spark-website git commit: Add 2.1.2 docs

2017-10-17 Thread holden
Repository: spark-website
Updated Branches:
  refs/heads/apache-asf-site [created] a6d9cbdef


http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/java/org/apache/spark/api/r/RRDD.html
--
diff --git a/site/docs/2.1.2/api/java/org/apache/spark/api/r/RRDD.html 
b/site/docs/2.1.2/api/java/org/apache/spark/api/r/RRDD.html
new file mode 100644
index 000..19fea40
--- /dev/null
+++ b/site/docs/2.1.2/api/java/org/apache/spark/api/r/RRDD.html
@@ -0,0 +1,2158 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+
+
+RRDD (Spark 2.1.2 JavaDoc)
+
+
+
+
+
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+
+
+
+Overview
+Package
+Class
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev 
Class
+Next 
Class
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+
+
+Detail:
+Field|
+Constr|
+Method
+
+
+
+
+
+
+
+
+org.apache.spark.api.r
+Class RRDDT
+
+
+
+Object
+
+
+org.apache.spark.rdd.RDDU
+
+
+org.apache.spark.api.r.BaseRRDDT,byte[]
+
+
+org.apache.spark.api.r.RRDDT
+
+
+
+
+
+
+
+
+
+
+
+All Implemented Interfaces:
+java.io.Serializable
+
+
+
+public class RRDDT
+extends BaseRRDDT,byte[]
+An RDD that stores serialized R objects as 
Array[Byte].
+See Also:Serialized
 Form
+
+
+
+
+
+
+
+
+
+
+
+Constructor Summary
+
+Constructors
+
+Constructor and Description
+
+
+RRDD(RDDTparent,
+byte[]func,
+Stringdeserializer,
+Stringserializer,
+byte[]packageNames,
+Object[]broadcastVars,
+scala.reflect.ClassTagTevidence$4)
+
+
+
+
+
+
+
+
+
+Method Summary
+
+Methods
+
+Modifier and Type
+Method and Description
+
+
+static RDDT
+$plus$plus(RDDTother)
+
+
+static UU
+aggregate(UzeroValue,
+ scala.Function2U,T,UseqOp,
+ scala.Function2U,U,UcombOp,
+ scala.reflect.ClassTagUevidence$30)
+
+
+JavaRDDbyte[]
+asJavaRDD()
+
+
+static RDDT
+cache()
+
+
+static URDDscala.Tuple2T,U
+cartesian(RDDUother,
+ scala.reflect.ClassTagUevidence$5)
+
+
+static void
+checkpoint()
+
+
+static RDDT
+coalesce(intnumPartitions,
+booleanshuffle,
+scala.OptionPartitionCoalescerpartitionCoalescer,
+scala.math.OrderingTord)
+
+
+static boolean
+coalesce$default$2()
+
+
+static scala.OptionPartitionCoalescer
+coalesce$default$3()
+
+
+static scala.math.OrderingT
+coalesce$default$4(intnumPartitions,
+  booleanshuffle,
+  scala.OptionPartitionCoalescerpartitionCoalescer)
+
+
+static Object
+collect()
+
+
+static URDDU
+collect(scala.PartialFunctionT,Uf,
+   scala.reflect.ClassTagUevidence$29)
+
+
+static 
scala.collection.IteratorU
+compute(Partitionpartition,
+   TaskContextcontext)
+
+
+static SparkContext
+context()
+
+
+static long
+count()
+
+
+static PartialResultBoundedDouble
+countApprox(longtimeout,
+   doubleconfidence)
+
+
+static double
+countApprox$default$2()
+
+
+static long
+countApproxDistinct(doublerelativeSD)
+
+
+static long
+countApproxDistinct(intp,
+   intsp)
+
+
+static double
+countApproxDistinct$default$1()
+
+
+static 
scala.collection.MapT,Object
+countByValue(scala.math.OrderingTord)
+
+
+static scala.math.OrderingT
+countByValue$default$1()
+
+
+static PartialResultscala.collection.MapT,BoundedDouble
+countByValueApprox(longtimeout,
+  doubleconfidence,
+  scala.math.OrderingTord)
+
+
+static double
+countByValueApprox$default$2()
+
+
+static scala.math.OrderingT
+countByValueApprox$default$3(longtimeout,
+doubleconfidence)
+
+
+static JavaRDDbyte[]
+createRDDFromArray(JavaSparkContextjsc,
+  byte[][]arr)
+Create an RRDD given a sequence of byte arrays.
+
+
+
+static JavaRDDbyte[]
+createRDDFromFile(JavaSparkContextjsc,
+ StringfileName,
+ intparallelism)
+Create an RRDD given a temporary file name.
+
+
+
+static JavaSparkContext
+createSparkContext(Stringmaster,
+  StringappName,
+  StringsparkHome,
+  String[]jars,
+  java.util.MapObject,ObjectsparkEnvirMap,
+  
java.util.MapObject,ObjectsparkExecutorEnvMap)
+
+
+static scala.collection.SeqDependency?
+dependencies()
+
+
+static RDDT
+distinct()
+
+
+static RDDT
+distinct(intnumPartitions,
+scala.math.OrderingTord)
+
+
+static scala.math.OrderingT
+distinct$default$2(intnumPartitions)
+
+
+static RDDT
+filter(scala.Function1T,Objectf)
+
+
+static T
+first()
+
+
+static URDDU

[31/51] [partial] spark-website git commit: Add 2.1.2 docs

2017-10-17 Thread holden
http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/java/index-all.html
--
diff --git a/site/docs/2.1.2/api/java/index-all.html 
b/site/docs/2.1.2/api/java/index-all.html
new file mode 100644
index 000..67505d4
--- /dev/null
+++ b/site/docs/2.1.2/api/java/index-all.html
@@ -0,0 +1,46481 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+
+
+Index (Spark 2.1.2 JavaDoc)
+
+
+
+
+
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+
+
+
+Overview
+Package
+Class
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev
+Next
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+
+
+$ABCDEFGHIJKLMNOPQRSTUVWXYZ_
+
+
+$
+
+$colon$bslash(B,
 Function2A, B, B) - Static method in class 
org.apache.spark.sql.types.StructType
+
+$colon$plus(B,
 CanBuildFromRepr, B, That) - Static method in class 
org.apache.spark.sql.types.StructType
+
+$div$colon(B,
 Function2B, A, B) - Static method in class 
org.apache.spark.sql.types.StructType
+
+$greater(A)
 - Static method in class org.apache.spark.sql.types.Decimal
+
+$greater(A)
 - Static method in class org.apache.spark.storage.RDDInfo
+
+$greater$eq(A)
 - Static method in class org.apache.spark.sql.types.Decimal
+
+$greater$eq(A)
 - Static method in class org.apache.spark.storage.RDDInfo
+
+$less(A) - 
Static method in class org.apache.spark.sql.types.Decimal
+
+$less(A) - 
Static method in class org.apache.spark.storage.RDDInfo
+
+$less$eq(A)
 - Static method in class org.apache.spark.sql.types.Decimal
+
+$less$eq(A)
 - Static method in class org.apache.spark.storage.RDDInfo
+
+$minus$greater(T)
 - Static method in class org.apache.spark.ml.param.DoubleParam
+
+$minus$greater(T)
 - Static method in class org.apache.spark.ml.param.FloatParam
+
+$plus$colon(B,
 CanBuildFromRepr, B, That) - Static method in class 
org.apache.spark.sql.types.StructType
+
+$plus$eq(T) - 
Static method in class org.apache.spark.Accumulator
+
+Deprecated.
+
+$plus$plus(RDDT)
 - Static method in class org.apache.spark.api.r.RRDD
+
+$plus$plus(RDDT)
 - Static method in class org.apache.spark.graphx.EdgeRDD
+
+$plus$plus(RDDT)
 - Static method in class org.apache.spark.graphx.impl.EdgeRDDImpl
+
+$plus$plus(RDDT)
 - Static method in class org.apache.spark.graphx.impl.VertexRDDImpl
+
+$plus$plus(RDDT)
 - Static method in class org.apache.spark.graphx.VertexRDD
+
+$plus$plus(RDDT)
 - Static method in class org.apache.spark.rdd.HadoopRDD
+
+$plus$plus(RDDT)
 - Static method in class org.apache.spark.rdd.JdbcRDD
+
+$plus$plus(RDDT)
 - Static method in class org.apache.spark.rdd.NewHadoopRDD
+
+$plus$plus(RDDT)
 - Static method in class org.apache.spark.rdd.PartitionPruningRDD
+
+$plus$plus(RDDT)
 - Static method in class org.apache.spark.rdd.UnionRDD
+
+$plus$plus(GenTraversableOnceB,
 CanBuildFromRepr, B, That) - Static method in class 
org.apache.spark.sql.types.StructType
+
+$plus$plus$colon(TraversableOnceB,
 CanBuildFromRepr, B, That) - Static method in class 
org.apache.spark.sql.types.StructType
+
+$plus$plus$colon(TraversableB,
 CanBuildFromRepr, B, That) - Static method in class 
org.apache.spark.sql.types.StructType
+
+$plus$plus$eq(R)
 - Static method in class org.apache.spark.Accumulator
+
+Deprecated.
+
+
+
+
+
+A
+
+abortJob(JobContext)
 - Method in class org.apache.spark.internal.io.FileCommitProtocol
+
+Aborts a job after the writes fail.
+
+abortJob(JobContext)
 - Method in class org.apache.spark.internal.io.HadoopMapReduceCommitProtocol
+
+abortTask(TaskAttemptContext)
 - Method in class org.apache.spark.internal.io.FileCommitProtocol
+
+Aborts a task after the writes have failed.
+
+abortTask(TaskAttemptContext)
 - Method in class org.apache.spark.internal.io.HadoopMapReduceCommitProtocol
+
+abs(Column)
 - Static method in class org.apache.spark.sql.functions
+
+Computes the absolute value.
+
+abs() - 
Method in class org.apache.spark.sql.types.Decimal
+
+absent() - 
Static method in class org.apache.spark.api.java.Optional
+
+AbsoluteError - Class in org.apache.spark.mllib.tree.loss
+
+:: DeveloperApi ::
+ Class for absolute error loss calculation (for regression).
+
+AbsoluteError()
 - Constructor for class org.apache.spark.mllib.tree.loss.AbsoluteError
+
+accept(Parsers)
 - Static method in class org.apache.spark.ml.feature.RFormulaParser
+
+accept(ES,
 Function1ES, ListObject) - Static method in class 
org.apache.spark.ml.feature.RFormulaParser
+
+accept(String,
 PartialFunctionObject, U) - Static method in class 
org.apache.spark.ml.feature.RFormulaParser
+
+acceptIf(Function1Object,
 

[39/51] [partial] spark-website git commit: Add 2.1.2 docs

2017-10-17 Thread holden
http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/R/sd.html
--
diff --git a/site/docs/2.1.2/api/R/sd.html b/site/docs/2.1.2/api/R/sd.html
new file mode 100644
index 000..1018604
--- /dev/null
+++ b/site/docs/2.1.2/api/R/sd.html
@@ -0,0 +1,85 @@
+http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd;>http://www.w3.org/1999/xhtml;>R: sd
+
+
+
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css;>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js";>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js";>
+hljs.initHighlightingOnLoad();
+
+
+sd {SparkR}R Documentation
+
+sd
+
+Description
+
+Aggregate function: alias for stddev_samp
+
+
+
+Usage
+
+
+sd(x, na.rm = FALSE)
+
+stddev(x)
+
+## S4 method for signature 'Column'
+sd(x)
+
+## S4 method for signature 'Column'
+stddev(x)
+
+
+
+Arguments
+
+
+x
+
+Column to compute on.
+
+na.rm
+
+currently not used.
+
+
+
+
+Note
+
+sd since 1.6.0
+
+stddev since 1.6.0
+
+
+
+See Also
+
+stddev_pop, stddev_samp
+
+Other agg_funcs: agg, avg,
+countDistinct, count,
+first, kurtosis,
+last, max,
+mean, min,
+skewness, stddev_pop,
+stddev_samp, sumDistinct,
+sum, var_pop,
+var_samp, var
+
+
+
+Examples
+
+## Not run: 
+##D stddev(df$c)
+##D select(df, stddev(df$age))
+##D agg(df, sd(df$age))
+## End(Not run)
+
+
+
+[Package SparkR version 2.1.2 
Index]
+

http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/R/second.html
--
diff --git a/site/docs/2.1.2/api/R/second.html 
b/site/docs/2.1.2/api/R/second.html
new file mode 100644
index 000..25cf1e6
--- /dev/null
+++ b/site/docs/2.1.2/api/R/second.html
@@ -0,0 +1,71 @@
+http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd;>http://www.w3.org/1999/xhtml;>R: second
+
+
+
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css;>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js";>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js";>
+hljs.initHighlightingOnLoad();
+
+
+second 
{SparkR}R Documentation
+
+second
+
+Description
+
+Extracts the seconds as an integer from a given date/timestamp/string.
+
+
+
+Usage
+
+
+second(x)
+
+## S4 method for signature 'Column'
+second(x)
+
+
+
+Arguments
+
+
+x
+
+Column to compute on.
+
+
+
+
+Note
+
+second since 1.5.0
+
+
+
+See Also
+
+Other datetime_funcs: add_months,
+date_add, date_format,
+date_sub, datediff,
+dayofmonth, dayofyear,
+from_unixtime,
+from_utc_timestamp, 
hour,
+last_day, minute,
+months_between, month,
+next_day, quarter,
+to_date, to_utc_timestamp,
+unix_timestamp, weekofyear,
+window, year
+
+
+
+Examples
+
+## Not run: second(df$c)
+
+
+
+[Package SparkR version 2.1.2 
Index]
+

http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/R/select.html
--
diff --git a/site/docs/2.1.2/api/R/select.html 
b/site/docs/2.1.2/api/R/select.html
new file mode 100644
index 000..82cc5b6
--- /dev/null
+++ b/site/docs/2.1.2/api/R/select.html
@@ -0,0 +1,150 @@
+http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd;>http://www.w3.org/1999/xhtml;>R: Select
+
+
+
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css;>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js";>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js";>
+hljs.initHighlightingOnLoad();
+
+
+select 
{SparkR}R Documentation
+
+Select
+
+Description
+
+Selects a set of columns with names or Column expressions.
+
+
+
+Usage
+
+
+select(x, col, ...)
+
+## S4 method for signature 'SparkDataFrame'
+x$name
+
+## S4 replacement method for signature 'SparkDataFrame'
+x$name - value
+
+## S4 method for signature 'SparkDataFrame,character'
+select(x, col, ...)
+
+## S4 method for signature 'SparkDataFrame,Column'
+select(x, col, ...)
+
+## S4 method for signature 'SparkDataFrame,list'
+select(x, col)
+
+
+
+Arguments
+
+
+x
+
+a SparkDataFrame.
+
+col
+
+a list of columns or single Column or name.
+
+...
+
+additional column(s) if only one column is specified in col.
+If more than one column is assigned in col, ...
+should be left empty.
+
+name
+
+name of a Column (without being wrapped by "").
+
+value
+
+a Column or an atomic vector in the length of 1 as literal value, or 
NULL.
+If NULL, the specified Column is dropped.
+
+
+
+
+Value
+
+A new SparkDataFrame with selected columns.
+
+
+
+Note
+
+$ since 1.4.0
+
+$- since 1.4.0
+
+select(SparkDataFrame, character) since 1.4.0
+
+select(SparkDataFrame, Column) since 1.4.0
+
+select(SparkDataFrame, list) since 1.4.0
+
+
+
+See Also
+
+Other SparkDataFrame functions: SparkDataFrame-class,
+agg, arrange,

[14/51] [partial] spark-website git commit: Add 2.1.2 docs

2017-10-17 Thread holden
http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/java/org/apache/spark/UnknownReason.html
--
diff --git a/site/docs/2.1.2/api/java/org/apache/spark/UnknownReason.html 
b/site/docs/2.1.2/api/java/org/apache/spark/UnknownReason.html
new file mode 100644
index 000..8feb256
--- /dev/null
+++ b/site/docs/2.1.2/api/java/org/apache/spark/UnknownReason.html
@@ -0,0 +1,352 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+
+
+UnknownReason (Spark 2.1.2 JavaDoc)
+
+
+
+
+
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+
+
+
+Overview
+Package
+Class
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev Class
+Next Class
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+
+
+Detail:
+Field|
+Constr|
+Method
+
+
+
+
+
+
+
+
+org.apache.spark
+Class UnknownReason
+
+
+
+Object
+
+
+org.apache.spark.UnknownReason
+
+
+
+
+
+
+
+
+public class UnknownReason
+extends Object
+:: DeveloperApi ::
+ We don't know why the task ended -- for example, because of a ClassNotFound 
exception when
+ deserializing the task result.
+
+
+
+
+
+
+
+
+
+
+
+Constructor Summary
+
+Constructors
+
+Constructor and Description
+
+
+UnknownReason()
+
+
+
+
+
+
+
+
+
+Method Summary
+
+Methods
+
+Modifier and Type
+Method and Description
+
+
+abstract static boolean
+canEqual(Objectthat)
+
+
+static boolean
+countTowardsTaskFailures()
+
+
+abstract static boolean
+equals(Objectthat)
+
+
+abstract static int
+productArity()
+
+
+abstract static Object
+productElement(intn)
+
+
+static 
scala.collection.IteratorObject
+productIterator()
+
+
+static String
+productPrefix()
+
+
+static String
+toErrorString()
+
+
+
+
+
+
+Methods inherited from classObject
+equals, getClass, hashCode, notify, notifyAll, toString, wait, wait, 
wait
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+Constructor Detail
+
+
+
+
+
+UnknownReason
+publicUnknownReason()
+
+
+
+
+
+
+
+
+
+Method Detail
+
+
+
+
+
+toErrorString
+public staticStringtoErrorString()
+
+
+
+
+
+
+
+countTowardsTaskFailures
+public staticbooleancountTowardsTaskFailures()
+
+
+
+
+
+
+
+canEqual
+public abstract staticbooleancanEqual(Objectthat)
+
+
+
+
+
+
+
+equals
+public abstract staticbooleanequals(Objectthat)
+
+
+
+
+
+
+
+productElement
+public abstract staticObjectproductElement(intn)
+
+
+
+
+
+
+
+productArity
+public abstract staticintproductArity()
+
+
+
+
+
+
+
+productIterator
+public 
staticscala.collection.IteratorObjectproductIterator()
+
+
+
+
+
+
+
+productPrefix
+public staticStringproductPrefix()
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+Overview
+Package
+Class
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev Class
+Next Class
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+
+
+Detail:
+Field|
+Constr|
+Method
+
+
+
+
+
+
+
+


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[45/51] [partial] spark-website git commit: Add 2.1.2 docs

2017-10-17 Thread holden
http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/R/date_format.html
--
diff --git a/site/docs/2.1.2/api/R/date_format.html 
b/site/docs/2.1.2/api/R/date_format.html
new file mode 100644
index 000..7f9ad3c
--- /dev/null
+++ b/site/docs/2.1.2/api/R/date_format.html
@@ -0,0 +1,87 @@
+http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd;>http://www.w3.org/1999/xhtml;>R: date_format
+
+
+
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css;>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js";>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js";>
+hljs.initHighlightingOnLoad();
+
+
+date_format {SparkR}R 
Documentation
+
+date_format
+
+Description
+
+Converts a date/timestamp/string to a value of string in the format 
specified by the date
+format given by the second argument.
+
+
+
+Usage
+
+
+date_format(y, x)
+
+## S4 method for signature 'Column,character'
+date_format(y, x)
+
+
+
+Arguments
+
+
+y
+
+Column to compute on.
+
+x
+
+date format specification.
+
+
+
+
+Details
+
+A pattern could be for instance 
+dd.MM. and could return a string like '18.03.1993'. All
+pattern letters of java.text.SimpleDateFormat can be used.
+
+Note: Use when ever possible specialized functions like year. 
These benefit from a
+specialized implementation.
+
+
+
+Note
+
+date_format since 1.5.0
+
+
+
+See Also
+
+Other datetime_funcs: add_months,
+date_add, date_sub,
+datediff, dayofmonth,
+dayofyear, from_unixtime,
+from_utc_timestamp, 
hour,
+last_day, minute,
+months_between, month,
+next_day, quarter,
+second, to_date,
+to_utc_timestamp,
+unix_timestamp, weekofyear,
+window, year
+
+
+
+Examples
+
+## Not run: date_format(df$t, MM/dd/yyy)
+
+
+
+[Package SparkR version 2.1.2 
Index]
+

http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/R/date_sub.html
--
diff --git a/site/docs/2.1.2/api/R/date_sub.html 
b/site/docs/2.1.2/api/R/date_sub.html
new file mode 100644
index 000..89d6661
--- /dev/null
+++ b/site/docs/2.1.2/api/R/date_sub.html
@@ -0,0 +1,75 @@
+http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd;>http://www.w3.org/1999/xhtml;>R: date_sub
+
+
+
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css;>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js";>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js";>
+hljs.initHighlightingOnLoad();
+
+
+date_sub 
{SparkR}R Documentation
+
+date_sub
+
+Description
+
+Returns the date that is x days before
+
+
+
+Usage
+
+
+date_sub(y, x)
+
+## S4 method for signature 'Column,numeric'
+date_sub(y, x)
+
+
+
+Arguments
+
+
+y
+
+Column to compute on
+
+x
+
+Number of days to substract
+
+
+
+
+Note
+
+date_sub since 1.5.0
+
+
+
+See Also
+
+Other datetime_funcs: add_months,
+date_add, date_format,
+datediff, dayofmonth,
+dayofyear, from_unixtime,
+from_utc_timestamp, 
hour,
+last_day, minute,
+months_between, month,
+next_day, quarter,
+second, to_date,
+to_utc_timestamp,
+unix_timestamp, weekofyear,
+window, year
+
+
+
+Examples
+
+## Not run: date_sub(df$d, 1)
+
+
+
+[Package SparkR version 2.1.2 
Index]
+

http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/R/datediff.html
--
diff --git a/site/docs/2.1.2/api/R/datediff.html 
b/site/docs/2.1.2/api/R/datediff.html
new file mode 100644
index 000..78ff84c
--- /dev/null
+++ b/site/docs/2.1.2/api/R/datediff.html
@@ -0,0 +1,75 @@
+http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd;>http://www.w3.org/1999/xhtml;>R: datediff
+
+
+
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css;>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js";>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js";>
+hljs.initHighlightingOnLoad();
+
+
+datediff 
{SparkR}R Documentation
+
+datediff
+
+Description
+
+Returns the number of days from start to end.
+
+
+
+Usage
+
+
+datediff(y, x)
+
+## S4 method for signature 'Column'
+datediff(y, x)
+
+
+
+Arguments
+
+
+y
+
+end Column to use.
+
+x
+
+start Column to use.
+
+
+
+
+Note
+
+datediff since 1.5.0
+
+
+
+See Also
+
+Other datetime_funcs: add_months,
+date_add, date_format,
+date_sub, dayofmonth,
+dayofyear, from_unixtime,
+from_utc_timestamp, 
hour,
+last_day, minute,
+months_between, month,
+next_day, quarter,
+second, to_date,
+to_utc_timestamp,
+unix_timestamp, weekofyear,
+window, year
+
+
+
+Examples
+
+## Not run: datediff(df$c, x)
+
+
+
+[Package SparkR version 2.1.2 
Index]
+

http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/R/dayofmonth.html

[03/51] [partial] spark-website git commit: Add 2.1.2 docs

2017-10-17 Thread holden
http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/java/org/apache/spark/api/java/function/VoidFunction.html
--
diff --git 
a/site/docs/2.1.2/api/java/org/apache/spark/api/java/function/VoidFunction.html 
b/site/docs/2.1.2/api/java/org/apache/spark/api/java/function/VoidFunction.html
new file mode 100644
index 000..8153556
--- /dev/null
+++ 
b/site/docs/2.1.2/api/java/org/apache/spark/api/java/function/VoidFunction.html
@@ -0,0 +1,219 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+
+
+VoidFunction (Spark 2.1.2 JavaDoc)
+
+
+
+
+
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+
+
+
+Overview
+Package
+Class
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev Class
+Next Class
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+
+
+Detail:
+Field|
+Constr|
+Method
+
+
+
+
+
+
+
+
+org.apache.spark.api.java.function
+Interface 
VoidFunctionT
+
+
+
+
+
+
+All Superinterfaces:
+java.io.Serializable
+
+
+
+public interface VoidFunctionT
+extends java.io.Serializable
+A function with no return value.
+
+
+
+
+
+
+
+
+
+
+
+Method Summary
+
+Methods
+
+Modifier and Type
+Method and Description
+
+
+void
+call(Tt)
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+Method Detail
+
+
+
+
+
+
+
+call
+voidcall(Tt)
+  throws Exception
+Throws:
+Exception
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+Overview
+Package
+Class
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev Class
+Next Class
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+
+
+Detail:
+Field|
+Constr|
+Method
+
+
+
+
+
+
+
+

http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/java/org/apache/spark/api/java/function/VoidFunction2.html
--
diff --git 
a/site/docs/2.1.2/api/java/org/apache/spark/api/java/function/VoidFunction2.html
 
b/site/docs/2.1.2/api/java/org/apache/spark/api/java/function/VoidFunction2.html
new file mode 100644
index 000..1dad476
--- /dev/null
+++ 
b/site/docs/2.1.2/api/java/org/apache/spark/api/java/function/VoidFunction2.html
@@ -0,0 +1,221 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+
+
+VoidFunction2 (Spark 2.1.2 JavaDoc)
+
+
+
+
+
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+
+
+
+Overview
+Package
+Class
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev Class
+Next Class
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+
+
+Detail:
+Field|
+Constr|
+Method
+
+
+
+
+
+
+
+
+org.apache.spark.api.java.function
+Interface 
VoidFunction2T1,T2
+
+
+
+
+
+
+All Superinterfaces:
+java.io.Serializable
+
+
+
+public interface VoidFunction2T1,T2
+extends java.io.Serializable
+A two-argument function that takes arguments of type T1 and 
T2 with no return value.
+
+
+
+
+
+
+
+
+
+
+
+Method Summary
+
+Methods
+
+Modifier and Type
+Method and Description
+
+
+void
+call(T1v1,
+T2v2)
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+Method Detail
+
+
+
+
+
+
+
+call
+voidcall(T1v1,
+T2v2)
+  throws Exception
+Throws:
+Exception
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+Overview
+Package
+Class
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev Class
+Next Class
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+
+
+Detail:
+Field|
+Constr|
+Method
+
+
+
+
+
+
+
+

http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/java/org/apache/spark/api/java/function/package-frame.html
--
diff --git 

[35/51] [partial] spark-website git commit: Add 2.1.2 docs

2017-10-17 Thread holden
http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/R/window.html
--
diff --git a/site/docs/2.1.2/api/R/window.html 
b/site/docs/2.1.2/api/R/window.html
new file mode 100644
index 000..fc41b25
--- /dev/null
+++ b/site/docs/2.1.2/api/R/window.html
@@ -0,0 +1,122 @@
+http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd;>http://www.w3.org/1999/xhtml;>R: window
+
+
+
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css;>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js";>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js";>
+hljs.initHighlightingOnLoad();
+
+
+window 
{SparkR}R Documentation
+
+window
+
+Description
+
+Bucketize rows into one or more time windows given a timestamp specifying 
column. Window
+starts are inclusive but the window ends are exclusive, e.g. 12:05 will be in 
the window
+[12:05,12:10) but not in [12:00,12:05). Windows can support microsecond 
precision. Windows in
+the order of months are not supported.
+
+
+
+Usage
+
+
+window(x, ...)
+
+## S4 method for signature 'Column'
+window(x, windowDuration, slideDuration = NULL,
+  startTime = NULL)
+
+
+
+Arguments
+
+
+x
+
+a time Column. Must be of TimestampType.
+
+...
+
+further arguments to be passed to or from other methods.
+
+windowDuration
+
+a string specifying the width of the window, e.g. '1 second',
+'1 day 12 hours', '2 minutes'. Valid interval strings are 'week',
+'day', 'hour', 'minute', 'second', 'millisecond', 'microsecond'. Note that
+the duration is a fixed length of time, and does not vary over time
+according to a calendar. For example, '1 day' always means 86,400,000
+milliseconds, not a calendar day.
+
+slideDuration
+
+a string specifying the sliding interval of the window. Same format as
+windowDuration. A new window will be generated every
+slideDuration. Must be less than or equal to
+the windowDuration. This duration is likewise absolute, and does 
not
+vary according to a calendar.
+
+startTime
+
+the offset with respect to 1970-01-01 00:00:00 UTC with which to start
+window intervals. For example, in order to have hourly tumbling windows
+that start 15 minutes past the hour, e.g. 12:15-13:15, 13:15-14:15... provide
+startTime as "15 minutes".
+
+
+
+
+Value
+
+An output column of struct called 'window' by default with the nested 
columns 'start'
+and 'end'.
+
+
+
+Note
+
+window since 2.0.0
+
+
+
+See Also
+
+Other datetime_funcs: add_months,
+date_add, date_format,
+date_sub, datediff,
+dayofmonth, dayofyear,
+from_unixtime,
+from_utc_timestamp, 
hour,
+last_day, minute,
+months_between, month,
+next_day, quarter,
+second, to_date,
+to_utc_timestamp,
+unix_timestamp, weekofyear,
+year
+
+
+
+Examples
+
+## Not run: 
+##D   # One minute windows every 15 seconds 10 seconds after the minute, e.g. 
09:00:10-09:01:10,
+##D   # 09:00:25-09:01:25, 09:00:40-09:01:40, ...
+##D   window(df$time, 1 minute, 15 seconds, 10 
seconds)
+##D 
+##D   # One minute tumbling windows 15 seconds after the minute, e.g. 
09:00:15-09:01:15,
+##D# 09:01:15-09:02:15...
+##D   window(df$time, 1 minute, startTime = 15 seconds)
+##D 
+##D   # Thirty-second windows every 10 seconds, e.g. 09:00:00-09:00:30, 
09:00:10-09:00:40, ...
+##D   window(df$time, 30 seconds, 10 seconds)
+## End(Not run)
+
+
+
+[Package SparkR version 2.1.2 
Index]
+

http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/R/windowOrderBy.html
--
diff --git a/site/docs/2.1.2/api/R/windowOrderBy.html 
b/site/docs/2.1.2/api/R/windowOrderBy.html
new file mode 100644
index 000..19a48fe
--- /dev/null
+++ b/site/docs/2.1.2/api/R/windowOrderBy.html
@@ -0,0 +1,71 @@
+http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd;>http://www.w3.org/1999/xhtml;>R: windowOrderBy
+
+
+
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css;>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js";>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js";>
+hljs.initHighlightingOnLoad();
+
+
+windowOrderBy {SparkR}R 
Documentation
+
+windowOrderBy
+
+Description
+
+Creates a WindowSpec with the ordering defined.
+
+
+
+Usage
+
+
+windowOrderBy(col, ...)
+
+## S4 method for signature 'character'
+windowOrderBy(col, ...)
+
+## S4 method for signature 'Column'
+windowOrderBy(col, ...)
+
+
+
+Arguments
+
+
+col
+
+A column name or Column by which rows are ordered within
+windows.
+
+...
+
+Optional column names or Columns in addition to col, by
+which rows are ordered within windows.
+
+
+
+
+Note
+
+windowOrderBy(character) since 2.0.0
+
+windowOrderBy(Column) since 2.0.0
+
+
+
+Examples
+
+## Not run: 
+##D   ws - windowOrderBy(key1, key2)
+##D   df1 - select(df, over(lead(value, 1), ws))
+##D 
+##D   ws - 

[28/51] [partial] spark-website git commit: Add 2.1.2 docs

2017-10-17 Thread holden
http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/java/org/apache/spark/Accumulable.html
--
diff --git a/site/docs/2.1.2/api/java/org/apache/spark/Accumulable.html 
b/site/docs/2.1.2/api/java/org/apache/spark/Accumulable.html
new file mode 100644
index 000..a54423b
--- /dev/null
+++ b/site/docs/2.1.2/api/java/org/apache/spark/Accumulable.html
@@ -0,0 +1,460 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+
+
+Accumulable (Spark 2.1.2 JavaDoc)
+
+
+
+
+
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+
+
+
+Overview
+Package
+Class
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev Class
+Next Class
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+
+
+Detail:
+Field|
+Constr|
+Method
+
+
+
+
+
+
+
+
+org.apache.spark
+Class AccumulableR,T
+
+
+
+Object
+
+
+org.apache.spark.AccumulableR,T
+
+
+
+
+
+
+
+All Implemented Interfaces:
+java.io.Serializable
+
+
+Direct Known Subclasses:
+Accumulator
+
+
+Deprecated.
+use AccumulatorV2. Since 2.0.0.
+
+
+public class AccumulableR,T
+extends Object
+implements java.io.Serializable
+A data type that can be accumulated, i.e. has a commutative 
and associative "add" operation,
+ but where the result type, R, may be different from the element 
type being added, T.
+ 
+ You must define how to add data, and how to merge two of these together.  For 
some data types,
+ such as a counter, these might be the same operation. In that case, you can 
use the simpler
+ Accumulator. They won't always be the same, 
though -- e.g., imagine you are
+ accumulating a set. You will add items to the set, and you will union two 
sets together.
+ 
+ Operations are not thread-safe.
+ 
+ param:  id ID of this accumulator; for internal use only.
+ param:  initialValue initial value of accumulator
+ param:  param helper object defining how to add elements of type 
R and T
+ param:  name human-readable name for use in Spark's web UI
+ param:  countFailedValues whether to accumulate values from failed tasks. 
This is set to true
+  for system and time metrics like serialization time 
or bytes spilled,
+  and false for things with absolute values like 
number of input rows.
+  This should be used for internal metrics only.
+See Also:Serialized 
Form
+
+
+
+
+
+
+
+
+
+
+
+Constructor Summary
+
+Constructors
+
+Constructor and Description
+
+
+Accumulable(RinitialValue,
+   AccumulableParamR,Tparam)
+Deprecated.
+
+
+
+
+
+
+
+
+
+
+Method Summary
+
+Methods
+
+Modifier and Type
+Method and Description
+
+
+void
+add(Tterm)
+Deprecated.
+Add more data to this accumulator / accumulable
+
+
+
+long
+id()
+Deprecated.
+
+
+
+R
+localValue()
+Deprecated.
+Get the current value of this accumulator from within a 
task.
+
+
+
+void
+merge(Rterm)
+Deprecated.
+Merge two accumulable objects together
+
+
+
+scala.OptionString
+name()
+Deprecated.
+
+
+
+void
+setValue(RnewValue)
+Deprecated.
+Set the accumulator's value.
+
+
+
+String
+toString()
+Deprecated.
+
+
+
+R
+value()
+Deprecated.
+Access the accumulator's current value; only allowed on 
driver.
+
+
+
+R
+zero()
+Deprecated.
+
+
+
+
+
+
+
+Methods inherited from classObject
+equals, getClass, hashCode, notify, notifyAll, wait, wait, 
wait
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+Constructor Detail
+
+
+
+
+
+
+
+Accumulable
+publicAccumulable(RinitialValue,
+   AccumulableParamR,Tparam)
+Deprecated.
+
+
+
+
+
+
+
+
+
+Method Detail
+
+
+
+
+
+id
+publiclongid()
+Deprecated.
+
+
+
+
+
+
+
+name
+publicscala.OptionStringname()
+Deprecated.
+
+
+
+
+
+
+
+zero
+publicRzero()
+Deprecated.
+
+
+
+
+
+
+
+
+
+add
+publicvoidadd(Tterm)
+Deprecated.
+Add more data to this accumulator / accumulable
+Parameters:term - 
the data to add
+
+
+
+
+
+
+
+
+
+merge
+publicvoidmerge(Rterm)
+Deprecated.
+Merge two accumulable objects together
+ 
+ Normally, a user will not want to use this version, but will instead call 
add.
+Parameters:term - 
the other R that will get merged with this
+
+
+
+
+
+
+
+value
+publicRvalue()
+Deprecated.
+Access the accumulator's current value; only allowed on 
driver.
+Returns:(undocumented)
+
+
+
+
+
+
+
+localValue
+publicRlocalValue()
+Deprecated.
+Get the current value of this accumulator from within a 
task.
+ 
+ This is NOT the global value of the accumulator.  To get the global value 
after a
+ completed operation on the dataset, call value.
+ 
+ The typical use of this method is to 

[43/51] [partial] spark-website git commit: Add 2.1.2 docs

2017-10-17 Thread holden
http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/R/hex.html
--
diff --git a/site/docs/2.1.2/api/R/hex.html b/site/docs/2.1.2/api/R/hex.html
new file mode 100644
index 000..ee4955b
--- /dev/null
+++ b/site/docs/2.1.2/api/R/hex.html
@@ -0,0 +1,79 @@
+http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd;>http://www.w3.org/1999/xhtml;>R: hex
+
+
+
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css;>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js";>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js";>
+hljs.initHighlightingOnLoad();
+
+
+hex 
{SparkR}R Documentation
+
+hex
+
+Description
+
+Computes hex value of the given column.
+
+
+
+Usage
+
+
+hex(x)
+
+## S4 method for signature 'Column'
+hex(x)
+
+
+
+Arguments
+
+
+x
+
+Column to compute on.
+
+
+
+
+Note
+
+hex since 1.5.0
+
+
+
+See Also
+
+Other math_funcs: acos, asin,
+atan2, atan,
+bin, bround,
+cbrt, ceil,
+conv, corr,
+cosh, cos,
+covar_pop, cov,
+expm1, exp,
+factorial, floor,
+hypot, log10,
+log1p, log2,
+log, pmod,
+rint, round,
+shiftLeft,
+shiftRightUnsigned,
+shiftRight, signum,
+sinh, sin,
+sqrt, tanh,
+tan, toDegrees,
+toRadians, unhex
+
+
+
+Examples
+
+## Not run: hex(df$c)
+
+
+
+[Package SparkR version 2.1.2 
Index]
+

http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/R/histogram.html
--
diff --git a/site/docs/2.1.2/api/R/histogram.html 
b/site/docs/2.1.2/api/R/histogram.html
new file mode 100644
index 000..9aedec9
--- /dev/null
+++ b/site/docs/2.1.2/api/R/histogram.html
@@ -0,0 +1,121 @@
+http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd;>http://www.w3.org/1999/xhtml;>R: Compute histogram 
statistics for given column
+
+
+
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css;>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js";>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js";>
+hljs.initHighlightingOnLoad();
+
+
+histogram 
{SparkR}R Documentation
+
+Compute histogram statistics for given column
+
+Description
+
+This function computes a histogram for a given SparkR Column.
+
+
+
+Usage
+
+
+## S4 method for signature 'SparkDataFrame,characterOrColumn'
+histogram(df, col, nbins = 10)
+
+
+
+Arguments
+
+
+df
+
+the SparkDataFrame containing the Column to build the histogram from.
+
+col
+
+the column as Character string or a Column to build the histogram from.
+
+nbins
+
+the number of bins (optional). Default value is 10.
+
+
+
+
+Value
+
+a data.frame with the histogram statistics, i.e., counts and centroids.
+
+
+
+Note
+
+histogram since 2.0.0
+
+
+
+See Also
+
+Other SparkDataFrame functions: SparkDataFrame-class,
+agg, arrange,
+as.data.frame, attach,
+cache, coalesce,
+collect, colnames,
+coltypes,
+createOrReplaceTempView,
+crossJoin, dapplyCollect,
+dapply, describe,
+dim, distinct,
+dropDuplicates, dropna,
+drop, dtypes,
+except, explain,
+filter, first,
+gapplyCollect, gapply,
+getNumPartitions, group_by,
+head, insertInto,
+intersect, isLocal,
+join, limit,
+merge, mutate,
+ncol, nrow,
+persist, printSchema,
+randomSplit, rbind,
+registerTempTable, rename,
+repartition, sample,
+saveAsTable, schema,
+selectExpr, select,
+showDF, show,
+storageLevel, str,
+subset, take,
+union, unpersist,
+withColumn, with,
+write.df, write.jdbc,
+write.json, write.orc,
+write.parquet, write.text
+
+
+
+Examples
+
+## Not run: 
+##D 
+##D # Create a SparkDataFrame from the Iris dataset
+##D irisDF - createDataFrame(iris)
+##D 
+##D # Compute histogram statistics
+##D histStats - histogram(irisDF, irisDF$Sepal_Length, nbins = 12)
+##D 
+##D # Once SparkR has computed the histogram statistics, the histogram can be
+##D # rendered using the ggplot2 library:
+##D 
+##D require(ggplot2)
+##D plot - ggplot(histStats, aes(x = centroids, y = counts)) +
+##D geom_bar(stat = identity) +
+##D xlab(Sepal_Length) + ylab(Frequency)
+## End(Not run)
+
+
+
+[Package SparkR version 2.1.2 
Index]
+

http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/R/hour.html
--
diff --git a/site/docs/2.1.2/api/R/hour.html b/site/docs/2.1.2/api/R/hour.html
new file mode 100644
index 000..9331aff
--- /dev/null
+++ b/site/docs/2.1.2/api/R/hour.html
@@ -0,0 +1,71 @@
+http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd;>http://www.w3.org/1999/xhtml;>R: hour
+
+
+
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css;>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js";>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js";>
+hljs.initHighlightingOnLoad();
+
+
+hour 

[12/51] [partial] spark-website git commit: Add 2.1.2 docs

2017-10-17 Thread holden
http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/java/org/apache/spark/api/java/JavaFutureAction.html
--
diff --git 
a/site/docs/2.1.2/api/java/org/apache/spark/api/java/JavaFutureAction.html 
b/site/docs/2.1.2/api/java/org/apache/spark/api/java/JavaFutureAction.html
new file mode 100644
index 000..073fb95
--- /dev/null
+++ b/site/docs/2.1.2/api/java/org/apache/spark/api/java/JavaFutureAction.html
@@ -0,0 +1,226 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+
+
+JavaFutureAction (Spark 2.1.2 JavaDoc)
+
+
+
+
+
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+
+
+
+Overview
+Package
+Class
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev 
Class
+Next 
Class
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+
+
+Detail:
+Field|
+Constr|
+Method
+
+
+
+
+
+
+
+
+org.apache.spark.api.java
+Interface 
JavaFutureActionT
+
+
+
+
+
+
+All Superinterfaces:
+java.util.concurrent.FutureT
+
+
+
+public interface JavaFutureActionT
+extends java.util.concurrent.FutureT
+
+
+
+
+
+
+
+
+
+
+
+Method Summary
+
+Methods
+
+Modifier and Type
+Method and Description
+
+
+java.util.ListInteger
+jobIds()
+Returns the job IDs run by the underlying async 
operation.
+
+
+
+
+
+
+
+Methods inherited from interfacejava.util.concurrent.Future
+cancel, get, get, isCancelled, isDone
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+Method Detail
+
+
+
+
+
+jobIds
+java.util.ListIntegerjobIds()
+Returns the job IDs run by the underlying async operation.
+
+ This returns the current snapshot of the job list. Certain operations may run 
multiple
+ jobs, so multiple calls to this method may return different lists.
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+Overview
+Package
+Class
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev 
Class
+Next 
Class
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+
+
+Detail:
+Field|
+Constr|
+Method
+
+
+
+
+
+
+
+

http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/java/org/apache/spark/api/java/JavaHadoopRDD.html
--
diff --git 
a/site/docs/2.1.2/api/java/org/apache/spark/api/java/JavaHadoopRDD.html 
b/site/docs/2.1.2/api/java/org/apache/spark/api/java/JavaHadoopRDD.html
new file mode 100644
index 000..60fc727
--- /dev/null
+++ b/site/docs/2.1.2/api/java/org/apache/spark/api/java/JavaHadoopRDD.html
@@ -0,0 +1,325 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+
+
+JavaHadoopRDD (Spark 2.1.2 JavaDoc)
+
+
+
+
+
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+
+
+
+Overview
+Package
+Class
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev 
Class
+Next 
Class
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+
+
+Detail:
+Field|
+Constr|
+Method
+
+
+
+
+
+
+
+
+org.apache.spark.api.java
+Class 
JavaHadoopRDDK,V
+
+
+
+Object
+
+
+org.apache.spark.api.java.JavaPairRDDK,V
+
+
+org.apache.spark.api.java.JavaHadoopRDDK,V
+
+
+
+
+
+
+
+
+
+All Implemented Interfaces:
+java.io.Serializable, JavaRDDLikescala.Tuple2K,V,JavaPairRDDK,V
+
+
+
+public class JavaHadoopRDDK,V
+extends JavaPairRDDK,V
+See Also:Serialized
 Form
+
+
+
+
+
+
+
+
+
+
+
+Constructor Summary
+
+Constructors
+
+Constructor and Description
+
+
+JavaHadoopRDD(HadoopRDDK,Vrdd,
+ scala.reflect.ClassTagKkClassTag,
+ scala.reflect.ClassTagVvClassTag)
+
+
+
+
+
+
+
+
+
+Method Summary
+
+Methods
+
+Modifier and Type
+Method and Description
+
+
+scala.reflect.ClassTagK
+kClassTag()
+
+
+RJavaRDDR
+mapPartitionsWithInputSplit(Function2org.apache.hadoop.mapred.InputSplit,java.util.Iteratorscala.Tuple2K,V,java.util.IteratorRf,
+   booleanpreservesPartitioning)
+Maps over a partition, providing the InputSplit that was 
used as the base of the 

[40/51] [partial] spark-website git commit: Add 2.1.2 docs

2017-10-17 Thread holden
http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/R/read.df.html
--
diff --git a/site/docs/2.1.2/api/R/read.df.html 
b/site/docs/2.1.2/api/R/read.df.html
new file mode 100644
index 000..fa64d11
--- /dev/null
+++ b/site/docs/2.1.2/api/R/read.df.html
@@ -0,0 +1,97 @@
+http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd;>http://www.w3.org/1999/xhtml;>R: Load a 
SparkDataFrame
+
+
+
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css;>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js";>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js";>
+hljs.initHighlightingOnLoad();
+
+
+read.df 
{SparkR}R Documentation
+
+Load a SparkDataFrame
+
+Description
+
+Returns the dataset in a data source as a SparkDataFrame
+
+
+
+Usage
+
+
+## Default S3 method:
+read.df(path = NULL, source = NULL, schema = NULL,
+  na.strings = "NA", ...)
+
+## Default S3 method:
+loadDF(path = NULL, source = NULL, schema = NULL, ...)
+
+
+
+Arguments
+
+
+path
+
+The path of files to load
+
+source
+
+The name of external data source
+
+schema
+
+The data schema defined in structType
+
+na.strings
+
+Default string value for NA when source is csv
+
+...
+
+additional external data source specific named properties.
+
+
+
+
+Details
+
+The data source is specified by the source and a set of 
options(...).
+If source is not specified, the default data source configured by
+spark.sql.sources.default will be used. 
+Similar to R read.csv, when source is csv, by 
default, a value of NA will be
+interpreted as NA.
+
+
+
+Value
+
+SparkDataFrame
+
+
+
+Note
+
+read.df since 1.4.0
+
+loadDF since 1.6.0
+
+
+
+Examples
+
+## Not run: 
+##D sparkR.session()
+##D df1 - read.df(path/to/file.json, source = json)
+##D schema - structType(structField(name, string),
+##D  structField(info, 
mapstring,double))
+##D df2 - read.df(mapTypeJsonPath, json, schema)
+##D df3 - loadDF(data/test_table, parquet, 
mergeSchema = true)
+## End(Not run)
+
+
+
+[Package SparkR version 2.1.2 
Index]
+

http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/R/read.jdbc.html
--
diff --git a/site/docs/2.1.2/api/R/read.jdbc.html 
b/site/docs/2.1.2/api/R/read.jdbc.html
new file mode 100644
index 000..f527c5b
--- /dev/null
+++ b/site/docs/2.1.2/api/R/read.jdbc.html
@@ -0,0 +1,105 @@
+http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd;>http://www.w3.org/1999/xhtml;>R: Create a SparkDataFrame 
representing the database table...
+
+
+
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css;>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js";>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js";>
+hljs.initHighlightingOnLoad();
+
+
+read.jdbc 
{SparkR}R Documentation
+
+Create a SparkDataFrame representing the database table accessible via 
JDBC URL
+
+Description
+
+Additional JDBC database connection properties can be set (...)
+
+
+
+Usage
+
+
+read.jdbc(url, tableName, partitionColumn = NULL, lowerBound = NULL,
+  upperBound = NULL, numPartitions = 0L, predicates = list(), ...)
+
+
+
+Arguments
+
+
+url
+
+JDBC database url of the form jdbc:subprotocol:subname
+
+tableName
+
+the name of the table in the external database
+
+partitionColumn
+
+the name of a column of integral type that will be used for partitioning
+
+lowerBound
+
+the minimum value of partitionColumn used to decide partition 
stride
+
+upperBound
+
+the maximum value of partitionColumn used to decide partition 
stride
+
+numPartitions
+
+the number of partitions, This, along with lowerBound 
(inclusive),
+upperBound (exclusive), form partition strides for generated WHERE
+clause expressions used to split the column partitionColumn 
evenly.
+This defaults to SparkContext.defaultParallelism when unset.
+
+predicates
+
+a list of conditions in the where clause; each one defines one partition
+
+...
+
+additional JDBC database connection named properties.
+
+
+
+
+Details
+
+Only one of partitionColumn or predicates should be set. Partitions of the 
table will be
+retrieved in parallel based on the numPartitions or by the 
predicates.
+
+Don't create too many partitions in parallel on a large cluster; otherwise 
Spark might crash
+your external database systems.
+
+
+
+Value
+
+SparkDataFrame
+
+
+
+Note
+
+read.jdbc since 2.0.0
+
+
+
+Examples
+
+## Not run: 
+##D sparkR.session()
+##D jdbcUrl - jdbc:mysql://localhost:3306/databasename
+##D df - read.jdbc(jdbcUrl, table, predicates = 
list(field=123), user = username)
+##D df2 - read.jdbc(jdbcUrl, table2, partitionColumn = 
index, lowerBound = 0,
+##D  upperBound = 1, user = username, password 
= password)
+## End(Not run)
+
+
+
+[Package 

[34/51] [partial] spark-website git commit: Add 2.1.2 docs

2017-10-17 Thread holden
http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/java/allclasses-frame.html
--
diff --git a/site/docs/2.1.2/api/java/allclasses-frame.html 
b/site/docs/2.1.2/api/java/allclasses-frame.html
new file mode 100644
index 000..90f5443
--- /dev/null
+++ b/site/docs/2.1.2/api/java/allclasses-frame.html
@@ -0,0 +1,1138 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+
+
+All Classes (Spark 2.1.2 JavaDoc)
+
+
+
+
+All Classes
+
+
+AbsoluteError
+Accumulable
+AccumulableInfo
+AccumulableInfo
+AccumulableParam
+Accumulator
+AccumulatorContext
+AccumulatorParam
+AccumulatorParam.DoubleAccumulatorParam$
+AccumulatorParam.FloatAccumulatorParam$
+AccumulatorParam.IntAccumulatorParam$
+AccumulatorParam.LongAccumulatorParam$
+AccumulatorParam.StringAccumulatorParam$
+AccumulatorV2
+AFTAggregator
+AFTCostFun
+AFTSurvivalRegression
+AFTSurvivalRegressionModel
+AggregatedDialect
+AggregatingEdgeContext
+Aggregator
+Aggregator
+Algo
+AllJobsCancelled
+AllReceiverIds
+ALS
+ALS
+ALS.InBlock$
+ALS.Rating
+ALS.Rating$
+ALS.RatingBlock$
+ALSModel
+AnalysisException
+And
+AnyDataType
+ApplicationAttemptInfo
+ApplicationInfo
+ApplicationsListResource
+ApplicationStatus
+ApplyInPlace
+AreaUnderCurve
+ArrayType
+AskPermissionToCommitOutput
+AssociationRules
+AssociationRules.Rule
+AsyncRDDActions
+Attribute
+AttributeGroup
+AttributeKeys
+AttributeType
+BaseRelation
+BaseRRDD
+BatchInfo
+BernoulliCellSampler
+BernoulliSampler
+Binarizer
+BinaryAttribute
+BinaryClassificationEvaluator
+BinaryClassificationMetrics
+BinaryLogisticRegressionSummary
+BinaryLogisticRegressionTrainingSummary
+BinarySample
+BinaryType
+BinomialBounds
+BisectingKMeans
+BisectingKMeans
+BisectingKMeansModel
+BisectingKMeansModel
+BisectingKMeansModel.SaveLoadV1_0$
+BisectingKMeansSummary
+BlacklistTracker
+BLAS
+BLAS
+BlockId
+BlockManagerId
+BlockManagerMessages
+BlockManagerMessages.BlockManagerHeartbeat
+BlockManagerMessages.BlockManagerHeartbeat$
+BlockManagerMessages.GetBlockStatus
+BlockManagerMessages.GetBlockStatus$
+BlockManagerMessages.GetExecutorEndpointRef
+BlockManagerMessages.GetExecutorEndpointRef$
+BlockManagerMessages.GetLocations
+BlockManagerMessages.GetLocations$
+BlockManagerMessages.GetLocationsMultipleBlockIds
+BlockManagerMessages.GetLocationsMultipleBlockIds$
+BlockManagerMessages.GetMatchingBlockIds
+BlockManagerMessages.GetMatchingBlockIds$
+BlockManagerMessages.GetMemoryStatus$
+BlockManagerMessages.GetPeers
+BlockManagerMessages.GetPeers$
+BlockManagerMessages.GetStorageStatus$
+BlockManagerMessages.HasCachedBlocks
+BlockManagerMessages.HasCachedBlocks$
+BlockManagerMessages.RegisterBlockManager
+BlockManagerMessages.RegisterBlockManager$
+BlockManagerMessages.RemoveBlock
+BlockManagerMessages.RemoveBlock$
+BlockManagerMessages.RemoveBroadcast
+BlockManagerMessages.RemoveBroadcast$
+BlockManagerMessages.RemoveExecutor
+BlockManagerMessages.RemoveExecutor$
+BlockManagerMessages.RemoveRdd
+BlockManagerMessages.RemoveRdd$
+BlockManagerMessages.RemoveShuffle
+BlockManagerMessages.RemoveShuffle$
+BlockManagerMessages.StopBlockManagerMaster$
+BlockManagerMessages.ToBlockManagerMaster
+BlockManagerMessages.ToBlockManagerSlave
+BlockManagerMessages.TriggerThreadDump$
+BlockManagerMessages.UpdateBlockInfo
+BlockManagerMessages.UpdateBlockInfo$
+BlockMatrix
+BlockNotFoundException
+BlockReplicationPolicy
+BlockStatus
+BlockUpdatedInfo
+BloomFilter
+BloomFilter.Version
+BooleanParam
+BooleanType
+BoostingStrategy
+BoundedDouble
+BreezeUtil
+Broadcast
+BroadcastBlockId
+Broker
+BucketedRandomProjectionLSH
+BucketedRandomProjectionLSHModel
+Bucketizer
+BufferReleasingInputStream
+BytecodeUtils
+ByteType
+CalendarIntervalType
+Catalog
+CatalystScan
+CategoricalSplit
+CausedBy
+CharType
+CheckpointReader
+CheckpointState
+ChiSqSelector
+ChiSqSelector
+ChiSqSelectorModel
+ChiSqSelectorModel
+ChiSqSelectorModel.SaveLoadV1_0$
+ChiSqTest
+ChiSqTest.Method
+ChiSqTest.Method$
+ChiSqTest.NullHypothesis$
+ChiSqTestResult
+CholeskyDecomposition
+ChunkedByteBufferInputStream
+ClassificationModel
+ClassificationModel
+Classifier
+CleanAccum
+CleanBroadcast
+CleanCheckpoint
+CleanRDD
+CleanShuffle
+CleanupTask
+CleanupTaskWeakReference
+ClosureCleaner
+ClusteringSummary
+CoarseGrainedClusterMessages
+CoarseGrainedClusterMessages.AddWebUIFilter
+CoarseGrainedClusterMessages.AddWebUIFilter$
+CoarseGrainedClusterMessages.GetExecutorLossReason
+CoarseGrainedClusterMessages.GetExecutorLossReason$
+CoarseGrainedClusterMessages.KillExecutors
+CoarseGrainedClusterMessages.KillExecutors$
+CoarseGrainedClusterMessages.KillTask
+CoarseGrainedClusterMessages.KillTask$
+CoarseGrainedClusterMessages.LaunchTask
+CoarseGrainedClusterMessages.LaunchTask$
+CoarseGrainedClusterMessages.RegisterClusterManager
+CoarseGrainedClusterMessages.RegisterClusterManager$
+CoarseGrainedClusterMessages.RegisteredExecutor$
+CoarseGrainedClusterMessages.RegisterExecutor

[18/51] [partial] spark-website git commit: Add 2.1.2 docs

2017-10-17 Thread holden
http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/java/org/apache/spark/SparkEnv.html
--
diff --git a/site/docs/2.1.2/api/java/org/apache/spark/SparkEnv.html 
b/site/docs/2.1.2/api/java/org/apache/spark/SparkEnv.html
new file mode 100644
index 000..bdb8d4d
--- /dev/null
+++ b/site/docs/2.1.2/api/java/org/apache/spark/SparkEnv.html
@@ -0,0 +1,478 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+
+
+SparkEnv (Spark 2.1.2 JavaDoc)
+
+
+
+
+
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+
+
+
+Overview
+Package
+Class
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev Class
+Next Class
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+
+
+Detail:
+Field|
+Constr|
+Method
+
+
+
+
+
+
+
+
+org.apache.spark
+Class SparkEnv
+
+
+
+Object
+
+
+org.apache.spark.SparkEnv
+
+
+
+
+
+
+
+
+public class SparkEnv
+extends Object
+:: DeveloperApi ::
+ Holds all the runtime environment objects for a running Spark instance 
(either master or worker),
+ including the serializer, RpcEnv, block manager, map output tracker, etc. 
Currently
+ Spark code finds the SparkEnv through a global variable, so all the threads 
can access the same
+ SparkEnv. It can be accessed by SparkEnv.get (e.g. after creating a 
SparkContext).
+ 
+ NOTE: This is not intended for external use. This is exposed for Shark and 
may be made private
+   in a future release.
+
+
+
+
+
+
+
+
+
+
+
+Constructor Summary
+
+Constructors
+
+Constructor and Description
+
+
+SparkEnv(StringexecutorId,
+org.apache.spark.rpc.RpcEnvrpcEnv,
+Serializerserializer,
+SerializerclosureSerializer,
+org.apache.spark.serializer.SerializerManagerserializerManager,
+org.apache.spark.MapOutputTrackermapOutputTracker,
+org.apache.spark.shuffle.ShuffleManagershuffleManager,
+org.apache.spark.broadcast.BroadcastManagerbroadcastManager,
+org.apache.spark.storage.BlockManagerblockManager,
+org.apache.spark.SecurityManagersecurityManager,
+org.apache.spark.metrics.MetricsSystemmetricsSystem,
+org.apache.spark.memory.MemoryManagermemoryManager,
+
org.apache.spark.scheduler.OutputCommitCoordinatoroutputCommitCoordinator,
+SparkConfconf)
+
+
+
+
+
+
+
+
+
+Method Summary
+
+Methods
+
+Modifier and Type
+Method and Description
+
+
+org.apache.spark.storage.BlockManager
+blockManager()
+
+
+org.apache.spark.broadcast.BroadcastManager
+broadcastManager()
+
+
+Serializer
+closureSerializer()
+
+
+SparkConf
+conf()
+
+
+String
+executorId()
+
+
+static SparkEnv
+get()
+Returns the SparkEnv.
+
+
+
+org.apache.spark.MapOutputTracker
+mapOutputTracker()
+
+
+org.apache.spark.memory.MemoryManager
+memoryManager()
+
+
+org.apache.spark.metrics.MetricsSystem
+metricsSystem()
+
+
+org.apache.spark.scheduler.OutputCommitCoordinator
+outputCommitCoordinator()
+
+
+org.apache.spark.SecurityManager
+securityManager()
+
+
+Serializer
+serializer()
+
+
+org.apache.spark.serializer.SerializerManager
+serializerManager()
+
+
+static void
+set(SparkEnve)
+
+
+org.apache.spark.shuffle.ShuffleManager
+shuffleManager()
+
+
+
+
+
+
+Methods inherited from classObject
+equals, getClass, hashCode, notify, notifyAll, toString, wait, wait, 
wait
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+Constructor Detail
+
+
+
+
+
+SparkEnv
+publicSparkEnv(StringexecutorId,
+org.apache.spark.rpc.RpcEnvrpcEnv,
+Serializerserializer,
+SerializerclosureSerializer,
+org.apache.spark.serializer.SerializerManagerserializerManager,
+org.apache.spark.MapOutputTrackermapOutputTracker,
+org.apache.spark.shuffle.ShuffleManagershuffleManager,
+org.apache.spark.broadcast.BroadcastManagerbroadcastManager,
+org.apache.spark.storage.BlockManagerblockManager,
+org.apache.spark.SecurityManagersecurityManager,
+org.apache.spark.metrics.MetricsSystemmetricsSystem,
+org.apache.spark.memory.MemoryManagermemoryManager,
+
org.apache.spark.scheduler.OutputCommitCoordinatoroutputCommitCoordinator,
+SparkConfconf)
+
+
+
+
+
+
+
+
+
+Method Detail
+
+
+
+
+
+set
+public staticvoidset(SparkEnve)
+
+
+
+
+
+
+
+get
+public staticSparkEnvget()
+Returns the SparkEnv.
+Returns:(undocumented)
+
+
+
+
+
+
+
+executorId
+publicStringexecutorId()
+
+
+
+
+
+
+
+serializer
+publicSerializerserializer()
+
+
+
+
+
+
+
+closureSerializer
+publicSerializerclosureSerializer()
+
+
+
+
+
+
+
+serializerManager

[50/51] [partial] spark-website git commit: Add 2.1.2 docs

2017-10-17 Thread holden
http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/README.md
--
diff --git a/site/docs/2.1.2/README.md b/site/docs/2.1.2/README.md
new file mode 100644
index 000..ffd3b57
--- /dev/null
+++ b/site/docs/2.1.2/README.md
@@ -0,0 +1,72 @@
+Welcome to the Spark documentation!
+
+This readme will walk you through navigating and building the Spark 
documentation, which is included
+here with the Spark source code. You can also find documentation specific to 
release versions of
+Spark at http://spark.apache.org/documentation.html.
+
+Read on to learn more about viewing documentation in plain text (i.e., 
markdown) or building the
+documentation yourself. Why build it yourself? So that you have the docs that 
corresponds to
+whichever version of Spark you currently have checked out of revision control.
+
+## Prerequisites
+The Spark documentation build uses a number of tools to build HTML docs and 
API docs in Scala,
+Python and R.
+
+You need to have 
[Ruby](https://www.ruby-lang.org/en/documentation/installation/) and
+[Python](https://docs.python.org/2/using/unix.html#getting-and-installing-the-latest-version-of-python)
+installed. Also install the following libraries:
+```sh
+$ sudo gem install jekyll jekyll-redirect-from pygments.rb
+$ sudo pip install Pygments
+# Following is needed only for generating API docs
+$ sudo pip install sphinx pypandoc
+$ sudo Rscript -e 'install.packages(c("knitr", "devtools", "roxygen2", 
"testthat", "rmarkdown"), repos="http://cran.stat.ucla.edu/;)'
+```
+(Note: If you are on a system with both Ruby 1.9 and Ruby 2.0 you may need to 
replace gem with gem2.0)
+
+## Generating the Documentation HTML
+
+We include the Spark documentation as part of the source (as opposed to using 
a hosted wiki, such as
+the github wiki, as the definitive documentation) to enable the documentation 
to evolve along with
+the source code and be captured by revision control (currently git). This way 
the code automatically
+includes the version of the documentation that is relevant regardless of which 
version or release
+you have checked out or downloaded.
+
+In this directory you will find textfiles formatted using Markdown, with an 
".md" suffix. You can
+read those text files directly if you want. Start with index.md.
+
+Execute `jekyll build` from the `docs/` directory to compile the site. 
Compiling the site with
+Jekyll will create a directory called `_site` containing index.html as well as 
the rest of the
+compiled files.
+
+$ cd docs
+$ jekyll build
+
+You can modify the default Jekyll build as follows:
+```sh
+# Skip generating API docs (which takes a while)
+$ SKIP_API=1 jekyll build
+
+# Serve content locally on port 4000
+$ jekyll serve --watch
+
+# Build the site with extra features used on the live page
+$ PRODUCTION=1 jekyll build
+```
+
+## API Docs (Scaladoc, Sphinx, roxygen2)
+
+You can build just the Spark scaladoc by running `build/sbt unidoc` from the 
SPARK_PROJECT_ROOT directory.
+
+Similarly, you can build just the PySpark docs by running `make html` from the
+SPARK_PROJECT_ROOT/python/docs directory. Documentation is only generated for 
classes that are listed as
+public in `__init__.py`. The SparkR docs can be built by running 
SPARK_PROJECT_ROOT/R/create-docs.sh.
+
+When you run `jekyll` in the `docs` directory, it will also copy over the 
scaladoc for the various
+Spark subprojects into the `docs` directory (and then also into the `_site` 
directory). We use a
+jekyll plugin to run `build/sbt unidoc` before building the site so if you 
haven't run it (recently) it
+may take some time as it generates all of the scaladoc.  The jekyll plugin 
also generates the
+PySpark docs using [Sphinx](http://sphinx-doc.org/).
+
+NOTE: To skip the step of building and copying over the Scala, Python, R API 
docs, run `SKIP_API=1
+jekyll`.

http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api.html
--
diff --git a/site/docs/2.1.2/api.html b/site/docs/2.1.2/api.html
new file mode 100644
index 000..2496122
--- /dev/null
+++ b/site/docs/2.1.2/api.html
@@ -0,0 +1,178 @@
+
+
+
+
+
+  
+
+
+
+Spark API Documentation - Spark 2.1.2 Documentation
+
+
+
+
+
+
+body {
+padding-top: 60px;
+padding-bottom: 40px;
+}
+
+
+
+
+
+
+
+
+
+
+
+
+  var _gaq = _gaq || [];
+  _gaq.push(['_setAccount', 'UA-32518208-2']);
+  _gaq.push(['_trackPageview']);
+
+  (function() {
+var ga = document.createElement('script'); ga.type = 
'text/javascript'; ga.async = true;
+ga.src = ('https:' == 

[02/51] [partial] spark-website git commit: Add 2.1.2 docs

2017-10-17 Thread holden
http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/java/org/apache/spark/api/r/BaseRRDD.html
--
diff --git a/site/docs/2.1.2/api/java/org/apache/spark/api/r/BaseRRDD.html 
b/site/docs/2.1.2/api/java/org/apache/spark/api/r/BaseRRDD.html
new file mode 100644
index 000..29034e8
--- /dev/null
+++ b/site/docs/2.1.2/api/java/org/apache/spark/api/r/BaseRRDD.html
@@ -0,0 +1,330 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+
+
+BaseRRDD (Spark 2.1.2 JavaDoc)
+
+
+
+
+
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+
+
+
+Overview
+Package
+Class
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev Class
+Next 
Class
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+
+
+Detail:
+Field|
+Constr|
+Method
+
+
+
+
+
+
+
+
+org.apache.spark.api.r
+Class BaseRRDDT,U
+
+
+
+Object
+
+
+org.apache.spark.rdd.RDDU
+
+
+org.apache.spark.api.r.BaseRRDDT,U
+
+
+
+
+
+
+
+
+
+All Implemented Interfaces:
+java.io.Serializable
+
+
+Direct Known Subclasses:
+PairwiseRRDD, RRDD, StringRRDD
+
+
+
+public abstract class BaseRRDDT,U
+extends RDDU
+See Also:Serialized
 Form
+
+
+
+
+
+
+
+
+
+
+
+Constructor Summary
+
+Constructors
+
+Constructor and Description
+
+
+BaseRRDD(RDDTparent,
+intnumPartitions,
+byte[]func,
+Stringdeserializer,
+Stringserializer,
+byte[]packageNames,
+BroadcastObject[]broadcastVars,
+scala.reflect.ClassTagTevidence$1,
+scala.reflect.ClassTagUevidence$2)
+
+
+
+
+
+
+
+
+
+Method Summary
+
+Methods
+
+Modifier and Type
+Method and Description
+
+
+scala.collection.IteratorU
+compute(Partitionpartition,
+   TaskContextcontext)
+:: DeveloperApi ::
+ Implemented by subclasses to compute a given partition.
+
+
+
+Partition[]
+getPartitions()
+Implemented by subclasses to return the set of partitions 
in this RDD.
+
+
+
+
+
+
+
+Methods inherited from classorg.apache.spark.rdd.RDD
+aggregate,
 cache, cartesian,
 checkpoint,
 coalesce,
 collect, 
collect,
 context, 
count, countApprox, countApproxDistinct,
 countApproxDistinct,
 countByValue,
 countByValueApprox,
 dependencies,
 distinct, distinct,
 doubleRDDToDoubleRDDFunctions,
 
 filter, first, flatMap,
 fold,
 foreach,
 foreachPartition,
 getCheckpointFile,
 getNumPartitions,
 getStorageLevel,
 glom, groupBy,
 groupBy,
 groupBy,
 id, intersection,
 intersection,
 intersection,
 isCheckpointed,
 isEmpty, 
iterator, keyBy,
 localCheckpoint,
 map,
 mapPartitions,
 mapPartitionsWithIndex,
 max,
 min,
 name, numericRDDToDoubleRDDFunctions, 
partitioner,
 partitions,
 persist, 
persist,
 pipe,
 pipe,
 pipe,
 preferredLocations,
 randomSplit, rddToAsyncRDDActions,
 rddToOrderedRDDFunctions,
 rddToPairRDDFunctions,
 rddToSequenceFileRDDFunctions,
 reduce,
 repartition, sample,
 saveAsObjectFile,
 saveAsTextFile,
 saveAsTextFile,
 setName,
 sortBy,
 sparkContext,
 subtract,
 subtract, subtract,
 take, takeOrdered,
 takeSample,
 toDebugString,
 toJavaRDD, 
toLocalIterator,
 top,
 toString, treeAggregate,
 treeReduce,
 union,
 unpersist,
 zip,
 zipPartitions,
 zipPartitions,
 zipPartitions,
 zipPartitions,
 zipPartitions,
 zipPartitions,
 zipWithIndex,
 zipWithUniqueId
+
+
+
+
+
+Methods inherited from classObject
+equals, getClass, hashCode, notify, notifyAll, wait, wait, 
wait
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+Constructor Detail
+
+
+
+
+
+BaseRRDD
+publicBaseRRDD(RDDTparent,
+intnumPartitions,
+byte[]func,
+Stringdeserializer,
+Stringserializer,
+byte[]packageNames,
+BroadcastObject[]broadcastVars,
+scala.reflect.ClassTagTevidence$1,
+scala.reflect.ClassTagUevidence$2)
+
+
+
+
+
+
+
+
+
+Method Detail
+
+
+
+
+
+getPartitions
+publicPartition[]getPartitions()
+Description copied from class:RDD
+Implemented by subclasses to return the set of partitions 
in this RDD. This method will only
+ be called once, so it is safe to implement a time-consuming computation in it.
+ 
+ The partitions in this array must satisfy the following property:
+   rdd.partitions.zipWithIndex.forall { case (partition, index) = 
partition.index == index }
+Returns:(undocumented)
+
+
+
+
+
+
+
+compute
+publicscala.collection.IteratorUcompute(Partitionpartition,
+   TaskContextcontext)
+Description copied from class:RDD
+:: DeveloperApi ::
+ Implemented by subclasses to compute a given partition.
+
+Specified by:
+computein
 classRDDU
+Parameters:partition - 
(undocumented)context 

[10/51] [partial] spark-website git commit: Add 2.1.2 docs

2017-10-17 Thread holden
http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/java/org/apache/spark/api/java/JavaPairRDD.html
--
diff --git 
a/site/docs/2.1.2/api/java/org/apache/spark/api/java/JavaPairRDD.html 
b/site/docs/2.1.2/api/java/org/apache/spark/api/java/JavaPairRDD.html
new file mode 100644
index 000..ac71f12
--- /dev/null
+++ b/site/docs/2.1.2/api/java/org/apache/spark/api/java/JavaPairRDD.html
@@ -0,0 +1,3597 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+
+
+JavaPairRDD (Spark 2.1.2 JavaDoc)
+
+
+
+
+
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+
+
+
+Overview
+Package
+Class
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev 
Class
+Next 
Class
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+
+
+Detail:
+Field|
+Constr|
+Method
+
+
+
+
+
+
+
+
+org.apache.spark.api.java
+Class JavaPairRDDK,V
+
+
+
+Object
+
+
+org.apache.spark.api.java.JavaPairRDDK,V
+
+
+
+
+
+
+
+All Implemented Interfaces:
+java.io.Serializable, JavaRDDLikescala.Tuple2K,V,JavaPairRDDK,V
+
+
+Direct Known Subclasses:
+JavaHadoopRDD, JavaNewHadoopRDD
+
+
+
+public class JavaPairRDDK,V
+extends Object
+See Also:Serialized
 Form
+
+
+
+
+
+
+
+
+
+
+
+Constructor Summary
+
+Constructors
+
+Constructor and Description
+
+
+JavaPairRDD(RDDscala.Tuple2K,Vrdd,
+   scala.reflect.ClassTagKkClassTag,
+   scala.reflect.ClassTagVvClassTag)
+
+
+
+
+
+
+
+
+
+Method Summary
+
+Methods
+
+Modifier and Type
+Method and Description
+
+
+static UU
+aggregate(UzeroValue,
+ Function2U,T,UseqOp,
+ Function2U,U,UcombOp)
+
+
+UJavaPairRDDK,U
+aggregateByKey(UzeroValue,
+  Function2U,V,UseqFunc,
+  Function2U,U,UcombFunc)
+Aggregate the values of each key, using given combine 
functions and a neutral "zero value".
+
+
+
+UJavaPairRDDK,U
+aggregateByKey(UzeroValue,
+  intnumPartitions,
+  Function2U,V,UseqFunc,
+  Function2U,U,UcombFunc)
+Aggregate the values of each key, using given combine 
functions and a neutral "zero value".
+
+
+
+UJavaPairRDDK,U
+aggregateByKey(UzeroValue,
+  Partitionerpartitioner,
+  Function2U,V,UseqFunc,
+  Function2U,U,UcombFunc)
+Aggregate the values of each key, using given combine 
functions and a neutral "zero value".
+
+
+
+JavaPairRDDK,V
+cache()
+Persist this RDD with the default storage level 
(MEMORY_ONLY).
+
+
+
+static UJavaPairRDDT,U
+cartesian(JavaRDDLikeU,?other)
+
+
+static void
+checkpoint()
+
+
+scala.reflect.ClassTagscala.Tuple2K,V
+classTag()
+
+
+JavaPairRDDK,V
+coalesce(intnumPartitions)
+Return a new RDD that is reduced into 
numPartitions partitions.
+
+
+
+JavaPairRDDK,V
+coalesce(intnumPartitions,
+booleanshuffle)
+Return a new RDD that is reduced into 
numPartitions partitions.
+
+
+
+WJavaPairRDDK,scala.Tuple2IterableV,IterableW
+cogroup(JavaPairRDDK,Wother)
+For each key k in this or other, 
return a resulting RDD that contains a tuple with the
+ list of values for that key in this as well as 
other.
+
+
+
+WJavaPairRDDK,scala.Tuple2IterableV,IterableW
+cogroup(JavaPairRDDK,Wother,
+   intnumPartitions)
+For each key k in this or other, 
return a resulting RDD that contains a tuple with the
+ list of values for that key in this as well as 
other.
+
+
+
+WJavaPairRDDK,scala.Tuple2IterableV,IterableW
+cogroup(JavaPairRDDK,Wother,
+   Partitionerpartitioner)
+For each key k in this or other, 
return a resulting RDD that contains a tuple with the
+ list of values for that key in this as well as 
other.
+
+
+
+W1,W2JavaPairRDDK,scala.Tuple3IterableV,IterableW1,IterableW2
+cogroup(JavaPairRDDK,W1other1,
+   JavaPairRDDK,W2other2)
+For each key k in this or other1 
or other2, return a resulting RDD that contains a
+ tuple with the list of values for that key in this, 
other1 and other2.
+
+
+
+W1,W2JavaPairRDDK,scala.Tuple3IterableV,IterableW1,IterableW2
+cogroup(JavaPairRDDK,W1other1,
+   JavaPairRDDK,W2other2,
+   intnumPartitions)
+For each key k in this or other1 
or other2, return a resulting RDD that contains a
+ tuple with the list of values for that key in this, 
other1 and other2.
+
+
+
+W1,W2,W3JavaPairRDDK,scala.Tuple4IterableV,IterableW1,IterableW2,IterableW3
+cogroup(JavaPairRDDK,W1other1,
+   JavaPairRDDK,W2other2,
+   JavaPairRDDK,W3other3)
+For each key k in this or other1 
or other2 or other3,
+ return a resulting RDD that contains a tuple with the list of values
+ for that key 

[36/51] [partial] spark-website git commit: Add 2.1.2 docs

2017-10-17 Thread holden
http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/R/summary.html
--
diff --git a/site/docs/2.1.2/api/R/summary.html 
b/site/docs/2.1.2/api/R/summary.html
new file mode 100644
index 000..d074b00
--- /dev/null
+++ b/site/docs/2.1.2/api/R/summary.html
@@ -0,0 +1,132 @@
+http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd;>http://www.w3.org/1999/xhtml;>R: summary
+
+
+
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css;>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js";>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js";>
+hljs.initHighlightingOnLoad();
+
+
+describe 
{SparkR}R Documentation
+
+summary
+
+Description
+
+Computes statistics for numeric and string columns.
+If no columns are given, this function computes statistics for all numerical 
or string columns.
+
+
+
+Usage
+
+
+describe(x, col, ...)
+
+summary(object, ...)
+
+## S4 method for signature 'SparkDataFrame,character'
+describe(x, col, ...)
+
+## S4 method for signature 'SparkDataFrame,ANY'
+describe(x)
+
+## S4 method for signature 'SparkDataFrame'
+summary(object, ...)
+
+
+
+Arguments
+
+
+x
+
+a SparkDataFrame to be computed.
+
+col
+
+a string of name.
+
+...
+
+additional expressions.
+
+object
+
+a SparkDataFrame to be summarized.
+
+
+
+
+Value
+
+A SparkDataFrame.
+
+
+
+Note
+
+describe(SparkDataFrame, character) since 1.4.0
+
+describe(SparkDataFrame) since 1.4.0
+
+summary(SparkDataFrame) since 1.5.0
+
+
+
+See Also
+
+Other SparkDataFrame functions: SparkDataFrame-class,
+agg, arrange,
+as.data.frame, attach,
+cache, coalesce,
+collect, colnames,
+coltypes,
+createOrReplaceTempView,
+crossJoin, dapplyCollect,
+dapply, dim,
+distinct, dropDuplicates,
+dropna, drop,
+dtypes, except,
+explain, filter,
+first, gapplyCollect,
+gapply, getNumPartitions,
+group_by, head,
+histogram, insertInto,
+intersect, isLocal,
+join, limit,
+merge, mutate,
+ncol, nrow,
+persist, printSchema,
+randomSplit, rbind,
+registerTempTable, rename,
+repartition, sample,
+saveAsTable, schema,
+selectExpr, select,
+showDF, show,
+storageLevel, str,
+subset, take,
+union, unpersist,
+withColumn, with,
+write.df, write.jdbc,
+write.json, write.orc,
+write.parquet, write.text
+
+
+
+Examples
+
+## Not run: 
+##D sparkR.session()
+##D path - path/to/file.json
+##D df - read.json(path)
+##D describe(df)
+##D describe(df, col1)
+##D describe(df, col1, col2)
+## End(Not run)
+
+
+
+[Package SparkR version 2.1.2 
Index]
+

http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/R/tableNames.html
--
diff --git a/site/docs/2.1.2/api/R/tableNames.html 
b/site/docs/2.1.2/api/R/tableNames.html
new file mode 100644
index 000..cf05272
--- /dev/null
+++ b/site/docs/2.1.2/api/R/tableNames.html
@@ -0,0 +1,61 @@
+http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd;>http://www.w3.org/1999/xhtml;>R: Table Names
+
+
+
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css;>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js";>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js";>
+hljs.initHighlightingOnLoad();
+
+
+tableNames 
{SparkR}R Documentation
+
+Table Names
+
+Description
+
+Returns the names of tables in the given database as an array.
+
+
+
+Usage
+
+
+## Default S3 method:
+tableNames(databaseName = NULL)
+
+
+
+Arguments
+
+
+databaseName
+
+name of the database
+
+
+
+
+Value
+
+a list of table names
+
+
+
+Note
+
+tableNames since 1.4.0
+
+
+
+Examples
+
+## Not run: 
+##D sparkR.session()
+##D tableNames(hive)
+## End(Not run)
+
+
+
+[Package SparkR version 2.1.2 
Index]
+

http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/R/tableToDF.html
--
diff --git a/site/docs/2.1.2/api/R/tableToDF.html 
b/site/docs/2.1.2/api/R/tableToDF.html
new file mode 100644
index 000..3164dae
--- /dev/null
+++ b/site/docs/2.1.2/api/R/tableToDF.html
@@ -0,0 +1,64 @@
+http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd;>http://www.w3.org/1999/xhtml;>R: Create a SparkDataFrame 
from a SparkSQL Table
+
+
+
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css;>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js";>
+https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js";>
+hljs.initHighlightingOnLoad();
+
+
+tableToDF 
{SparkR}R Documentation
+
+Create a SparkDataFrame from a SparkSQL Table
+
+Description
+
+Returns the specified Table as a SparkDataFrame.  The Table must have 
already been registered
+in the SparkSession.
+
+
+
+Usage
+
+
+tableToDF(tableName)
+
+
+
+Arguments
+
+
+tableName
+
+The SparkSQL Table to 

[04/51] [partial] spark-website git commit: Add 2.1.2 docs

2017-10-17 Thread holden
http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/java/org/apache/spark/api/java/function/Function0.html
--
diff --git 
a/site/docs/2.1.2/api/java/org/apache/spark/api/java/function/Function0.html 
b/site/docs/2.1.2/api/java/org/apache/spark/api/java/function/Function0.html
new file mode 100644
index 000..35cf9d7
--- /dev/null
+++ b/site/docs/2.1.2/api/java/org/apache/spark/api/java/function/Function0.html
@@ -0,0 +1,217 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+
+
+Function0 (Spark 2.1.2 JavaDoc)
+
+
+
+
+
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+
+
+
+Overview
+Package
+Class
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev Class
+Next Class
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+
+
+Detail:
+Field|
+Constr|
+Method
+
+
+
+
+
+
+
+
+org.apache.spark.api.java.function
+Interface Function0R
+
+
+
+
+
+
+All Superinterfaces:
+java.io.Serializable
+
+
+
+public interface Function0R
+extends java.io.Serializable
+A zero-argument function that returns an R.
+
+
+
+
+
+
+
+
+
+
+
+Method Summary
+
+Methods
+
+Modifier and Type
+Method and Description
+
+
+R
+call()
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+Method Detail
+
+
+
+
+
+call
+Rcall()
+   throws Exception
+Throws:
+Exception
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+Overview
+Package
+Class
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev Class
+Next Class
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+
+
+Detail:
+Field|
+Constr|
+Method
+
+
+
+
+
+
+
+

http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/java/org/apache/spark/api/java/function/Function2.html
--
diff --git 
a/site/docs/2.1.2/api/java/org/apache/spark/api/java/function/Function2.html 
b/site/docs/2.1.2/api/java/org/apache/spark/api/java/function/Function2.html
new file mode 100644
index 000..ef10f45
--- /dev/null
+++ b/site/docs/2.1.2/api/java/org/apache/spark/api/java/function/Function2.html
@@ -0,0 +1,221 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+
+
+Function2 (Spark 2.1.2 JavaDoc)
+
+
+
+
+
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+
+
+
+Overview
+Package
+Class
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev Class
+Next Class
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+
+
+Detail:
+Field|
+Constr|
+Method
+
+
+
+
+
+
+
+
+org.apache.spark.api.java.function
+Interface 
Function2T1,T2,R
+
+
+
+
+
+
+All Superinterfaces:
+java.io.Serializable
+
+
+
+public interface Function2T1,T2,R
+extends java.io.Serializable
+A two-argument function that takes arguments of type T1 and 
T2 and returns an R.
+
+
+
+
+
+
+
+
+
+
+
+Method Summary
+
+Methods
+
+Modifier and Type
+Method and Description
+
+
+R
+call(T1v1,
+T2v2)
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+Method Detail
+
+
+
+
+
+
+
+call
+Rcall(T1v1,
+ T2v2)
+   throws Exception
+Throws:
+Exception
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+Overview
+Package
+Class
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev Class
+Next Class
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+
+
+Detail:
+Field|
+Constr|
+Method
+
+
+
+
+
+
+
+

http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/java/org/apache/spark/api/java/function/Function3.html
--
diff --git 
a/site/docs/2.1.2/api/java/org/apache/spark/api/java/function/Function3.html 

[25/51] [partial] spark-website git commit: Add 2.1.2 docs

2017-10-17 Thread holden
http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/java/org/apache/spark/ComplexFutureAction.html
--
diff --git a/site/docs/2.1.2/api/java/org/apache/spark/ComplexFutureAction.html 
b/site/docs/2.1.2/api/java/org/apache/spark/ComplexFutureAction.html
new file mode 100644
index 000..85f6c0d
--- /dev/null
+++ b/site/docs/2.1.2/api/java/org/apache/spark/ComplexFutureAction.html
@@ -0,0 +1,489 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+
+
+ComplexFutureAction (Spark 2.1.2 JavaDoc)
+
+
+
+
+
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+
+
+
+Overview
+Package
+Class
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev 
Class
+Next Class
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+
+
+Detail:
+Field|
+Constr|
+Method
+
+
+
+
+
+
+
+
+org.apache.spark
+Class 
ComplexFutureActionT
+
+
+
+Object
+
+
+org.apache.spark.ComplexFutureActionT
+
+
+
+
+
+
+
+All Implemented Interfaces:
+FutureActionT, 
scala.concurrent.AwaitableT, scala.concurrent.FutureT
+
+
+
+public class ComplexFutureActionT
+extends Object
+implements FutureActionT
+A FutureAction for actions 
that could trigger multiple Spark jobs. Examples include take,
+ takeSample. Cancellation works by setting the cancelled flag to true and 
cancelling any pending
+ jobs.
+
+
+
+
+
+
+
+
+
+
+
+Nested Class Summary
+
+
+
+
+Nested classes/interfaces inherited from 
interfacescala.concurrent.Future
+scala.concurrent.Future.InternalCallbackExecutor$
+
+
+
+
+
+
+
+
+Constructor Summary
+
+Constructors
+
+Constructor and Description
+
+
+ComplexFutureAction(scala.Function1JobSubmitter,scala.concurrent.FutureTrun)
+
+
+
+
+
+
+
+
+
+Method Summary
+
+Methods
+
+Modifier and Type
+Method and Description
+
+
+void
+cancel()
+Cancels the execution of this action.
+
+
+
+boolean
+isCancelled()
+Returns whether the action has been cancelled.
+
+
+
+boolean
+isCompleted()
+Returns whether the action has already been completed with 
a value or an exception.
+
+
+
+scala.collection.SeqObject
+jobIds()
+Returns the job IDs run by the underlying async 
operation.
+
+
+
+Uvoid
+onComplete(scala.Function1scala.util.TryT,Ufunc,
+  scala.concurrent.ExecutionContextexecutor)
+When this action is completed, either through an exception, 
or a value, applies the provided
+ function.
+
+
+
+ComplexFutureActionT
+ready(scala.concurrent.duration.DurationatMost,
+ scala.concurrent.CanAwaitpermit)
+Blocks until this action completes.
+
+
+
+T
+result(scala.concurrent.duration.DurationatMost,
+  scala.concurrent.CanAwaitpermit)
+Awaits and returns the result (of type T) of this 
action.
+
+
+
+scala.Optionscala.util.TryT
+value()
+The value of this Future.
+
+
+
+
+
+
+
+Methods inherited from classObject
+equals, getClass, hashCode, notify, notifyAll, toString, wait, wait, 
wait
+
+
+
+
+
+Methods inherited from interfaceorg.apache.spark.FutureAction
+get
+
+
+
+
+
+Methods inherited from interfacescala.concurrent.Future
+andThen, collect, failed, fallbackTo, filter, flatMap, foreach, map, 
mapTo, onFailure, onSuccess, recover, recoverWith, transform, withFilter, 
zip
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+Constructor Detail
+
+
+
+
+
+ComplexFutureAction
+publicComplexFutureAction(scala.Function1JobSubmitter,scala.concurrent.FutureTrun)
+
+
+
+
+
+
+
+
+
+Method Detail
+
+
+
+
+
+cancel
+publicvoidcancel()
+Description copied from interface:FutureAction
+Cancels the execution of this action.
+
+Specified by:
+cancelin
 interfaceFutureActionT
+
+
+
+
+
+
+
+
+isCancelled
+publicbooleanisCancelled()
+Description copied from interface:FutureAction
+Returns whether the action has been cancelled.
+
+Specified by:
+isCancelledin
 interfaceFutureActionT
+Returns:(undocumented)
+
+
+
+
+
+
+
+ready
+publicComplexFutureActionTready(scala.concurrent.duration.DurationatMost,
+   scala.concurrent.CanAwaitpermit)
+ throws InterruptedException,
+java.util.concurrent.TimeoutException
+Description copied from interface:FutureAction
+Blocks until this action completes.
+ 
+
+Specified by:
+readyin
 interfaceFutureActionT
+Specified by:
+readyin 
interfacescala.concurrent.AwaitableT
+Parameters:atMost - 
maximum wait time, which may be negative (no waiting is done), Duration.Inf
+   for unbounded waiting, or a finite positive 
durationpermit - (undocumented)
+Returns:this FutureAction
+Throws:
+InterruptedException

[27/51] [partial] spark-website git commit: Add 2.1.2 docs

2017-10-17 Thread holden
http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/java/org/apache/spark/AccumulatorParam.IntAccumulatorParam$.html
--
diff --git 
a/site/docs/2.1.2/api/java/org/apache/spark/AccumulatorParam.IntAccumulatorParam$.html
 
b/site/docs/2.1.2/api/java/org/apache/spark/AccumulatorParam.IntAccumulatorParam$.html
new file mode 100644
index 000..8a8d442
--- /dev/null
+++ 
b/site/docs/2.1.2/api/java/org/apache/spark/AccumulatorParam.IntAccumulatorParam$.html
@@ -0,0 +1,365 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+
+
+AccumulatorParam.IntAccumulatorParam$ (Spark 2.1.2 JavaDoc)
+
+
+
+
+
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+
+
+
+Overview
+Package
+Class
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev 
Class
+Next 
Class
+
+
+Frames
+No 
Frames
+
+
+All Classes
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+
+
+Detail:
+Field|
+Constr|
+Method
+
+
+
+
+
+
+
+
+org.apache.spark
+Class 
AccumulatorParam.IntAccumulatorParam$
+
+
+
+Object
+
+
+org.apache.spark.AccumulatorParam.IntAccumulatorParam$
+
+
+
+
+
+
+
+All Implemented Interfaces:
+java.io.Serializable, AccumulableParamObject,Object, AccumulatorParamObject
+
+
+Enclosing interface:
+AccumulatorParamT
+
+
+Deprecated.
+use AccumulatorV2. Since 2.0.0.
+
+
+public static class AccumulatorParam.IntAccumulatorParam$
+extends Object
+implements AccumulatorParamObject
+See Also:Serialized
 Form
+
+
+
+
+
+
+
+
+
+
+
+Nested Class Summary
+
+
+
+
+Nested classes/interfaces inherited from 
interfaceorg.apache.spark.AccumulatorParam
+AccumulatorParam.DoubleAccumulatorParam$, 
AccumulatorParam.FloatAccumulatorParam$, 
AccumulatorParam.IntAccumulatorParam$, AccumulatorParam.LongAccumulatorParam$, 
AccumulatorParam.StringAccumulatorParam$
+
+
+
+
+
+
+
+
+Field Summary
+
+Fields
+
+Modifier and Type
+Field and Description
+
+
+static AccumulatorParam.IntAccumulatorParam$
+MODULE$
+Deprecated.
+Static reference to the singleton instance of this Scala 
object.
+
+
+
+
+
+
+
+
+
+
+Constructor Summary
+
+Constructors
+
+Constructor and Description
+
+
+AccumulatorParam.IntAccumulatorParam$()
+Deprecated.
+
+
+
+
+
+
+
+
+
+
+Method Summary
+
+Methods
+
+Modifier and Type
+Method and Description
+
+
+int
+addInPlace(intt1,
+  intt2)
+Deprecated.
+
+
+
+int
+zero(intinitialValue)
+Deprecated.
+
+
+
+
+
+
+
+Methods inherited from classObject
+equals, getClass, hashCode, notify, notifyAll, toString, wait, wait, 
wait
+
+
+
+
+
+Methods inherited from interfaceorg.apache.spark.AccumulatorParam
+addAccumulator
+
+
+
+
+
+Methods inherited from interfaceorg.apache.spark.AccumulableParam
+addInPlace,
 zero
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+Field Detail
+
+
+
+
+
+MODULE$
+public static finalAccumulatorParam.IntAccumulatorParam$ 
MODULE$
+Deprecated.
+Static reference to the singleton instance of this Scala 
object.
+
+
+
+
+
+
+
+
+
+Constructor Detail
+
+
+
+
+
+AccumulatorParam.IntAccumulatorParam$
+publicAccumulatorParam.IntAccumulatorParam$()
+Deprecated.
+
+
+
+
+
+
+
+
+
+Method Detail
+
+
+
+
+
+addInPlace
+publicintaddInPlace(intt1,
+ intt2)
+Deprecated.
+
+
+
+
+
+
+
+zero
+publicintzero(intinitialValue)
+Deprecated.
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+Overview
+Package
+Class
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev 
Class
+Next 
Class
+
+
+Frames
+No 
Frames
+
+
+All Classes
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+
+
+Detail:
+Field|
+Constr|
+Method
+
+
+
+
+
+
+
+

http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/java/org/apache/spark/AccumulatorParam.LongAccumulatorParam$.html
--
diff --git 
a/site/docs/2.1.2/api/java/org/apache/spark/AccumulatorParam.LongAccumulatorParam$.html
 
b/site/docs/2.1.2/api/java/org/apache/spark/AccumulatorParam.LongAccumulatorParam$.html
new file mode 100644
index 000..e2cd2c9
--- /dev/null
+++ 
b/site/docs/2.1.2/api/java/org/apache/spark/AccumulatorParam.LongAccumulatorParam$.html
@@ -0,0 +1,365 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+
+
+AccumulatorParam.LongAccumulatorParam$ (Spark 2.1.2 JavaDoc)
+
+
+
+
+
+  var _gaq = _gaq || [];
+  _gaq.push(['_setAccount', 'UA-32518208-2']);
+  _gaq.push(['_trackPageview']);
+  (function() {
+var ga = document.createElement('script'); ga.type = 'text/javascript'; 
ga.async = true;
+ga.src = ('https:' == document.location.protocol ? 'https://ssl' : 
'http://www') + '.google-analytics.com/ga.js';
+var s = document.getElementsByTagName('script')[0]; 
s.parentNode.insertBefore(ga, s);
+  })();
+
+  
+  function trackOutboundLink(link, category, action) {
+try {
+  _gaq.push(['_trackEvent', category , action]);
+} catch(err){}
+
+setTimeout(function() {
+  document.location.href = link.href;
+}, 100);
+  }
+  
+
+  
+  
+
+
+
+
+https://code.jquery.com/jquery.js";>
+https://netdna.bootstrapcdn.com/bootstrap/3.0.3/js/bootstrap.min.js";>
+
+
+
+
+
+
+  
+
+  
+  
+  Lightning-fast cluster computing
+  
+
+  
+
+
+
+  
+  
+
+  Toggle navigation
+  
+  
+  
+
+  
+
+  
+  
+
+  Download
+  
+
+  Libraries 
+
+
+  SQL and DataFrames
+  Spark Streaming
+  MLlib (machine learning)
+  GraphX (graph)
+  
+  Third-Party 
Projects
+
+  
+  
+
+  Documentation 
+
+
+  Latest Release (Spark 2.2.0)
+  Older Versions and Other 
Resources
+  Frequently Asked Questions
+
+  
+  Examples
+  
+
+  Community 
+
+
+  Mailing Lists  Resources
+  Contributing to Spark
+  Improvement Proposals 
(SPIP)
+  https://issues.apache.org/jira/browse/SPARK;>Issue 
Tracker
+

[1/3] spark-website git commit: Update the website for Spark 2.1.2 and add linkts to API documentation and clarify release JIRA process.

2017-10-17 Thread holden
Repository: spark-website
Updated Branches:
  refs/heads/asf-site 5e04ca053 -> bdb87e97c


http://git-wip-us.apache.org/repos/asf/spark-website/blob/bdb87e97/site/releases/spark-release-1-2-0.html
--
diff --git a/site/releases/spark-release-1-2-0.html 
b/site/releases/spark-release-1-2-0.html
index 45a9b17..98ba92a 100644
--- a/site/releases/spark-release-1-2-0.html
+++ b/site/releases/spark-release-1-2-0.html
@@ -161,6 +161,9 @@
   Latest News
   
 
+  Spark 2.1.2 
released
+  (Oct 09, 2017)
+
   Spark 
Summit Europe (October 24-26th, 2017, Dublin, Ireland) agenda posted
   (Aug 28, 2017)
 
@@ -170,9 +173,6 @@
   Spark 2.1.1 
released
   (May 02, 2017)
 
-  Spark 
Summit (June 5-7th, 2017, San Francisco) agenda posted
-  (Mar 31, 2017)
-
   
   Archive
 

http://git-wip-us.apache.org/repos/asf/spark-website/blob/bdb87e97/site/releases/spark-release-1-2-1.html
--
diff --git a/site/releases/spark-release-1-2-1.html 
b/site/releases/spark-release-1-2-1.html
index 6103993..6383c66 100644
--- a/site/releases/spark-release-1-2-1.html
+++ b/site/releases/spark-release-1-2-1.html
@@ -161,6 +161,9 @@
   Latest News
   
 
+  Spark 2.1.2 
released
+  (Oct 09, 2017)
+
   Spark 
Summit Europe (October 24-26th, 2017, Dublin, Ireland) agenda posted
   (Aug 28, 2017)
 
@@ -170,9 +173,6 @@
   Spark 2.1.1 
released
   (May 02, 2017)
 
-  Spark 
Summit (June 5-7th, 2017, San Francisco) agenda posted
-  (Mar 31, 2017)
-
   
   Archive
 

http://git-wip-us.apache.org/repos/asf/spark-website/blob/bdb87e97/site/releases/spark-release-1-2-2.html
--
diff --git a/site/releases/spark-release-1-2-2.html 
b/site/releases/spark-release-1-2-2.html
index bd2ca9f..48a11b1 100644
--- a/site/releases/spark-release-1-2-2.html
+++ b/site/releases/spark-release-1-2-2.html
@@ -161,6 +161,9 @@
   Latest News
   
 
+  Spark 2.1.2 
released
+  (Oct 09, 2017)
+
   Spark 
Summit Europe (October 24-26th, 2017, Dublin, Ireland) agenda posted
   (Aug 28, 2017)
 
@@ -170,9 +173,6 @@
   Spark 2.1.1 
released
   (May 02, 2017)
 
-  Spark 
Summit (June 5-7th, 2017, San Francisco) agenda posted
-  (Mar 31, 2017)
-
   
   Archive
 

http://git-wip-us.apache.org/repos/asf/spark-website/blob/bdb87e97/site/releases/spark-release-1-3-0.html
--
diff --git a/site/releases/spark-release-1-3-0.html 
b/site/releases/spark-release-1-3-0.html
index 008f31b..fb22f53 100644
--- a/site/releases/spark-release-1-3-0.html
+++ b/site/releases/spark-release-1-3-0.html
@@ -161,6 +161,9 @@
   Latest News
   
 
+  Spark 2.1.2 
released
+  (Oct 09, 2017)
+
   Spark 
Summit Europe (October 24-26th, 2017, Dublin, Ireland) agenda posted
   (Aug 28, 2017)
 
@@ -170,9 +173,6 @@
   Spark 2.1.1 
released
   (May 02, 2017)
 
-  Spark 
Summit (June 5-7th, 2017, San Francisco) agenda posted
-  (Mar 31, 2017)
-
   
   Archive
 

http://git-wip-us.apache.org/repos/asf/spark-website/blob/bdb87e97/site/releases/spark-release-1-3-1.html
--
diff --git a/site/releases/spark-release-1-3-1.html 
b/site/releases/spark-release-1-3-1.html
index 8ce1da5..886826f 100644
--- a/site/releases/spark-release-1-3-1.html
+++ b/site/releases/spark-release-1-3-1.html
@@ -161,6 +161,9 @@
   Latest News
   
 
+  Spark 2.1.2 
released
+  (Oct 09, 2017)
+
   Spark 
Summit Europe (October 24-26th, 2017, Dublin, Ireland) agenda posted
   (Aug 28, 2017)
 
@@ -170,9 +173,6 @@
   Spark 2.1.1 
released
   (May 02, 2017)
 
-  Spark 
Summit (June 5-7th, 2017, San Francisco) agenda posted
-  (Mar 31, 2017)
-
   
   Archive
 

http://git-wip-us.apache.org/repos/asf/spark-website/blob/bdb87e97/site/releases/spark-release-1-4-0.html
--
diff --git a/site/releases/spark-release-1-4-0.html 
b/site/releases/spark-release-1-4-0.html
index e8c1448..d86635f 100644
--- a/site/releases/spark-release-1-4-0.html
+++ b/site/releases/spark-release-1-4-0.html
@@ -161,6 +161,9 @@
   Latest News
   
 
+  Spark 2.1.2 
released
+  (Oct 09, 2017)
+
   Spark 
Summit Europe (October 24-26th, 

[3/3] spark-website git commit: Update the website for Spark 2.1.2 and add linkts to API documentation and clarify release JIRA process.

2017-10-17 Thread holden
Update the website for Spark 2.1.2 and add linkts to API documentation and 
clarify release JIRA process.

update for release 2.1.2

doc

add release on jira


Project: http://git-wip-us.apache.org/repos/asf/spark-website/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark-website/commit/bdb87e97
Tree: http://git-wip-us.apache.org/repos/asf/spark-website/tree/bdb87e97
Diff: http://git-wip-us.apache.org/repos/asf/spark-website/diff/bdb87e97

Branch: refs/heads/asf-site
Commit: bdb87e97c006e8ca98b2b2d2554a9f07c26d494c
Parents: 5e04ca0
Author: Felix Cheung 
Authored: Sun Oct 15 13:40:12 2017 -0700
Committer: Holden Karau 
Committed: Tue Oct 17 22:57:04 2017 -0700

--
 documentation.md|   2 +
 js/downloads.js |   1 +
 news/_posts/2017-10-09-spark-2-1-2-released.md  |  15 ++
 release-process.md  |  28 ++-
 .../_posts/2017-10-09-spark-release-2-1-2.md|  18 ++
 security.md |  10 +-
 site/committers.html|   6 +-
 site/community.html |   6 +-
 site/contributing.html  |   6 +-
 site/developer-tools.html   |   6 +-
 site/docs/2.0.2/latest  |   1 -
 site/documentation.html |   8 +-
 site/downloads.html |   6 +-
 site/examples.html  |   6 +-
 site/faq.html   |   6 +-
 site/graphx/index.html  |   6 +-
 site/improvement-proposals.html |   6 +-
 site/index.html |   6 +-
 site/js/downloads.js|   1 +
 site/mailing-lists.html |   6 +-
 site/mllib/index.html   |   6 +-
 site/news/amp-camp-2013-registration-ope.html   |   6 +-
 .../news/announcing-the-first-spark-summit.html |   6 +-
 .../news/fourth-spark-screencast-published.html |   6 +-
 site/news/index.html|  15 +-
 site/news/nsdi-paper.html   |   6 +-
 site/news/one-month-to-spark-summit-2015.html   |   6 +-
 .../proposals-open-for-spark-summit-east.html   |   6 +-
 ...registration-open-for-spark-summit-east.html |   6 +-
 .../news/run-spark-and-shark-on-amazon-emr.html |   6 +-
 site/news/spark-0-6-1-and-0-5-2-released.html   |   6 +-
 site/news/spark-0-6-2-released.html |   6 +-
 site/news/spark-0-7-0-released.html |   6 +-
 site/news/spark-0-7-2-released.html |   6 +-
 site/news/spark-0-7-3-released.html |   6 +-
 site/news/spark-0-8-0-released.html |   6 +-
 site/news/spark-0-8-1-released.html |   6 +-
 site/news/spark-0-9-0-released.html |   6 +-
 site/news/spark-0-9-1-released.html |   6 +-
 site/news/spark-0-9-2-released.html |   6 +-
 site/news/spark-1-0-0-released.html |   6 +-
 site/news/spark-1-0-1-released.html |   6 +-
 site/news/spark-1-0-2-released.html |   6 +-
 site/news/spark-1-1-0-released.html |   6 +-
 site/news/spark-1-1-1-released.html |   6 +-
 site/news/spark-1-2-0-released.html |   6 +-
 site/news/spark-1-2-1-released.html |   6 +-
 site/news/spark-1-2-2-released.html |   6 +-
 site/news/spark-1-3-0-released.html |   6 +-
 site/news/spark-1-4-0-released.html |   6 +-
 site/news/spark-1-4-1-released.html |   6 +-
 site/news/spark-1-5-0-released.html |   6 +-
 site/news/spark-1-5-1-released.html |   6 +-
 site/news/spark-1-5-2-released.html |   6 +-
 site/news/spark-1-6-0-released.html |   6 +-
 site/news/spark-1-6-1-released.html |   6 +-
 site/news/spark-1-6-2-released.html |   6 +-
 site/news/spark-1-6-3-released.html |   6 +-
 site/news/spark-2-0-0-released.html |   6 +-
 site/news/spark-2-0-1-released.html |   6 +-
 site/news/spark-2-0-2-released.html |   6 +-
 site/news/spark-2-1-0-released.html |   6 +-
 site/news/spark-2-1-1-released.html |   6 +-
 site/news/spark-2-1-2-released.html | 223 ++
 site/news/spark-2-2-0-released.html |   6 +-
 site/news/spark-2.0.0-preview.html  |   6 +-
 .../spark-accepted-into-apache-incubator.html   |   6 +-
 site/news/spark-and-shark-in-the-news.html  |   6 +-
 site/news/spark-becomes-tlp.html|   6 +-
 site/news/spark-featured-in-wired.html  |   6 +-
 .../spark-mailing-lists-moving-to-apache.html   |   6 +-
 site/news/spark-meetups.html|   6 +-
 

<    1   2