[41/51] [partial] spark-website git commit: Add 2.1.2 docs
http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6155a89/site/docs/2.1.2/api/R/ncol.html -- diff --git a/site/docs/2.1.2/api/R/ncol.html b/site/docs/2.1.2/api/R/ncol.html new file mode 100644 index 000..8677389 --- /dev/null +++ b/site/docs/2.1.2/api/R/ncol.html @@ -0,0 +1,97 @@ +http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd";>http://www.w3.org/1999/xhtml";>R: Returns the number of columns in a SparkDataFrame + + + +https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css";> +https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js";> +https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js";> +hljs.initHighlightingOnLoad(); + + +ncol {SparkR}R Documentation + +Returns the number of columns in a SparkDataFrame + +Description + +Returns the number of columns in a SparkDataFrame + + + +Usage + + +## S4 method for signature 'SparkDataFrame' +ncol(x) + + + +Arguments + + +x + +a SparkDataFrame + + + + +Note + +ncol since 1.5.0 + + + +See Also + +Other SparkDataFrame functions: SparkDataFrame-class, +agg, arrange, +as.data.frame, attach, +cache, coalesce, +collect, colnames, +coltypes, +createOrReplaceTempView, +crossJoin, dapplyCollect, +dapply, describe, +dim, distinct, +dropDuplicates, dropna, +drop, dtypes, +except, explain, +filter, first, +gapplyCollect, gapply, +getNumPartitions, group_by, +head, histogram, +insertInto, intersect, +isLocal, join, +limit, merge, +mutate, nrow, +persist, printSchema, +randomSplit, rbind, +registerTempTable, rename, +repartition, sample, +saveAsTable, schema, +selectExpr, select, +showDF, show, +storageLevel, str, +subset, take, +union, unpersist, +withColumn, with, +write.df, write.jdbc, +write.json, write.orc, +write.parquet, write.text + + + +Examples + +## Not run: +##D sparkR.session() +##D path <- "path/to/file.json" +##D df <- read.json(path) +##D ncol(df) +## End(Not run) + + + +[Package SparkR version 2.1.2 Index] + http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6155a89/site/docs/2.1.2/api/R/negate.html -- diff --git a/site/docs/2.1.2/api/R/negate.html b/site/docs/2.1.2/api/R/negate.html new file mode 100644 index 000..cc0e06d --- /dev/null +++ b/site/docs/2.1.2/api/R/negate.html @@ -0,0 +1,67 @@ +http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd";>http://www.w3.org/1999/xhtml";>R: negate + + + +https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css";> +https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js";> +https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js";> +hljs.initHighlightingOnLoad(); + + +negate {SparkR}R Documentation + +negate + +Description + +Unary minus, i.e. negate the expression. + + + +Usage + + +negate(x) + +## S4 method for signature 'Column' +negate(x) + + + +Arguments + + +x + +Column to compute on. + + + + +Note + +negate since 1.5.0 + + + +See Also + +Other normal_funcs: abs, +bitwiseNOT, coalesce, +column, expr, +greatest, ifelse, +isnan, least, +lit, nanvl, +randn, rand, +struct, when + + + +Examples + +## Not run: negate(df$c) + + + +[Package SparkR version 2.1.2 Index] + http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6155a89/site/docs/2.1.2/api/R/next_day.html -- diff --git a/site/docs/2.1.2/api/R/next_day.html b/site/docs/2.1.2/api/R/next_day.html new file mode 100644 index 000..736d7e4 --- /dev/null +++ b/site/docs/2.1.2/api/R/next_day.html @@ -0,0 +1,89 @@ +http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd";>http://www.w3.org/1999/xhtml";>R: next_day + + + +https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css";> +https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js";> +https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js";> +hljs.initHighlightingOnLoad(); + + +next_day {SparkR}R Documentation + +next_day + +Description + +Given a date column, returns the first date which is later than the value of the date column +that is on the specified day of the week. + + + +Usage + + +next_day(y, x) + +## S4 method for signature 'Column,character' +next_day(y, x) + + + +Arguments + + +y + +Column to compute on. + +x + +Day of the week string. + + + + +Details + +For example, next_day('2015-07-27', "Sunday") returns 2015-08-02 because that is the first +Sunday after 2015-07-27. + +Day of the week parameter is case insensitive, and accepts first three or two characters: +"Mon", "Tue", "Wed", "Thu", "Fri", "Sat", "Sun". + + + +Note + +next_day since 1.5.0 + + + +See Also + +Other datetime_funcs: add_months, +date_add, date_format, +date_sub, datediff, +dayofmonth, dayofyear, +from_unixtime, +from_utc_timestamp, hour, +last_day, minute, +months_between, month, +quarter, secon
[41/51] [partial] spark-website git commit: Add 2.1.2 docs
http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/R/ncol.html -- diff --git a/site/docs/2.1.2/api/R/ncol.html b/site/docs/2.1.2/api/R/ncol.html new file mode 100644 index 000..8677389 --- /dev/null +++ b/site/docs/2.1.2/api/R/ncol.html @@ -0,0 +1,97 @@ +http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd";>http://www.w3.org/1999/xhtml";>R: Returns the number of columns in a SparkDataFrame + + + +https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css";> +https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js";> +https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js";> +hljs.initHighlightingOnLoad(); + + +ncol {SparkR}R Documentation + +Returns the number of columns in a SparkDataFrame + +Description + +Returns the number of columns in a SparkDataFrame + + + +Usage + + +## S4 method for signature 'SparkDataFrame' +ncol(x) + + + +Arguments + + +x + +a SparkDataFrame + + + + +Note + +ncol since 1.5.0 + + + +See Also + +Other SparkDataFrame functions: SparkDataFrame-class, +agg, arrange, +as.data.frame, attach, +cache, coalesce, +collect, colnames, +coltypes, +createOrReplaceTempView, +crossJoin, dapplyCollect, +dapply, describe, +dim, distinct, +dropDuplicates, dropna, +drop, dtypes, +except, explain, +filter, first, +gapplyCollect, gapply, +getNumPartitions, group_by, +head, histogram, +insertInto, intersect, +isLocal, join, +limit, merge, +mutate, nrow, +persist, printSchema, +randomSplit, rbind, +registerTempTable, rename, +repartition, sample, +saveAsTable, schema, +selectExpr, select, +showDF, show, +storageLevel, str, +subset, take, +union, unpersist, +withColumn, with, +write.df, write.jdbc, +write.json, write.orc, +write.parquet, write.text + + + +Examples + +## Not run: +##D sparkR.session() +##D path <- "path/to/file.json" +##D df <- read.json(path) +##D ncol(df) +## End(Not run) + + + +[Package SparkR version 2.1.2 Index] + http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/R/negate.html -- diff --git a/site/docs/2.1.2/api/R/negate.html b/site/docs/2.1.2/api/R/negate.html new file mode 100644 index 000..cc0e06d --- /dev/null +++ b/site/docs/2.1.2/api/R/negate.html @@ -0,0 +1,67 @@ +http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd";>http://www.w3.org/1999/xhtml";>R: negate + + + +https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css";> +https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js";> +https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js";> +hljs.initHighlightingOnLoad(); + + +negate {SparkR}R Documentation + +negate + +Description + +Unary minus, i.e. negate the expression. + + + +Usage + + +negate(x) + +## S4 method for signature 'Column' +negate(x) + + + +Arguments + + +x + +Column to compute on. + + + + +Note + +negate since 1.5.0 + + + +See Also + +Other normal_funcs: abs, +bitwiseNOT, coalesce, +column, expr, +greatest, ifelse, +isnan, least, +lit, nanvl, +randn, rand, +struct, when + + + +Examples + +## Not run: negate(df$c) + + + +[Package SparkR version 2.1.2 Index] + http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6d9cbde/site/docs/2.1.2/api/R/next_day.html -- diff --git a/site/docs/2.1.2/api/R/next_day.html b/site/docs/2.1.2/api/R/next_day.html new file mode 100644 index 000..736d7e4 --- /dev/null +++ b/site/docs/2.1.2/api/R/next_day.html @@ -0,0 +1,89 @@ +http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd";>http://www.w3.org/1999/xhtml";>R: next_day + + + +https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css";> +https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js";> +https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js";> +hljs.initHighlightingOnLoad(); + + +next_day {SparkR}R Documentation + +next_day + +Description + +Given a date column, returns the first date which is later than the value of the date column +that is on the specified day of the week. + + + +Usage + + +next_day(y, x) + +## S4 method for signature 'Column,character' +next_day(y, x) + + + +Arguments + + +y + +Column to compute on. + +x + +Day of the week string. + + + + +Details + +For example, next_day('2015-07-27', "Sunday") returns 2015-08-02 because that is the first +Sunday after 2015-07-27. + +Day of the week parameter is case insensitive, and accepts first three or two characters: +"Mon", "Tue", "Wed", "Thu", "Fri", "Sat", "Sun". + + + +Note + +next_day since 1.5.0 + + + +See Also + +Other datetime_funcs: add_months, +date_add, date_format, +date_sub, datediff, +dayofmonth, dayofyear, +from_unixtime, +from_utc_timestamp, hour, +last_day, minute, +months_between, month, +quarter, secon
[41/51] [partial] spark-website git commit: Add 2.1.2 docs
http://git-wip-us.apache.org/repos/asf/spark-website/blob/0b563c84/site/docs/2.1.2/api/R/ncol.html -- diff --git a/site/docs/2.1.2/api/R/ncol.html b/site/docs/2.1.2/api/R/ncol.html new file mode 100644 index 000..8677389 --- /dev/null +++ b/site/docs/2.1.2/api/R/ncol.html @@ -0,0 +1,97 @@ +http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd";>http://www.w3.org/1999/xhtml";>R: Returns the number of columns in a SparkDataFrame + + + +https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css";> +https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js";> +https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js";> +hljs.initHighlightingOnLoad(); + + +ncol {SparkR}R Documentation + +Returns the number of columns in a SparkDataFrame + +Description + +Returns the number of columns in a SparkDataFrame + + + +Usage + + +## S4 method for signature 'SparkDataFrame' +ncol(x) + + + +Arguments + + +x + +a SparkDataFrame + + + + +Note + +ncol since 1.5.0 + + + +See Also + +Other SparkDataFrame functions: SparkDataFrame-class, +agg, arrange, +as.data.frame, attach, +cache, coalesce, +collect, colnames, +coltypes, +createOrReplaceTempView, +crossJoin, dapplyCollect, +dapply, describe, +dim, distinct, +dropDuplicates, dropna, +drop, dtypes, +except, explain, +filter, first, +gapplyCollect, gapply, +getNumPartitions, group_by, +head, histogram, +insertInto, intersect, +isLocal, join, +limit, merge, +mutate, nrow, +persist, printSchema, +randomSplit, rbind, +registerTempTable, rename, +repartition, sample, +saveAsTable, schema, +selectExpr, select, +showDF, show, +storageLevel, str, +subset, take, +union, unpersist, +withColumn, with, +write.df, write.jdbc, +write.json, write.orc, +write.parquet, write.text + + + +Examples + +## Not run: +##D sparkR.session() +##D path <- "path/to/file.json" +##D df <- read.json(path) +##D ncol(df) +## End(Not run) + + + +[Package SparkR version 2.1.2 Index] + http://git-wip-us.apache.org/repos/asf/spark-website/blob/0b563c84/site/docs/2.1.2/api/R/negate.html -- diff --git a/site/docs/2.1.2/api/R/negate.html b/site/docs/2.1.2/api/R/negate.html new file mode 100644 index 000..cc0e06d --- /dev/null +++ b/site/docs/2.1.2/api/R/negate.html @@ -0,0 +1,67 @@ +http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd";>http://www.w3.org/1999/xhtml";>R: negate + + + +https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css";> +https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js";> +https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js";> +hljs.initHighlightingOnLoad(); + + +negate {SparkR}R Documentation + +negate + +Description + +Unary minus, i.e. negate the expression. + + + +Usage + + +negate(x) + +## S4 method for signature 'Column' +negate(x) + + + +Arguments + + +x + +Column to compute on. + + + + +Note + +negate since 1.5.0 + + + +See Also + +Other normal_funcs: abs, +bitwiseNOT, coalesce, +column, expr, +greatest, ifelse, +isnan, least, +lit, nanvl, +randn, rand, +struct, when + + + +Examples + +## Not run: negate(df$c) + + + +[Package SparkR version 2.1.2 Index] + http://git-wip-us.apache.org/repos/asf/spark-website/blob/0b563c84/site/docs/2.1.2/api/R/next_day.html -- diff --git a/site/docs/2.1.2/api/R/next_day.html b/site/docs/2.1.2/api/R/next_day.html new file mode 100644 index 000..736d7e4 --- /dev/null +++ b/site/docs/2.1.2/api/R/next_day.html @@ -0,0 +1,89 @@ +http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd";>http://www.w3.org/1999/xhtml";>R: next_day + + + +https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css";> +https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js";> +https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js";> +hljs.initHighlightingOnLoad(); + + +next_day {SparkR}R Documentation + +next_day + +Description + +Given a date column, returns the first date which is later than the value of the date column +that is on the specified day of the week. + + + +Usage + + +next_day(y, x) + +## S4 method for signature 'Column,character' +next_day(y, x) + + + +Arguments + + +y + +Column to compute on. + +x + +Day of the week string. + + + + +Details + +For example, next_day('2015-07-27', "Sunday") returns 2015-08-02 because that is the first +Sunday after 2015-07-27. + +Day of the week parameter is case insensitive, and accepts first three or two characters: +"Mon", "Tue", "Wed", "Thu", "Fri", "Sat", "Sun". + + + +Note + +next_day since 1.5.0 + + + +See Also + +Other datetime_funcs: add_months, +date_add, date_format, +date_sub, datediff, +dayofmonth, dayofyear, +from_unixtime, +from_utc_timestamp, hour, +last_day, minute, +months_between, month, +quarter, secon