Github user dongjoon-hyun commented on a diff in the pull request:
https://github.com/apache/spark/pull/13765#discussion_r67610142
--- Diff:
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/dsl/package.scala ---
@@ -374,6 +374,9 @@ package object dsl {
case
GitHub user dongjoon-hyun opened a pull request:
https://github.com/apache/spark/pull/13768
Add `spark_partition_id` in SparkR
## What changes were proposed in this pull request?
This PR adds `spark_partition_id` virtual column function in SparkR for API
parity
Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/13765
Hi, @cloud-fan .
Could you review this optimizer?
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does
Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/13768
Hi, @davies .
Could you review this PR?
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have
GitHub user dongjoon-hyun opened a pull request:
https://github.com/apache/spark/pull/13765
[SPARK-16052][SQL] Add CollapseRepartitionBy optimizer
## What changes were proposed in this pull request?
This issue adds a new optimizer, `CollapseRepartitionBy`, which is similar
Github user dongjoon-hyun commented on a diff in the pull request:
https://github.com/apache/spark/pull/13765#discussion_r67611572
--- Diff: python/pyspark/sql/dataframe.py ---
@@ -451,10 +451,10 @@ def repartition(self, numPartitions, *cols
Github user dongjoon-hyun commented on a diff in the pull request:
https://github.com/apache/spark/pull/13765#discussion_r67610137
--- Diff: python/pyspark/sql/dataframe.py ---
@@ -451,10 +451,10 @@ def repartition(self, numPartitions, *cols
GitHub user dongjoon-hyun opened a pull request:
https://github.com/apache/spark/pull/13763
[SPARK-16051][R] Add `read.orc/write.orc` to SparkR
## What changes were proposed in this pull request?
This issue adds `read.orc/write.orc` to SparkR for API parity.
## How
Github user dongjoon-hyun commented on a diff in the pull request:
https://github.com/apache/spark/pull/13403#discussion_r67597228
--- Diff: core/src/main/scala/org/apache/spark/util/StatCounter.scala ---
@@ -125,9 +128,12 @@ class StatCounter(values: TraversableOnce[Double
Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/13730
Yep. The case still exists for `parquet/csv` and I updated the cases.
The previous `text` case changes like the following and looks legitimate.
```
scala
Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/13734
Hi, @shivaram and @felixcheung .
Now the document is updated with master and shows merged notes correctly. I
manually checked all the merged notes and used function signatures
Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/13730
Hi, @rxin .
Could you review this PR ?
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have
Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/13403
Thank you so much for your review, @srowen !
I updated the PR according to your comments.
---
If your project is set up for it, you can reply to this email and have your
reply appear
Github user dongjoon-hyun commented on a diff in the pull request:
https://github.com/apache/spark/pull/13403#discussion_r67597273
--- Diff: core/src/test/scala/org/apache/spark/PartitioningSuite.scala ---
@@ -244,6 +244,10 @@ class PartitioningSuite extends SparkFunSuite
Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/13635
Oh, great!!
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so
Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/13730
Oh, sorry. The master was changed.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/13730
I will recheck this PR again.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
GitHub user dongjoon-hyun opened a pull request:
https://github.com/apache/spark/pull/13774
[SPARK-16059][R] Add `monotonically_increasing_id` function in SparkR
## What changes were proposed in this pull request?
This PR adds `monotonically_increasing_id` column function
GitHub user dongjoon-hyun opened a pull request:
https://github.com/apache/spark/pull/13870
[SPARK-16165][SQL] Fix the update logic for
InMemoryTableScanExec.readBatches
## What changes were proposed in this pull request?
Currently, `readBatches` accumulator
Github user dongjoon-hyun commented on a diff in the pull request:
https://github.com/apache/spark/pull/13403#discussion_r68188989
--- Diff: core/src/main/scala/org/apache/spark/rdd/DoubleRDDFunctions.scala
---
@@ -74,6 +74,22 @@ class DoubleRDDFunctions(self: RDD[Double]) extends
Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/13872
Sure, I fully agree with your view. That's the declarative language.
However, we can provide more *natural* order as a default order like in
this PR. As you see, without considering
Github user dongjoon-hyun commented on a diff in the pull request:
https://github.com/apache/spark/pull/13403#discussion_r68184739
--- Diff: core/src/main/scala/org/apache/spark/rdd/DoubleRDDFunctions.scala
---
@@ -74,6 +74,22 @@ class DoubleRDDFunctions(self: RDD[Double]) extends
GitHub user dongjoon-hyun opened a pull request:
https://github.com/apache/spark/pull/13872
[SPARK-16164][SQL] Filter pushdown should keep the ordering in the logical
plan
## What changes were proposed in this pull request?
Chris McCubbin reported a bug when he used
Github user dongjoon-hyun commented on a diff in the pull request:
https://github.com/apache/spark/pull/13403#discussion_r68185536
--- Diff: core/src/main/scala/org/apache/spark/rdd/DoubleRDDFunctions.scala
---
@@ -74,6 +74,22 @@ class DoubleRDDFunctions(self: RDD[Double]) extends
Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/13872
For any conclusion, thank you for review, @mengxr and @liancheng !
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your
Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/13730
Hi, @tdas .
Could you give me some advice for the direction about how to change this PR?
---
If your project is set up for it, you can reply to this email and have your
reply appear
Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/13403
Thank you for everything, @srowen , @mengxr , @rxin .
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does
Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/13872
I think I had better change the title of this PR. (I just copied from the
JIRA.)
Does that will reduce your concern a little bit?
---
If your project is set up for it, you can reply
Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/13870
Hi, @liancheng .
Could you review this PR, too?
This was initially introduced in your
https://github.com/apache/spark/commit/74049249abb952ad061c0e221c22ff894a9e9c8d#diff
Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/13854
+1 @hvanhovell .
Since Jenkins does not run scala-2.10 compilation, I ran the build locally
on Ubuntu/JDK7/Scala 2.10 (just for double-check for this PR.)
---
If your project is set
Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/13763
Thank you, @felixcheung !
By the way, unfortunately, `DataFrameReader.scala` provides ORC and Parquet
feature differently.
For ORC, we can accept only one path now
Github user dongjoon-hyun commented on a diff in the pull request:
https://github.com/apache/spark/pull/13768#discussion_r67671989
--- Diff: R/pkg/R/generics.R ---
@@ -1126,6 +1126,10 @@ setGeneric("sort_array", function(x, asc = TRUE) {
standardGeneric(&
Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/13734
By the way, before discussing further, should I retarget this PR for Spark
2.1.0 ?
I think this PR misses the deadline a little bit.
---
If your project is set up for it, you can reply
Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/13763
Thank you so much, @felixcheung !
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/13734
It seems you're concerning multiple issue. I'll focus on the same versions
issue first.
The principle of this PR is simply adding `since` tag for all exposed
functions.
IMHO
Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/13763
Actually, for the ORC, the reason I didn't try to get multiple file is the
API consistently.
Scala/Python also only supports single ORC, so R should does.
I didn't dig futher, but I
Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/13768
Thank you for merging, @shivaram !
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/13403
Ping~
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so
Github user dongjoon-hyun commented on a diff in the pull request:
https://github.com/apache/spark/pull/13763#discussion_r67725053
--- Diff: R/pkg/inst/tests/testthat/test_sparkSQL.R ---
@@ -1667,6 +1668,25 @@ test_that("mutate(), transform(), rename() and
Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/13734
For other important issue about `see also`, all the previous doc look like
that.
http://spark.apache.org/docs/1.6.0/api/R/approxCountDistinct.html
http://spark.apache.org/docs
Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/13786
Hi, @shivaram , @felixcheung .
This is the up-to-date `pivot` PR.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your
GitHub user dongjoon-hyun opened a pull request:
https://github.com/apache/spark/pull/13786
[SPARK-15294][R] Add `pivot` to SparkR
## What changes were proposed in this pull request?
This PR adds `pivot` function to SparkR for API parity. Since this PR is
based on https
Github user dongjoon-hyun commented on a diff in the pull request:
https://github.com/apache/spark/pull/13109#discussion_r67811795
--- Diff: R/pkg/R/stats.R ---
@@ -134,9 +129,7 @@ setMethod("freqItems", signature(x = "SparkDataFrame",
cols = "character&qu
Github user dongjoon-hyun commented on a diff in the pull request:
https://github.com/apache/spark/pull/13109#discussion_r67812243
--- Diff: R/pkg/R/stats.R ---
@@ -134,9 +129,7 @@ setMethod("freqItems", signature(x = "SparkDataFrame",
cols = "character&qu
Github user dongjoon-hyun commented on a diff in the pull request:
https://github.com/apache/spark/pull/13109#discussion_r67812187
--- Diff: R/pkg/R/stats.R ---
@@ -134,9 +129,7 @@ setMethod("freqItems", signature(x = "SparkDataFrame",
cols = "character&qu
Github user dongjoon-hyun commented on a diff in the pull request:
https://github.com/apache/spark/pull/13109#discussion_r67812909
--- Diff: R/pkg/R/stats.R ---
@@ -134,9 +129,7 @@ setMethod("freqItems", signature(x = "SparkDataFrame",
cols = "character&qu
Github user dongjoon-hyun commented on a diff in the pull request:
https://github.com/apache/spark/pull/13798#discussion_r67813477
--- Diff: R/pkg/R/DataFrame.R ---
@@ -606,10 +607,10 @@ setMethod("unpersist",
#'
#' The following options for repartition ar
Github user dongjoon-hyun commented on a diff in the pull request:
https://github.com/apache/spark/pull/13768#discussion_r67750057
--- Diff: R/pkg/R/functions.R ---
@@ -1179,6 +1179,27 @@ setMethod("soundex",
column(jc)
})
Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/13782
LGTM. It's that all?
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled
Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/13734
Thank you, @shivaram !
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled
Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/13774
Thank you, @shivaram and @felixcheung !
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have
Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/13295
This will be really useful. I'll make this PR up-to-date and adds credit
description for @mhnatiuk .
---
If your project is set up for it, you can reply to this email and have your
reply
Github user dongjoon-hyun commented on a diff in the pull request:
https://github.com/apache/spark/pull/13768#discussion_r67744103
--- Diff: R/pkg/R/functions.R ---
@@ -1179,6 +1179,27 @@ setMethod("soundex",
Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/13768
Thank you, @shivaram .
According to your advice and #13394 , I fixed the title convention.
That's all for this PR.
(For my other PRs, I will fix like that, too.)
---
If your
Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/13734
Thank you for narrowing the scope. Indeed, this seems to be an experimental
attempt.
I rebased to resolved the conflict. As @felixcheung mentioned, the use of
function signature
Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/13786
Thank you always, @felixcheung !
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
Github user dongjoon-hyun commented on a diff in the pull request:
https://github.com/apache/spark/pull/13109#discussion_r67803750
--- Diff: R/pkg/R/stats.R ---
@@ -19,7 +19,8 @@
setOldClass("jobj")
-#' crosstab
+#' @title SparkDataFrame statistic
Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/13786
Thank you, @shivaram !
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled
Github user dongjoon-hyun commented on a diff in the pull request:
https://github.com/apache/spark/pull/13109#discussion_r67806331
--- Diff: R/pkg/R/stats.R ---
@@ -19,7 +19,8 @@
setOldClass("jobj")
-#' crosstab
+#' @title SparkDataFrame statistic
Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/13109
In line 333 of `functions.R`, `@rdname covar_pop` -> `@rdname cov`?
```
#' covar_pop
#'
#' Compute the population covariance between two expressions.
#'
#' @rdn
Github user dongjoon-hyun commented on a diff in the pull request:
https://github.com/apache/spark/pull/13109#discussion_r67809806
--- Diff: R/pkg/R/generics.R ---
@@ -430,19 +430,19 @@ setGeneric("coltypes<-", function(x, value) {
standardGeneric("coltypes<-
Github user dongjoon-hyun commented on a diff in the pull request:
https://github.com/apache/spark/pull/13109#discussion_r67806157
--- Diff: R/pkg/R/stats.R ---
@@ -19,7 +19,8 @@
setOldClass("jobj")
-#' crosstab
+#' @title SparkDataFrame statistic
Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/13109
Oh, the root cause exists in `generics.R`. Nice catch!
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does
Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/13109
Yes. Indeed, we had better keep each function on own RD generally.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your
Github user dongjoon-hyun commented on a diff in the pull request:
https://github.com/apache/spark/pull/13786#discussion_r67801232
--- Diff: R/pkg/R/group.R ---
@@ -129,6 +129,48 @@ methods <- c("avg", "max", "mean", "min", "s
Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/13684
Thank you, @shivaram and @sun-rui .
Now, it's ready for review again.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well
Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/13684
Yes, @sun-rui . I realigned the parameter comment.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does
GitHub user dongjoon-hyun opened a pull request:
https://github.com/apache/spark/pull/13714
[SPARK-15996][R] Fix R dataframe example by removing deprecated functions
## What changes were proposed in this pull request?
Currently, R dataframe example fails like the following
Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/13643
Hi, @srowen .
Could you review and merge this PR please?
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project
Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/13714
Hi, @shivaram , @felixcheung , @sun-rui .
Could you review this PR?
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your
Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/13870
No problem! And, thank you for attention! :)
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have
Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/13870
Oh, thank you for review, @davies .
But, sorry. I'm not sure what you mean. Do you mean there is a reason that
`readBatches` should be `0` when the option
Github user dongjoon-hyun commented on a diff in the pull request:
https://github.com/apache/spark/pull/13887#discussion_r68380170
--- Diff:
sql/core/src/main/scala/org/apache/spark/sql/execution/columnar/InMemoryTableScanExec.scala
---
@@ -79,6 +79,11 @@ private[sql] case class
Github user dongjoon-hyun commented on a diff in the pull request:
https://github.com/apache/spark/pull/13887#discussion_r68381115
--- Diff:
sql/core/src/main/scala/org/apache/spark/sql/execution/columnar/InMemoryTableScanExec.scala
---
@@ -79,6 +79,11 @@ private[sql] case class
GitHub user dongjoon-hyun opened a pull request:
https://github.com/apache/spark/pull/13876
[SPARK-16174][SQL] Add RemoveLiteralRepetitionFromIn optimizer
## What changes were proposed in this pull request?
This PR adds an optimizer to remove the duplicated literals from
Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/13887
Hi, @cloud-fan .
I updated the PR. IMO,
- InSet is used for large size of `IN` .
- This PR is used for small size of `IN`.
---
If your project is set up for it, you can reply
Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/13887
Thank you, @cloud-fan !
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled
GitHub user dongjoon-hyun opened a pull request:
https://github.com/apache/spark/pull/13887
[SPARK-16186][SQL] Support partition batch pruning with `IN` predicate in
InMemoryTableScanExec
## What changes were proposed in this pull request?
One of the most frequent usage
Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/13887
cc @rxin , @davies , @cloud-fan .
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/13786
Hi, @Div333 .
@mhnatiuk is right.
For the binary and document, Spark 2.0 is very close to release. You had
better wait. :)
---
If your project is set up for it, you can reply
Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/13887
Although I decided to make this PR after observing TPC-DS queries, I will
definitely update this PR if there are another useful scenarios.
---
If your project is set up for it, you can reply
Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/13730
Rebased.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so
Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/13876
Hi, @rxin .
Could you review this PR again when you have some time?
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your
Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/13765
Rebased.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so
Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/13887
Thank you for your review and valuable improvement ideas, @davies . Let me
rephrase about your ideas,
1. For `IN` with single expression, we definitely had better improve
Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/13734
Thank you, @felixcheung . I removed the note on `dataFrame`.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project
Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/13734
Hi, @shivaram and @felixcheung .
Now, PR and the generated HTML site are up-to-date again. (For HTML site,
you may need to refresh.)
---
If your project is set up for it, you can reply
Github user dongjoon-hyun commented on a diff in the pull request:
https://github.com/apache/spark/pull/13751#discussion_r67604389
--- Diff: docs/sparkr.md ---
@@ -14,29 +14,24 @@ supports operations like selection, filtering,
aggregation etc. (similar to R da
[dplyr](https
Github user dongjoon-hyun commented on a diff in the pull request:
https://github.com/apache/spark/pull/13751#discussion_r67604494
--- Diff: docs/sparkr.md ---
@@ -14,29 +14,24 @@ supports operations like selection, filtering,
aggregation etc. (similar to R da
[dplyr](https
Github user dongjoon-hyun commented on a diff in the pull request:
https://github.com/apache/spark/pull/13486#discussion_r67578803
--- Diff:
sql/core/src/test/scala/org/apache/spark/sql/streaming/test/DataFrameReaderWriterSuite.scala
---
@@ -572,4 +572,16 @@ class
Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/13730
Hi, @tdas .
Could you review this PR again when you have some time?
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your
Github user dongjoon-hyun commented on a diff in the pull request:
https://github.com/apache/spark/pull/13751#discussion_r67604590
--- Diff: docs/sparkr.md ---
@@ -158,20 +152,19 @@ write.df(people, path="people.parquet",
source="parquet", mode="overwr
Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/13486
Oh, I see. I will fix tonight.
Thank you, @tdas !
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does
Github user dongjoon-hyun commented on a diff in the pull request:
https://github.com/apache/spark/pull/13721#discussion_r67457577
--- Diff: R/pkg/R/DataFrame.R ---
@@ -2884,3 +2884,39 @@ setMethod("write.jdbc",
write <- callJMethod(write,
Github user dongjoon-hyun commented on a diff in the pull request:
https://github.com/apache/spark/pull/13721#discussion_r67457739
--- Diff: R/pkg/inst/tests/testthat/test_sparkSQL.R ---
@@ -2264,6 +2264,14 @@ test_that("createDataFrame sqlContext parameter
backward compatib
Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/13684
Thank you, @shivaram !
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled
Github user dongjoon-hyun commented on a diff in the pull request:
https://github.com/apache/spark/pull/13721#discussion_r67457591
--- Diff: R/pkg/inst/tests/testthat/test_sparkSQL.R ---
@@ -2264,6 +2264,14 @@ test_that("createDataFrame sqlContext parameter
backward compatib
Github user dongjoon-hyun commented on a diff in the pull request:
https://github.com/apache/spark/pull/13751#discussion_r67604919
--- Diff: docs/sparkr.md ---
@@ -91,17 +86,17 @@ The following options can be set in `sparkEnvir` with
`sparkR.init` from RStudio
Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/13751
@felixcheung LGTM excepts a few comments.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have
Github user dongjoon-hyun commented on a diff in the pull request:
https://github.com/apache/spark/pull/13751#discussion_r67604853
--- Diff: docs/sparkr.md ---
@@ -113,16 +108,15 @@ head(df)
### From Data Sources
-SparkR supports operating on a variety of data
201 - 300 of 7331 matches
Mail list logo