Github user SparkQA commented on the issue:
https://github.com/apache/spark/pull/14523
**[Test build #63319 has
started](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/63319/consoleFull)**
for PR 14523 at commit
GitHub user gatorsmile opened a pull request:
https://github.com/apache/spark/pull/14523
[SPARK-16936] [SQL] Case Sensitivity Support for Refresh Temp Table
### What changes were proposed in this pull request?
Currently, the `refreshTable` API is always case sensitive.
Github user HyukjinKwon commented on the issue:
https://github.com/apache/spark/pull/14491
@ganeshchand Ah, I see, right. `Dataset` is `DataFrame` so you didn't
change? I skimmed through the list I provided and it seems these are all and it
seems `structured_network_wordcount.py` is
Github user kiszk commented on a diff in the pull request:
https://github.com/apache/spark/pull/13680#discussion_r73795810
--- Diff:
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/ScalaReflection.scala
---
@@ -101,6 +101,8 @@ object ScalaReflection extends
Github user NarineK commented on the issue:
https://github.com/apache/spark/pull/14431
It seems that, currently, in SparkR the `GroupedData` which represents
scala's GroupedData object doesn't have any information about the grouping
keys. `RelationalGroupedDataset` has a private
Github user SparkQA commented on the issue:
https://github.com/apache/spark/pull/13680
**[Test build #63318 has
started](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/63318/consoleFull)**
for PR 13680 at commit
Github user HyukjinKwon commented on the issue:
https://github.com/apache/spark/pull/14509
Actually, `inferSchema` in CSV would be CSV-datasource specific option in
order to allow read the headers as column names but to avoid infer the schema.
---
If your project is set up for it,
Github user HyukjinKwon commented on the issue:
https://github.com/apache/spark/pull/14509
If my understanding is correct, JSON one does have have `inferSchema`
option.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If
Github user AmplabJenkins commented on the issue:
https://github.com/apache/spark/pull/14522
Merged build finished. Test PASSed.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
Github user AmplabJenkins commented on the issue:
https://github.com/apache/spark/pull/14522
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/63317/
Test PASSed.
---
Github user SparkQA commented on the issue:
https://github.com/apache/spark/pull/14522
**[Test build #63317 has
finished](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/63317/consoleFull)**
for PR 14522 at commit
Github user felixcheung commented on a diff in the pull request:
https://github.com/apache/spark/pull/14258#discussion_r73794933
--- Diff: R/pkg/R/install.R ---
@@ -0,0 +1,230 @@
+#
+# Licensed to the Apache Software Foundation (ASF) under one or more
+# contributor
Github user felixcheung commented on a diff in the pull request:
https://github.com/apache/spark/pull/14258#discussion_r73794892
--- Diff: R/pkg/R/install.R ---
@@ -0,0 +1,230 @@
+#
+# Licensed to the Apache Software Foundation (ASF) under one or more
+# contributor
Github user felixcheung commented on a diff in the pull request:
https://github.com/apache/spark/pull/14258#discussion_r73794861
--- Diff: R/pkg/R/sparkR.R ---
@@ -365,6 +365,23 @@ sparkR.session <- function(
}
overrideEnvs(sparkConfigMap, paramMap)
}
+
Github user felixcheung commented on a diff in the pull request:
https://github.com/apache/spark/pull/14258#discussion_r73794831
--- Diff: R/pkg/R/install.R ---
@@ -0,0 +1,230 @@
+#
+# Licensed to the Apache Software Foundation (ASF) under one or more
+# contributor
Github user felixcheung commented on a diff in the pull request:
https://github.com/apache/spark/pull/14258#discussion_r73794815
--- Diff: R/pkg/R/sparkR.R ---
@@ -365,6 +365,23 @@ sparkR.session <- function(
}
overrideEnvs(sparkConfigMap, paramMap)
}
+
Github user asfgit closed the pull request at:
https://github.com/apache/spark/pull/14510
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is
Github user felixcheung commented on the issue:
https://github.com/apache/spark/pull/14258
I think we should go ahead with this and get some usage from the community
if we could as early as possible.
LGTM - we could see if we could improve on how to detect if running from
shell
Github user JoshRosen commented on the issue:
https://github.com/apache/spark/pull/14510
I'm going to merge this to master, branch-2.0, and branch-1.6. I have a
followup patch to add configuration options for controlling the "remove
application that has experienced too many
Github user SparkQA commented on the issue:
https://github.com/apache/spark/pull/14522
**[Test build #63317 has
started](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/63317/consoleFull)**
for PR 14522 at commit
GitHub user junyangq opened a pull request:
https://github.com/apache/spark/pull/14522
[Spark-16508][SparkR] Split docs for arrange and orderBy methods
## What changes were proposed in this pull request?
This PR splits arrange and orderBy methods according to their
Github user AmplabJenkins commented on the issue:
https://github.com/apache/spark/pull/14521
Merged build finished. Test PASSed.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
Github user AmplabJenkins commented on the issue:
https://github.com/apache/spark/pull/14521
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/63316/
Test PASSed.
---
Github user SparkQA commented on the issue:
https://github.com/apache/spark/pull/14521
**[Test build #63316 has
finished](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/63316/consoleFull)**
for PR 14521 at commit
Github user debasish83 commented on the issue:
https://github.com/apache/spark/pull/12574
I will take a pass at the PR as well..
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
Github user debasish83 commented on the issue:
https://github.com/apache/spark/pull/12574
@MLnick I recently visited IBM STC but unfortunately missed you on the
meeting...we discussed about the ML/MLlib changes for matrix factorization...
---
If your project is set up for it, you
Github user JoshRosen commented on a diff in the pull request:
https://github.com/apache/spark/pull/14469#discussion_r73794005
--- Diff: python/pyspark/sql/session.py ---
@@ -384,17 +384,15 @@ def _createFromLocal(self, data, schema):
if schema is None or
Github user SparkQA commented on the issue:
https://github.com/apache/spark/pull/14469
**[Test build #3205 has
finished](https://amplab.cs.berkeley.edu/jenkins/job/NewSparkPullRequestBuilder/3205/consoleFull)**
for PR 14469 at commit
Github user JoshRosen commented on the issue:
https://github.com/apache/spark/pull/14469
This looks pretty good to me overall but I have a couple of clarification
questions regarding some of the doc changes.
---
If your project is set up for it, you can reply to this email and have
Github user JoshRosen commented on a diff in the pull request:
https://github.com/apache/spark/pull/14469#discussion_r73793929
--- Diff: python/pyspark/sql/types.py ---
@@ -582,6 +582,8 @@ def toInternal(self, obj):
else:
if isinstance(obj, dict):
Github user JoshRosen commented on a diff in the pull request:
https://github.com/apache/spark/pull/14469#discussion_r73793910
--- Diff: python/pyspark/sql/tests.py ---
@@ -411,6 +411,21 @@ def test_infer_schema_to_local(self):
df3 = self.spark.createDataFrame(rdd,
Github user JoshRosen commented on a diff in the pull request:
https://github.com/apache/spark/pull/14469#discussion_r73793890
--- Diff: python/pyspark/sql/session.py ---
@@ -432,14 +430,9 @@ def createDataFrame(self, data, schema=None,
samplingRatio=None):
Github user JoshRosen commented on a diff in the pull request:
https://github.com/apache/spark/pull/14469#discussion_r73793841
--- Diff: python/pyspark/sql/session.py ---
@@ -432,14 +430,9 @@ def createDataFrame(self, data, schema=None,
samplingRatio=None):
Github user JoshRosen commented on a diff in the pull request:
https://github.com/apache/spark/pull/14469#discussion_r73793829
--- Diff: python/pyspark/sql/context.py ---
@@ -253,6 +254,8 @@ def createDataFrame(self, data, schema=None,
samplingRatio=None):
If it's
Github user SparkQA commented on the issue:
https://github.com/apache/spark/pull/14469
**[Test build #3205 has
started](https://amplab.cs.berkeley.edu/jenkins/job/NewSparkPullRequestBuilder/3205/consoleFull)**
for PR 14469 at commit
Github user SparkQA commented on the issue:
https://github.com/apache/spark/pull/14521
**[Test build #63316 has
started](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/63316/consoleFull)**
for PR 14521 at commit
GitHub user gatorsmile opened a pull request:
https://github.com/apache/spark/pull/14521
[SPARK-16935] [SQL] Verification of Function-related ExternalCatalog APIs
### What changes were proposed in this pull request?
Function-related `HiveExternalCatalog` APIs do not have enough
Github user shivaram commented on a diff in the pull request:
https://github.com/apache/spark/pull/14489#discussion_r73789927
--- Diff: R/pkg/R/DataFrame.R ---
@@ -41,7 +41,7 @@ setOldClass("structType")
#'\dontrun{
#' sparkR.session()
#' df <-
Github user zasdfgbnm closed the pull request at:
https://github.com/apache/spark/pull/14231
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is
Github user AmplabJenkins commented on the issue:
https://github.com/apache/spark/pull/14520
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/63315/
Test PASSed.
---
Github user AmplabJenkins commented on the issue:
https://github.com/apache/spark/pull/14520
Merged build finished. Test PASSed.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
Github user SparkQA commented on the issue:
https://github.com/apache/spark/pull/14520
**[Test build #63315 has
finished](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/63315/consoleFull)**
for PR 14520 at commit
Github user eyalfa commented on a diff in the pull request:
https://github.com/apache/spark/pull/1#discussion_r73788331
--- Diff:
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/parser/AstBuilder.scala
---
@@ -1101,7 +1101,7 @@ class AstBuilder extends
Github user WeichenXu123 commented on the issue:
https://github.com/apache/spark/pull/14520
Oh..its another algorithm and there are several different details so in
order to make it clear I create a separated PR to discuss it , thanks!
---
If your project is set up for it, you can
Github user srowen commented on the issue:
https://github.com/apache/spark/pull/14520
Let's put this into https://github.com/apache/spark/pull/14109
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does
Github user srowen commented on the issue:
https://github.com/apache/spark/pull/14519
Let's put this into https://github.com/apache/spark/pull/14109
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does
Github user SparkQA commented on the issue:
https://github.com/apache/spark/pull/14520
**[Test build #63315 has
started](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/63315/consoleFull)**
for PR 14520 at commit
Github user WeichenXu123 commented on a diff in the pull request:
https://github.com/apache/spark/pull/14519#discussion_r73787877
--- Diff:
mllib/src/main/scala/org/apache/spark/ml/regression/AFTSurvivalRegression.scala
---
@@ -583,19 +591,22 @@ private class AFTAggregator(
Github user WeichenXu123 commented on the issue:
https://github.com/apache/spark/pull/14520
cc @sethah @yanboliang
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and
GitHub user WeichenXu123 opened a pull request:
https://github.com/apache/spark/pull/14520
[SPARK-16934][ML][MLLib] Improve LogisticCostFun to avoid redundant
serielization
## What changes were proposed in this pull request?
Improve LogisticCostFun, replace closure var
Github user yanboliang commented on the issue:
https://github.com/apache/spark/pull/14519
cc @sethah @dbtsai
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes
Github user AmplabJenkins commented on the issue:
https://github.com/apache/spark/pull/14519
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/63314/
Test PASSed.
---
Github user AmplabJenkins commented on the issue:
https://github.com/apache/spark/pull/14519
Merged build finished. Test PASSed.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
Github user SparkQA commented on the issue:
https://github.com/apache/spark/pull/14519
**[Test build #63314 has
finished](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/63314/consoleFull)**
for PR 14519 at commit
Github user SparkQA commented on the issue:
https://github.com/apache/spark/pull/14519
**[Test build #63314 has
started](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/63314/consoleFull)**
for PR 14519 at commit
GitHub user yanboliang opened a pull request:
https://github.com/apache/spark/pull/14519
[SPARK-16933] [ML] Fix AFTAggregator in AFTSurvivalRegression serializes
unnecessary data.
## What changes were proposed in this pull request?
Similar to ```LeastSquaresAggregator``` in
Github user yanboliang commented on the issue:
https://github.com/apache/spark/pull/14109
The current fix for broadcast variable destroy is ok. LGTM. Thanks!
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your
Github user AmplabJenkins commented on the issue:
https://github.com/apache/spark/pull/14504
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/63313/
Test PASSed.
---
Github user AmplabJenkins commented on the issue:
https://github.com/apache/spark/pull/14504
Merged build finished. Test PASSed.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
Github user SparkQA commented on the issue:
https://github.com/apache/spark/pull/14504
**[Test build #63313 has
finished](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/63313/consoleFull)**
for PR 14504 at commit
Github user SparkQA commented on the issue:
https://github.com/apache/spark/pull/14504
**[Test build #63313 has
started](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/63313/consoleFull)**
for PR 14504 at commit
Github user AmplabJenkins commented on the issue:
https://github.com/apache/spark/pull/14175
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/63312/
Test PASSed.
---
Github user AmplabJenkins commented on the issue:
https://github.com/apache/spark/pull/14175
Build finished. Test PASSed.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled
Github user SparkQA commented on the issue:
https://github.com/apache/spark/pull/14175
**[Test build #63312 has
finished](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/63312/consoleFull)**
for PR 14175 at commit
Github user yanboliang commented on the issue:
https://github.com/apache/spark/pull/14392
@felixcheung @junyangq Any thoughts?
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
Github user biglobster commented on a diff in the pull request:
https://github.com/apache/spark/pull/14477#discussion_r73782824
--- Diff: docs/sql-programming-guide.md ---
@@ -790,6 +790,15 @@ Configuration of Parquet can be done using the
`setConf` method on `SparkSession
Github user SparkQA commented on the issue:
https://github.com/apache/spark/pull/14175
**[Test build #63312 has
started](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/63312/consoleFull)**
for PR 14175 at commit
Github user sun-rui commented on the issue:
https://github.com/apache/spark/pull/14175
@mgummelt, regression test case added. Not sure it is the expected one.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your
Github user srowen commented on a diff in the pull request:
https://github.com/apache/spark/pull/14502#discussion_r73781961
--- Diff: core/src/main/scala/org/apache/spark/rdd/JdbcRDD.scala ---
@@ -79,14 +79,19 @@ class JdbcRDD[T: ClassTag](
val conn = getConnection()
Github user AmplabJenkins commented on the issue:
https://github.com/apache/spark/pull/14502
Merged build finished. Test FAILed.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
Github user AmplabJenkins commented on the issue:
https://github.com/apache/spark/pull/14502
Test FAILed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/63311/
Test FAILed.
---
Github user SparkQA commented on the issue:
https://github.com/apache/spark/pull/14502
**[Test build #63311 has
finished](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/63311/consoleFull)**
for PR 14502 at commit
Github user SparkQA commented on the issue:
https://github.com/apache/spark/pull/14502
**[Test build #63311 has
started](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/63311/consoleFull)**
for PR 14502 at commit
Github user srowen commented on the issue:
https://github.com/apache/spark/pull/14502
Jenkins test this please
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes
Github user AmplabJenkins commented on the issue:
https://github.com/apache/spark/pull/14504
Merged build finished. Test FAILed.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
Github user AmplabJenkins commented on the issue:
https://github.com/apache/spark/pull/14504
Test FAILed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/63310/
Test FAILed.
---
Github user SparkQA commented on the issue:
https://github.com/apache/spark/pull/14504
**[Test build #63310 has
finished](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/63310/consoleFull)**
for PR 14504 at commit
77 matches
Mail list logo