Github user SparkQA commented on the issue:
https://github.com/apache/spark/pull/13403
**[Test build #61106 has
finished](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/61106/consoleFull)**
for PR 13403 at commit
[`bb12a7f`](https://github.com/apache/spark/commit/
Github user srowen commented on the issue:
https://github.com/apache/spark/pull/13624
I think this is correct and can see why it's faster.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this
Github user srowen commented on a diff in the pull request:
https://github.com/apache/spark/pull/13624#discussion_r68205142
--- Diff:
mllib/src/main/scala/org/apache/spark/ml/tree/impl/GradientBoostedTrees.scala
---
@@ -205,31 +205,31 @@ private[spark] object GradientBoostedTrees
Github user lianhuiwang commented on a diff in the pull request:
https://github.com/apache/spark/pull/13494#discussion_r68204608
--- Diff:
sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/FileSourceStrategy.scala
---
@@ -109,108 +111,45 @@ private[sql] object Fil
Github user AmplabJenkins commented on the issue:
https://github.com/apache/spark/pull/13874
Merged build finished. Test FAILed.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
e
Github user SparkQA commented on the issue:
https://github.com/apache/spark/pull/13874
**[Test build #61110 has
finished](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/61110/consoleFull)**
for PR 13874 at commit
[`98027fc`](https://github.com/apache/spark/commit/
Github user AmplabJenkins commented on the issue:
https://github.com/apache/spark/pull/13874
Test FAILed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/61110/
Test FAILed.
---
Github user srowen commented on a diff in the pull request:
https://github.com/apache/spark/pull/13624#discussion_r68203172
--- Diff:
mllib/src/main/scala/org/apache/spark/ml/tree/impl/GradientBoostedTrees.scala
---
@@ -205,31 +205,31 @@ private[spark] object GradientBoostedTrees
Github user srowen commented on a diff in the pull request:
https://github.com/apache/spark/pull/13624#discussion_r68202993
--- Diff:
mllib/src/main/scala/org/apache/spark/ml/tree/impl/GradientBoostedTrees.scala
---
@@ -205,31 +205,31 @@ private[spark] object GradientBoostedTrees
Github user srowen commented on a diff in the pull request:
https://github.com/apache/spark/pull/13624#discussion_r68202942
--- Diff:
mllib/src/main/scala/org/apache/spark/ml/tree/impl/GradientBoostedTrees.scala
---
@@ -205,31 +205,31 @@ private[spark] object GradientBoostedTrees
Github user SparkQA commented on the issue:
https://github.com/apache/spark/pull/13874
**[Test build #61110 has
started](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/61110/consoleFull)**
for PR 13874 at commit
[`98027fc`](https://github.com/apache/spark/commit/9
Github user cloud-fan commented on the issue:
https://github.com/apache/spark/pull/13874
cc @hvanhovell
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or
GitHub user cloud-fan opened a pull request:
https://github.com/apache/spark/pull/13874
[SQL][minor] ParserUtils.operationNotAllowed should throw exception directly
## What changes were proposed in this pull request?
It's weird that `ParserUtils.operationNotAllowed` returns
Github user SparkQA commented on the issue:
https://github.com/apache/spark/pull/13865
**[Test build #61109 has
started](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/61109/consoleFull)**
for PR 13865 at commit
[`ebed01e`](https://github.com/apache/spark/commit/e
Github user AmplabJenkins commented on the issue:
https://github.com/apache/spark/pull/13872
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/61105/
Test PASSed.
---
Github user AmplabJenkins commented on the issue:
https://github.com/apache/spark/pull/13872
Merged build finished. Test PASSed.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
e
Github user SparkQA commented on the issue:
https://github.com/apache/spark/pull/13872
**[Test build #61105 has
finished](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/61105/consoleFull)**
for PR 13872 at commit
[`2b21fd7`](https://github.com/apache/spark/commit/
Github user cloud-fan commented on the issue:
https://github.com/apache/spark/pull/13756
I'm thinking about if it's possible to concentrate error checking logics at
one place for table creation. For example, we check duplicated table column
names at parser for SQL
statement(https://g
Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/13872
For any conclusion, thank you for review, @mengxr and @liancheng !
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your proje
Github user AmplabJenkins commented on the issue:
https://github.com/apache/spark/pull/13839
Merged build finished. Test PASSed.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
e
Github user AmplabJenkins commented on the issue:
https://github.com/apache/spark/pull/13839
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/61102/
Test PASSed.
---
Github user SparkQA commented on the issue:
https://github.com/apache/spark/pull/13839
**[Test build #61102 has
finished](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/61102/consoleFull)**
for PR 13839 at commit
[`b4d9279`](https://github.com/apache/spark/commit/
Github user SparkQA commented on the issue:
https://github.com/apache/spark/pull/13873
**[Test build #61108 has
started](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/61108/consoleFull)**
for PR 13873 at commit
[`093a9fa`](https://github.com/apache/spark/commit/0
Github user AmplabJenkins commented on the issue:
https://github.com/apache/spark/pull/13836
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/61101/
Test PASSed.
---
GitHub user ueshin opened a pull request:
https://github.com/apache/spark/pull/13873
[SPARK-16167][SQL] RowEncoder should preserve array/map type nullability.
## What changes were proposed in this pull request?
Currently `RowEncoder` doesn't preserve nullability of `ArrayTyp
Github user AmplabJenkins commented on the issue:
https://github.com/apache/spark/pull/13859
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/61107/
Test PASSed.
---
Github user AmplabJenkins commented on the issue:
https://github.com/apache/spark/pull/13836
Merged build finished. Test PASSed.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
e
Github user AmplabJenkins commented on the issue:
https://github.com/apache/spark/pull/13859
Merged build finished. Test PASSed.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
e
Github user SparkQA commented on the issue:
https://github.com/apache/spark/pull/13859
**[Test build #61107 has
finished](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/61107/consoleFull)**
for PR 13859 at commit
[`c445b93`](https://github.com/apache/spark/commit/
Github user SparkQA commented on the issue:
https://github.com/apache/spark/pull/13836
**[Test build #61101 has
finished](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/61101/consoleFull)**
for PR 13836 at commit
[`8d2dea7`](https://github.com/apache/spark/commit/
Github user maropu commented on the issue:
https://github.com/apache/spark/pull/13847
I'm now checking failed tests...
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and
Github user tnachen commented on a diff in the pull request:
https://github.com/apache/spark/pull/13143#discussion_r68194509
--- Diff:
core/src/main/scala/org/apache/spark/scheduler/cluster/mesos/MesosSchedulerUtils.scala
---
@@ -120,14 +120,25 @@ private[mesos] trait MesosSchedul
Github user cloud-fan commented on a diff in the pull request:
https://github.com/apache/spark/pull/13494#discussion_r68193963
--- Diff:
sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/FileSourceStrategy.scala
---
@@ -109,108 +111,45 @@ private[sql] object FileS
Github user AmplabJenkins commented on the issue:
https://github.com/apache/spark/pull/13870
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/61100/
Test PASSed.
---
Github user SparkQA commented on the issue:
https://github.com/apache/spark/pull/13870
**[Test build #61100 has
finished](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/61100/consoleFull)**
for PR 13870 at commit
[`b1a80dd`](https://github.com/apache/spark/commit/
Github user AmplabJenkins commented on the issue:
https://github.com/apache/spark/pull/13870
Merged build finished. Test PASSed.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
e
Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/13872
I think I had better change the title of this PR. (I just copied from the
JIRA.)
Does that will reduce your concern a little bit?
---
If your project is set up for it, you can reply to th
Github user AmplabJenkins commented on the issue:
https://github.com/apache/spark/pull/13756
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/61099/
Test PASSed.
---
Github user AmplabJenkins commented on the issue:
https://github.com/apache/spark/pull/13756
Merged build finished. Test PASSed.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
e
Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/13872
Sure, I fully agree with your view. That's the declarative language.
However, we can provide more *natural* order as a default order like in
this PR. As you see, without considering th
Github user SparkQA commented on the issue:
https://github.com/apache/spark/pull/13756
**[Test build #61099 has
finished](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/61099/consoleFull)**
for PR 13756 at commit
[`24edb5f`](https://github.com/apache/spark/commit/
Github user liancheng commented on the issue:
https://github.com/apache/spark/pull/13872
@dongjoon-hyun Thanks for the work! However, I think the optimizer should
have the freedom to reorder predicate evaluation order. For example, we may
evaluate cheap predicates first in order to sh
Github user AmplabJenkins commented on the issue:
https://github.com/apache/spark/pull/13858
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/61098/
Test PASSed.
---
Github user AmplabJenkins commented on the issue:
https://github.com/apache/spark/pull/13858
Merged build finished. Test PASSed.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
e
Github user SparkQA commented on the issue:
https://github.com/apache/spark/pull/13858
**[Test build #61098 has
finished](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/61098/consoleFull)**
for PR 13858 at commit
[`b497dc9`](https://github.com/apache/spark/commit/
Github user srowen commented on the issue:
https://github.com/apache/spark/pull/13868
Yes, this is not the change discussed in the JIRA. The best way forward
seems to be to replace attempts to make a `file:` URI manually from a string
with use of `File.toURI` or something from Java 7'
Github user lianhuiwang commented on the issue:
https://github.com/apache/spark/pull/13494
@cloud-fan Yes, I think what you said is right. as Hive/Prestodb, if
queries that did some functions (example: MIN/MAX) or distinct aggregates on
partition column and the value of config 'spark.
Github user mengxr commented on a diff in the pull request:
https://github.com/apache/spark/pull/13841#discussion_r68191189
--- Diff:
mllib/src/main/scala/org/apache/spark/ml/classification/LogisticRegression.scala
---
@@ -674,12 +674,12 @@ object LogisticRegressionModel extends
Github user SparkQA commented on the issue:
https://github.com/apache/spark/pull/13859
**[Test build #61107 has
started](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/61107/consoleFull)**
for PR 13859 at commit
[`c445b93`](https://github.com/apache/spark/commit/c
Github user mengxr commented on a diff in the pull request:
https://github.com/apache/spark/pull/13844#discussion_r68191096
--- Diff:
mllib/src/main/scala/org/apache/spark/ml/feature/MinMaxScaler.scala ---
@@ -232,7 +233,9 @@ object MinMaxScalerModel extends
MLReadable[MinMaxScale
Github user mengxr commented on a diff in the pull request:
https://github.com/apache/spark/pull/13859#discussion_r68190859
--- Diff: mllib/src/main/scala/org/apache/spark/mllib/package-info.java ---
@@ -16,6 +16,26 @@
*/
/**
- * Spark's machine learning library
Github user SparkQA commented on the issue:
https://github.com/apache/spark/pull/13403
**[Test build #61106 has
started](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/61106/consoleFull)**
for PR 13403 at commit
[`bb12a7f`](https://github.com/apache/spark/commit/b
Github user lianhuiwang commented on a diff in the pull request:
https://github.com/apache/spark/pull/13494#discussion_r68190320
--- Diff:
sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/FileSourceStrategy.scala
---
@@ -109,108 +111,45 @@ private[sql] object Fil
Github user MLnick commented on the issue:
https://github.com/apache/spark/pull/13859
Looks good subject to @hhbyyh comment above
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
Github user dongjoon-hyun commented on a diff in the pull request:
https://github.com/apache/spark/pull/13403#discussion_r68188989
--- Diff: core/src/main/scala/org/apache/spark/rdd/DoubleRDDFunctions.scala
---
@@ -74,6 +74,22 @@ class DoubleRDDFunctions(self: RDD[Double]) extends
Github user mengxr commented on the issue:
https://github.com/apache/spark/pull/13872
cc: @liancheng
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if
Github user srowen commented on a diff in the pull request:
https://github.com/apache/spark/pull/13403#discussion_r68188408
--- Diff: core/src/main/scala/org/apache/spark/rdd/DoubleRDDFunctions.scala
---
@@ -74,6 +74,22 @@ class DoubleRDDFunctions(self: RDD[Double]) extends
Loggin
Github user SparkQA commented on the issue:
https://github.com/apache/spark/pull/13872
**[Test build #61105 has
started](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/61105/consoleFull)**
for PR 13872 at commit
[`2b21fd7`](https://github.com/apache/spark/commit/2
Github user renozhang commented on a diff in the pull request:
https://github.com/apache/spark/pull/13836#discussion_r68188071
--- Diff: python/pyspark/context.py ---
@@ -156,7 +156,7 @@ def _do_init(self, master, appName, sparkHome, pyFiles,
environment, batchSize,
se
GitHub user dongjoon-hyun opened a pull request:
https://github.com/apache/spark/pull/13872
[SPARK-16164][SQL] Filter pushdown should keep the ordering in the logical
plan
## What changes were proposed in this pull request?
Chris McCubbin reported a bug when he used StringI
Github user HyukjinKwon commented on the issue:
https://github.com/apache/spark/pull/13771
Hi @davies, it seems related codes were written by you. Would this be a
meaningful change?
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub
Github user cloud-fan commented on the issue:
https://github.com/apache/spark/pull/13494
hi @lianhuiwang , thanks for working on it!
The overall idea LGTM, we should elimiante unnecessary file scan if only
partition columns are read. However, the current implementation looks n
Github user cloud-fan commented on a diff in the pull request:
https://github.com/apache/spark/pull/13494#discussion_r68186202
--- Diff:
sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/FileSourceStrategy.scala
---
@@ -109,108 +111,45 @@ private[sql] object FileS
Github user MLnick commented on the issue:
https://github.com/apache/spark/pull/13745
@GayathriMurali couple final comments, then I think it's good to go. Thanks!
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your pr
Github user AmplabJenkins commented on the issue:
https://github.com/apache/spark/pull/13871
Test FAILed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/61104/
Test FAILed.
---
Github user SparkQA commented on the issue:
https://github.com/apache/spark/pull/13871
**[Test build #61104 has
finished](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/61104/consoleFull)**
for PR 13871 at commit
[`ecdf2b8`](https://github.com/apache/spark/commit/
Github user dongjoon-hyun commented on a diff in the pull request:
https://github.com/apache/spark/pull/13403#discussion_r68185536
--- Diff: core/src/main/scala/org/apache/spark/rdd/DoubleRDDFunctions.scala
---
@@ -74,6 +74,22 @@ class DoubleRDDFunctions(self: RDD[Double]) extends
Github user AmplabJenkins commented on the issue:
https://github.com/apache/spark/pull/13871
Merged build finished. Test FAILed.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
e
Github user SparkQA commented on the issue:
https://github.com/apache/spark/pull/13871
**[Test build #61104 has
started](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/61104/consoleFull)**
for PR 13871 at commit
[`ecdf2b8`](https://github.com/apache/spark/commit/e
Github user srowen commented on a diff in the pull request:
https://github.com/apache/spark/pull/13699#discussion_r68184716
--- Diff:
core/src/main/java/org/apache/spark/shuffle/sort/UnsafeShuffleWriter.java ---
@@ -122,6 +123,8 @@ public UnsafeShuffleWriter(
this.taskCont
Github user dongjoon-hyun commented on a diff in the pull request:
https://github.com/apache/spark/pull/13403#discussion_r68184739
--- Diff: core/src/main/scala/org/apache/spark/rdd/DoubleRDDFunctions.scala
---
@@ -74,6 +74,22 @@ class DoubleRDDFunctions(self: RDD[Double]) extends
Github user AmplabJenkins commented on the issue:
https://github.com/apache/spark/pull/13758
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/61097/
Test PASSed.
---
GitHub user davies opened a pull request:
https://github.com/apache/spark/pull/13871
[SPARK-16163] [SQL] Cache the statistics for logical plans
## What changes were proposed in this pull request?
This calculation of statistics is not trivial anymore, it could be very
slow o
Github user davies commented on the issue:
https://github.com/apache/spark/pull/13871
cc @liancheng
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if
Github user AmplabJenkins commented on the issue:
https://github.com/apache/spark/pull/13758
Merged build finished. Test PASSed.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
e
Github user SparkQA commented on the issue:
https://github.com/apache/spark/pull/13758
**[Test build #61097 has
finished](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/61097/consoleFull)**
for PR 13758 at commit
[`1f1d77c`](https://github.com/apache/spark/commit/
Github user SparkQA commented on the issue:
https://github.com/apache/spark/pull/13839
**[Test build #61102 has
started](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/61102/consoleFull)**
for PR 13839 at commit
[`b4d9279`](https://github.com/apache/spark/commit/b
Github user srowen commented on a diff in the pull request:
https://github.com/apache/spark/pull/13403#discussion_r68184371
--- Diff: core/src/main/scala/org/apache/spark/rdd/DoubleRDDFunctions.scala
---
@@ -74,6 +74,22 @@ class DoubleRDDFunctions(self: RDD[Double]) extends
Loggin
Github user SparkQA commented on the issue:
https://github.com/apache/spark/pull/13834
**[Test build #61103 has
started](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/61103/consoleFull)**
for PR 13834 at commit
[`04c8637`](https://github.com/apache/spark/commit/0
Github user ScrapCodes commented on a diff in the pull request:
https://github.com/apache/spark/pull/13839#discussion_r68184305
--- Diff: R/pkg/R/DataFrame.R ---
@@ -194,7 +195,13 @@ setMethod("isLocal",
setMethod("showDF",
signature(x = "SparkDataFrame"),
Github user srowen commented on the issue:
https://github.com/apache/spark/pull/13834
Jenkins retest this please
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishe
401 - 481 of 481 matches
Mail list logo