Github user AmplabJenkins commented on the pull request:
https://github.com/apache/spark/pull/3401#issuecomment-65178375
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/24
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/3401#issuecomment-65178372
[Test build #24016 has
finished](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/24016/consoleFull)
for PR 3401 at commit
[`07700cf`](https://gith
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/3441#issuecomment-65177621
[Test build #24019 has
started](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/24019/consoleFull)
for PR 3441 at commit
[`f6a587f`](https://githu
Github user AmplabJenkins commented on the pull request:
https://github.com/apache/spark/pull/2576#issuecomment-65177482
Test FAILed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/24
Github user liancheng commented on a diff in the pull request:
https://github.com/apache/spark/pull/3441#discussion_r21138093
--- Diff: sql/core/src/main/scala/org/apache/spark/sql/SQLConf.scala ---
@@ -134,12 +133,6 @@ private[sql] trait SQLConf {
getConf(DEFAULT_SIZE_IN_B
Github user AmplabJenkins commented on the pull request:
https://github.com/apache/spark/pull/3443#issuecomment-65177298
Test FAILed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/24
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/3443#issuecomment-65177296
[Test build #24018 has
finished](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/24018/consoleFull)
for PR 3443 at commit
[`36236a5`](https://gith
Github user pwendell commented on the pull request:
https://github.com/apache/spark/pull/3542#issuecomment-65177161
LGTM
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled a
Github user AmplabJenkins commented on the pull request:
https://github.com/apache/spark/pull/3543#issuecomment-65176806
Can one of the admins verify this patch?
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your pro
Github user koeninger closed the pull request at:
https://github.com/apache/spark/pull/3102
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is e
Github user adrian-wang commented on a diff in the pull request:
https://github.com/apache/spark/pull/3528#discussion_r21137835
--- Diff:
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/arithmetic.scala
---
@@ -185,11 +184,11 @@ case class BitwiseNot(child: E
Github user koeninger commented on the pull request:
https://github.com/apache/spark/pull/3102#issuecomment-65176731
Yes, the new hadoop config documentation is just documenting the behavior
of SparkHadoopUtil.scala lines 95-100
Sorry about the branch situation, I was unclear
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/3443#issuecomment-65176683
[Test build #24018 has
started](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/24018/consoleFull)
for PR 3443 at commit
[`36236a5`](https://githu
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/3526#issuecomment-65176665
[Test build #24017 has
started](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/24017/consoleFull)
for PR 3526 at commit
[`d446e14`](https://githu
GitHub user koeninger opened a pull request:
https://github.com/apache/spark/pull/3543
Closes SPARK-4229 Create hadoop configuration in a consistent way
You can merge this pull request into a Git repository by running:
$ git pull https://github.com/koeninger/spark-1 SPARK-4229
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/3526#issuecomment-65176378
had modified according @marmbrus commentï¼)
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your proje
Github user rxin commented on the pull request:
https://github.com/apache/spark/pull/3531#issuecomment-65176150
Actually one more question before you make big changes: executeCollect
should be called most of the time (if you run a sql query). In what cases did
you run into this proble
Github user rxin commented on the pull request:
https://github.com/apache/spark/pull/3531#issuecomment-65176091
BTW it doesn't have to be a new operator. Can also just add a flag to Limit.
---
If your project is set up for it, you can reply to this email and have your
reply appear on
Github user scwf commented on the pull request:
https://github.com/apache/spark/pull/3249#issuecomment-65176098
Hi @ravipesala, can you rebase this PR
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does n
Github user scwf commented on the pull request:
https://github.com/apache/spark/pull/3531#issuecomment-65175802
Yes, we can not change to coalesce here, I agree with you of coalesce's
advantages situation, and i will try to do the optimization with coalesce for
no filter. Thanks;)
Github user AmplabJenkins commented on the pull request:
https://github.com/apache/spark/pull/3249#issuecomment-65175714
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/24
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/3249#issuecomment-65175709
[Test build #24010 has
finished](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/24010/consoleFull)
for PR 3249 at commit
[`59dfab5`](https://gith
Github user rxin commented on the pull request:
https://github.com/apache/spark/pull/3470#issuecomment-65175457
That definitely seems like the better option to me. It makes the options
universally lower case.
---
If your project is set up for it, you can reply to this email and have
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/3500#issuecomment-65175355
[Test build #24013 has
finished](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/24013/consoleFull)
for PR 3500 at commit
[`21cf624`](https://gith
Github user AmplabJenkins commented on the pull request:
https://github.com/apache/spark/pull/3500#issuecomment-65175358
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/24
Github user AmplabJenkins commented on the pull request:
https://github.com/apache/spark/pull/3486#issuecomment-65175297
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/24
Github user marmbrus commented on the pull request:
https://github.com/apache/spark/pull/3470#issuecomment-65175244
Yeah, that's what I was thinking when I said case insensitive map.
On Dec 1, 2014 6:33 PM, "Reynold Xin" wrote:
> How about moving the toLowerCase into the g
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/3486#issuecomment-65175291
[Test build #24012 has
finished](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/24012/consoleFull)
for PR 3486 at commit
[`501f284`](https://gith
Github user scwf commented on the pull request:
https://github.com/apache/spark/pull/3470#issuecomment-65175161
Yes, We can implement a case insensitive map and in this map's ```get```
function we use toLowercase
---
If your project is set up for it, you can reply to this email and
Github user rxin commented on the pull request:
https://github.com/apache/spark/pull/3531#issuecomment-65175014
I think it is too risky to do this this way right now. It seems to me the
main advantage of using coalesce is when you have a huge number of partitions
without a highly sele
Github user rxin commented on the pull request:
https://github.com/apache/spark/pull/3470#issuecomment-65174831
How about moving the toLowerCase into the get funciton itself?
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well
Github user scwf commented on the pull request:
https://github.com/apache/spark/pull/3531#issuecomment-65174711
Yes, i also realize this, it will not be always faster, since ```coalesce
(1)``` will lead to run with a single thread.
---
If your project is set up for it, you can reply
Github user rxin commented on the pull request:
https://github.com/apache/spark/pull/3531#issuecomment-65174393
@scwf I am not sure if this is a good idea in general. Think about a highly
selective filter, e.g.
```sql
select * from every_body_in_the_world where company="Databri
Github user AmplabJenkins commented on the pull request:
https://github.com/apache/spark/pull/3542#issuecomment-65174162
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/24
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/3542#issuecomment-65174154
[Test build #24008 has
finished](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/24008/consoleFull)
for PR 3542 at commit
[`a8fc43b`](https://gith
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/3401#issuecomment-65174072
[Test build #24016 has
started](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/24016/consoleFull)
for PR 3401 at commit
[`07700cf`](https://githu
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/3486#issuecomment-65173450
[Test build #24011 has
finished](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/24011/consoleFull)
for PR 3486 at commit
[`864fc47`](https://gith
Github user AmplabJenkins commented on the pull request:
https://github.com/apache/spark/pull/3486#issuecomment-65173451
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/24
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/3401#issuecomment-65172981
[Test build #24015 has
finished](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/24015/consoleFull)
for PR 3401 at commit
[`bc5f4cb`](https://gith
Github user AmplabJenkins commented on the pull request:
https://github.com/apache/spark/pull/3348#issuecomment-65172953
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/24
Github user AmplabJenkins commented on the pull request:
https://github.com/apache/spark/pull/3401#issuecomment-65172988
Test FAILed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/24
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/3348#issuecomment-65172947
[Test build #24007 has
finished](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/24007/consoleFull)
for PR 3348 at commit
[`4a31ca7`](https://gith
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/3401#issuecomment-65172365
[Test build #24015 has
started](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/24015/consoleFull)
for PR 3401 at commit
[`bc5f4cb`](https://githu
Github user sarutak commented on a diff in the pull request:
https://github.com/apache/spark/pull/3433#discussion_r21136072
--- Diff:
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/SqlParser.scala ---
@@ -362,7 +362,7 @@ class SqlParser extends AbstractSparkSQLParser {
Github user wangxiaojing commented on a diff in the pull request:
https://github.com/apache/spark/pull/3442#discussion_r21136073
--- Diff: sql/core/src/test/scala/org/apache/spark/sql/JoinSuite.scala ---
@@ -377,4 +378,39 @@ class JoinSuite extends QueryTest with
BeforeAndAfterEach
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/2576#issuecomment-65172020
[Test build #24014 has
started](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/24014/consoleFull)
for PR 2576 at commit
[`601d242`](https://githu
Github user scwf commented on the pull request:
https://github.com/apache/spark/pull/3414#issuecomment-65171947
Ok, if @liancheng fix them, i will close this two PR.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your
Github user sarutak commented on a diff in the pull request:
https://github.com/apache/spark/pull/3401#discussion_r21135915
--- Diff: sql/core/src/test/scala/org/apache/spark/sql/DslQuerySuite.scala
---
@@ -282,4 +283,72 @@ class DslQuerySuite extends QueryTest {
(1, "1"
Github user scwf closed the pull request at:
https://github.com/apache/spark/pull/3352
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enable
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/3395#issuecomment-65171371
[Test build #24009 has
finished](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/24009/consoleFull)
for PR 3395 at commit
[`53c3952`](https://gith
Github user AmplabJenkins commented on the pull request:
https://github.com/apache/spark/pull/3395#issuecomment-65171373
Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/24
Github user scwf closed the pull request at:
https://github.com/apache/spark/pull/3416
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enable
Github user scwf commented on the pull request:
https://github.com/apache/spark/pull/3416#issuecomment-65171217
This have be fixed by @liancheng in #3506. to close this.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If
Github user scwf commented on the pull request:
https://github.com/apache/spark/pull/3531#issuecomment-65170151
I tested with limit number 5000. I am testing more for this. I do not think
limit number has big affect.
---
If your project is set up for it, you can reply to this email a
Github user marmbrus commented on a diff in the pull request:
https://github.com/apache/spark/pull/2589#discussion_r21135116
--- Diff:
sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/HiveTableScanSuite.scala
---
@@ -68,5 +76,5 @@ class HiveTableScanSuite extends HiveCo
Github user marmbrus commented on the pull request:
https://github.com/apache/spark/pull/2589#issuecomment-65169782
Sorry for the delay. I like the new approach of only doing this in local
mode for testing. LGTM with one minor comment.
---
If your project is set up for it, you can
Github user asfgit closed the pull request at:
https://github.com/apache/spark/pull/1737
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enab
Github user asfgit closed the pull request at:
https://github.com/apache/spark/pull/2029
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enab
Github user asfgit closed the pull request at:
https://github.com/apache/spark/pull/2386
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enab
Github user josephsu commented on the pull request:
https://github.com/apache/spark/pull/1737#issuecomment-65169706
no problem. thanks for heads up!
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does no
Github user asfgit closed the pull request at:
https://github.com/apache/spark/pull/1612
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enab
Github user asfgit closed the pull request at:
https://github.com/apache/spark/pull/2997
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enab
Github user asfgit closed the pull request at:
https://github.com/apache/spark/pull/2252
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enab
Github user asfgit closed the pull request at:
https://github.com/apache/spark/pull/2723
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enab
Github user marmbrus commented on the pull request:
https://github.com/apache/spark/pull/1875#issuecomment-65169491
Any update here? I think this would be a great feature to have, but
perhaps we should close this issue until it is ready to review (to make the
size of the PR queue a l
Github user kozanitis commented on the pull request:
https://github.com/apache/spark/pull/2939#issuecomment-65169392
Yeah I see your point... I will revise the design as soon as I find some
time to do some more clever parsing of the existing predicates
---
If your project is set up f
Github user tsudukim commented on the pull request:
https://github.com/apache/spark/pull/3500#issuecomment-65169336
@sryza and @tgravescs Thank you for your review. I removed them. Only
`spark.yarn.queue` is added.
---
If your project is set up for it, you can reply to this email and
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/3500#issuecomment-65169098
[Test build #24013 has
started](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/24013/consoleFull)
for PR 3500 at commit
[`21cf624`](https://githu
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/3486#issuecomment-65168894
[Test build #24012 has
started](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/24012/consoleFull)
for PR 3486 at commit
[`501f284`](https://githu
Github user marmbrus commented on the pull request:
https://github.com/apache/spark/pull/2386#issuecomment-65168822
Thanks for working on this, however, since its a duplicate I think we
should probably close this issue and continue any discussion on #1875.
---
If your project is set
Github user marmbrus commented on the pull request:
https://github.com/apache/spark/pull/2252#issuecomment-65168543
Thanks for working on this, but we are trying to clean up the PR queue (in
order to make it easier for us to review). Thus, I think we should close this
issue for now an
Github user marmbrus commented on the pull request:
https://github.com/apache/spark/pull/1737#issuecomment-65168627
Thanks for working on this, but we are trying to clean up the PR queue (in
order to make it easier for us to review). Thus, I think we should close this
issue for now an
Github user marmbrus commented on the pull request:
https://github.com/apache/spark/pull/2186#issuecomment-65168502
ping @JoshRosen, trying to decrease the number of open SQL PRs ;)
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub
Github user yhuai commented on a diff in the pull request:
https://github.com/apache/spark/pull/3173#discussion_r21134546
--- Diff:
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/plans/physical/partitioning.scala
---
@@ -163,6 +177,37 @@ case class HashPartitioning(expr
Github user marmbrus commented on a diff in the pull request:
https://github.com/apache/spark/pull/3414#discussion_r21134536
--- Diff:
sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveContext.scala ---
@@ -83,6 +83,14 @@ class HiveContext(sc: SparkContext) extends
SQLContext(
Github user marmbrus commented on a diff in the pull request:
https://github.com/apache/spark/pull/3441#discussion_r21134521
--- Diff: sql/core/src/main/scala/org/apache/spark/sql/SQLConf.scala ---
@@ -134,12 +133,6 @@ private[sql] trait SQLConf {
getConf(DEFAULT_SIZE_IN_BY
Github user chenghao-intel commented on the pull request:
https://github.com/apache/spark/pull/3531#issuecomment-65167920
Is there a assumption, the LIMIT number is quite small?
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as w
Github user yhuai commented on a diff in the pull request:
https://github.com/apache/spark/pull/3173#discussion_r21134250
--- Diff:
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/plans/physical/partitioning.scala
---
@@ -163,6 +177,37 @@ case class HashPartitioning(expr
Github user yhuai commented on a diff in the pull request:
https://github.com/apache/spark/pull/3173#discussion_r21134197
--- Diff:
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/plans/physical/partitioning.scala
---
@@ -59,6 +59,20 @@ case class ClusteredDistribution(c
Github user ksakellis commented on a diff in the pull request:
https://github.com/apache/spark/pull/3486#discussion_r21134086
--- Diff:
core/src/main/scala/org/apache/spark/executor/CoarseGrainedExecutorBackend.scala
---
@@ -50,10 +50,16 @@ private[spark] class CoarseGrainedExecut
Github user scwf commented on the pull request:
https://github.com/apache/spark/pull/3531#issuecomment-65167159
hi @marmbrus, the old version also eliminate the parallelism to 1 by a
shuffledRDD, the diff is this PR using ```coalesce``` to the same thing but
avoid shuffle(which will w
Github user manishamde commented on the pull request:
https://github.com/apache/spark/pull/3461#issuecomment-65167028
@jkbradley The GBDT sections looks good to me but the subsection on
Comparison with RFs could possibly be moved towards the end. It breaks the flow
in my opinion.
--
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/3486#issuecomment-65166835
[Test build #24011 has
started](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/24011/consoleFull)
for PR 3486 at commit
[`864fc47`](https://githu
Github user marmbrus commented on a diff in the pull request:
https://github.com/apache/spark/pull/2802#discussion_r21133711
--- Diff: sql/hive/src/main/scala/org/apache/spark/sql/hive/hiveUdfs.scala
---
@@ -172,6 +177,8 @@ private[hive] case class
HiveGenericUdf(functionClassName
Github user marmbrus commented on a diff in the pull request:
https://github.com/apache/spark/pull/2802#discussion_r21133683
--- Diff: sql/hive/src/main/scala/org/apache/spark/sql/hive/hiveUdfs.scala
---
@@ -165,6 +165,11 @@ private[hive] case class
HiveGenericUdf(functionClassNam
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/3249#issuecomment-65166584
[Test build #24010 has
started](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/24010/consoleFull)
for PR 3249 at commit
[`59dfab5`](https://githu
Github user manishamde commented on a diff in the pull request:
https://github.com/apache/spark/pull/3461#discussion_r21133672
--- Diff: docs/mllib-decision-tree.md ---
@@ -103,36 +106,73 @@ and the resulting `$M-1$` split candidates are
considered.
### Stopping rule
Github user vanzin commented on the pull request:
https://github.com/apache/spark/pull/3409#issuecomment-65166307
@zhzhan using this new config to set memory parameters seems like a recipe
for confusion. If memory is specified both in `spark.yarn.am.extraJavaOptions`
and `spark.driver
Github user marmbrus commented on the pull request:
https://github.com/apache/spark/pull/3249#issuecomment-65166234
ok to test
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
ena
Github user marmbrus commented on the pull request:
https://github.com/apache/spark/pull/#issuecomment-65166049
@liancheng, I know you are doing a bunch of refactoring on these tests.
Can you please comment on the approach here? Perhaps it would be better to add
these after you
Github user marmbrus commented on the pull request:
https://github.com/apache/spark/pull/3173#issuecomment-65165934
/cc @yhuai for the changes to our partitioning API.
I also agree with @chenghao-intel that we probably want to keep
ShuffleHashJoin and also that we need some pe
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/3395#issuecomment-65165713
[Test build #24009 has
started](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/24009/consoleFull)
for PR 3395 at commit
[`53c3952`](https://githu
Github user marmbrus commented on a diff in the pull request:
https://github.com/apache/spark/pull/3133#discussion_r21133097
--- Diff:
sql/core/src/main/scala/org/apache/spark/sql/execution/joins/BroadcastHashJoin.scala
---
@@ -42,6 +42,15 @@ case class BroadcastHashJoin(
Github user marmbrus commented on the pull request:
https://github.com/apache/spark/pull/3395#issuecomment-65165293
ok to test
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
ena
Github user asfgit closed the pull request at:
https://github.com/apache/spark/pull/3132
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enab
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/3542#issuecomment-65165236
[Test build #24008 has
started](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/24008/consoleFull)
for PR 3542 at commit
[`a8fc43b`](https://githu
Github user marmbrus commented on a diff in the pull request:
https://github.com/apache/spark/pull/3132#discussion_r21132851
--- Diff:
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/SparkSQLParser.scala
---
@@ -97,10 +97,10 @@ class SqlLexical(val keywords: Seq[String])
Github user marmbrus commented on the pull request:
https://github.com/apache/spark/pull/3102#issuecomment-65164900
Sorry for the delay here. A few comments: can you open the PR against
master instead of a specific branch and also merge with master?
The new hadoop config docu
GitHub user kayousterhout opened a pull request:
https://github.com/apache/spark/pull/3542
[SPARK-4686] Link to allowed master URLs is broken
The link points to the old scala programming guide; it should point to the
submitting applications page.
You can merge this pull request int
Github user marmbrus commented on the pull request:
https://github.com/apache/spark/pull/2939#issuecomment-65164578
Thanks for working on this, my biggest concern is adding new syntax to SQL
since it is something we will have to support forever. Would it instead be
possible to expres
101 - 200 of 504 matches
Mail list logo