Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/1398#issuecomment-48869958
QA tests have started for PR 1398. This patch merges cleanly. View
progress:
https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/16616/consoleFull
---
If
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/1399#issuecomment-48869966
QA results for PR 1399:- This patch FAILED unit tests.- This patch
merges cleanly- This patch adds the following public classes
(experimental):class HiveThriftServer2(hiv
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/1399#issuecomment-48869960
QA tests have started for PR 1399. This patch merges cleanly. View
progress:
https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/16615/consoleFull
---
If
Github user li-zhihui commented on the pull request:
https://github.com/apache/spark/pull/900#issuecomment-48869936
@tgravescs add a commit according to comments.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your pr
GitHub user liancheng opened a pull request:
https://github.com/apache/spark/pull/1399
[SPARK-2410][SQL] Cherry picked Hive Thrift/JDBC server
JIRA issue: [SPARK-2410](https://issues.apache.org/jira/browse/SPARK-2410)
Cherry picked the Hive Thrift/JDBC server from
[branch-1
GitHub user ueshin opened a pull request:
https://github.com/apache/spark/pull/1398
[SPARK-2467] Revert SparkBuild to publish-local to both .m2 and .ivy2.
You can merge this pull request into a Git repository by running:
$ git pull https://github.com/ueshin/apache-spark issues
Github user adrian-wang closed the pull request at:
https://github.com/apache/spark/pull/1397
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is
Github user adrian-wang commented on the pull request:
https://github.com/apache/spark/pull/1397#issuecomment-48869638
OK, I'll close this. Thank you Reynold!
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your projec
Github user rxin commented on the pull request:
https://github.com/apache/spark/pull/1397#issuecomment-48869513
That one already has a main method. Perhaps best to leave this here since
it just starts a connection manager.
---
If your project is set up for it, you can reply to this e
Github user adrian-wang commented on the pull request:
https://github.com/apache/spark/pull/1397#issuecomment-48867685
There's an object in `ConnectionManagerTest.scala` in this package, maybe
move these codes there?
---
If your project is set up for it, you can reply to this email a
Github user rxin commented on the pull request:
https://github.com/apache/spark/pull/1397#issuecomment-48867494
Maybe a better change would be just adding some inline comment explaining
they are used for benchmarks.
---
If your project is set up for it, you can reply to this email an
Github user rxin commented on the pull request:
https://github.com/apache/spark/pull/1397#issuecomment-48867486
I think those are used to do manual testing for performance benchmarks, so
probably best to leave them there.
---
If your project is set up for it, you can reply to this em
Github user andrewor14 commented on the pull request:
https://github.com/apache/spark/pull/1056#issuecomment-48867225
Hi @sryza, I left a couple of comments. In general, I think this patch can
be simplified by using akka only for the driver-executor heartbeats. We should
also clarify
Github user manishamde commented on the pull request:
https://github.com/apache/spark/pull/886#issuecomment-48867199
Thanks Evan. I have compared to scikit-learn on the covertype dataset and
the results looked similar.
---
If your project is set up for it, you can reply to this emai
Github user manishamde commented on a diff in the pull request:
https://github.com/apache/spark/pull/886#discussion_r14865144
--- Diff:
mllib/src/main/scala/org/apache/spark/mllib/tree/DecisionTree.scala ---
@@ -768,104 +973,157 @@ object DecisionTree extends Serializable with
Log
Github user andrewor14 commented on a diff in the pull request:
https://github.com/apache/spark/pull/1056#discussion_r14865082
--- Diff: core/src/main/scala/org/apache/spark/executor/Executor.scala ---
@@ -341,4 +345,47 @@ private[spark] class Executor(
}
}
Github user andrewor14 commented on a diff in the pull request:
https://github.com/apache/spark/pull/1056#discussion_r14865059
--- Diff:
core/src/main/scala/org/apache/spark/ui/jobs/JobProgressListener.scala ---
@@ -197,33 +189,71 @@ class JobProgressListener(conf: SparkConf) exten
Github user andrewor14 commented on a diff in the pull request:
https://github.com/apache/spark/pull/1056#discussion_r14864999
--- Diff:
core/src/main/scala/org/apache/spark/ui/jobs/JobProgressListener.scala ---
@@ -197,33 +189,71 @@ class JobProgressListener(conf: SparkConf) exten
Github user andrewor14 commented on a diff in the pull request:
https://github.com/apache/spark/pull/1056#discussion_r14864962
--- Diff:
core/src/main/scala/org/apache/spark/storage/BlockManagerMasterActor.scala ---
@@ -230,6 +229,10 @@ class BlockManagerMasterActor(val isLocal: Bo
Github user andrewor14 commented on a diff in the pull request:
https://github.com/apache/spark/pull/1056#discussion_r14864951
--- Diff: core/src/main/scala/org/apache/spark/HeartbeatReceiver.scala ---
@@ -0,0 +1,33 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF)
Github user andrewor14 commented on a diff in the pull request:
https://github.com/apache/spark/pull/1056#discussion_r14864946
--- Diff: core/src/main/scala/org/apache/spark/HeartbeatReceiver.scala ---
@@ -0,0 +1,33 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF)
Github user andrewor14 commented on a diff in the pull request:
https://github.com/apache/spark/pull/1056#discussion_r14864917
--- Diff:
core/src/main/scala/org/apache/spark/storage/BlockManagerMasterActor.scala ---
@@ -129,7 +128,7 @@ class BlockManagerMasterActor(val isLocal: Boo
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/1385#discussion_r14864765
--- Diff: core/src/main/scala/org/apache/spark/SparkContext.scala ---
@@ -552,17 +552,10 @@ class SparkContext(config: SparkConf) extends Logging
{
va
Github user rxin commented on a diff in the pull request:
https://github.com/apache/spark/pull/1385#discussion_r14864755
--- Diff:
sql/hive/src/main/scala/org/apache/spark/sql/hive/TableReader.scala ---
@@ -206,17 +202,10 @@ class HadoopTableReader(@transient _tableDesc:
TableDesc
Github user andrewor14 commented on a diff in the pull request:
https://github.com/apache/spark/pull/1056#discussion_r14864751
--- Diff:
core/src/main/scala/org/apache/spark/storage/BlockManagerMasterActor.scala ---
@@ -52,25 +52,24 @@ class BlockManagerMasterActor(val isLocal: Boo
Github user andrewor14 commented on a diff in the pull request:
https://github.com/apache/spark/pull/1056#discussion_r14864612
--- Diff:
core/src/main/scala/org/apache/spark/scheduler/SparkListener.scala ---
@@ -158,6 +161,11 @@ trait SparkListener {
* Called when the appli
Github user andrewor14 commented on a diff in the pull request:
https://github.com/apache/spark/pull/1056#discussion_r14864590
--- Diff:
core/src/main/scala/org/apache/spark/scheduler/EventLoggingListener.scala ---
@@ -81,13 +81,16 @@ private[spark] class EventLoggingListener(
Github user andrewor14 commented on a diff in the pull request:
https://github.com/apache/spark/pull/1056#discussion_r14864551
--- Diff: core/src/main/scala/org/apache/spark/scheduler/DAGScheduler.scala
---
@@ -37,8 +36,15 @@ import org.apache.spark._
import org.apache.spark.e
Github user andrewor14 commented on a diff in the pull request:
https://github.com/apache/spark/pull/1056#discussion_r14864526
--- Diff: core/src/main/scala/org/apache/spark/executor/Executor.scala ---
@@ -341,4 +345,47 @@ private[spark] class Executor(
}
}
Github user andrewor14 commented on a diff in the pull request:
https://github.com/apache/spark/pull/1056#discussion_r14864470
--- Diff: core/src/main/scala/org/apache/spark/executor/Executor.scala ---
@@ -341,4 +345,47 @@ private[spark] class Executor(
}
}
Github user andrewor14 commented on a diff in the pull request:
https://github.com/apache/spark/pull/1056#discussion_r14864464
--- Diff:
core/src/main/scala/org/apache/spark/executor/ExecutorBackend.scala ---
@@ -20,6 +20,7 @@ package org.apache.spark.executor
import java.nio.
Github user andrewor14 commented on a diff in the pull request:
https://github.com/apache/spark/pull/1056#discussion_r14864449
--- Diff: core/src/main/scala/org/apache/spark/executor/Executor.scala ---
@@ -341,4 +345,47 @@ private[spark] class Executor(
}
}
Github user andrewor14 commented on a diff in the pull request:
https://github.com/apache/spark/pull/1056#discussion_r14864427
--- Diff:
core/src/main/scala/org/apache/spark/executor/CoarseGrainedExecutorBackend.scala
---
@@ -32,6 +32,9 @@ import org.apache.spark.deploy.worker.Wor
Github user andrewor14 commented on a diff in the pull request:
https://github.com/apache/spark/pull/1056#discussion_r14864424
--- Diff: core/src/main/scala/org/apache/spark/HeartbeatReceiver.scala ---
@@ -0,0 +1,33 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF)
Github user andrewor14 commented on the pull request:
https://github.com/apache/spark/pull/1056#issuecomment-48865504
Jenkins, test this please
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/1397#issuecomment-48864486
QA results for PR 1397:- This patch PASSES unit tests.- This patch
merges cleanly- This patch adds no public classesFor more
information see test
ouptut:https://amplab.c
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/1387#issuecomment-48863186
QA results for PR 1387:- This patch PASSES unit tests.- This patch
merges cleanly- This patch adds no public classesFor more
information see test
ouptut:https://amplab.c
Github user mengxr commented on the pull request:
https://github.com/apache/spark/pull/1269#issuecomment-48863082
Jenkins, add to whitelist.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have th
Github user mengxr commented on the pull request:
https://github.com/apache/spark/pull/1269#issuecomment-48863087
Jenkins, test this please.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have th
Github user mengxr commented on a diff in the pull request:
https://github.com/apache/spark/pull/1311#discussion_r14863059
--- Diff: docs/mllib-linear-methods.md ---
@@ -242,7 +242,96 @@ Similarly, you can use replace `SVMWithSGD` by
All of MLlib's methods use Java-friendly typ
Github user mengxr commented on a diff in the pull request:
https://github.com/apache/spark/pull/1311#discussion_r14863048
--- Diff: docs/mllib-optimization.md ---
@@ -263,7 +267,110 @@ println("Loss of each step in training process")
loss.foreach(println)
println("Area un
Github user mengxr commented on a diff in the pull request:
https://github.com/apache/spark/pull/1311#discussion_r14863037
--- Diff: docs/mllib-optimization.md ---
@@ -263,7 +267,110 @@ println("Loss of each step in training process")
loss.foreach(println)
println("Area un
Github user mengxr commented on a diff in the pull request:
https://github.com/apache/spark/pull/1311#discussion_r14863023
--- Diff: docs/mllib-optimization.md ---
@@ -263,7 +267,110 @@ println("Loss of each step in training process")
loss.foreach(println)
println("Area un
Github user mengxr commented on the pull request:
https://github.com/apache/spark/pull/1311#issuecomment-48862778
@miccagiann I made one pass through the example code. Besides inline
comments:
1. We moved mllib's data to `data/mllib` in #1394 . Could you please update
the pat
Github user aarondav commented on a diff in the pull request:
https://github.com/apache/spark/pull/1385#discussion_r14862987
--- Diff: core/src/main/scala/org/apache/spark/SparkContext.scala ---
@@ -552,17 +552,10 @@ class SparkContext(config: SparkConf) extends Logging
{
Github user aarondav commented on a diff in the pull request:
https://github.com/apache/spark/pull/1385#discussion_r14862963
--- Diff: core/src/main/scala/org/apache/spark/rdd/HadoopRDD.scala ---
@@ -128,25 +123,13 @@ class HadoopRDD[K, V](
// Returns a JobConf that wil
Github user mengxr commented on a diff in the pull request:
https://github.com/apache/spark/pull/1311#discussion_r14862952
--- Diff: docs/mllib-optimization.md ---
@@ -263,7 +267,110 @@ println("Loss of each step in training process")
loss.foreach(println)
println("Area un
Github user mengxr commented on a diff in the pull request:
https://github.com/apache/spark/pull/1311#discussion_r14862947
--- Diff: docs/mllib-linear-methods.md ---
@@ -338,7 +427,74 @@ and
[`LassoWithSGD`](api/scala/index.html#org.apache.spark.mllib.regression.Lass
All of ML
Github user chenghao-intel commented on a diff in the pull request:
https://github.com/apache/spark/pull/1390#discussion_r14862941
--- Diff:
sql/hive/src/main/scala/org/apache/spark/sql/hive/TableReader.scala ---
@@ -157,21 +161,60 @@ class HadoopTableReader(@transient _tableDesc:
Github user mengxr commented on a diff in the pull request:
https://github.com/apache/spark/pull/1311#discussion_r14862939
--- Diff: docs/mllib-linear-methods.md ---
@@ -338,7 +427,74 @@ and
[`LassoWithSGD`](api/scala/index.html#org.apache.spark.mllib.regression.Lass
All of ML
Github user mengxr commented on a diff in the pull request:
https://github.com/apache/spark/pull/1311#discussion_r14862917
--- Diff: docs/mllib-dimensionality-reduction.md ---
@@ -57,10 +57,57 @@ val U: RowMatrix = svd.U // The U factor is a RowMatrix.
val s: Vector = svd.s //
Github user mengxr commented on a diff in the pull request:
https://github.com/apache/spark/pull/1311#discussion_r14862920
--- Diff: docs/mllib-linear-methods.md ---
@@ -242,7 +242,96 @@ Similarly, you can use replace `SVMWithSGD` by
All of MLlib's methods use Java-friendly typ
Github user mengxr commented on a diff in the pull request:
https://github.com/apache/spark/pull/1311#discussion_r14862918
--- Diff: docs/mllib-dimensionality-reduction.md ---
@@ -91,4 +138,51 @@ val pc: Matrix = mat.computePrincipalComponents(10) //
Principal components are
v
Github user mengxr commented on a diff in the pull request:
https://github.com/apache/spark/pull/1311#discussion_r14862921
--- Diff: docs/mllib-linear-methods.md ---
@@ -242,7 +242,96 @@ Similarly, you can use replace `SVMWithSGD` by
All of MLlib's methods use Java-friendly typ
Github user mengxr commented on a diff in the pull request:
https://github.com/apache/spark/pull/1311#discussion_r14862916
--- Diff: docs/mllib-collaborative-filtering.md ---
@@ -99,7 +99,88 @@ val model = ALS.trainImplicit(ratings, rank,
numIterations, alpha)
All of MLlib's m
Github user mengxr commented on a diff in the pull request:
https://github.com/apache/spark/pull/1311#discussion_r14862914
--- Diff: docs/mllib-collaborative-filtering.md ---
@@ -99,7 +99,88 @@ val model = ALS.trainImplicit(ratings, rank,
numIterations, alpha)
All of MLlib's m
Github user mengxr commented on a diff in the pull request:
https://github.com/apache/spark/pull/1311#discussion_r14862915
--- Diff: docs/mllib-collaborative-filtering.md ---
@@ -99,7 +99,88 @@ val model = ALS.trainImplicit(ratings, rank,
numIterations, alpha)
All of MLlib's m
Github user mengxr commented on a diff in the pull request:
https://github.com/apache/spark/pull/1311#discussion_r14862912
--- Diff: docs/mllib-collaborative-filtering.md ---
@@ -99,7 +99,88 @@ val model = ALS.trainImplicit(ratings, rank,
numIterations, alpha)
All of MLlib's m
Github user mengxr commented on a diff in the pull request:
https://github.com/apache/spark/pull/1311#discussion_r14862910
--- Diff: docs/mllib-collaborative-filtering.md ---
@@ -99,7 +99,88 @@ val model = ALS.trainImplicit(ratings, rank,
numIterations, alpha)
All of MLlib's m
Github user mengxr commented on a diff in the pull request:
https://github.com/apache/spark/pull/1311#discussion_r14862911
--- Diff: docs/mllib-collaborative-filtering.md ---
@@ -99,7 +99,88 @@ val model = ALS.trainImplicit(ratings, rank,
numIterations, alpha)
All of MLlib's m
Github user mengxr commented on a diff in the pull request:
https://github.com/apache/spark/pull/1311#discussion_r14862909
--- Diff: docs/mllib-clustering.md ---
@@ -69,7 +69,54 @@ println("Within Set Sum of Squared Errors = " + WSSSE)
All of MLlib's methods use Java-friendly t
Github user mengxr commented on a diff in the pull request:
https://github.com/apache/spark/pull/1311#discussion_r14862906
--- Diff: docs/mllib-clustering.md ---
@@ -69,7 +69,54 @@ println("Within Set Sum of Squared Errors = " + WSSSE)
All of MLlib's methods use Java-friendly t
Github user mengxr commented on a diff in the pull request:
https://github.com/apache/spark/pull/1311#discussion_r14862907
--- Diff: docs/mllib-clustering.md ---
@@ -69,7 +69,54 @@ println("Within Set Sum of Squared Errors = " + WSSSE)
All of MLlib's methods use Java-friendly t
Github user scwf commented on the pull request:
https://github.com/apache/spark/pull/1385#issuecomment-48861711
@rxin and @aarondav, yeah ï¼the master branch deadlocks, it seems locks of
#1273 and Hadoop-10456 lead to the problem. when run hivesql self join sql---
hql("SELECT t1.a, t
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/1397#issuecomment-48861087
QA tests have started for PR 1397. This patch merges cleanly. View
progress:
https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/16612/consoleFull
---
If
GitHub user adrian-wang opened a pull request:
https://github.com/apache/spark/pull/1397
remove not used test in src/main
Maybe I should put that back in some test suite?
You can merge this pull request into a Git repository by running:
$ git pull https://github.com/adrian-wang
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/1377#issuecomment-48860618
QA results for PR 1377:- This patch PASSES unit tests.- This patch
merges cleanly- This patch adds no public classesFor more
information see test
ouptut:https://amplab.c
Github user asfgit closed the pull request at:
https://github.com/apache/spark/pull/1394
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enab
Github user yhuai commented on the pull request:
https://github.com/apache/spark/pull/1390#issuecomment-48860420
In general, I suggest adding more comments to explain what we are doing at
here because this part of code is pretty Hive-specific.
---
If your project is set up for it, yo
Github user mengxr commented on the pull request:
https://github.com/apache/spark/pull/1394#issuecomment-48860407
@srowen This looks good to me and thank you for updating the docs as well!
---
If your project is set up for it, you can reply to this email and have your
reply appear on
Github user yhuai commented on a diff in the pull request:
https://github.com/apache/spark/pull/1390#discussion_r14862338
--- Diff:
sql/hive/src/main/scala/org/apache/spark/sql/hive/TableReader.scala ---
@@ -157,21 +161,60 @@ class HadoopTableReader(@transient _tableDesc:
TableDes
Github user yhuai commented on a diff in the pull request:
https://github.com/apache/spark/pull/1390#discussion_r14862300
--- Diff:
sql/hive/src/main/scala/org/apache/spark/sql/hive/TableReader.scala ---
@@ -157,21 +161,60 @@ class HadoopTableReader(@transient _tableDesc:
TableDes
Github user yhuai commented on a diff in the pull request:
https://github.com/apache/spark/pull/1390#discussion_r14862289
--- Diff:
sql/hive/src/main/scala/org/apache/spark/sql/hive/TableReader.scala ---
@@ -157,21 +161,60 @@ class HadoopTableReader(@transient _tableDesc:
TableDes
Github user yhuai commented on the pull request:
https://github.com/apache/spark/pull/1390#issuecomment-48860018
@chenghao-intel I am not sure I understand your comment on column pruning.
I think for a Hive table, we should use `ColumnProjectionUtils` to set needed
columns. So, RCFile
Github user lirui-intel commented on the pull request:
https://github.com/apache/spark/pull/1313#issuecomment-48859854
This looks good to me :)
Just a reminder that when TaskSchedulerImpl calls
TaskSetManager.resourceOffer, the maxLocality (changed to preferredLocality in
this PR)
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/1387#issuecomment-48859861
QA tests have started for PR 1387. This patch merges cleanly. View
progress:
https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/16611/consoleFull
---
If
Github user chenghao-intel commented on the pull request:
https://github.com/apache/spark/pull/1390#issuecomment-48859842
And as the Hive SerDe actually provides the feature of `lazy` parsing,
hence during the converting of `raw object` to `Row`, we need to support the
column pruning
Github user chenghao-intel commented on the pull request:
https://github.com/apache/spark/pull/1390#issuecomment-48859675
The code looks good to me. However, I think we can avoid the work around
solution (de-serializing (with partition serde) and then serialize (with table
serde) agai
Github user aarondav commented on the pull request:
https://github.com/apache/spark/pull/1259#issuecomment-48859674
If we actually want people to get information out of all those numbers, can
we consider using a human readable format such as `Task(stageId = 1, taskId =
5, attempt = 0)
Github user jerryshao commented on the pull request:
https://github.com/apache/spark/pull/1210#issuecomment-48859519
Hi Matei, thanks a lot for your review, I will change the code according to
your comments.
---
If your project is set up for it, you can reply to this email and have y
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/1377#issuecomment-48857093
QA tests have started for PR 1377. This patch merges cleanly. View
progress:
https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/16610/consoleFull
---
If
Github user marmbrus commented on the pull request:
https://github.com/apache/spark/pull/1377#issuecomment-48857036
test this please
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this featu
Github user miccagiann commented on the pull request:
https://github.com/apache/spark/pull/1311#issuecomment-48855958
Hello guys,
I have provided Java examples for the following documentation files:
mllib-clustering.md
mllib-collaborative-filtering.md
mllib-dimensio
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/1392#issuecomment-48855862
QA results for PR 1392:- This patch FAILED unit tests.- This patch
merges cleanly- This patch adds no public classesFor more
information see test
ouptut:https://amplab.c
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/1392#issuecomment-48854002
QA tests have started for PR 1392. This patch merges cleanly. View
progress:
https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/16609/consoleFull
---
If
Github user andrewor14 commented on the pull request:
https://github.com/apache/spark/pull/1392#issuecomment-48853888
Jenkins, test this please
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/1396#issuecomment-48853489
QA results for PR 1396:- This patch FAILED unit tests.- This patch
merges cleanly- This patch adds no public classesFor more
information see test
ouptut:https://amplab.c
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/1393#issuecomment-48852700
QA results for PR 1393:- This patch PASSES unit tests.- This patch
merges cleanly- This patch adds the following public classes
(experimental):case class Rating(user: Lon
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/1396#issuecomment-48851429
QA tests have started for PR 1396. This patch merges cleanly. View
progress:
https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/16608/consoleFull
---
If
GitHub user marmbrus opened a pull request:
https://github.com/apache/spark/pull/1396
[SQL] Whitelist more Hive tests.
You can merge this pull request into a Git repository by running:
$ git pull https://github.com/marmbrus/spark moreTests
Alternatively you can review and app
Github user rxin commented on the pull request:
https://github.com/apache/spark/pull/1351#issuecomment-48851059
Note that there are multiple problems. We can solve the problem of out of
memory by simply limiting the length of a record. Ideally, csvRDD(RDD[String])
should just be one e
Github user AmplabJenkins commented on the pull request:
https://github.com/apache/spark/pull/1395#issuecomment-48851025
Can one of the admins verify this patch?
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your pro
GitHub user staple opened a pull request:
https://github.com/apache/spark/pull/1395
[SPARK-546] Add full outer join to RDD and DStream.
You can merge this pull request into a Git repository by running:
$ git pull https://github.com/staple/spark SPARK-546
Alternatively you can
Github user falaki commented on the pull request:
https://github.com/apache/spark/pull/1351#issuecomment-48850882
This is not a bad idea, especially considering that a file can be split
across partitions. @marmbrus you suggested this feature. What do you think
about Reynold's suggesti
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/1393#issuecomment-48850704
QA tests have started for PR 1393. This patch merges cleanly. View
progress:
https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/16607/consoleFull
---
If
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/1360#issuecomment-48850708
QA results for PR 1360:- This patch PASSES unit tests.- This patch
merges cleanly- This patch adds no public classesFor more
information see test
ouptut:https://amplab.c
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/1394#issuecomment-48849070
QA results for PR 1394:- This patch PASSES unit tests.- This patch
merges cleanly- This patch adds no public classesFor more
information see test
ouptut:https://amplab.c
Github user mridulm commented on the pull request:
https://github.com/apache/spark/pull/1313#issuecomment-48849034
Hi @CodingCat looks good to me.
My only doubt, which we discussed last, was whether we want to
differentiate between tasks which have no locations at all vs tasks whic
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/1393#issuecomment-48848503
QA results for PR 1393:- This patch FAILED unit tests.- This patch
merges cleanly- This patch adds the following public classes
(experimental):case class Rating(user: Lon
Github user andrewor14 commented on the pull request:
https://github.com/apache/spark/pull/1259#issuecomment-48848292
Hi @rxin, I took a pass over the patch and the changes mostly look good. On
a higher level point, I notice that we log this pattern `0.0:4.0 (TID 4 ...)`
quite often,
1 - 100 of 163 matches
Mail list logo