Github user andrewor14 commented on the pull request:
https://github.com/apache/spark/pull/1482#issuecomment-54671967
Just so I understand, by the time we enter this code block we would have
already logged an OOM message, so it's confusing if we log more messages.
However, I think thi
Github user tgravescs commented on the pull request:
https://github.com/apache/spark/pull/2268#issuecomment-54671834
@pwendell do you just want to run the set version and commit it or do you
want to do it through this jira?
---
If your project is set up for it, you can reply to thi
Github user andrewor14 commented on the pull request:
https://github.com/apache/spark/pull/1106#issuecomment-54670726
@WangTaoTheTonic I looked at this more and I think it will actually be
slower with the new changes. Before this patch we shuffle all the workers only
once, but here we
Github user davies commented on a diff in the pull request:
https://github.com/apache/spark/pull/2287#discussion_r17192524
--- Diff: python/pyspark/cloudpickle.py ---
@@ -691,13 +699,13 @@ def save_file(self, obj):
tmpfile.close()
if tst != '':
Github user sarutak commented on the pull request:
https://github.com/apache/spark/pull/2250#issuecomment-54669961
Thanks, I modified the description.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does n
Github user mrocklin commented on the pull request:
https://github.com/apache/spark/pull/2288#issuecomment-54670005
done
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled a
Github user rxin commented on the pull request:
https://github.com/apache/spark/pull/2282#issuecomment-54669948
Jenkins, test this please.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this
Github user davies commented on the pull request:
https://github.com/apache/spark/pull/2288#issuecomment-54669643
The title is a bit confusing, would you mind to change it to "Use new-style
classes in PySpark" ?
The patch looks good to me, thanks!
---
If your project is set
Github user davies commented on the pull request:
https://github.com/apache/spark/pull/2144#issuecomment-54669121
Jenkins, test this please.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have th
GitHub user davies opened a pull request:
https://github.com/apache/spark/pull/2292
[SPARK-3430] [PySpark] [Doc] generate PySpark API docs using Sphinx
Using Sphinx to generate API docs for PySpark.
requirement: Sphinx
```
$ cd docs/api/python/
$ make html
GitHub user liancheng opened a pull request:
https://github.com/apache/spark/pull/2291
[SPARK-3421][SQL] Allows arbitrary character in StructField.name
`StructField.toString` now quotes the `name` field and escapes backslashes
and double quotes within the string. The `DataType` pars
Github user liancheng commented on a diff in the pull request:
https://github.com/apache/spark/pull/2291#discussion_r17191419
--- Diff:
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/types/dataTypes.scala
---
@@ -65,7 +70,7 @@ object DataType extends RegexParsers {
Github user JoshRosen commented on the pull request:
https://github.com/apache/spark/pull/2250#issuecomment-54667475
Thanks for submitting this! I noticed that #1067 is an old PR addressing a
similar issue. If your PR subsumes that one, which I think that it does, could
you add `clo
Github user JoshRosen commented on the pull request:
https://github.com/apache/spark/pull/1106#issuecomment-54665758
I agree that this seems like a bit of a rare corner-case.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well
Github user JoshRosen commented on the pull request:
https://github.com/apache/spark/pull/1106#issuecomment-54665441
Can you give this PR a more descriptive title? "Optimize the schedule
procedure in Master" sounds like it could describe many different changes, so
it's kind of hard t
Github user shaneknapp commented on the pull request:
https://github.com/apache/spark/pull/2289#issuecomment-54665233
Jenkins, test this please.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not hav
Github user asfgit closed the pull request at:
https://github.com/apache/spark/pull/2270
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enab
Github user JoshRosen commented on the pull request:
https://github.com/apache/spark/pull/2290#issuecomment-54664634
Woah, I didn't know optparse had `%default`. Cool!
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If y
GitHub user nchammas opened a pull request:
https://github.com/apache/spark/pull/2290
[EC2] don't duplicate default values
This PR makes two minor changes to the `spark-ec2` script:
1. The script's input parameter default values are duplicated into the help
text. This is un
Github user JoshRosen commented on the pull request:
https://github.com/apache/spark/pull/1482#issuecomment-54664325
This seems reasonable to me. /cc @andrewor14 for another pair of eyes.
To recap [some discussion on the
JIRA](https://issues.apache.org/jira/browse/SPARK-2491)
GitHub user shaneknapp opened a pull request:
https://github.com/apache/spark/pull/2289
TEST ONLY DO NOT MERGE
TEST ONLY DO NOT MERGE
You can merge this pull request into a Git repository by running:
$ git pull https://github.com/shaneknapp/spark sknapptest
Alternatively you c
Github user xinyunh commented on the pull request:
https://github.com/apache/spark/pull/2099#issuecomment-54663384
Sorry, I forgot
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
Github user JoshRosen commented on a diff in the pull request:
https://github.com/apache/spark/pull/2287#discussion_r17188918
--- Diff: python/pyspark/tests.py ---
@@ -180,6 +180,22 @@ def tearDown(self):
self.sc.stop()
sys.path = self._old_sys_path
Github user bbejeck commented on the pull request:
https://github.com/apache/spark/pull/2227#issuecomment-54662763
> Feels to me like it would be better to fix this in
Utils.memoryStringToMb. That way all code using it benefits.
I thought the same thing, but I was not sure abo
Github user JoshRosen commented on a diff in the pull request:
https://github.com/apache/spark/pull/2287#discussion_r17188799
--- Diff: python/pyspark/tests.py ---
@@ -180,6 +180,22 @@ def tearDown(self):
self.sc.stop()
sys.path = self._old_sys_path
Github user JoshRosen commented on a diff in the pull request:
https://github.com/apache/spark/pull/2287#discussion_r17188600
--- Diff: python/pyspark/tests.py ---
@@ -180,6 +180,22 @@ def tearDown(self):
self.sc.stop()
sys.path = self._old_sys_path
Github user bbejeck commented on the pull request:
https://github.com/apache/spark/pull/2227#issuecomment-54662304
Josh,
Thanks for the heads up on testing with environment variables. I will look
at the PR and make the required changes to the test.
---
If your project is s
Github user JoshRosen commented on a diff in the pull request:
https://github.com/apache/spark/pull/2287#discussion_r17188526
--- Diff: python/pyspark/tests.py ---
@@ -180,6 +180,22 @@ def tearDown(self):
self.sc.stop()
sys.path = self._old_sys_path
Github user wardviaene commented on the pull request:
https://github.com/apache/spark/pull/2287#issuecomment-54661969
Hi @JoshRosen
I added a test script in this pull request. The sys.stderr in a class
triggers the bug.
---
If your project is set up for it, you can reply to
Github user andrewor14 commented on the pull request:
https://github.com/apache/spark/pull/2229#issuecomment-54661731
retest this please
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this f
Github user andrewor14 commented on the pull request:
https://github.com/apache/spark/pull/1616#issuecomment-54661638
Do we need to clean up the new cache files we created? Or is that handled
automatically somewhere
---
If your project is set up for it, you can reply to this email an
Github user JoshRosen commented on a diff in the pull request:
https://github.com/apache/spark/pull/1616#discussion_r17188168
--- Diff: core/src/main/scala/org/apache/spark/util/Utils.scala ---
@@ -313,14 +313,74 @@ private[spark] object Utils extends Logging {
}
Github user andrewor14 commented on a diff in the pull request:
https://github.com/apache/spark/pull/1616#discussion_r17188080
--- Diff: core/src/main/scala/org/apache/spark/util/Utils.scala ---
@@ -313,14 +313,74 @@ private[spark] object Utils extends Logging {
}
Github user andrewor14 commented on a diff in the pull request:
https://github.com/apache/spark/pull/1616#discussion_r17188055
--- Diff: core/src/main/scala/org/apache/spark/util/Utils.scala ---
@@ -313,14 +313,74 @@ private[spark] object Utils extends Logging {
}
Github user JoshRosen commented on the pull request:
https://github.com/apache/spark/pull/2270#issuecomment-54660424
Looks good to me, too. Thanks for fixing this!
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your
Github user davies commented on the pull request:
https://github.com/apache/spark/pull/2259#issuecomment-54657265
Jenkins, retest this please.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have
Github user davies commented on the pull request:
https://github.com/apache/spark/pull/2270#issuecomment-54657139
This patch looks good to me.
@JoshRosen could you help to re-visit this?
---
If your project is set up for it, you can reply to this email and have your
reply app
Github user JoshRosen commented on a diff in the pull request:
https://github.com/apache/spark/pull/2227#discussion_r17185052
--- Diff:
core/src/test/scala/org/apache/spark/deploy/worker/WorkerArgumentsTest.scala ---
@@ -0,0 +1,60 @@
+/*
+ * Licensed to the Apache Software
Github user davies commented on a diff in the pull request:
https://github.com/apache/spark/pull/2270#discussion_r17184963
--- Diff: bin/pyspark ---
@@ -85,6 +85,8 @@ export PYSPARK_SUBMIT_ARGS
# For pyspark tests
if [[ -n "$SPARK_TESTING" ]]; then
+ unset YARN_
Github user JoshRosen commented on a diff in the pull request:
https://github.com/apache/spark/pull/2227#discussion_r17184945
--- Diff:
core/src/test/scala/org/apache/spark/deploy/worker/WorkerArgumentsTest.scala ---
@@ -0,0 +1,60 @@
+/*
+ * Licensed to the Apache Software
Github user vanzin commented on the pull request:
https://github.com/apache/spark/pull/2275#issuecomment-54655288
Oops. Thanks for fixing it.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have t
Github user vanzin commented on the pull request:
https://github.com/apache/spark/pull/2227#issuecomment-54655129
Feels to me like it would be better to fix this in
`Utils.memoryStringToMb`. That way all code using it benefits.
As for the behavior of that method, maybe it shou
Github user JoshRosen commented on the pull request:
https://github.com/apache/spark/pull/2227#issuecomment-54654828
Jenkins, this is ok to test.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not ha
Github user davies commented on the pull request:
https://github.com/apache/spark/pull/2144#issuecomment-54654638
Jenkins, test this please.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have th
Github user davies commented on the pull request:
https://github.com/apache/spark/pull/2144#issuecomment-54654606
@mateiz @JoshRosen @mattf run-tests will try to run tests for spark core
and sql with PyPy.
One known issue is that serialization of array in PyPy is similar to
P
Github user JoshRosen commented on the pull request:
https://github.com/apache/spark/pull/2277#issuecomment-54654607
Jenkins, retest this please. (Not sure if Jenkins is programmed to listen
to @nchammas or not...)
---
If your project is set up for it, you can reply to this email an
Github user sarutak commented on a diff in the pull request:
https://github.com/apache/spark/pull/2270#discussion_r17184086
--- Diff: bin/pyspark ---
@@ -85,6 +85,8 @@ export PYSPARK_SUBMIT_ARGS
# For pyspark tests
if [[ -n "$SPARK_TESTING" ]]; then
+ unset YARN
Github user vanzin commented on the pull request:
https://github.com/apache/spark/pull/2276#issuecomment-54653232
No, alpha means pre-branch-2 hadoop (I think, Hadoop branching is not
exactly an exact science). Anyway, there are stable releases without YARN-1203.
So that probably shou
Github user benoyantony commented on the pull request:
https://github.com/apache/spark/pull/2276#issuecomment-54652578
Sure. I'll do both.
Does Alpha corresponds to Hadoop versions before YARN-1203 ? As you know,
before YARN-1203, we cannot pass AM URLS with scheme.
---
If you
Github user vanzin commented on the pull request:
https://github.com/apache/spark/pull/899#issuecomment-54652280
@zeodtr does this compile with anything < hadoop 2.4? If it doesn't, this
is a no-go.
---
If your project is set up for it, you can reply to this email and have your
reply
Github user davies commented on a diff in the pull request:
https://github.com/apache/spark/pull/2270#discussion_r17182845
--- Diff: bin/pyspark ---
@@ -85,6 +85,8 @@ export PYSPARK_SUBMIT_ARGS
# For pyspark tests
if [[ -n "$SPARK_TESTING" ]]; then
+ unset YARN_
Github user vanzin commented on the pull request:
https://github.com/apache/spark/pull/2257#issuecomment-54648499
Jenkins, test this please.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have th
Github user tgravescs commented on a diff in the pull request:
https://github.com/apache/spark/pull/2276#discussion_r17180863
--- Diff:
yarn/alpha/src/main/scala/org/apache/spark/deploy/yarn/YarnRMClientImpl.scala
---
@@ -96,7 +96,7 @@ private class YarnRMClientImpl(args:
Applica
Github user JoshRosen commented on the pull request:
https://github.com/apache/spark/pull/2280#issuecomment-54641296
It looks like `sql.py` overrides the default `persist()`, so you might want
to update it there, too. LGTM otherwise.
---
If your project is set up for it, you can rep
Github user mrocklin commented on the pull request:
https://github.com/apache/spark/pull/2288#issuecomment-54639853
Done
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled a
Github user mrocklin commented on the pull request:
https://github.com/apache/spark/pull/2288#issuecomment-54639788
```
mrocklin@notebook:~/workspace/spark$ git grep "^class \w*:"
mrocklin@notebook:~/workspace/spark$
```
---
If your project is set up for it, you
Github user nchammas commented on the pull request:
https://github.com/apache/spark/pull/2277#issuecomment-54638477
Jenkinshenck, could you test this please?
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project
Github user mrocklin commented on the pull request:
https://github.com/apache/spark/pull/2288#issuecomment-54638388
Sure. Next time I find a few free minutes.
On Fri, Sep 5, 2014 at 8:04 AM, Josh Rosen wrote:
> Also, do you mind opening a JIRA ticket on
> ht
Github user bbejeck commented on the pull request:
https://github.com/apache/spark/pull/2227#issuecomment-54637031
Did any of the admin had chance to check it out? Let me know if you want me
to modify anything in it?
---
If your project is set up for it, you can reply to this email a
Github user JoshRosen commented on the pull request:
https://github.com/apache/spark/pull/2288#issuecomment-54636685
Also, do you mind opening a JIRA ticket on
https://issues.apache.org/jira/browse/SPARK and editing the title of your pull
request to reference it, e.g. `[SPARK-] Us
Github user JoshRosen commented on the pull request:
https://github.com/apache/spark/pull/2288#issuecomment-54636484
Good catch! While you're at it, are there any other old-style classes in
PySpark that should be made into new-style ones?
---
If your project is set up for it, you ca
Github user JoshRosen commented on the pull request:
https://github.com/apache/spark/pull/2287#issuecomment-54636076
Hi @wardviaene,
Do you have an example program that reproduces this bug? We should
probably add it as a regression test (see `python/pyspark/tests.py` for
exa
Github user asfgit closed the pull request at:
https://github.com/apache/spark/pull/2275
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enab
Github user asfgit closed the pull request at:
https://github.com/apache/spark/pull/2185
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enab
GitHub user mrocklin opened a pull request:
https://github.com/apache/spark/pull/2288
pyspark.sql.SQLContext is new-style class
Tiny PR making SQLContext a new-style class. This allows various type
logic to work more effectively
```Python
In [1]: import pyspark
Github user tgravescs commented on the pull request:
https://github.com/apache/spark/pull/2253#issuecomment-54634933
Jenkins, test this please
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have
Github user tgravescs commented on the pull request:
https://github.com/apache/spark/pull/2283#issuecomment-54634308
I don't think this is really necessary as I see the value of the Filesystem
one as a public api now and changing its value would break compatibility, but
I'm ok with it
Github user nchammas commented on the pull request:
https://github.com/apache/spark/pull/2117#issuecomment-54634193
@ash211 Thank you for explaining that.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project do
Github user sarutak commented on the pull request:
https://github.com/apache/spark/pull/2283#issuecomment-54631454
Jenkins, retest this please .
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not hav
Github user tgravescs commented on the pull request:
https://github.com/apache/spark/pull/2286#issuecomment-54630202
Thanks for working on this, I've been meaning to fix this for a while.
Could you also please file a jira and link them. The header of the pr
should include ji
Github user tgravescs commented on a diff in the pull request:
https://github.com/apache/spark/pull/2286#discussion_r17174246
--- Diff: core/src/main/scala/org/apache/spark/SecurityManager.scala ---
@@ -162,7 +162,7 @@ private[spark] class SecurityManager(sparkConf:
SparkConf) exte
GitHub user wardviaene opened a pull request:
https://github.com/apache/spark/pull/2287
[SPARK-3415] [PySpark] removes SerializingAdapter code
This code removes the SerializingAdapter code that was copied from PiCloud
You can merge this pull request into a Git repository by running:
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/2286#issuecomment-54610134
[QA tests have
finished](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/19804/consoleFull)
for PR 2286 at commit
[`cf973a1`](https://github.com/a
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/2285#issuecomment-54609772
[QA tests have
finished](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/19803/consoleFull)
for PR 2285 at commit
[`24f3381`](https://github.com/a
Github user mubarak commented on the pull request:
https://github.com/apache/spark/pull/1723#issuecomment-54606650
Jenkins, this is ok to test.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have
Github user mubarak commented on the pull request:
https://github.com/apache/spark/pull/1723#issuecomment-54606563
@tdas
Can you please review? Thanks
![screen shot 2014-09-05 at 1 42 28
am](https://cloud.githubusercontent.com/assets/668134/4163160/b9b9b538-34e3-11e4-9fae-0e70
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/2286#issuecomment-54604796
[QA tests have
started](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/19804/consoleFull)
for PR 2286 at commit
[`cf973a1`](https://github.com/ap
Github user SparkQA commented on the pull request:
https://github.com/apache/spark/pull/2285#issuecomment-54604364
[QA tests have
started](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/19803/consoleFull)
for PR 2285 at commit
[`24f3381`](https://github.com/ap
GitHub user ash211 opened a pull request:
https://github.com/apache/spark/pull/2286
Don't include the empty string "" as a defaultAclUser
Changes logging from
```
14/09/05 02:01:08 INFO SecurityManager: Changing view acls to: aash,
14/09/05 02:01:08 INFO SecurityMana
Github user sarutak commented on the pull request:
https://github.com/apache/spark/pull/2283#issuecomment-54602306
test this please.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this featu
Github user cloud-fan commented on the pull request:
https://github.com/apache/spark/pull/2230#issuecomment-54601682
I'm not sure how to modify `lazy val resolved` in `GetField` since it
handles not only StructType now. Currently I just removed the type check. What
do you think? @marm
Github user li-zhihui commented on the pull request:
https://github.com/apache/spark/pull/1616#issuecomment-54600822
@JoshRosen @andrewor14
I use url.hashCode + timestamp as cachedFileName,
I believe it is impossible that existing url.hashCode collision
and timestamp collision at
Github user cloud-fan commented on the pull request:
https://github.com/apache/spark/pull/2230#issuecomment-54601247
@marmbrus Seems hive parser will pass something like "a.b.c..." to
`LogicalPlan`, so I have to roll back(and I changed `dotExpressionHeader` to
`ident "." ident {"." id
Github user ScrapCodes commented on a diff in the pull request:
https://github.com/apache/spark/pull/2285#discussion_r17162225
--- Diff: dev/mima ---
@@ -25,11 +25,15 @@ FWDIR="$(cd `dirname $0`/..; pwd)"
cd "$FWDIR"
echo -e "q\n" | sbt/sbt oldDeps/update
+rm -f
Github user chenghao-intel commented on the pull request:
https://github.com/apache/spark/pull/2284#issuecomment-54597712
test this please.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have thi
Github user ash211 commented on the pull request:
https://github.com/apache/spark/pull/1588#issuecomment-54598213
Yep good to close -- we can refer to the ticket in the future if it comes
back up
---
If your project is set up for it, you can reply to this email and have your
reply ap
Github user ScrapCodes commented on a diff in the pull request:
https://github.com/apache/spark/pull/2285#discussion_r17161635
--- Diff: project/SparkBuild.scala ---
@@ -187,7 +187,7 @@ object OldDeps {
Some("org.apache.spark" % fullId % "1.0.0")
}
- def o
Github user ScrapCodes commented on a diff in the pull request:
https://github.com/apache/spark/pull/2285#discussion_r17161831
--- Diff: dev/mima ---
@@ -25,12 +25,16 @@ FWDIR="$(cd `dirname $0`/..; pwd)"
cd "$FWDIR"
echo -e "q\n" | sbt/sbt oldDeps/update
+rm -f
Github user ScrapCodes commented on a diff in the pull request:
https://github.com/apache/spark/pull/2285#discussion_r17161813
--- Diff: dev/mima ---
@@ -25,12 +25,16 @@ FWDIR="$(cd `dirname $0`/..; pwd)"
cd "$FWDIR"
echo -e "q\n" | sbt/sbt oldDeps/update
+rm -f
Github user ash211 commented on the pull request:
https://github.com/apache/spark/pull/2281#issuecomment-54597321
I don't see that contract in the API documented in the Scaladoc for the
method:
```
588 /**
589* Return a new RDD by applying a function to each pa
Github user ScrapCodes commented on the pull request:
https://github.com/apache/spark/pull/2194#issuecomment-54598665
@rxin There is a reason and (workaround type of)fix for this on #2285.
---
If your project is set up for it, you can reply to this email and have your
reply appear on
GitHub user ScrapCodes opened a pull request:
https://github.com/apache/spark/pull/2285
Fix for false positives reported by mima on PR 2194.
You can merge this pull request into a Git repository by running:
$ git pull https://github.com/ScrapCodes/spark-1 mima-fix
Alternative
Github user rxin commented on the pull request:
https://github.com/apache/spark/pull/2281#issuecomment-54596928
The correct assumption is to not reuse objects. However, in Spark SQL we
exploited the implementation of the old shuffle behavior (which serializes each
row object immediate
Github user ash211 commented on the pull request:
https://github.com/apache/spark/pull/2281#issuecomment-54596774
What's the implication here for other client code of the Spark API? It
looks like there are mutability concerns in whether you can save a reference to
the object you get
Github user ash211 commented on the pull request:
https://github.com/apache/spark/pull/2117#issuecomment-54596256
Regarding the merge, I'm guessing this is too late to land in the Spark 1.1
release. Is it a candidate for a backport to a 1.1.x?
---
If your project is set up for it, y
Github user ash211 commented on the pull request:
https://github.com/apache/spark/pull/2117#issuecomment-54596140
@nchammas I'm guessing your OOM issue is unrelated to this one.
```
a = sc.parallelize(["Nick", "John", "Bob"])
a = a.repartition(24000)
a.keyBy(lambda
Github user sarutak commented on the pull request:
https://github.com/apache/spark/pull/2283#issuecomment-54595905
Jenkins, test this please.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have t
GitHub user chenghao-intel opened a pull request:
https://github.com/apache/spark/pull/2284
[SPARK-3412] [SQL] Add 3 missing types for Row API
`BinaryType`, `DecimalType` and `TimestampType` are missing in the Row API.
You can merge this pull request into a Git repository by running
Github user rxin commented on the pull request:
https://github.com/apache/spark/pull/2282#issuecomment-54594495
Jenkins, test this please.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this
Github user mengxr commented on the pull request:
https://github.com/apache/spark/pull/2125#issuecomment-54593766
Jenkins, retest this please.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have
401 - 500 of 500 matches
Mail list logo