Repository: spark
Updated Branches:
refs/heads/branch-1.3 cee6d0877 -> ba8352c76
[SPARK-6247][SQL] Fix resolution of ambiguous joins caused by new aliases
We need to handle ambiguous `exprId`s that are produced by new aliases as well
as those caused by leaf nodes (`MultiInstanceRelation`).
A
Repository: spark
Updated Branches:
refs/heads/master a6ee2f794 -> 357900311
[SPARK-6247][SQL] Fix resolution of ambiguous joins caused by new aliases
We need to handle ambiguous `exprId`s that are produced by new aliases as well
as those caused by leaf nodes (`MultiInstanceRelation`).
Attem
Repository: spark
Updated Branches:
refs/heads/master 78cb08a5d -> a6ee2f794
[SPARK-5651][SQL] Add input64 in blacklist and add test suit for create table
within backticks
Now spark version is only support
```create table table_in_database_creation.test1 as select * from src limit
1;``` in H
Repository: spark
Updated Branches:
refs/heads/master 5c80643d1 -> 78cb08a5d
[SPARK-5404] [SQL] Update the default statistic number
By default, the statistic for logical plan with multiple children is quite
aggressive, and those statistic are quite critical for the join optimization,
hence w
Repository: spark
Updated Branches:
refs/heads/master a012e0863 -> 5c80643d1
[SPARK-5908][SQL] Resolve UdtfsAlias when only single Alias is used
`ResolveUdtfsAlias` in `hiveUdfs` only considers the `HiveGenericUdtf` with
multiple alias. When only single alias is used with `HiveGenericUdtf`, t
Repository: spark
Updated Branches:
refs/heads/branch-1.3 3ea38bc3d -> cee6d0877
[SPARK-6383][SQL]Fixed compiler and errors in Dataframe examples
Author: Tijo Thomas
Closes #5068 from tijoparacka/fix_sql_dataframe_example and squashes the
following commits:
6953ac1 [Tijo Thomas] Handled Ja
Repository: spark
Updated Branches:
refs/heads/master dc9c9196d -> a012e0863
[SPARK-6383][SQL]Fixed compiler and errors in Dataframe examples
Author: Tijo Thomas
Closes #5068 from tijoparacka/fix_sql_dataframe_example and squashes the
following commits:
6953ac1 [Tijo Thomas] Handled Java a
Repository: spark
Updated Branches:
refs/heads/master 4633a87b8 -> dc9c9196d
[SPARK-6366][SQL] In Python API, the default save mode for save and saveAsTable
should be "error" instead of "append".
https://issues.apache.org/jira/browse/SPARK-6366
Author: Yin Huai
Closes #5053 from yhuai/SPAR
Repository: spark
Updated Branches:
refs/heads/branch-1.3 9d88f0cbd -> 3ea38bc3d
[SPARK-6366][SQL] In Python API, the default save mode for save and saveAsTable
should be "error" instead of "append".
https://issues.apache.org/jira/browse/SPARK-6366
Author: Yin Huai
Closes #5053 from yhuai/
Repository: spark
Updated Branches:
refs/heads/branch-1.3 476c4e117 -> 9d88f0cbd
[SPARK-6330] [SQL] Add a test case for SPARK-6330
When getting file statuses, create file system from each path instead of a
single one from hadoop configuration.
Author: Pei-Lun Lee
Closes #5039 from ypcat/sp
Repository: spark
Updated Branches:
refs/heads/master c94d06264 -> 4633a87b8
[SPARK-6330] [SQL] Add a test case for SPARK-6330
When getting file statuses, create file system from each path instead of a
single one from hadoop configuration.
Author: Pei-Lun Lee
Closes #5039 from ypcat/spark-
Repository: spark
Updated Branches:
refs/heads/master d9f3e0168 -> c94d06264
[SPARK-6226][MLLIB] add save/load in PySpark's KMeansModel
Use `_py2java` and `_java2py` to convert Python model to/from Java model.
yinxusen
Author: Xiangrui Meng
Closes #5049 from mengxr/SPARK-6226-mengxr and sq
Repository: spark
Updated Branches:
refs/heads/master 4cca3917d -> d9f3e0168
[SPARK-6336] LBFGS should document what convergenceTol means
LBFGS uses convergence tolerance. This value should be written in document as
an argument.
Author: lewuathe
Closes #5033 from Lewuathe/SPARK-6336 and sq
Repository: spark
Updated Branches:
refs/heads/branch-1.3 ac0e7cc7f -> 476c4e117
[SPARK-6336] LBFGS should document what convergenceTol means
LBFGS uses convergence tolerance. This value should be written in document as
an argument.
Author: lewuathe
Closes #5033 from Lewuathe/SPARK-6336 an
Repository: spark
Updated Branches:
refs/heads/branch-1.3 febb12308 -> ac0e7cc7f
[SPARK-6365] jetty-security also needed for SPARK_PREPEND_CLASSES to work
https://issues.apache.org/jira/browse/SPARK-6365
thanks vanzin for helping me figure this out
Author: Imran Rashid
Closes #5071 from sq
Repository: spark
Updated Branches:
refs/heads/branch-1.3 29e39e178 -> febb12308
[SPARK-6313] Add config option to disable file locks/fetchFile cache to ...
...support NFS mounts.
This is a work around for now with the goal to find a more permanent solution.
https://issues.apache.org/jira/bro
Repository: spark
Updated Branches:
refs/heads/branch-1.2 9ebd6f12e -> a2a94a154
[SPARK-6313] Add config option to disable file locks/fetchFile cache to ...
...support NFS mounts.
This is a work around for now with the goal to find a more permanent solution.
https://issues.apache.org/jira/bro
Repository: spark
Updated Branches:
refs/heads/master 0f673c21f -> 4cca3917d
[SPARK-6313] Add config option to disable file locks/fetchFile cache to ...
...support NFS mounts.
This is a work around for now with the goal to find a more permanent solution.
https://issues.apache.org/jira/browse/
Repository: spark
Updated Branches:
refs/heads/branch-1.3 95f8d1c51 -> 29e39e178
[SPARK-3266] Use intermediate abstract classes to fix type erasure issues in
Java APIs
This PR addresses a Scala compiler bug
([SI-8905](https://issues.scala-lang.org/browse/SI-8905)) that was breaking
some of
Repository: spark
Updated Branches:
refs/heads/master e9f22c612 -> 0f673c21f
[SPARK-3266] Use intermediate abstract classes to fix type erasure issues in
Java APIs
This PR addresses a Scala compiler bug
([SI-8905](https://issues.scala-lang.org/browse/SI-8905)) that was breaking
some of the
Repository: spark
Updated Branches:
refs/heads/master c928796ad -> e9f22c612
[SPARK-6365] jetty-security also needed for SPARK_PREPEND_CLASSES to work
https://issues.apache.org/jira/browse/SPARK-6365
thanks vanzin for helping me figure this out
Author: Imran Rashid
Closes #5052 from squito
Repository: spark
Updated Branches:
refs/heads/branch-1.3 426816b5c -> 95f8d1c51
[SPARK-6331] Load new master URL if present when recovering streaming context
from checkpoint
In streaming driver recovery, when the SparkConf is reconstructed based on the
checkpointed configuration, it recover
Repository: spark
Updated Branches:
refs/heads/master e26db9be4 -> c928796ad
[SPARK-6331] Load new master URL if present when recovering streaming context
from checkpoint
In streaming driver recovery, when the SparkConf is reconstructed based on the
checkpointed configuration, it recovers th
Repository: spark
Updated Branches:
refs/heads/master 005d1c5f2 -> e26db9be4
[docs] [SPARK-4820] Spark build encounters "File name too long" on some
encrypted filesystems
Added a note instructing users how to build Spark in an encrypted file system.
Author: Theodore Vasiloudis
Closes #5041
Repository: spark
Updated Branches:
refs/heads/master 25f35806e -> 005d1c5f2
[SPARK-6269] [CORE] Use ScalaRunTime's array methods instead of
java.lang.reflect.Array in size estimation
This patch switches the usage of java.lang.reflect.Array in Size estimation to
using scala's RunTime array-g
[SPARK-4011] tighten the visibility of the members in Master/Worker class
https://issues.apache.org/jira/browse/SPARK-4011
Currently, most of the members in Master/Worker are with public accessibility.
We might wish to tighten the accessibility of them
a bit more discussion is here:
https://gi
Repository: spark
Updated Branches:
refs/heads/master b2d8c0222 -> 25f35806e
http://git-wip-us.apache.org/repos/asf/spark/blob/25f35806/core/src/main/scala/org/apache/spark/deploy/worker/Worker.scala
--
diff --git a/core/src/ma
27 matches
Mail list logo