Repository: spark
Updated Branches:
refs/heads/master 8d3e2414d - 418f38d92
[SPARK-3355][Core]: Allow running maven tests in run-tests
Added an AMPLAB_JENKINS_BUILD_TOOL env. variable to allow differentiation
between maven and sbt build / test suites. The only issue I found with this is
Repository: spark
Updated Branches:
refs/heads/branch-1.3 317694ccf - bb1661872
[SPARK-6074] [sql] Package pyspark sql bindings.
This is needed for the SQL bindings to work on Yarn.
Author: Marcelo Vanzin van...@cloudera.com
Closes #4822 from vanzin/SPARK-6074 and squashes the following
Repository: spark
Updated Branches:
refs/heads/master 2df5f1f00 - fd8d283ee
[SPARK-6074] [sql] Package pyspark sql bindings.
This is needed for the SQL bindings to work on Yarn.
Author: Marcelo Vanzin van...@cloudera.com
Closes #4822 from vanzin/SPARK-6074 and squashes the following
of this change?
Author: Sean Owen so...@cloudera.com
Closes #4833 from srowen/SPARK-5628.2 and squashes the following commits:
e723911 [Sean Owen] Backport of https://github.com/apache/spark/pull/4414 to
1.2: Add version option to spark-ec2
Project: http://git-wip-us.apache.org/repos/asf/spark
Repository: spark
Updated Branches:
refs/heads/master 0c9a8eaed - 975643c25
[SPARK-6118] making package name of deploy.worker.CommandUtils and
deploy.CommandUtilsSuite consistent
https://issues.apache.org/jira/browse/SPARK-6118
I found that the object CommandUtils is placed under
Repository: spark
Updated Branches:
refs/heads/master 975643c25 - e359794ce
[SPARK-6138][CORE][minor] enhance the `toArray` method in `SizeTrackingVector`
Use array copy instead of `Iterator#toArray` to make it more efficient.
Author: Wenchen Fan cloud0...@outlook.com
Closes #4825 from
Repository: spark
Updated Branches:
refs/heads/master 934876741 - 757b14b86
[SPARK-5943][Streaming] Update the test to use new API to reduce the warning
Author: Saisai Shao saisai.s...@intel.com
Closes #4722 from jerryshao/SPARK-5943 and squashes the following commits:
1b01233 [Saisai Shao]
Repository: spark
Updated Branches:
refs/heads/branch-1.3 f172387dd - 67b7f7929
[SPARK-5943][Streaming] Update the test to use new API to reduce the warning
Author: Saisai Shao saisai.s...@intel.com
Closes #4722 from jerryshao/SPARK-5943 and squashes the following commits:
1b01233 [Saisai
Repository: spark
Updated Branches:
refs/heads/master 757b14b86 - 242d49584
[SPARK-5724] fix the misconfiguration in AkkaUtils
https://issues.apache.org/jira/browse/SPARK-5724
In AkkaUtil, we set several failure detector related the parameters as following
```
al akkaConf =
Repository: spark
Updated Branches:
refs/heads/master 105791e35 - c5ba975ee
[Spark-5708] Add Slf4jSink to Spark Metrics
Add Slf4jSink to Spark Metrics using Coda Hale's SlfjReporter.
This sends metrics to log4j, allowing spark users to reuse log4j pipeline for
metrics collection.
Reviewed
Repository: spark
Updated Branches:
refs/heads/master d641fbb39 - d51ed263e
[SPARK-5666][streaming][MQTT streaming] some trivial fixes
modified to adhere to accepted coding standards as pointed by tdas in PR #3844
Author: prabs prabsma...@gmail.com
Author: Prabeesh K prabsma...@gmail.com
Repository: spark
Updated Branches:
refs/heads/master d51ed263e - 5b8480e03
[GraphX] fixing 3 typos in the graphx programming guide
Corrected 3 Typos in the GraphX programming guide. I hope this is the correct
way to contribute.
Author: Benedikt Linse benedikt.li...@gmail.com
Closes #4766
Repository: spark
Updated Branches:
refs/heads/master 5b8480e03 - dd077abf2
[SPARK-5771] Number of Cores in Completed Applications of Standalone Master Web
Page always be 0 if sc.stop() is called
In Standalone mode, the number of cores in Completed Applications of the Master
Web Page will
Repository: spark
Updated Branches:
refs/heads/branch-1.3 04d3b328f - eed7389cf
SPARK-5669 [BUILD] Reverse exclusion of JBLAS libs for 1.3
CC mengxr
Author: Sean Owen so...@cloudera.com
Closes #4715 from srowen/SPARK-5669.3 and squashes the following commits:
b27ffa9 [Sean Owen] Reverse
Repository: spark
Updated Branches:
refs/heads/branch-1.3 76e3e6527 - c5a5c6f61
[DOCS] Fix typo in API for custom InputFormats based on the ânewâ MapReduce
API
This looks like a simple typo ```SparkContext.newHadoopRDD``` instead of
```SparkContext.newAPIHadoopRDD``` as in actual
Repository: spark
Updated Branches:
refs/heads/master 46462ff25 - a7f903902
[DOCS] Fix typo in API for custom InputFormats based on the ânewâ MapReduce
API
This looks like a simple typo ```SparkContext.newHadoopRDD``` instead of
```SparkContext.newAPIHadoopRDD``` as in actual
Repository: spark
Updated Branches:
refs/heads/master a7f903902 - 275b1bef8
[DataFrame] [Typo] Fix the typo
Author: Cheng Hao hao.ch...@intel.com
Closes #4717 from chenghao-intel/typo1 and squashes the following commits:
858d7b0 [Cheng Hao] update the typo
Project:
Repository: spark
Updated Branches:
refs/heads/branch-1.3 4186dd3dd - f172387dd
[EXAMPLES] fix typo.
Author: Makoto Fukuhara fuku...@gmail.com
Closes #4724 from fukuo33/fix-typo and squashes the following commits:
8c806b9 [Makoto Fukuhara] fix typo.
(cherry picked from commit
Repository: spark
Updated Branches:
refs/heads/master 95cd643aa - 934876741
[EXAMPLES] fix typo.
Author: Makoto Fukuhara fuku...@gmail.com
Closes #4724 from fukuo33/fix-typo and squashes the following commits:
8c806b9 [Makoto Fukuhara] fix typo.
Project:
Repository: spark
Updated Branches:
refs/heads/master 192e42a29 - df3d559b3
[SPARK-5801] [core] Avoid creating nested directories.
Cache the value of the local root dirs to use for storing local data,
so that the same directories are reused.
Also, to avoid an extra level of nesting, use a
Repository: spark
Updated Branches:
refs/heads/master 5b0a42cb1 - e15532471
[MLlib] fix typo
fix typo: it should be default: instead of default;
Author: Jacky Li jack...@users.noreply.github.com
Closes #4713 from jackylk/patch-10 and squashes the following commits:
15daf2e [Jacky Li]
Repository: spark
Updated Branches:
refs/heads/master e43139f40 - 51a6f9097
[SPARK-5914] to run spark-submit requiring only user perm on windows
Because windows on-default does not grant read permission to jars except to
admin, spark-submit would fail with ClassNotFound exception if user
Repository: spark
Updated Branches:
refs/heads/master b38dec2ff - e60ad2f4c
SPARK-6045 RecordWriter should be checked against null in PairRDDFunctio...
...ns#saveAsNewAPIHadoopDataset
Author: tedyu yuzhih...@gmail.com
Closes #4794 from tedyu/master and squashes the following commits:
Repository: spark
Updated Branches:
refs/heads/master dda4dedca - 8cb23a1f9
[SPARK-5313][Project Infra]: Create simple framework for highlighting changes
introduced in a PR
Built a simple framework with a `dev/tests` directory to house all pull request
related tests. I've moved the two
Author: Sean Owen so...@cloudera.com
Closes #5148 from srowen/SPARK-6480 and squashes the following commits:
974a0a0 [Sean Owen] Additional test of huge ranges, and a few more comments
(and comment fixes)
23ec01e [Sean Owen] Fix fastBucketFunction for histogram() to handle edge
conditions more
: Sean Owen so...@cloudera.com
Closes #5148 from srowen/SPARK-6480 and squashes the following commits:
974a0a0 [Sean Owen] Additional test of huge ranges, and a few more comments
(and comment fixes)
23ec01e [Sean Owen] Fix fastBucketFunction for histogram() to handle edge
conditions more correctly
Author: Sean Owen so...@cloudera.com
Closes #5148 from srowen/SPARK-6480 and squashes the following commits:
974a0a0 [Sean Owen] Additional test of huge ranges, and a few more comments
(and comment fixes)
23ec01e [Sean Owen] Fix fastBucketFunction for histogram() to handle edge
conditions more
Repository: spark
Updated Branches:
refs/heads/master 82701ee25 - b0d884f04
[SPARK-6560][CORE] Do not suppress exceptions from writer.write.
If there is a failure in the Hadoop backend while calling
writer.write, we should remember this original exception,
and try to call writer.close(), but
Repository: spark
Updated Branches:
refs/heads/master b0d884f04 - b52c7f9fc
[MLLIB] Remove println in LogisticRegression.scala
There's no corresponding printing in linear regression. Here was my previous PR
(something weird happened and I can't reopen it)
Repository: spark
Updated Branches:
refs/heads/master c23ba81b8 - 2c43ea38e
[SPARK-6492][CORE] SparkContext.stop() can deadlock when
DAGSchedulerEventProcessLoop dies
I've added a timeout and retry loop around the SparkContext shutdown code that
should fix this deadlock. If a SparkContext
Repository: spark
Updated Branches:
refs/heads/master 512a2f191 - dc6dff248
[Minor][SQL] Fix typo
Just fix a typo.
Author: Liang-Chi Hsieh vii...@gmail.com
Closes #5352 from viirya/fix_a_typo and squashes the following commits:
303b2d2 [Liang-Chi Hsieh] Fix typo.
Project:
Repository: spark
Updated Branches:
refs/heads/branch-1.3 8e5caa227 - 2954468b0
SPARK-6878 [CORE] Fix for sum on empty RDD fails with exception
Author: Erik van Oosten evanoos...@ebay.com
Closes #5489 from erikvanoosten/master and squashes the following commits:
1c91954 [Erik van Oosten]
Repository: spark
Updated Branches:
refs/heads/branch-1.2 daec1c635 - 899ffdcc0
SPARK-6878 [CORE] Fix for sum on empty RDD fails with exception
Author: Erik van Oosten evanoos...@ebay.com
Closes #5489 from erikvanoosten/master and squashes the following commits:
1c91954 [Erik van Oosten]
Repository: spark
Updated Branches:
refs/heads/master 77eeb10fd - 628a72f70
[SPARK-6731] Bump version of apache commons-math3
Version 3.1.1 is two years old and the newer version includes
approximate percentile statistics (among other things).
Author: Punyashloka Biswal
Repository: spark
Updated Branches:
refs/heads/master 628a72f70 - 51b306b93
SPARK-6878 [CORE] Fix for sum on empty RDD fails with exception
Author: Erik van Oosten evanoos...@ebay.com
Closes #5489 from erikvanoosten/master and squashes the following commits:
1c91954 [Erik van Oosten]
Repository: spark
Updated Branches:
refs/heads/master baa83a9a6 - 6d0749cae
[SPARK-7087] [BUILD] Fix path issue change version script
Author: Tijo Thomas tijopara...@gmail.com
Closes #5656 from tijoparacka/FIX_PATHISSUE_CHANGE_VERSION_SCRIPT and squashes
the following commits:
ab4f4b1
Repository: spark
Updated Branches:
refs/heads/master 3e91cc273 - baa83a9a6
[SPARK-6879] [HISTORYSERVER] check if app is completed before clean it up
https://issues.apache.org/jira/browse/SPARK-6879
Use `applications` to replace `FileStatus`, and check if the app is completed
before clean
Repository: spark
Updated Branches:
refs/heads/master 73db132bf - 336f7f537
[SPARK-7037] [CORE] Inconsistent behavior for non-spark config properties in
spark-shell and spark-submit
When specifying non-spark properties (i.e. names don't start with spark.) in
the command line and config
Repository: spark
Updated Branches:
refs/heads/master 8fbd45c74 - 0424da68d
[SPARK-6963][CORE]Flaky test: o.a.s.ContextCleanerSuite automatically cleanup
checkpoint
cc andrewor14
Author: GuoQiang Li wi...@qq.com
Closes #5548 from witgo/SPARK-6963 and squashes the following commits:
/4690)
cc andrewor14, srowen
Author: Nishkam Ravi nr...@cloudera.com
Author: nishkamravi2 nishkamr...@gmail.com
Author: nravi nr...@c1704.halxg.cloudera.com
Closes #5672 from nishkamravi2/master_nravi and squashes the following commits:
0f1abd0 [nishkamravi2] Update Utils.scala
474e3bf
Repository: spark
Updated Branches:
refs/heads/master aa6966ff3 - a11c8683c
[SPARK-7092] Update spark scala version to 2.11.6
Author: Prashant Sharma prashan...@imaginea.com
Closes #5662 from ScrapCodes/SPARK-7092/scala-update-2.11.6 and squashes the
following commits:
58cf4f9 [Prashant
Repository: spark
Updated Branches:
refs/heads/master f5473c2bb - 9a5bbe05f
[MINOR] [MLLIB] Refactor toString method in MLLIB
1. predict(predict.toString) has already output prefix âpredictâ thus
itâs duplicated to print , predict = again
2. there are some extra spaces
Author: Alain
Repository: spark
Updated Branches:
refs/heads/master c035c0f2d - c25ca7c5a
SPARK-3276 Added a new configuration spark.streaming.minRememberDuration
SPARK-3276 Added a new configuration parameter
``spark.streaming.minRememberDuration``, with a default value of 1 minute.
So that when a Spark
Repository: spark
Updated Branches:
refs/heads/master b063a61b9 - e72c16e30
[SPARK-6014] [core] Revamp Spark shutdown hooks, fix shutdown races.
This change adds some new utility code to handle shutdown hooks in
Spark. The main goal is to take advantage of Hadoop 2.x's API for
shutdown hooks,
Repository: spark
Updated Branches:
refs/heads/master 59b7cfc41 - cca9905b9
update the deprecated CountMinSketchMonoid function to TopPctCMS function
http://twitter.github.io/algebird/index.html#com.twitter.algebird.legacy.CountMinSketchMonoid$
The CountMinSketchMonoid has been deprecated
Repository: spark
Updated Branches:
refs/heads/branch-1.3 1526f9c29 - 9909ad3e0
[doc][mllib] Fix typo of the page title in Isotonic regression documents
* Fix the page title in Isotonic regression documents (Naive Bayes - Isotonic
regression)
* Add a newline character at the end of the file
Repository: spark
Updated Branches:
refs/heads/branch-1.3 8549ff4f0 - 948f2f635
[SPARK-6985][streaming] Receiver maxRate over 1000 causes a StackOverflowError
A simple truncation in integer division (on rates over 1000 messages / second)
causes the existing implementation to sleep for 0
Repository: spark
Updated Branches:
refs/heads/master 1f2f723b0 - 5fea3e5c3
[SPARK-6985][streaming] Receiver maxRate over 1000 causes a StackOverflowError
A simple truncation in integer division (on rates over 1000 messages / second)
causes the existing implementation to sleep for 0
Repository: spark
Updated Branches:
refs/heads/master caf0136ec - 438859eb7
[SPARK-6122] [CORE] Upgrade tachyon-client version to 0.6.3
This is a reopening of #4867.
A short summary of the issues resolved from the previous PR:
1. HTTPClient version mismatch: Selenium (used for UI tests)
Repository: spark
Updated Branches:
refs/heads/master 7f3b3b7eb - 75905c57c
[SPARK-7100] [MLLIB] Fix persisted RDD leak in GradientBoostTrees
This fixes a leak of a persisted RDD where GradientBoostTrees can call persist
but never unpersists.
Jira:
srowen/SPARK-7168 and squashes the following commits:
98a8947 [Sean Owen] Make install, deploy plugin versions explicit
4ecf3b2 [Sean Owen] Update Maven build plugin versions and centralize plugin
version management
Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git
Repository: spark
Updated Branches:
refs/heads/master b9de9e040 - 8e1c00dbf
[SPARK-6738] [CORE] Improve estimate the size of a large array
Currently, SizeEstimator.visitArray is not correct in the follow case,
```
array size 200,
elem has the share object
```
when I add a debug log in
Repository: spark
Updated Branches:
refs/heads/master ab5adb7a9 - 62888a4de
[SPARK-7162] [YARN] Launcher error in yarn-client
jira: https://issues.apache.org/jira/browse/SPARK-7162
Author: GuoQiang Li wi...@qq.com
Closes #5716 from witgo/SPARK-7162 and squashes the following commits:
Repository: spark
Updated Branches:
refs/heads/branch-1.3 81de30ae5 - d13080aa2
[SPARK-7103] Fix crash with SparkContext.union when RDD has no partitioner
Added a check to the SparkContext.union method to check that a partitioner is
defined on all RDDs when instantiating a
Repository: spark
Updated Branches:
refs/heads/master 8e1c00dbf - 5d45e1f60
[SPARK-3090] [CORE] Stop SparkContext if user forgets to.
Set up a shutdown hook to try to stop the Spark context in
case the user forgets to do it. The main effect is that any
open logs files are flushed and closed,
Repository: spark
Updated Branches:
refs/heads/master ca9f4ebb8 - b9de9e040
[SPARK-7103] Fix crash with SparkContext.union when RDD has no partitioner
Added a check to the SparkContext.union method to check that a partitioner is
defined on all RDDs when instantiating a
of Guava
Author: Sean Owen so...@cloudera.com
Closes #5703 from srowen/SPARK-7145 and squashes the following commits:
21fbe03 [Sean Owen] Remove use of commons-lang in favor of commons-lang3
classes; remove commons-io use in favor of Guava
Project: http://git-wip-us.apache.org/repos/asf/spark
Repository: spark
Updated Branches:
refs/heads/branch-1.3 d7269493d - 98ac39d2f
[SPARK-6954] [YARN] ExecutorAllocationManager can end up requesting a ne...
...gative n...
...umber of executors
Author: Sandy Ryza sandycloudera.com
Closes #5704 from sryza/sandy-spark-6954 and squashes the
Repository: spark
Updated Branches:
refs/heads/master ecc6eb50a - 856a571ef
[SPARK-3444] Fix typo in Dataframes.py introduced in []
Author: Dean Chen deanch...@gmail.com
Closes #5866 from deanchen/patch-1 and squashes the following commits:
0028bc4 [Dean Chen] Fix typo in Dataframes.py
Repository: spark
Updated Branches:
refs/heads/master 856a571ef - da303526e
[SPARK-7323] [SPARK CORE] Use insertAll instead of insert while merging
combiners in reducer
Author: Mridul Muralidharan mrid...@yahoo-inc.com
Closes #5862 from mridulm/optimize_aggregator and squashes the following
Repository: spark
Updated Branches:
refs/heads/branch-1.3 98ac39d2f - 702356c5d
[SPARK-7323] [SPARK CORE] Use insertAll instead of insert while merging
combiners in reducer
Author: Mridul Muralidharan mrid...@yahoo-inc.com
Closes #5862 from mridulm/optimize_aggregator and squashes the
Repository: spark
Updated Branches:
refs/heads/master bfcd528d6 - 82c8c37c0
[MINOR] [HIVE] Fix QueryPartitionSuite.
At least in the version of Hive I tested on, the test was deleting
a temp directory generated by Hive instead of one containing partition
data. So fix the filter to only
to srowen and mateiz
Author: Ye Xianjin advance...@gmail.com
Closes #4783 from advancedxy/SPARK-6030 and squashes the following commits:
c4dcb41 [Ye Xianjin] Add super.beforeEach in the beforeEach method to make the
trait stackable.. Remove useless leading whitespace.
3f80640 [Ye Xianjin] The size
Repository: spark
Updated Branches:
refs/heads/master 5d6b90d93 - ea841efc5
[SPARK-7255] [STREAMING] [DOCUMENTATION] Added documentation for
spark.streaming.kafka.maxRetries
Added documentation for spark.streaming.kafka.maxRetries
Author: BenFradet benjamin.fra...@gmail.com
Closes #5808
Repository: spark
Updated Branches:
refs/heads/master ea841efc5 - 49549d5a1
[SPARK-7031] [THRIFTSERVER] let thrift server take SPARK_DAEMON_MEMORY and
SPARK_DAEMON_JAVA_OPTS
We should let Thrift Server take these two parameters as it is a daemon. And it
is better to read driver-related
0.23, but also don't support Hadoop 0.23 anymore AFAICT.
There are no builds or releases for it.
In fact, on a related note, refs to CDH3 (Hadoop 0.20.2) should be removed as
this certainly isn't supported either.
Author: Sean Owen so...@cloudera.com
Closes #5863 from srowen/SPARK-7302
Repository: spark
Updated Branches:
refs/heads/master f53a48827 - 7b5dd3e3c
[SPARK-7281] [YARN] Add option to set AM's lib path in client mode.
Author: Marcelo Vanzin van...@cloudera.com
Closes #5813 from vanzin/SPARK-7281 and squashes the following commits:
1cb6f42 [Marcelo Vanzin]
Repository: spark
Updated Branches:
refs/heads/master c6d9a4294 - f53a48827
[SPARK-7213] [YARN] Check for read permissions before copying a Hadoop config
file
Author: Nishkam Ravi nr...@cloudera.com
Author: nishkamravi2 nishkamr...@gmail.com
Author: nravi nr...@c1704.halxg.cloudera.com
Repository: spark
Updated Branches:
refs/heads/master 7630213ca - c8c481da1
Limit help option regex
Added word-boundary delimiters so that embedded text such as -h within
command line options and values doesn't trigger the usage script and exit.
Author: Chris Biow chris.b...@10gen.com
Repository: spark
Updated Branches:
refs/heads/branch-1.3 f64b99421 - d7269493d
Limit help option regex
Added word-boundary delimiters so that embedded text such as -h within
command line options and values doesn't trigger the usage script and exit.
Author: Chris Biow chris.b...@10gen.com
. Standalone UIs are not affected.
Author: Sean Owen so...@cloudera.com
Closes #5837 from srowen/SPARK-6846.2 and squashes the following commits:
c17c386 [Sean Owen] Partial undoing of SPARK-6846; YARN AM proxy won't forward
POSTs, so go back to GET for kill links in Spark UI. Standalone UIs
Repository: spark
Updated Branches:
refs/heads/branch-1.2 6fd74d87c - c0bd415bd
[SPARK-7181] [CORE] fix inifite loop in Externalsorter's mergeWithAggregation
see [SPARK-7181](https://issues.apache.org/jira/browse/SPARK-7181).
Author: Qiping Li liqiping1...@gmail.com
Closes #5737 from
Repository: spark
Updated Branches:
refs/heads/master d7dbce8f7 - 7f4b58373
[SPARK-7181] [CORE] fix inifite loop in Externalsorter's mergeWithAggregation
see [SPARK-7181](https://issues.apache.org/jira/browse/SPARK-7181).
Author: Qiping Li liqiping1...@gmail.com
Closes #5737 from
Repository: spark
Updated Branches:
refs/heads/branch-1.3 bf4ca12f7 - 5b893bd60
[SPARK-7181] [CORE] fix inifite loop in Externalsorter's mergeWithAggregation
see [SPARK-7181](https://issues.apache.org/jira/browse/SPARK-7181).
Author: Qiping Li liqiping1...@gmail.com
Closes #5737 from
Repository: spark
Updated Branches:
refs/heads/master 7f4b58373 - 3fc6cfd07
[SPARK-7155] [CORE] Allow newAPIHadoopFile to support comma-separated list of
files as input
See JIRA: https://issues.apache.org/jira/browse/SPARK-7155
SparkContext's newAPIHadoopFile() does not support
Repository: spark
Updated Branches:
refs/heads/branch-1.3 5b893bd60 - 3a41a1327
[SPARK-7155] [CORE] Allow newAPIHadoopFile to support comma-separated list of
files as input
See JIRA: https://issues.apache.org/jira/browse/SPARK-7155
SparkContext's newAPIHadoopFile() does not support
Repository: spark
Updated Branches:
refs/heads/branch-1.3 3bce87ebd - ec196ab1c
[SPARK-5529] [CORE] Add expireDeadHosts in HeartbeatReceiver
If a blockManager has not send heartBeat more than 120s,
BlockManagerMasterActor will remove it. But coarseGrainedSchedulerBackend can
only remove
Repository: spark
Updated Branches:
refs/heads/master d49735800 - b83091ae4
[MINOR] Minor update for document
Two minor doc errors in `BytesToBytesMap` and `UnsafeRow`.
Author: Liang-Chi Hsieh vii...@gmail.com
Closes #5906 from viirya/minor_doc and squashes the following commits:
27f9089
Repository: spark
Updated Branches:
refs/heads/branch-1.4 ff8b44995 - d28832299
[MINOR] Minor update for document
Two minor doc errors in `BytesToBytesMap` and `UnsafeRow`.
Author: Liang-Chi Hsieh vii...@gmail.com
Closes #5906 from viirya/minor_doc and squashes the following commits:
Repository: spark
Updated Branches:
refs/heads/branch-1.4 93af96a2f - 0327ca2b2
[SPARK-5112] Expose SizeEstimator as a developer api
The best way to size the amount of memory consumption your dataset will
require is to create an RDD, put it into cache, and look at the SparkContext
logs on
Repository: spark
Updated Branches:
refs/heads/master fc8feaa8e - 4222da68d
[SPARK-5112] Expose SizeEstimator as a developer api
The best way to size the amount of memory consumption your dataset will
require is to create an RDD, put it into cache, and look at the SparkContext
logs on your
Author: srowen
Date: Wed May 6 06:45:28 2015
New Revision: 1677934
URL: http://svn.apache.org/r1677934
Log:
Fix 'seemlessly' typo
Modified:
spark/site/sql/index.html
Modified: spark/site/sql/index.html
URL:
http://svn.apache.org/viewvc/spark/site/sql/index.html?rev=1677934r1=1677933r2
Author: srowen
Date: Wed May 6 06:42:44 2015
New Revision: 1677933
URL: http://svn.apache.org/r1677933
Log:
SPARK-6889 point to new contribution guidelines
Modified:
spark/site/faq.html
Modified: spark/site/faq.html
URL:
http://svn.apache.org/viewvc/spark/site/faq.html?rev=1677933r1
Repository: spark
Updated Branches:
refs/heads/branch-1.4 d28832299 - 6f35dac50
[SPARK-5074] [CORE] [TESTS] Fix the flakey test 'run shuffle with map stage
failure' in DAGSchedulerSuite
Test failure:
Repository: spark
Updated Branches:
refs/heads/branch-1.3 b34b5bdf5 - 9278b7a6a
[SPARK-5074] [CORE] [TESTS] Fix the flakey test 'run shuffle with map stage
failure' in DAGSchedulerSuite
Test failure:
Repository: spark
Updated Branches:
refs/heads/master b83091ae4 - 5ffc73e68
[SPARK-5074] [CORE] [TESTS] Fix the flakey test 'run shuffle with map stage
failure' in DAGSchedulerSuite
Test failure:
Repository: spark
Updated Branches:
refs/heads/master 9d250e64d - d4cb38aeb
[MLLIB] [TREE] Verify size of input rdd 0 when building meta data
Require non empty input rdd such that we can take the first labeledpoint and
get the feature size
Author: Alain a...@usc.edu
Author: a...@usc.edu
Repository: spark
Updated Branches:
refs/heads/master 5ffc73e68 - c6d1efba2
[SPARK-7350] [STREAMING] [WEBUI] Attach the Streaming tab when calling
ssc.start()
It's meaningless to display the Streaming tab before `ssc.start()`. So we
should attach it in the `ssc.start` method.
Author:
Repository: spark
Updated Branches:
refs/heads/branch-1.4 6f35dac50 - 49923f7ca
[SPARK-7350] [STREAMING] [WEBUI] Attach the Streaming tab when calling
ssc.start()
It's meaningless to display the Streaming tab before `ssc.start()`. So we
should attach it in the `ssc.start` method.
Author:
Repository: spark
Updated Branches:
refs/heads/branch-1.4 bb5872f2d - f5e9678e3
[SPARK-7392] [CORE] bugfix: Kryo buffer size cannot be larger than 2M
Author: Zhang, Liye liye.zh...@intel.com
Closes #5934 from liyezhang556520/kryoBufSize and squashes the following
commits:
5707e04 [Zhang,
Repository: spark
Updated Branches:
refs/heads/master 2242ab31e - d70a07689
[SPARK-7326] [STREAMING] Performing window() on a WindowedDStream doesn't work
all the time
tdas
https://issues.apache.org/jira/browse/SPARK-7326
The problem most likely resides in DStream.slice() implementation,
Repository: spark
Updated Branches:
refs/heads/master 4222da68d - 51f462003
[SPARK-7357] Improving HBaseTest example
Author: Jihong MA linlin200...@gmail.com
Closes #5904 from JihongMA/SPARK-7357 and squashes the following commits:
7d6153a [Jihong MA] SPARK-7357 Improving HBaseTest example
Repository: spark
Updated Branches:
refs/heads/master e9b16e67c - da738cffa
[MINOR] Renamed variables in SparkKMeans.scala, LocalKMeans.scala and kmeans.py
to simplify readability
With the previous syntax it could look like that the reduceByKey sums
separately abscissas and ordinates of
Repository: spark
Updated Branches:
refs/heads/master da738cffa - c5790a2f7
[MINOR] [BUILD] Declare ivy dependency in root pom.
Without this, any dependency that pulls ivy transitively may override
the version and potentially cause issue. In my machine, the hive tests
were pulling an old
Repository: spark
Updated Branches:
refs/heads/branch-1.4 21ed108fa - 5160437c1
[MINOR] [BUILD] Declare ivy dependency in root pom.
Without this, any dependency that pulls ivy transitively may override
the version and potentially cause issue. In my machine, the hive tests
were pulling an old
Repository: spark
Updated Branches:
refs/heads/branch-1.3 702356c5d - b34b5bdf5
[MINOR] [BUILD] Declare ivy dependency in root pom.
Without this, any dependency that pulls ivy transitively may override
the version and potentially cause issue. In my machine, the hive tests
were pulling an old
Repository: spark
Updated Branches:
refs/heads/master a8ea09683 - 91dc3dfd7
[MINOR] [DOCS] Fix the link to test building info on the wiki
Author: Sean Owen so...@cloudera.com
Closes #6063 from srowen/FixRunningTestsLink and squashes the following commits:
db62018 [Sean Owen] Fix the link
Repository: spark
Updated Branches:
refs/heads/branch-1.4 25972d371 - b0460f414
[SPARK-7438] [SPARK CORE] Fixed validation of relativeSD in countApproxDistinct
Author: Vinod K C vinod...@huawei.com
Closes #5974 from vinodkc/fix_countApproxDistinct_Validation and squashes the
following
Repository: spark
Updated Branches:
refs/heads/master 299262384 - dda6d9f40
[SPARK-7438] [SPARK CORE] Fixed validation of relativeSD in countApproxDistinct
Author: Vinod K C vinod...@huawei.com
Closes #5974 from vinodkc/fix_countApproxDistinct_Validation and squashes the
following commits:
Repository: spark
Updated Branches:
refs/heads/master dda6d9f40 - 12b95abc7
[SPARK-7403] [WEBUI] Link URL in objects on Timeline View is wrong in case of
running on YARN
When we use Spark on YARN and have AllJobPage via ResourceManager's proxy, the
link URL in objects which represent each
201 - 300 of 4820 matches
Mail list logo