spark git commit: [SPARK-3355][Core]: Allow running maven tests in run-tests

2015-03-04 Thread srowen
Repository: spark Updated Branches: refs/heads/master 8d3e2414d - 418f38d92 [SPARK-3355][Core]: Allow running maven tests in run-tests Added an AMPLAB_JENKINS_BUILD_TOOL env. variable to allow differentiation between maven and sbt build / test suites. The only issue I found with this is

spark git commit: [SPARK-6074] [sql] Package pyspark sql bindings.

2015-03-01 Thread srowen
Repository: spark Updated Branches: refs/heads/branch-1.3 317694ccf - bb1661872 [SPARK-6074] [sql] Package pyspark sql bindings. This is needed for the SQL bindings to work on Yarn. Author: Marcelo Vanzin van...@cloudera.com Closes #4822 from vanzin/SPARK-6074 and squashes the following

spark git commit: [SPARK-6074] [sql] Package pyspark sql bindings.

2015-03-01 Thread srowen
Repository: spark Updated Branches: refs/heads/master 2df5f1f00 - fd8d283ee [SPARK-6074] [sql] Package pyspark sql bindings. This is needed for the SQL bindings to work on Yarn. Author: Marcelo Vanzin van...@cloudera.com Closes #4822 from vanzin/SPARK-6074 and squashes the following

spark git commit: SPARK-5628 [EC2] Backport: Add version option to spark-ec2

2015-03-01 Thread srowen
of this change? Author: Sean Owen so...@cloudera.com Closes #4833 from srowen/SPARK-5628.2 and squashes the following commits: e723911 [Sean Owen] Backport of https://github.com/apache/spark/pull/4414 to 1.2: Add version option to spark-ec2 Project: http://git-wip-us.apache.org/repos/asf/spark

spark git commit: [SPARK-6118] making package name of deploy.worker.CommandUtils and deploy.CommandUtilsSuite consistent

2015-03-03 Thread srowen
Repository: spark Updated Branches: refs/heads/master 0c9a8eaed - 975643c25 [SPARK-6118] making package name of deploy.worker.CommandUtils and deploy.CommandUtilsSuite consistent https://issues.apache.org/jira/browse/SPARK-6118 I found that the object CommandUtils is placed under

spark git commit: [SPARK-6138][CORE][minor] enhance the `toArray` method in `SizeTrackingVector`

2015-03-03 Thread srowen
Repository: spark Updated Branches: refs/heads/master 975643c25 - e359794ce [SPARK-6138][CORE][minor] enhance the `toArray` method in `SizeTrackingVector` Use array copy instead of `Iterator#toArray` to make it more efficient. Author: Wenchen Fan cloud0...@outlook.com Closes #4825 from

spark git commit: [SPARK-5943][Streaming] Update the test to use new API to reduce the warning

2015-02-23 Thread srowen
Repository: spark Updated Branches: refs/heads/master 934876741 - 757b14b86 [SPARK-5943][Streaming] Update the test to use new API to reduce the warning Author: Saisai Shao saisai.s...@intel.com Closes #4722 from jerryshao/SPARK-5943 and squashes the following commits: 1b01233 [Saisai Shao]

spark git commit: [SPARK-5943][Streaming] Update the test to use new API to reduce the warning

2015-02-23 Thread srowen
Repository: spark Updated Branches: refs/heads/branch-1.3 f172387dd - 67b7f7929 [SPARK-5943][Streaming] Update the test to use new API to reduce the warning Author: Saisai Shao saisai.s...@intel.com Closes #4722 from jerryshao/SPARK-5943 and squashes the following commits: 1b01233 [Saisai

spark git commit: [SPARK-5724] fix the misconfiguration in AkkaUtils

2015-02-23 Thread srowen
Repository: spark Updated Branches: refs/heads/master 757b14b86 - 242d49584 [SPARK-5724] fix the misconfiguration in AkkaUtils https://issues.apache.org/jira/browse/SPARK-5724 In AkkaUtil, we set several failure detector related the parameters as following ``` al akkaConf =

spark git commit: [Spark-5708] Add Slf4jSink to Spark Metrics

2015-02-24 Thread srowen
Repository: spark Updated Branches: refs/heads/master 105791e35 - c5ba975ee [Spark-5708] Add Slf4jSink to Spark Metrics Add Slf4jSink to Spark Metrics using Coda Hale's SlfjReporter. This sends metrics to log4j, allowing spark users to reuse log4j pipeline for metrics collection. Reviewed

spark git commit: [SPARK-5666][streaming][MQTT streaming] some trivial fixes

2015-02-25 Thread srowen
Repository: spark Updated Branches: refs/heads/master d641fbb39 - d51ed263e [SPARK-5666][streaming][MQTT streaming] some trivial fixes modified to adhere to accepted coding standards as pointed by tdas in PR #3844 Author: prabs prabsma...@gmail.com Author: Prabeesh K prabsma...@gmail.com

spark git commit: [GraphX] fixing 3 typos in the graphx programming guide

2015-02-25 Thread srowen
Repository: spark Updated Branches: refs/heads/master d51ed263e - 5b8480e03 [GraphX] fixing 3 typos in the graphx programming guide Corrected 3 Typos in the GraphX programming guide. I hope this is the correct way to contribute. Author: Benedikt Linse benedikt.li...@gmail.com Closes #4766

spark git commit: [SPARK-5771] Number of Cores in Completed Applications of Standalone Master Web Page always be 0 if sc.stop() is called

2015-02-25 Thread srowen
Repository: spark Updated Branches: refs/heads/master 5b8480e03 - dd077abf2 [SPARK-5771] Number of Cores in Completed Applications of Standalone Master Web Page always be 0 if sc.stop() is called In Standalone mode, the number of cores in Completed Applications of the Master Web Page will

spark git commit: SPARK-5669 [BUILD] Reverse exclusion of JBLAS libs for 1.3

2015-02-22 Thread srowen
Repository: spark Updated Branches: refs/heads/branch-1.3 04d3b328f - eed7389cf SPARK-5669 [BUILD] Reverse exclusion of JBLAS libs for 1.3 CC mengxr Author: Sean Owen so...@cloudera.com Closes #4715 from srowen/SPARK-5669.3 and squashes the following commits: b27ffa9 [Sean Owen] Reverse

spark git commit: [DOCS] Fix typo in API for custom InputFormats based on the “new” MapReduce API

2015-02-22 Thread srowen
Repository: spark Updated Branches: refs/heads/branch-1.3 76e3e6527 - c5a5c6f61 [DOCS] Fix typo in API for custom InputFormats based on the “new” MapReduce API This looks like a simple typo ```SparkContext.newHadoopRDD``` instead of ```SparkContext.newAPIHadoopRDD``` as in actual

spark git commit: [DOCS] Fix typo in API for custom InputFormats based on the “new” MapReduce API

2015-02-22 Thread srowen
Repository: spark Updated Branches: refs/heads/master 46462ff25 - a7f903902 [DOCS] Fix typo in API for custom InputFormats based on the “new” MapReduce API This looks like a simple typo ```SparkContext.newHadoopRDD``` instead of ```SparkContext.newAPIHadoopRDD``` as in actual

spark git commit: [DataFrame] [Typo] Fix the typo

2015-02-22 Thread srowen
Repository: spark Updated Branches: refs/heads/master a7f903902 - 275b1bef8 [DataFrame] [Typo] Fix the typo Author: Cheng Hao hao.ch...@intel.com Closes #4717 from chenghao-intel/typo1 and squashes the following commits: 858d7b0 [Cheng Hao] update the typo Project:

spark git commit: [EXAMPLES] fix typo.

2015-02-23 Thread srowen
Repository: spark Updated Branches: refs/heads/branch-1.3 4186dd3dd - f172387dd [EXAMPLES] fix typo. Author: Makoto Fukuhara fuku...@gmail.com Closes #4724 from fukuo33/fix-typo and squashes the following commits: 8c806b9 [Makoto Fukuhara] fix typo. (cherry picked from commit

spark git commit: [EXAMPLES] fix typo.

2015-02-23 Thread srowen
Repository: spark Updated Branches: refs/heads/master 95cd643aa - 934876741 [EXAMPLES] fix typo. Author: Makoto Fukuhara fuku...@gmail.com Closes #4724 from fukuo33/fix-typo and squashes the following commits: 8c806b9 [Makoto Fukuhara] fix typo. Project:

spark git commit: [SPARK-5801] [core] Avoid creating nested directories.

2015-02-26 Thread srowen
Repository: spark Updated Branches: refs/heads/master 192e42a29 - df3d559b3 [SPARK-5801] [core] Avoid creating nested directories. Cache the value of the local root dirs to use for storing local data, so that the same directories are reused. Also, to avoid an extra level of nesting, use a

spark git commit: [MLlib] fix typo

2015-02-21 Thread srowen
Repository: spark Updated Branches: refs/heads/master 5b0a42cb1 - e15532471 [MLlib] fix typo fix typo: it should be default: instead of default; Author: Jacky Li jack...@users.noreply.github.com Closes #4713 from jackylk/patch-10 and squashes the following commits: 15daf2e [Jacky Li]

spark git commit: [SPARK-5914] to run spark-submit requiring only user perm on windows

2015-02-26 Thread srowen
Repository: spark Updated Branches: refs/heads/master e43139f40 - 51a6f9097 [SPARK-5914] to run spark-submit requiring only user perm on windows Because windows on-default does not grant read permission to jars except to admin, spark-submit would fail with ClassNotFound exception if user

spark git commit: SPARK-6045 RecordWriter should be checked against null in PairRDDFunctio...

2015-02-26 Thread srowen
Repository: spark Updated Branches: refs/heads/master b38dec2ff - e60ad2f4c SPARK-6045 RecordWriter should be checked against null in PairRDDFunctio... ...ns#saveAsNewAPIHadoopDataset Author: tedyu yuzhih...@gmail.com Closes #4794 from tedyu/master and squashes the following commits:

spark git commit: [SPARK-5313][Project Infra]: Create simple framework for highlighting changes introduced in a PR

2015-03-19 Thread srowen
Repository: spark Updated Branches: refs/heads/master dda4dedca - 8cb23a1f9 [SPARK-5313][Project Infra]: Create simple framework for highlighting changes introduced in a PR Built a simple framework with a `dev/tests` directory to house all pull request related tests. I've moved the two

spark git commit: SPARK-6480 [CORE] histogram() bucket function is wrong in some simple edge cases

2015-03-26 Thread srowen
Author: Sean Owen so...@cloudera.com Closes #5148 from srowen/SPARK-6480 and squashes the following commits: 974a0a0 [Sean Owen] Additional test of huge ranges, and a few more comments (and comment fixes) 23ec01e [Sean Owen] Fix fastBucketFunction for histogram() to handle edge conditions more

spark git commit: SPARK-6480 [CORE] histogram() bucket function is wrong in some simple edge cases

2015-03-26 Thread srowen
: Sean Owen so...@cloudera.com Closes #5148 from srowen/SPARK-6480 and squashes the following commits: 974a0a0 [Sean Owen] Additional test of huge ranges, and a few more comments (and comment fixes) 23ec01e [Sean Owen] Fix fastBucketFunction for histogram() to handle edge conditions more correctly

spark git commit: SPARK-6480 [CORE] histogram() bucket function is wrong in some simple edge cases

2015-03-26 Thread srowen
Author: Sean Owen so...@cloudera.com Closes #5148 from srowen/SPARK-6480 and squashes the following commits: 974a0a0 [Sean Owen] Additional test of huge ranges, and a few more comments (and comment fixes) 23ec01e [Sean Owen] Fix fastBucketFunction for histogram() to handle edge conditions more

spark git commit: [SPARK-6560][CORE] Do not suppress exceptions from writer.write.

2015-04-03 Thread srowen
Repository: spark Updated Branches: refs/heads/master 82701ee25 - b0d884f04 [SPARK-6560][CORE] Do not suppress exceptions from writer.write. If there is a failure in the Hadoop backend while calling writer.write, we should remember this original exception, and try to call writer.close(), but

spark git commit: [MLLIB] Remove println in LogisticRegression.scala

2015-04-03 Thread srowen
Repository: spark Updated Branches: refs/heads/master b0d884f04 - b52c7f9fc [MLLIB] Remove println in LogisticRegression.scala There's no corresponding printing in linear regression. Here was my previous PR (something weird happened and I can't reopen it)

spark git commit: [SPARK-6492][CORE] SparkContext.stop() can deadlock when DAGSchedulerEventProcessLoop dies

2015-04-03 Thread srowen
Repository: spark Updated Branches: refs/heads/master c23ba81b8 - 2c43ea38e [SPARK-6492][CORE] SparkContext.stop() can deadlock when DAGSchedulerEventProcessLoop dies I've added a timeout and retry loop around the SparkContext shutdown code that should fix this deadlock. If a SparkContext

spark git commit: [Minor][SQL] Fix typo

2015-04-03 Thread srowen
Repository: spark Updated Branches: refs/heads/master 512a2f191 - dc6dff248 [Minor][SQL] Fix typo Just fix a typo. Author: Liang-Chi Hsieh vii...@gmail.com Closes #5352 from viirya/fix_a_typo and squashes the following commits: 303b2d2 [Liang-Chi Hsieh] Fix typo. Project:

spark git commit: SPARK-6878 [CORE] Fix for sum on empty RDD fails with exception

2015-04-14 Thread srowen
Repository: spark Updated Branches: refs/heads/branch-1.3 8e5caa227 - 2954468b0 SPARK-6878 [CORE] Fix for sum on empty RDD fails with exception Author: Erik van Oosten evanoos...@ebay.com Closes #5489 from erikvanoosten/master and squashes the following commits: 1c91954 [Erik van Oosten]

spark git commit: SPARK-6878 [CORE] Fix for sum on empty RDD fails with exception

2015-04-14 Thread srowen
Repository: spark Updated Branches: refs/heads/branch-1.2 daec1c635 - 899ffdcc0 SPARK-6878 [CORE] Fix for sum on empty RDD fails with exception Author: Erik van Oosten evanoos...@ebay.com Closes #5489 from erikvanoosten/master and squashes the following commits: 1c91954 [Erik van Oosten]

spark git commit: [SPARK-6731] Bump version of apache commons-math3

2015-04-14 Thread srowen
Repository: spark Updated Branches: refs/heads/master 77eeb10fd - 628a72f70 [SPARK-6731] Bump version of apache commons-math3 Version 3.1.1 is two years old and the newer version includes approximate percentile statistics (among other things). Author: Punyashloka Biswal

spark git commit: SPARK-6878 [CORE] Fix for sum on empty RDD fails with exception

2015-04-14 Thread srowen
Repository: spark Updated Branches: refs/heads/master 628a72f70 - 51b306b93 SPARK-6878 [CORE] Fix for sum on empty RDD fails with exception Author: Erik van Oosten evanoos...@ebay.com Closes #5489 from erikvanoosten/master and squashes the following commits: 1c91954 [Erik van Oosten]

spark git commit: [SPARK-7087] [BUILD] Fix path issue change version script

2015-04-23 Thread srowen
Repository: spark Updated Branches: refs/heads/master baa83a9a6 - 6d0749cae [SPARK-7087] [BUILD] Fix path issue change version script Author: Tijo Thomas tijopara...@gmail.com Closes #5656 from tijoparacka/FIX_PATHISSUE_CHANGE_VERSION_SCRIPT and squashes the following commits: ab4f4b1

spark git commit: [SPARK-6879] [HISTORYSERVER] check if app is completed before clean it up

2015-04-23 Thread srowen
Repository: spark Updated Branches: refs/heads/master 3e91cc273 - baa83a9a6 [SPARK-6879] [HISTORYSERVER] check if app is completed before clean it up https://issues.apache.org/jira/browse/SPARK-6879 Use `applications` to replace `FileStatus`, and check if the app is completed before clean

spark git commit: [SPARK-7037] [CORE] Inconsistent behavior for non-spark config properties in spark-shell and spark-submit

2015-04-23 Thread srowen
Repository: spark Updated Branches: refs/heads/master 73db132bf - 336f7f537 [SPARK-7037] [CORE] Inconsistent behavior for non-spark config properties in spark-shell and spark-submit When specifying non-spark properties (i.e. names don't start with spark.) in the command line and config

spark git commit: [SPARK-6963][CORE]Flaky test: o.a.s.ContextCleanerSuite automatically cleanup checkpoint

2015-04-19 Thread srowen
Repository: spark Updated Branches: refs/heads/master 8fbd45c74 - 0424da68d [SPARK-6963][CORE]Flaky test: o.a.s.ContextCleanerSuite automatically cleanup checkpoint cc andrewor14 Author: GuoQiang Li wi...@qq.com Closes #5548 from witgo/SPARK-6963 and squashes the following commits:

spark git commit: [SPARK-6014] [CORE] [HOTFIX] Add try-catch block around ShutDownHook

2015-04-25 Thread srowen
/4690) cc andrewor14, srowen Author: Nishkam Ravi nr...@cloudera.com Author: nishkamravi2 nishkamr...@gmail.com Author: nravi nr...@c1704.halxg.cloudera.com Closes #5672 from nishkamravi2/master_nravi and squashes the following commits: 0f1abd0 [nishkamravi2] Update Utils.scala 474e3bf

spark git commit: [SPARK-7092] Update spark scala version to 2.11.6

2015-04-25 Thread srowen
Repository: spark Updated Branches: refs/heads/master aa6966ff3 - a11c8683c [SPARK-7092] Update spark scala version to 2.11.6 Author: Prashant Sharma prashan...@imaginea.com Closes #5662 from ScrapCodes/SPARK-7092/scala-update-2.11.6 and squashes the following commits: 58cf4f9 [Prashant

spark git commit: [MINOR] [MLLIB] Refactor toString method in MLLIB

2015-04-26 Thread srowen
Repository: spark Updated Branches: refs/heads/master f5473c2bb - 9a5bbe05f [MINOR] [MLLIB] Refactor toString method in MLLIB 1. predict(predict.toString) has already output prefix “predict” thus it’s duplicated to print , predict = again 2. there are some extra spaces Author: Alain

spark git commit: SPARK-3276 Added a new configuration spark.streaming.minRememberDuration

2015-04-21 Thread srowen
Repository: spark Updated Branches: refs/heads/master c035c0f2d - c25ca7c5a SPARK-3276 Added a new configuration spark.streaming.minRememberDuration SPARK-3276 Added a new configuration parameter ``spark.streaming.minRememberDuration``, with a default value of 1 minute. So that when a Spark

spark git commit: [SPARK-6014] [core] Revamp Spark shutdown hooks, fix shutdown races.

2015-04-21 Thread srowen
Repository: spark Updated Branches: refs/heads/master b063a61b9 - e72c16e30 [SPARK-6014] [core] Revamp Spark shutdown hooks, fix shutdown races. This change adds some new utility code to handle shutdown hooks in Spark. The main goal is to take advantage of Hadoop 2.x's API for shutdown hooks,

spark git commit: update the deprecated CountMinSketchMonoid function to TopPctCMS function

2015-04-25 Thread srowen
Repository: spark Updated Branches: refs/heads/master 59b7cfc41 - cca9905b9 update the deprecated CountMinSketchMonoid function to TopPctCMS function http://twitter.github.io/algebird/index.html#com.twitter.algebird.legacy.CountMinSketchMonoid$ The CountMinSketchMonoid has been deprecated

spark git commit: [doc][mllib] Fix typo of the page title in Isotonic regression documents

2015-04-19 Thread srowen
Repository: spark Updated Branches: refs/heads/branch-1.3 1526f9c29 - 9909ad3e0 [doc][mllib] Fix typo of the page title in Isotonic regression documents * Fix the page title in Isotonic regression documents (Naive Bayes - Isotonic regression) * Add a newline character at the end of the file

spark git commit: [SPARK-6985][streaming] Receiver maxRate over 1000 causes a StackOverflowError

2015-04-21 Thread srowen
Repository: spark Updated Branches: refs/heads/branch-1.3 8549ff4f0 - 948f2f635 [SPARK-6985][streaming] Receiver maxRate over 1000 causes a StackOverflowError A simple truncation in integer division (on rates over 1000 messages / second) causes the existing implementation to sleep for 0

spark git commit: [SPARK-6985][streaming] Receiver maxRate over 1000 causes a StackOverflowError

2015-04-21 Thread srowen
Repository: spark Updated Branches: refs/heads/master 1f2f723b0 - 5fea3e5c3 [SPARK-6985][streaming] Receiver maxRate over 1000 causes a StackOverflowError A simple truncation in integer division (on rates over 1000 messages / second) causes the existing implementation to sleep for 0

spark git commit: [SPARK-6122] [CORE] Upgrade tachyon-client version to 0.6.3

2015-04-24 Thread srowen
Repository: spark Updated Branches: refs/heads/master caf0136ec - 438859eb7 [SPARK-6122] [CORE] Upgrade tachyon-client version to 0.6.3 This is a reopening of #4867. A short summary of the issues resolved from the previous PR: 1. HTTPClient version mismatch: Selenium (used for UI tests)

spark git commit: [SPARK-7100] [MLLIB] Fix persisted RDD leak in GradientBoostTrees

2015-04-28 Thread srowen
Repository: spark Updated Branches: refs/heads/master 7f3b3b7eb - 75905c57c [SPARK-7100] [MLLIB] Fix persisted RDD leak in GradientBoostTrees This fixes a leak of a persisted RDD where GradientBoostTrees can call persist but never unpersists. Jira:

spark git commit: [SPARK-7168] [BUILD] Update plugin versions in Maven build and centralize versions

2015-04-28 Thread srowen
srowen/SPARK-7168 and squashes the following commits: 98a8947 [Sean Owen] Make install, deploy plugin versions explicit 4ecf3b2 [Sean Owen] Update Maven build plugin versions and centralize plugin version management Project: http://git-wip-us.apache.org/repos/asf/spark/repo Commit: http://git

spark git commit: [SPARK-6738] [CORE] Improve estimate the size of a large array

2015-04-27 Thread srowen
Repository: spark Updated Branches: refs/heads/master b9de9e040 - 8e1c00dbf [SPARK-6738] [CORE] Improve estimate the size of a large array Currently, SizeEstimator.visitArray is not correct in the follow case, ``` array size 200, elem has the share object ``` when I add a debug log in

spark git commit: [SPARK-7162] [YARN] Launcher error in yarn-client

2015-04-27 Thread srowen
Repository: spark Updated Branches: refs/heads/master ab5adb7a9 - 62888a4de [SPARK-7162] [YARN] Launcher error in yarn-client jira: https://issues.apache.org/jira/browse/SPARK-7162 Author: GuoQiang Li wi...@qq.com Closes #5716 from witgo/SPARK-7162 and squashes the following commits:

spark git commit: [SPARK-7103] Fix crash with SparkContext.union when RDD has no partitioner

2015-04-27 Thread srowen
Repository: spark Updated Branches: refs/heads/branch-1.3 81de30ae5 - d13080aa2 [SPARK-7103] Fix crash with SparkContext.union when RDD has no partitioner Added a check to the SparkContext.union method to check that a partitioner is defined on all RDDs when instantiating a

spark git commit: [SPARK-3090] [CORE] Stop SparkContext if user forgets to.

2015-04-27 Thread srowen
Repository: spark Updated Branches: refs/heads/master 8e1c00dbf - 5d45e1f60 [SPARK-3090] [CORE] Stop SparkContext if user forgets to. Set up a shutdown hook to try to stop the Spark context in case the user forgets to do it. The main effect is that any open logs files are flushed and closed,

spark git commit: [SPARK-7103] Fix crash with SparkContext.union when RDD has no partitioner

2015-04-27 Thread srowen
Repository: spark Updated Branches: refs/heads/master ca9f4ebb8 - b9de9e040 [SPARK-7103] Fix crash with SparkContext.union when RDD has no partitioner Added a check to the SparkContext.union method to check that a partitioner is defined on all RDDs when instantiating a

spark git commit: [SPARK-7145] [CORE] commons-lang (2.x) classes used instead of commons-lang3 (3.x); commons-io used without dependency

2015-04-27 Thread srowen
of Guava Author: Sean Owen so...@cloudera.com Closes #5703 from srowen/SPARK-7145 and squashes the following commits: 21fbe03 [Sean Owen] Remove use of commons-lang in favor of commons-lang3 classes; remove commons-io use in favor of Guava Project: http://git-wip-us.apache.org/repos/asf/spark

spark git commit: [SPARK-6954] [YARN] ExecutorAllocationManager can end up requesting a ne...

2015-05-02 Thread srowen
Repository: spark Updated Branches: refs/heads/branch-1.3 d7269493d - 98ac39d2f [SPARK-6954] [YARN] ExecutorAllocationManager can end up requesting a ne... ...gative n... ...umber of executors Author: Sandy Ryza sandycloudera.com Closes #5704 from sryza/sandy-spark-6954 and squashes the

spark git commit: [SPARK-3444] Fix typo in Dataframes.py introduced in []

2015-05-02 Thread srowen
Repository: spark Updated Branches: refs/heads/master ecc6eb50a - 856a571ef [SPARK-3444] Fix typo in Dataframes.py introduced in [] Author: Dean Chen deanch...@gmail.com Closes #5866 from deanchen/patch-1 and squashes the following commits: 0028bc4 [Dean Chen] Fix typo in Dataframes.py

spark git commit: [SPARK-7323] [SPARK CORE] Use insertAll instead of insert while merging combiners in reducer

2015-05-02 Thread srowen
Repository: spark Updated Branches: refs/heads/master 856a571ef - da303526e [SPARK-7323] [SPARK CORE] Use insertAll instead of insert while merging combiners in reducer Author: Mridul Muralidharan mrid...@yahoo-inc.com Closes #5862 from mridulm/optimize_aggregator and squashes the following

spark git commit: [SPARK-7323] [SPARK CORE] Use insertAll instead of insert while merging combiners in reducer

2015-05-02 Thread srowen
Repository: spark Updated Branches: refs/heads/branch-1.3 98ac39d2f - 702356c5d [SPARK-7323] [SPARK CORE] Use insertAll instead of insert while merging combiners in reducer Author: Mridul Muralidharan mrid...@yahoo-inc.com Closes #5862 from mridulm/optimize_aggregator and squashes the

spark git commit: [MINOR] [HIVE] Fix QueryPartitionSuite.

2015-05-02 Thread srowen
Repository: spark Updated Branches: refs/heads/master bfcd528d6 - 82c8c37c0 [MINOR] [HIVE] Fix QueryPartitionSuite. At least in the version of Hive I tested on, the test was deleting a temp directory generated by Hive instead of one containing partition data. So fix the filter to only

spark git commit: [SPARK-6030] [CORE] Using simulated field layout method to compute class shellSize

2015-05-02 Thread srowen
to srowen and mateiz Author: Ye Xianjin advance...@gmail.com Closes #4783 from advancedxy/SPARK-6030 and squashes the following commits: c4dcb41 [Ye Xianjin] Add super.beforeEach in the beforeEach method to make the trait stackable.. Remove useless leading whitespace. 3f80640 [Ye Xianjin] The size

spark git commit: [SPARK-7255] [STREAMING] [DOCUMENTATION] Added documentation for spark.streaming.kafka.maxRetries

2015-05-02 Thread srowen
Repository: spark Updated Branches: refs/heads/master 5d6b90d93 - ea841efc5 [SPARK-7255] [STREAMING] [DOCUMENTATION] Added documentation for spark.streaming.kafka.maxRetries Added documentation for spark.streaming.kafka.maxRetries Author: BenFradet benjamin.fra...@gmail.com Closes #5808

spark git commit: [SPARK-7031] [THRIFTSERVER] let thrift server take SPARK_DAEMON_MEMORY and SPARK_DAEMON_JAVA_OPTS

2015-05-02 Thread srowen
Repository: spark Updated Branches: refs/heads/master ea841efc5 - 49549d5a1 [SPARK-7031] [THRIFTSERVER] let thrift server take SPARK_DAEMON_MEMORY and SPARK_DAEMON_JAVA_OPTS We should let Thrift Server take these two parameters as it is a daemon. And it is better to read driver-related

spark git commit: [SPARK-7302] [DOCS] SPARK building documentation still mentions building for yarn 0.23

2015-05-03 Thread srowen
0.23, but also don't support Hadoop 0.23 anymore AFAICT. There are no builds or releases for it. In fact, on a related note, refs to CDH3 (Hadoop 0.20.2) should be removed as this certainly isn't supported either. Author: Sean Owen so...@cloudera.com Closes #5863 from srowen/SPARK-7302

spark git commit: [SPARK-7281] [YARN] Add option to set AM's lib path in client mode.

2015-05-01 Thread srowen
Repository: spark Updated Branches: refs/heads/master f53a48827 - 7b5dd3e3c [SPARK-7281] [YARN] Add option to set AM's lib path in client mode. Author: Marcelo Vanzin van...@cloudera.com Closes #5813 from vanzin/SPARK-7281 and squashes the following commits: 1cb6f42 [Marcelo Vanzin]

spark git commit: [SPARK-7213] [YARN] Check for read permissions before copying a Hadoop config file

2015-05-01 Thread srowen
Repository: spark Updated Branches: refs/heads/master c6d9a4294 - f53a48827 [SPARK-7213] [YARN] Check for read permissions before copying a Hadoop config file Author: Nishkam Ravi nr...@cloudera.com Author: nishkamravi2 nishkamr...@gmail.com Author: nravi nr...@c1704.halxg.cloudera.com

spark git commit: Limit help option regex

2015-05-01 Thread srowen
Repository: spark Updated Branches: refs/heads/master 7630213ca - c8c481da1 Limit help option regex Added word-boundary delimiters so that embedded text such as -h within command line options and values doesn't trigger the usage script and exit. Author: Chris Biow chris.b...@10gen.com

spark git commit: Limit help option regex

2015-05-01 Thread srowen
Repository: spark Updated Branches: refs/heads/branch-1.3 f64b99421 - d7269493d Limit help option regex Added word-boundary delimiters so that embedded text such as -h within command line options and values doesn't trigger the usage script and exit. Author: Chris Biow chris.b...@10gen.com

spark git commit: [SPARK-6846] [WEBUI] [HOTFIX] return to GET for kill link in UI since YARN AM won't proxy POST

2015-05-01 Thread srowen
. Standalone UIs are not affected. Author: Sean Owen so...@cloudera.com Closes #5837 from srowen/SPARK-6846.2 and squashes the following commits: c17c386 [Sean Owen] Partial undoing of SPARK-6846; YARN AM proxy won't forward POSTs, so go back to GET for kill links in Spark UI. Standalone UIs

spark git commit: [SPARK-7181] [CORE] fix inifite loop in Externalsorter's mergeWithAggregation

2015-04-29 Thread srowen
Repository: spark Updated Branches: refs/heads/branch-1.2 6fd74d87c - c0bd415bd [SPARK-7181] [CORE] fix inifite loop in Externalsorter's mergeWithAggregation see [SPARK-7181](https://issues.apache.org/jira/browse/SPARK-7181). Author: Qiping Li liqiping1...@gmail.com Closes #5737 from

spark git commit: [SPARK-7181] [CORE] fix inifite loop in Externalsorter's mergeWithAggregation

2015-04-29 Thread srowen
Repository: spark Updated Branches: refs/heads/master d7dbce8f7 - 7f4b58373 [SPARK-7181] [CORE] fix inifite loop in Externalsorter's mergeWithAggregation see [SPARK-7181](https://issues.apache.org/jira/browse/SPARK-7181). Author: Qiping Li liqiping1...@gmail.com Closes #5737 from

spark git commit: [SPARK-7181] [CORE] fix inifite loop in Externalsorter's mergeWithAggregation

2015-04-29 Thread srowen
Repository: spark Updated Branches: refs/heads/branch-1.3 bf4ca12f7 - 5b893bd60 [SPARK-7181] [CORE] fix inifite loop in Externalsorter's mergeWithAggregation see [SPARK-7181](https://issues.apache.org/jira/browse/SPARK-7181). Author: Qiping Li liqiping1...@gmail.com Closes #5737 from

spark git commit: [SPARK-7155] [CORE] Allow newAPIHadoopFile to support comma-separated list of files as input

2015-04-29 Thread srowen
Repository: spark Updated Branches: refs/heads/master 7f4b58373 - 3fc6cfd07 [SPARK-7155] [CORE] Allow newAPIHadoopFile to support comma-separated list of files as input See JIRA: https://issues.apache.org/jira/browse/SPARK-7155 SparkContext's newAPIHadoopFile() does not support

spark git commit: [SPARK-7155] [CORE] Allow newAPIHadoopFile to support comma-separated list of files as input

2015-04-29 Thread srowen
Repository: spark Updated Branches: refs/heads/branch-1.3 5b893bd60 - 3a41a1327 [SPARK-7155] [CORE] Allow newAPIHadoopFile to support comma-separated list of files as input See JIRA: https://issues.apache.org/jira/browse/SPARK-7155 SparkContext's newAPIHadoopFile() does not support

spark git commit: [SPARK-5529] [CORE] Add expireDeadHosts in HeartbeatReceiver

2015-04-30 Thread srowen
Repository: spark Updated Branches: refs/heads/branch-1.3 3bce87ebd - ec196ab1c [SPARK-5529] [CORE] Add expireDeadHosts in HeartbeatReceiver If a blockManager has not send heartBeat more than 120s, BlockManagerMasterActor will remove it. But coarseGrainedSchedulerBackend can only remove

spark git commit: [MINOR] Minor update for document

2015-05-05 Thread srowen
Repository: spark Updated Branches: refs/heads/master d49735800 - b83091ae4 [MINOR] Minor update for document Two minor doc errors in `BytesToBytesMap` and `UnsafeRow`. Author: Liang-Chi Hsieh vii...@gmail.com Closes #5906 from viirya/minor_doc and squashes the following commits: 27f9089

spark git commit: [MINOR] Minor update for document

2015-05-05 Thread srowen
Repository: spark Updated Branches: refs/heads/branch-1.4 ff8b44995 - d28832299 [MINOR] Minor update for document Two minor doc errors in `BytesToBytesMap` and `UnsafeRow`. Author: Liang-Chi Hsieh vii...@gmail.com Closes #5906 from viirya/minor_doc and squashes the following commits:

spark git commit: [SPARK-5112] Expose SizeEstimator as a developer api

2015-05-05 Thread srowen
Repository: spark Updated Branches: refs/heads/branch-1.4 93af96a2f - 0327ca2b2 [SPARK-5112] Expose SizeEstimator as a developer api The best way to size the amount of memory consumption your dataset will require is to create an RDD, put it into cache, and look at the SparkContext logs on

spark git commit: [SPARK-5112] Expose SizeEstimator as a developer api

2015-05-05 Thread srowen
Repository: spark Updated Branches: refs/heads/master fc8feaa8e - 4222da68d [SPARK-5112] Expose SizeEstimator as a developer api The best way to size the amount of memory consumption your dataset will require is to create an RDD, put it into cache, and look at the SparkContext logs on your

svn commit: r1677934 - /spark/site/sql/index.html

2015-05-06 Thread srowen
Author: srowen Date: Wed May 6 06:45:28 2015 New Revision: 1677934 URL: http://svn.apache.org/r1677934 Log: Fix 'seemlessly' typo Modified: spark/site/sql/index.html Modified: spark/site/sql/index.html URL: http://svn.apache.org/viewvc/spark/site/sql/index.html?rev=1677934r1=1677933r2

svn commit: r1677933 - /spark/site/faq.html

2015-05-06 Thread srowen
Author: srowen Date: Wed May 6 06:42:44 2015 New Revision: 1677933 URL: http://svn.apache.org/r1677933 Log: SPARK-6889 point to new contribution guidelines Modified: spark/site/faq.html Modified: spark/site/faq.html URL: http://svn.apache.org/viewvc/spark/site/faq.html?rev=1677933r1

spark git commit: [SPARK-5074] [CORE] [TESTS] Fix the flakey test 'run shuffle with map stage failure' in DAGSchedulerSuite

2015-05-05 Thread srowen
Repository: spark Updated Branches: refs/heads/branch-1.4 d28832299 - 6f35dac50 [SPARK-5074] [CORE] [TESTS] Fix the flakey test 'run shuffle with map stage failure' in DAGSchedulerSuite Test failure:

spark git commit: [SPARK-5074] [CORE] [TESTS] Fix the flakey test 'run shuffle with map stage failure' in DAGSchedulerSuite

2015-05-05 Thread srowen
Repository: spark Updated Branches: refs/heads/branch-1.3 b34b5bdf5 - 9278b7a6a [SPARK-5074] [CORE] [TESTS] Fix the flakey test 'run shuffle with map stage failure' in DAGSchedulerSuite Test failure:

spark git commit: [SPARK-5074] [CORE] [TESTS] Fix the flakey test 'run shuffle with map stage failure' in DAGSchedulerSuite

2015-05-05 Thread srowen
Repository: spark Updated Branches: refs/heads/master b83091ae4 - 5ffc73e68 [SPARK-5074] [CORE] [TESTS] Fix the flakey test 'run shuffle with map stage failure' in DAGSchedulerSuite Test failure:

spark git commit: [MLLIB] [TREE] Verify size of input rdd 0 when building meta data

2015-05-05 Thread srowen
Repository: spark Updated Branches: refs/heads/master 9d250e64d - d4cb38aeb [MLLIB] [TREE] Verify size of input rdd 0 when building meta data Require non empty input rdd such that we can take the first labeledpoint and get the feature size Author: Alain a...@usc.edu Author: a...@usc.edu

spark git commit: [SPARK-7350] [STREAMING] [WEBUI] Attach the Streaming tab when calling ssc.start()

2015-05-05 Thread srowen
Repository: spark Updated Branches: refs/heads/master 5ffc73e68 - c6d1efba2 [SPARK-7350] [STREAMING] [WEBUI] Attach the Streaming tab when calling ssc.start() It's meaningless to display the Streaming tab before `ssc.start()`. So we should attach it in the `ssc.start` method. Author:

spark git commit: [SPARK-7350] [STREAMING] [WEBUI] Attach the Streaming tab when calling ssc.start()

2015-05-05 Thread srowen
Repository: spark Updated Branches: refs/heads/branch-1.4 6f35dac50 - 49923f7ca [SPARK-7350] [STREAMING] [WEBUI] Attach the Streaming tab when calling ssc.start() It's meaningless to display the Streaming tab before `ssc.start()`. So we should attach it in the `ssc.start` method. Author:

spark git commit: [SPARK-7392] [CORE] bugfix: Kryo buffer size cannot be larger than 2M

2015-05-08 Thread srowen
Repository: spark Updated Branches: refs/heads/branch-1.4 bb5872f2d - f5e9678e3 [SPARK-7392] [CORE] bugfix: Kryo buffer size cannot be larger than 2M Author: Zhang, Liye liye.zh...@intel.com Closes #5934 from liyezhang556520/kryoBufSize and squashes the following commits: 5707e04 [Zhang,

spark git commit: [SPARK-7326] [STREAMING] Performing window() on a WindowedDStream doesn't work all the time

2015-05-11 Thread srowen
Repository: spark Updated Branches: refs/heads/master 2242ab31e - d70a07689 [SPARK-7326] [STREAMING] Performing window() on a WindowedDStream doesn't work all the time tdas https://issues.apache.org/jira/browse/SPARK-7326 The problem most likely resides in DStream.slice() implementation,

spark git commit: [SPARK-7357] Improving HBaseTest example

2015-05-05 Thread srowen
Repository: spark Updated Branches: refs/heads/master 4222da68d - 51f462003 [SPARK-7357] Improving HBaseTest example Author: Jihong MA linlin200...@gmail.com Closes #5904 from JihongMA/SPARK-7357 and squashes the following commits: 7d6153a [Jihong MA] SPARK-7357 Improving HBaseTest example

spark git commit: [MINOR] Renamed variables in SparkKMeans.scala, LocalKMeans.scala and kmeans.py to simplify readability

2015-05-05 Thread srowen
Repository: spark Updated Branches: refs/heads/master e9b16e67c - da738cffa [MINOR] Renamed variables in SparkKMeans.scala, LocalKMeans.scala and kmeans.py to simplify readability With the previous syntax it could look like that the reduceByKey sums separately abscissas and ordinates of

spark git commit: [MINOR] [BUILD] Declare ivy dependency in root pom.

2015-05-05 Thread srowen
Repository: spark Updated Branches: refs/heads/master da738cffa - c5790a2f7 [MINOR] [BUILD] Declare ivy dependency in root pom. Without this, any dependency that pulls ivy transitively may override the version and potentially cause issue. In my machine, the hive tests were pulling an old

spark git commit: [MINOR] [BUILD] Declare ivy dependency in root pom.

2015-05-05 Thread srowen
Repository: spark Updated Branches: refs/heads/branch-1.4 21ed108fa - 5160437c1 [MINOR] [BUILD] Declare ivy dependency in root pom. Without this, any dependency that pulls ivy transitively may override the version and potentially cause issue. In my machine, the hive tests were pulling an old

spark git commit: [MINOR] [BUILD] Declare ivy dependency in root pom.

2015-05-05 Thread srowen
Repository: spark Updated Branches: refs/heads/branch-1.3 702356c5d - b34b5bdf5 [MINOR] [BUILD] Declare ivy dependency in root pom. Without this, any dependency that pulls ivy transitively may override the version and potentially cause issue. In my machine, the hive tests were pulling an old

spark git commit: [MINOR] [DOCS] Fix the link to test building info on the wiki

2015-05-11 Thread srowen
Repository: spark Updated Branches: refs/heads/master a8ea09683 - 91dc3dfd7 [MINOR] [DOCS] Fix the link to test building info on the wiki Author: Sean Owen so...@cloudera.com Closes #6063 from srowen/FixRunningTestsLink and squashes the following commits: db62018 [Sean Owen] Fix the link

spark git commit: [SPARK-7438] [SPARK CORE] Fixed validation of relativeSD in countApproxDistinct

2015-05-09 Thread srowen
Repository: spark Updated Branches: refs/heads/branch-1.4 25972d371 - b0460f414 [SPARK-7438] [SPARK CORE] Fixed validation of relativeSD in countApproxDistinct Author: Vinod K C vinod...@huawei.com Closes #5974 from vinodkc/fix_countApproxDistinct_Validation and squashes the following

spark git commit: [SPARK-7438] [SPARK CORE] Fixed validation of relativeSD in countApproxDistinct

2015-05-09 Thread srowen
Repository: spark Updated Branches: refs/heads/master 299262384 - dda6d9f40 [SPARK-7438] [SPARK CORE] Fixed validation of relativeSD in countApproxDistinct Author: Vinod K C vinod...@huawei.com Closes #5974 from vinodkc/fix_countApproxDistinct_Validation and squashes the following commits:

spark git commit: [SPARK-7403] [WEBUI] Link URL in objects on Timeline View is wrong in case of running on YARN

2015-05-09 Thread srowen
Repository: spark Updated Branches: refs/heads/master dda6d9f40 - 12b95abc7 [SPARK-7403] [WEBUI] Link URL in objects on Timeline View is wrong in case of running on YARN When we use Spark on YARN and have AllJobPage via ResourceManager's proxy, the link URL in objects which represent each

<    1   2   3   4   5   6   7   8   9   10   >