[GitHub] spark issue #12951: [SPARK-15176][Core] Add maxShares setting to Pools

2016-10-13 Thread njwhite
Github user njwhite commented on the issue: https://github.com/apache/spark/pull/12951 Actually, @kayousterhout - I'm not entirely sure what you expect for the semantics of maxShares in general. Maybe a worked example would help: if I have a pool X with 5 running tasks from Taskset

[GitHub] spark issue #12951: [SPARK-15176][Core] Add maxShares setting to Pools

2016-10-13 Thread njwhite
Github user njwhite commented on the issue: https://github.com/apache/spark/pull/12951 @kayousterhout minShares is a configuration parameter for the fair scheduler algorithm only - what would the semantics of a maxShares setting for the FIFO algorithm be? --- If your project is set

[GitHub] spark issue #12951: [SPARK-15176][Core] Add maxShares setting to Pools

2016-07-15 Thread njwhite
Github user njwhite commented on the issue: https://github.com/apache/spark/pull/12951 ping? --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so

[GitHub] spark issue #12951: [SPARK-15176][Core] Add maxShares setting to Pools

2016-07-04 Thread njwhite
Github user njwhite commented on the issue: https://github.com/apache/spark/pull/12951 Hi @kayousterhout - I've renamed all references to `maxRunningTasks` and updated the Markdown documentation in the repo. Is this OK? Thanks - --- If your project is set up for it, you can reply

[GitHub] spark issue #12951: [SPARK-15176][Core] Add maxShares setting to Pools

2016-06-13 Thread njwhite
Github user njwhite commented on the issue: https://github.com/apache/spark/pull/12951 @squito is this OK? --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so

[GitHub] spark issue #12951: [SPARK-15176][Core] Add maxShares setting to Pools

2016-06-06 Thread njwhite
Github user njwhite commented on the issue: https://github.com/apache/spark/pull/12951 @squito thanks - I've expanded the `Scheduler respects maxRunningTasks setting of its pool` test to cover the cases you mention (and a couple of others). --- If your project is set up for it, you

[GitHub] spark pull request: [SPARK-15176][Core] Add maxShares setting to P...

2016-05-28 Thread njwhite
Github user njwhite commented on the pull request: https://github.com/apache/spark/pull/12951#issuecomment-97641 Thanks @markhamstra! The Jenkins build failed because a single test, `ExternalAppendOnlyMapSuite#spilling with compression`, failed. It seems unrelated (and passes

[GitHub] spark pull request: [SPARK-15176][Core] Add maxShares setting to P...

2016-05-27 Thread njwhite
Github user njwhite commented on the pull request: https://github.com/apache/spark/pull/12951#issuecomment-222175368 Thanks @squito; I've renamed the setting to `maxRunningTasks` and added the tests you asked for to `TaskSetManagerSuite`. I've also added support (& t

[GitHub] spark pull request: [SPARK-15176][Core] Add maxShares setting to P...

2016-05-20 Thread njwhite
Github user njwhite commented on the pull request: https://github.com/apache/spark/pull/12951#issuecomment-220555299 Thanks for the review @squito - I've commented on the JIRA about why this feature would be useful. As for the implementation - maybe "maxShare" is the

[GitHub] spark pull request: [SPARK-15176][Core] Add maxShares setting to P...

2016-05-06 Thread njwhite
Github user njwhite commented on the pull request: https://github.com/apache/spark/pull/12951#issuecomment-217498019 @HyukjinKwon I've run `run-tests` and fixed all the style issues. Could you take another look? Thanks - --- If your project is set up for it, you can reply

[GitHub] spark pull request: [SPARK-15176][Core] Add maxShares setting to P...

2016-05-06 Thread njwhite
GitHub user njwhite opened a pull request: https://github.com/apache/spark/pull/12951 [SPARK-15176][Core] Add maxShares setting to Pools ## What changes were proposed in this pull request? Help guarantee resource availablity by (hierarchically) limiting the amount of tasks

[GitHub] spark pull request: [SPARK-14859][PYSPARK] Make Lambda Serializer ...

2016-04-25 Thread njwhite
Github user njwhite commented on the pull request: https://github.com/apache/spark/pull/12620#issuecomment-214210492 @davies I'm using this to use the "dill" serializer, as it can pickle more things (and allows more fine-grained control) than the cloud-pickle serializer.

[GitHub] spark pull request: [SPARK-14859][PYSPARK] Make Lambda Serializer ...

2016-04-22 Thread njwhite
GitHub user njwhite opened a pull request: https://github.com/apache/spark/pull/12620 [SPARK-14859][PYSPARK] Make Lambda Serializer Configurable ## What changes were proposed in this pull request? Store the serializer that we should use to serialize RDD transformation