[GitHub] spark pull request: Fixed the number of worker thread

2014-07-29 Thread pwendell
Github user pwendell commented on the pull request: https://github.com/apache/spark/pull/1485#issuecomment-50545882 Hey there - as Aaron said, the executors should never have more than N tasks active if there are N cores. I think there might be a bug causing this. So I'd recommend we

[GitHub] spark pull request: Fixed the number of worker thread

2014-07-29 Thread asfgit
Github user asfgit closed the pull request at: https://github.com/apache/spark/pull/1485 --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is

[GitHub] spark pull request: Fixed the number of worker thread

2014-07-19 Thread fireflyc
Github user fireflyc commented on the pull request: https://github.com/apache/spark/pull/1485#issuecomment-49501533 My program is spark streaming over Hadoop yarn.It work for user click stream. I read code,number of worker threads and block? --- If your project is set up for it,

[GitHub] spark pull request: Fixed the number of worker thread

2014-07-19 Thread aarondav
Github user aarondav commented on the pull request: https://github.com/apache/spark/pull/1485#issuecomment-49526386 @fireflyc Spark should not be scheduling more than N concurrent tasks on an Executor. It appears that the tasks may be returning success but then don't actually return

[GitHub] spark pull request: Fixed the number of worker thread

2014-07-18 Thread fireflyc
GitHub user fireflyc opened a pull request: https://github.com/apache/spark/pull/1485 Fixed the number of worker thread There are a lot of input Block cause too many Worker threads and will load all data.So it should be to control the number of Worker threads You can merge this

[GitHub] spark pull request: Fixed the number of worker thread

2014-07-18 Thread AmplabJenkins
Github user AmplabJenkins commented on the pull request: https://github.com/apache/spark/pull/1485#issuecomment-49443851 Can one of the admins verify this patch? --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your

[GitHub] spark pull request: Fixed the number of worker thread

2014-07-18 Thread srowen
Github user srowen commented on the pull request: https://github.com/apache/spark/pull/1485#issuecomment-49444796 Slightly bigger point: both the 'fixed' and 'cached' executors from `Executors` have some drawbacks: - 'fixed' always keeps the given number of threads active

[GitHub] spark pull request: Fixed the number of worker thread

2014-07-18 Thread aarondav
Github user aarondav commented on the pull request: https://github.com/apache/spark/pull/1485#issuecomment-49494194 The tasks launched on an Executor are controlled by the DAGScheduler, and should not exceed the number of cores that executor is advertising. In what situation have you

[GitHub] spark pull request: Fixed the number of worker thread

2014-07-18 Thread fireflyc
Github user fireflyc commented on the pull request: https://github.com/apache/spark/pull/1485#issuecomment-49495043 My application have 1000+ Worker Threads.