[GitHub] spark pull request: Add timeout for fetch file

2014-03-09 Thread asfgit
Github user asfgit closed the pull request at: https://github.com/apache/spark/pull/98 --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is

[GitHub] spark pull request: Add timeout for fetch file

2014-03-07 Thread guojc
GitHub user guojc opened a pull request: https://github.com/apache/spark/pull/98 Add timeout for fetch file Currently, when fetch a file, the connection's connect timeout and read timeout is based on the default jvm setting, in this change, I change it to use

[GitHub] spark pull request: Add timeout for fetch file

2014-03-07 Thread AmplabJenkins
Github user AmplabJenkins commented on the pull request: https://github.com/apache/spark/pull/98#issuecomment-37033983 Can one of the admins verify this patch? --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your

[GitHub] spark pull request: Add timeout for fetch file

2014-03-07 Thread AmplabJenkins
Github user AmplabJenkins commented on the pull request: https://github.com/apache/spark/pull/98#issuecomment-37052716 Merged build started. --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have

[GitHub] spark pull request: Add timeout for fetch file

2014-03-07 Thread AmplabJenkins
Github user AmplabJenkins commented on the pull request: https://github.com/apache/spark/pull/98#issuecomment-37052715 Merged build triggered. --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not

[GitHub] spark pull request: Add timeout for fetch file

2014-03-07 Thread pwendell
Github user pwendell commented on the pull request: https://github.com/apache/spark/pull/98#issuecomment-37052776 @guojc hey I'm wondering - if the default is -1 (unlimited, no timeout) then why is it removing your task set due to failure? If there is no timeout then won't it just

[GitHub] spark pull request: Add timeout for fetch file

2014-03-07 Thread guojc
Github user guojc commented on the pull request: https://github.com/apache/spark/pull/98#issuecomment-37054016 I'm not sure the behavior of default -1, as in http://docs.oracle.com/javase/7/docs/api/java/net/URLConnection.html#setReadTimeout%28int%29 says 0 is for infinity. But we do