[GitHub] spark pull request #18093: [WIP][SPARK-20774][SQL] Cancel all jobs when Quer...

2017-08-14 Thread liyichao
Github user liyichao closed the pull request at: https://github.com/apache/spark/pull/18093 --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature

[GitHub] spark issue #18070: [SPARK-20713][Spark Core] Convert CommitDenied to TaskKi...

2017-08-03 Thread liyichao
Github user liyichao commented on the issue: https://github.com/apache/spark/pull/18070 Oh, I did not notice that, since @nlyu follows up, I will close this pr now. --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your

[GitHub] spark pull request #18070: [SPARK-20713][Spark Core] Convert CommitDenied to...

2017-08-03 Thread liyichao
Github user liyichao closed the pull request at: https://github.com/apache/spark/pull/18070 --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature

[GitHub] spark issue #18070: [SPARK-20713][Spark Core] Convert CommitDenied to TaskKi...

2017-08-02 Thread liyichao
Github user liyichao commented on the issue: https://github.com/apache/spark/pull/18070 I will update the pr in a day. --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled

[GitHub] spark issue #18093: [WIP][SPARK-20774][SQL] Cancel all jobs when QueryExecti...

2017-07-23 Thread liyichao
Github user liyichao commented on the issue: https://github.com/apache/spark/pull/18093 Sorry about that, I will test it when I have time. --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have

[GitHub] spark pull request #18092: [SPARK-20640][CORE]Make rpc timeout and retry for...

2017-06-18 Thread liyichao
Github user liyichao commented on a diff in the pull request: https://github.com/apache/spark/pull/18092#discussion_r122621671 --- Diff: core/src/test/scala/org/apache/spark/storage/BlockManagerSuite.scala --- @@ -1281,6 +1286,61 @@ class BlockManagerSuite extends SparkFunSuite

[GitHub] spark pull request #18092: [SPARK-20640][CORE]Make rpc timeout and retry for...

2017-06-18 Thread liyichao
Github user liyichao commented on a diff in the pull request: https://github.com/apache/spark/pull/18092#discussion_r122620600 --- Diff: core/src/test/scala/org/apache/spark/storage/BlockManagerSuite.scala --- @@ -1281,6 +1286,59 @@ class BlockManagerSuite extends SparkFunSuite

[GitHub] spark pull request #18092: [SPARK-20640][CORE]Make rpc timeout and retry for...

2017-06-18 Thread liyichao
Github user liyichao commented on a diff in the pull request: https://github.com/apache/spark/pull/18092#discussion_r122620196 --- Diff: core/src/test/scala/org/apache/spark/storage/BlockManagerSuite.scala --- @@ -1281,6 +1286,59 @@ class BlockManagerSuite extends SparkFunSuite

[GitHub] spark issue #18092: [SPARK-20640][CORE]Make rpc timeout and retry for shuffl...

2017-06-18 Thread liyichao
Github user liyichao commented on the issue: https://github.com/apache/spark/pull/18092 ping @jiangxb1987 --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so

[GitHub] spark pull request #18144: [SPARK-20912][SQL] Allow column name in map funct...

2017-06-14 Thread liyichao
Github user liyichao closed the pull request at: https://github.com/apache/spark/pull/18144 --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature

[GitHub] spark issue #18144: [SPARK-20912][SQL] Allow column name in map functions.

2017-06-14 Thread liyichao
Github user liyichao commented on the issue: https://github.com/apache/spark/pull/18144 As the idea is not that good, this is closed. --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have

[GitHub] spark pull request #18084: [SPARK-19900][core]Remove driver when relaunching...

2017-06-14 Thread liyichao
Github user liyichao commented on a diff in the pull request: https://github.com/apache/spark/pull/18084#discussion_r121902338 --- Diff: core/src/test/scala/org/apache/spark/deploy/master/MasterSuite.scala --- @@ -588,6 +633,70 @@ class MasterSuite extends SparkFunSuite

[GitHub] spark issue #18084: [SPARK-19900][core]Remove driver when relaunching.

2017-06-12 Thread liyichao
Github user liyichao commented on the issue: https://github.com/apache/spark/pull/18084 OK, another scenario: * driver with driverId1 started on worker1 * worker1 lost * master add driverId1 to waitingDrivers * worker1 reconnects and sends DriverStateChanged

[GitHub] spark issue #18084: [SPARK-19900][core]Remove driver when relaunching.

2017-06-12 Thread liyichao
Github user liyichao commented on the issue: https://github.com/apache/spark/pull/18084 Hi, add a workerId may not work. For example, this scenario: * driver with driverId1 started on worker1 * worker1 lost * master add driverId1 to waitingDrivers * worker1

[GitHub] spark issue #18092: [SPARK-20640][CORE]Make rpc timeout and retry for shuffl...

2017-06-12 Thread liyichao
Github user liyichao commented on the issue: https://github.com/apache/spark/pull/18092 @JoshRosen Could you please see the failed test? It seems unrelated to this pr. --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well

[GitHub] spark pull request #18092: [SPARK-20640][CORE]Make rpc timeout and retry for...

2017-06-11 Thread liyichao
Github user liyichao commented on a diff in the pull request: https://github.com/apache/spark/pull/18092#discussion_r121281612 --- Diff: core/src/test/scala/org/apache/spark/storage/BlockManagerSuite.scala --- @@ -1281,6 +1285,57 @@ class BlockManagerSuite extends SparkFunSuite

[GitHub] spark issue #18070: [SPARK-20713][Spark Core] Convert CommitDenied to TaskKi...

2017-06-02 Thread liyichao
Github user liyichao commented on the issue: https://github.com/apache/spark/pull/18070 How about Letting TaskCommitDenied and TaskKilled extend a same trait (for example, TaskKilledReason)? This way when accounting metrics, TaskCommitDenied and TaskKilled are all contributing

[GitHub] spark issue #18084: [SPARK-19900][core]Remove driver when relaunching.

2017-05-31 Thread liyichao
Github user liyichao commented on the issue: https://github.com/apache/spark/pull/18084 ping @jiangxb1987 --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so

[GitHub] spark issue #18070: [SPARK-20713][Spark Core] Convert CommitDenied to TaskKi...

2017-05-31 Thread liyichao
Github user liyichao commented on the issue: https://github.com/apache/spark/pull/18070 ping @tgravescs --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so

[GitHub] spark issue #18092: [SPARK-20640][CORE]Make rpc timeout and retry for shuffl...

2017-05-31 Thread liyichao
Github user liyichao commented on the issue: https://github.com/apache/spark/pull/18092 ping @JoshRosen --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so

[GitHub] spark pull request #18129: [SPARK-20365][YARN] Remove LocalSchem when add pa...

2017-05-31 Thread liyichao
Github user liyichao commented on a diff in the pull request: https://github.com/apache/spark/pull/18129#discussion_r119309080 --- Diff: resource-managers/yarn/src/test/scala/org/apache/spark/deploy/yarn/ClientSuite.scala --- @@ -116,15 +116,16 @@ class ClientSuite extends

[GitHub] spark pull request #18144: [SPARK-20912][SQL] Allow column name in map funct...

2017-05-30 Thread liyichao
GitHub user liyichao opened a pull request: https://github.com/apache/spark/pull/18144 [SPARK-20912][SQL] Allow column name in map functions. ## What changes were proposed in this pull request? `map` function only accepts Column values only. It'd be very helpful to have

[GitHub] spark pull request #18129: Remove LocalSchem when add path to ClassPath.

2017-05-27 Thread liyichao
GitHub user liyichao opened a pull request: https://github.com/apache/spark/pull/18129 Remove LocalSchem when add path to ClassPath. ## What changes were proposed in this pull request? In Spark on YARN, when configuring "spark.yarn.jars" with local jars (jar

[GitHub] spark issue #18084: [SPARK-19900][core]Remove driver when relaunching.

2017-05-26 Thread liyichao
Github user liyichao commented on the issue: https://github.com/apache/spark/pull/18084 Hi, I've thought more thoroughly about this. The main state involved here is Master.workers, Master.idToWorker, and WorkerInfo.drivers. Say `driverId1` runs on Worker A. Assume

[GitHub] spark issue #18084: [SPARK-19900][core]Remove driver when relaunching.

2017-05-26 Thread liyichao
Github user liyichao commented on the issue: https://github.com/apache/spark/pull/18084 Thanks for the reply. I have add some more tests to verify the state of master and worker after relaunching. I will try think about if there are ways to reuse the old driver struct

[GitHub] spark pull request #18092: [SPARK-20640][CORE]Make rpc timeout and retry for...

2017-05-26 Thread liyichao
Github user liyichao commented on a diff in the pull request: https://github.com/apache/spark/pull/18092#discussion_r118671187 --- Diff: core/src/main/scala/org/apache/spark/storage/BlockManager.scala --- @@ -170,11 +170,17 @@ private[spark] class BlockManager( // service

[GitHub] spark issue #18092: [SPARK-20640][CORE]Make rpc timeout and retry for shuffl...

2017-05-25 Thread liyichao
Github user liyichao commented on the issue: https://github.com/apache/spark/pull/18092 Sorry, I thought it not necessary to duplicate message in JIRA, thanks for the suggestion. PR is updated. As to the test plan, the modification seems straightforward, and I can not think

[GitHub] spark pull request #18084: [SPARK-19900][core]Remove driver when relaunching...

2017-05-25 Thread liyichao
Github user liyichao commented on a diff in the pull request: https://github.com/apache/spark/pull/18084#discussion_r118436939 --- Diff: core/src/test/scala/org/apache/spark/deploy/master/MasterSuite.scala --- @@ -499,4 +500,103 @@ class MasterSuite extends SparkFunSuite

[GitHub] spark pull request #18084: [SPARK-19900][core]Remove driver when relaunching...

2017-05-25 Thread liyichao
Github user liyichao commented on a diff in the pull request: https://github.com/apache/spark/pull/18084#discussion_r118424700 --- Diff: core/src/main/scala/org/apache/spark/deploy/master/Master.scala --- @@ -796,9 +796,12 @@ private[deploy] class Master

[GitHub] spark pull request #18093: [SPARK-20774][SQL] Cancel all jobs when QueryExec...

2017-05-24 Thread liyichao
GitHub user liyichao opened a pull request: https://github.com/apache/spark/pull/18093 [SPARK-20774][SQL] Cancel all jobs when QueryExection throws. see https://issues.apache.org/jira/browse/SPARK-20774?filter=12340455 ## What changes were proposed in this pull request

[GitHub] spark pull request #18092: Make rpc timeout and retry for shuffle registrati...

2017-05-24 Thread liyichao
GitHub user liyichao opened a pull request: https://github.com/apache/spark/pull/18092 Make rpc timeout and retry for shuffle registration configurable. ## What changes were proposed in this pull request? As title said ## How was this patch tested

[GitHub] spark pull request #18084: [SPARK-19900][core]Remove driver when relaunching...

2017-05-24 Thread liyichao
GitHub user liyichao opened a pull request: https://github.com/apache/spark/pull/18084 [SPARK-19900][core]Remove driver when relaunching. This is https://github.com/apache/spark/pull/17888 . cc @cloud-fan @jiangxb1987 You can merge this pull request into a Git repository

[GitHub] spark issue #17888: [SPARK-19900][core]Remove driver when relaunching.

2017-05-24 Thread liyichao
Github user liyichao commented on the issue: https://github.com/apache/spark/pull/17888 Sorry, it seems I make a mistake when rebase. I will open another pr. --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your

[GitHub] spark pull request #17888: [SPARK-19900][core]Remove driver when relaunching...

2017-05-24 Thread liyichao
Github user liyichao closed the pull request at: https://github.com/apache/spark/pull/17888 --- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature

[GitHub] spark issue #17888: [SPARK-19900][core]Remove driver when relaunching.

2017-05-23 Thread liyichao
Github user liyichao commented on the issue: https://github.com/apache/spark/pull/17888 Thanks for reviewing. Basically, the problem is that when relaunching a driver and later the original driver reconnect, there will be an application which does not have a corresponding driver. I

[GitHub] spark pull request #18070: [SPARK-20713][Spark Core] Convert CommitDenied to...

2017-05-23 Thread liyichao
Github user liyichao commented on a diff in the pull request: https://github.com/apache/spark/pull/18070#discussion_r118050668 --- Diff: core/src/main/scala/org/apache/spark/executor/Executor.scala --- @@ -338,6 +340,9 @@ private[spark] class Executor

[GitHub] spark pull request #18070: Convert CommitDenied to TaskKilled.

2017-05-23 Thread liyichao
GitHub user liyichao opened a pull request: https://github.com/apache/spark/pull/18070 Convert CommitDenied to TaskKilled. ## What changes were proposed in this pull request? In executor, `CommitDeniedException` is converted to `TaskKilledException` to avoid

[GitHub] spark pull request #17888: [SPARK-19900][core]Remove driver when relaunching...

2017-05-07 Thread liyichao
GitHub user liyichao opened a pull request: https://github.com/apache/spark/pull/17888 [SPARK-19900][core]Remove driver when relaunching. ## What changes were proposed in this pull request? * remove failed apps when worker down * do not reuse driver id when relaunching