Github user jiangxb1987 commented on the issue:

    https://github.com/apache/spark/pull/20347
  
    My major concern is that, if there is a existing `SparkContext`, some confs 
you set may not take effect, as described in `SparkContext.getOrCreate()`. It's 
hard to enumerate the use cases but I'm sure there are some that pass in 
specific confs to create a new `JavaSparkContext`, so I tend to keep the 
current behavior here.
    
    On the other hand, the following comment copyed from the comment of the 
class `JavaSparkContext`:
    ```
     * Only one SparkContext may be active per JVM.  You must `stop()` the 
active SparkContext before
     * creating a new one.  This limitation may eventually be removed; see 
SPARK-2243 for more details.
    ```
    If that is the case, there should be no active `SparkContext` before we 
initiate the `JavaSparkContext`, so the change doesn't bring any advantage in 
that means.


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org

Reply via email to