[ https://issues.apache.org/jira/browse/HIVE-12538?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15034578#comment-15034578 ]
Jimmy Xiang commented on HIVE-12538: ------------------------------------ bq. Not quite follow.Is there anything special in operation conf for SparkSession? And when to set "isSparkConfigUpdated =false " ? We can set it to false for the session level conf only. So this flag in the operation level is totaly ignored, all the time. Things are a little tricky actually. If we use the session level conf, we could miss some non-spark-related settings in the operation level conf. If we use the operation level conf, we could miss some spark-related settings in the session level conf. Instead of just maintaining a isSparkConfigUpdated flag, probably, we should have a separate map to store such changed spark-related settings temporarily. This map can be reset upon SparkUtilities#getSparkSession() is invoked. > After set spark related config, SparkSession never get reused > ------------------------------------------------------------- > > Key: HIVE-12538 > URL: https://issues.apache.org/jira/browse/HIVE-12538 > Project: Hive > Issue Type: Bug > Components: Spark > Affects Versions: 1.3.0 > Reporter: Nemon Lou > Assignee: Nemon Lou > Attachments: HIVE-12538.1.patch, HIVE-12538.patch > > > Hive on Spark yarn-cluster mode. > After setting "set spark.yarn.queue=QueueA;" , > run the query "select count(*) from test" 3 times and you will find 3 > different yarn applications. > Two of the yarn applications in FINISHED & SUCCEEDED state,and one in RUNNING > & UNDEFINED state waiting for next work. > And if you submit one more "select count(*) from test" ,the third one will be > in FINISHED & SUCCEEDED state and a new yarn application will start up. -- This message was sent by Atlassian JIRA (v6.3.4#6332)