-----------------------------------------------------------
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/30055/#review68686
-----------------------------------------------------------


Szehon, do you try to transfer RSC configuration through RemoteDriver --conf 
option? The command generated in SparkClientImpl should looks like: 
SparkSubmit --properties-file /tmp/spark-submit.1267525585014474423.properties 
--class org.apache.hive.spark.client.RemoteDriver 
/usr/lib/hive-0.15.0/lib/hive-exec-0.15.0-SNAPSHOT.jar --remote-host node14-4 
--remote-port 38136 --conf hive.spark....=... --conf hive.spark...=...
It's quite strage that SparkSubmit handle it's child main class's arguments.

- chengxiang li


On Jan. 19, 2015, 11:16 p.m., Szehon Ho wrote:
> 
> -----------------------------------------------------------
> This is an automatically generated e-mail. To reply, visit:
> https://reviews.apache.org/r/30055/
> -----------------------------------------------------------
> 
> (Updated Jan. 19, 2015, 11:16 p.m.)
> 
> 
> Review request for hive and chengxiang li.
> 
> 
> Bugs: HIVE-9337
>     https://issues.apache.org/jira/browse/HIVE-9337
> 
> 
> Repository: hive-git
> 
> 
> Description
> -------
> 
> This change allows the Remote Spark Driver's properties to be set dynamically 
> via Hive configuration (ie, set commands).
> 
> Went through the Remote Spark Driver's properties and added them to HiveConf, 
> fixing the descriptions so that they're more clear in a global context with 
> other Hive properties.  Also fixed a bug in description that stated default 
> value of max message size is 10MB, should read 50MB.  One open question is 
> that I did not move 'hive.spark.log.dir' as I could not find where it was 
> read, and did not know if its still being used somewhere?
> 
> The passing of these properties between client (Hive) and RemoteSparkDriver 
> is done via the properties file.  One note is that these properties have to 
> be appended with 'spark', as SparkConf only accepts those.  I tried a long 
> time to pass them via 'conf' but found that it won't work (see 
> SparkSubmitArguments.scala).  It may be possible to pass them each as another 
> argument (like --hive.spark.XXX=YYY), but I think its more scalable to do it 
> via properties file.
> 
> On the Remote Spark Driver side, I kept the defensive logic to provide a 
> default value in case the conf object doesn't contain the property.  This may 
> occur if a prop is unset. For this, I had to instantiate a HiveConf on that 
> process to get the default value, as some of the timeout props need a 
> hiveConf instance to do calculation on.
> 
> 
> Diffs
> -----
> 
>   common/src/java/org/apache/hadoop/hive/conf/HiveConf.java 068c962 
>   
> ql/src/java/org/apache/hadoop/hive/ql/exec/spark/HiveSparkClientFactory.java 
> 334c191 
>   
> spark-client/src/main/java/org/apache/hive/spark/client/SparkClientImpl.java 
> 865e03e 
>   spark-client/src/main/java/org/apache/hive/spark/client/rpc/Rpc.java 
> ac71ae9 
>   
> spark-client/src/main/java/org/apache/hive/spark/client/rpc/RpcConfiguration.java
>  5a826ba 
>   spark-client/src/test/java/org/apache/hive/spark/client/rpc/TestRpc.java 
> a2dd3e6 
> 
> Diff: https://reviews.apache.org/r/30055/diff/
> 
> 
> Testing
> -------
> 
> 
> Thanks,
> 
> Szehon Ho
> 
>

Reply via email to