Hi All,

We are pulling the data from oracle tables and writing them using partitions as 
parquet files, this is daily process it works fine till 18th day (18 days load 
works fine), however on 19 th day load the data frame load process hangs and 
load action called more than once, if we remove the data and just run for 19th 
day it loads the data successfully, why the load fails for 19th day in APPEND 
mode where as the 19th day works fine. On Spark UI we can see first load job 
takes around 5 min and duplicate load jobs just takes few seconds, we are stuck 
with this we want to process 60 days of data.

∂
Thank you 
Anil Langote


> On Apr 20, 2016, at 1:12 PM, Wei Chen <wei.chen.ri...@gmail.com> wrote:
> 
> Found it. In case someone else if looking for this:
> cvModel.bestModel.asInstanceOf[org.apache.spark.ml.classification.LogisticRegressionModel].weights
> 
> On Tue, Apr 19, 2016 at 1:12 PM, Wei Chen <wei.chen.ri...@gmail.com 
> <mailto:wei.chen.ri...@gmail.com>> wrote:
> Hi All,
> 
> I am using the example of model selection via cross-validation from the 
> documentation here: http://spark.apache.org/docs/latest/ml-guide.html 
> <http://spark.apache.org/docs/latest/ml-guide.html>. After I get the 
> "cvModel", I would like to see the weights for each feature for the best 
> logistic regression model. I've been looking at the methods and attributes of 
> this "cvModel" and "cvModel.bestModel" and still can't figure out where these 
> weights are referred. It must be somewhere since we can use "cvModel" to 
> transform a new dataframe. Your help is much appreciated.
> 
> 
> Thank you,
> Wei
> 
> 
> 
> -- 
> Wei Chen, Ph.D.
> Astronomer and Data Scientist
> Phone: (832)646-7124
> Email: wei.chen.ri...@gmail.com <mailto:wei.chen.ri...@gmail.com>
> LinkedIn: https://www.linkedin.com/in/weichen1984 
> <https://www.linkedin.com/in/weichen1984>

Reply via email to