[ 
https://issues.apache.org/jira/browse/SPARK-4766?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Joseph K. Bradley updated SPARK-4766:
-------------------------------------
    Description: 
Currently, in spark.ml, both Transformers and Estimators extend the same Params 
classes.  There should be one Params class for the Transformer and one for the 
Estimator, where the Estimator params class extends the Transformer one.

E.g., it is weird to be able to do:
{code}
val model: LogisticRegressionModel = ...
model.getMaxIter()
{code}

(This is the only case where this happens currently, but it is worth setting a 
precedent.)

  was:
Currently, in spark.ml, both Transformers and Estimators extend the same Params 
classes.  There should be one Params class for the Transformer and one for the 
Estimator, where the Estimator params class extends the Transformer one.

E.g., it is weird to be able to do:
{code}
val model: LogisticRegressionModel = ...
model.getMaxIter()
{code}



> ML Estimator Params should subclass Transformer Params
> ------------------------------------------------------
>
>                 Key: SPARK-4766
>                 URL: https://issues.apache.org/jira/browse/SPARK-4766
>             Project: Spark
>          Issue Type: Improvement
>          Components: ML
>    Affects Versions: 1.2.0
>            Reporter: Joseph K. Bradley
>
> Currently, in spark.ml, both Transformers and Estimators extend the same 
> Params classes.  There should be one Params class for the Transformer and one 
> for the Estimator, where the Estimator params class extends the Transformer 
> one.
> E.g., it is weird to be able to do:
> {code}
> val model: LogisticRegressionModel = ...
> model.getMaxIter()
> {code}
> (This is the only case where this happens currently, but it is worth setting 
> a precedent.)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to