Hi Klaus, There is no ideal method but some workaround. Train model in Spark cluster or YARN cluster, then use RDD.saveAsTextFile to store this model which include weights and intercept to HDFS. Load weights file and intercept file from HDFS, construct a GLM model, and then run model.predict() method, you can get what you want.
The Spark community also have some ongoing work about export model with PMML. 2014-12-10 18:32 GMT+08:00 Simon Chan <simonc...@gmail.com>: > Hi Klaus, > > PredictionIO is an open source product based on Spark MLlib for exactly > this purpose. > This is the tutorial for classification in particular: > http://docs.prediction.io/classification/quickstart/ > > You can add custom serving logics and retrieve prediction result through > REST API/SDKs at other places. > > Simon > > > On Wed, Dec 10, 2014 at 2:25 AM, Klausen Schaefersinho < > klaus.schaef...@gmail.com> wrote: > >> Hi, >> >> >> I would like to use Spark to train a model, but use the model in some >> other place,, e.g. a servelt to do some classification in real time. >> >> What is the best way to do this? Can I just copy I model file or >> something and load it in the servelt? Can anybody point me to a good >> tutorial? >> >> >> Cheers, >> >> >> Klaus >> >> >> >> -- >> “Overfitting” is not about an excessive amount of physical exercise... >> > >