Mind if I ask what 1.3/1.4 ML features that you are looking for?

On Saturday, June 20, 2015, Debasish Das <debasish.da...@gmail.com> wrote:

> After getting used to Scala, writing Java is too much work :-)
>
> I am looking for scala based project that's using netty at its core (spray
> is one example).
>
> prediction.io is an option but that also looks quite complicated and not
> using all the ML features that got added in 1.3/1.4
>
> Velox built on top of ML / Keystone ML pipeline API and that's useful but
> it is still using javax servlets which is not netty based.
>
> On Sat, Jun 20, 2015 at 10:25 AM, Sandy Ryza <sandy.r...@cloudera.com
> <javascript:_e(%7B%7D,'cvml','sandy.r...@cloudera.com');>> wrote:
>
>> Oops, that link was for Oryx 1. Here's the repo for Oryx 2:
>> https://github.com/OryxProject/oryx
>>
>> On Sat, Jun 20, 2015 at 10:20 AM, Sandy Ryza <sandy.r...@cloudera.com
>> <javascript:_e(%7B%7D,'cvml','sandy.r...@cloudera.com');>> wrote:
>>
>>> Hi Debasish,
>>>
>>> The Oryx project (https://github.com/cloudera/oryx), which is Apache 2
>>> licensed, contains a model server that can serve models built with MLlib.
>>>
>>> -Sandy
>>>
>>> On Sat, Jun 20, 2015 at 8:00 AM, Charles Earl <charles.ce...@gmail.com
>>> <javascript:_e(%7B%7D,'cvml','charles.ce...@gmail.com');>> wrote:
>>>
>>>> Is velox NOT open source?
>>>>
>>>>
>>>> On Saturday, June 20, 2015, Debasish Das <debasish.da...@gmail.com
>>>> <javascript:_e(%7B%7D,'cvml','debasish.da...@gmail.com');>> wrote:
>>>>
>>>>> Hi,
>>>>>
>>>>> The demo of end-to-end ML pipeline including the model server
>>>>> component at Spark Summit was really cool.
>>>>>
>>>>> I was wondering if the Model Server component is based upon Velox or
>>>>> it uses a completely different architecture.
>>>>>
>>>>> https://github.com/amplab/velox-modelserver
>>>>>
>>>>> We are looking for an open source version of model server to build
>>>>> upon.
>>>>>
>>>>> Thanks.
>>>>> Deb
>>>>>
>>>>>
>>>>
>>>> --
>>>> - Charles
>>>>
>>>
>>>
>>
>

-- 
Donald Szeto
PredictionIO

Reply via email to