not all things unfortunately map gracefully into algebra. But hopefully
some of the whole can still be.

I am even a little bit worried that we may develop almost too much (is
there such thing) of ML before we have a chance to cyrstallize data frames
and perhaps dictionary discussions. these are more tools to keep abstracted.

I just don't want Mahout to be yet-another mllib. I shudder every time
somebody says "we want to create a Spark version of (an|the) algorithm".  I
know it will be creating wrong talking points for somebody anxious to draw
parallels.


On Sun, Apr 13, 2014 at 10:51 PM, Sebastian Schelter <s...@apache.org> wrote:

> Andy, that would be awesome. Have you had a look at our new scala DSL [1]?
> Does it offer enough constructs for you to rewrite your implementation with
> it?
>
> --sebastian
>
>
> [1] https://mahout.apache.org/users/sparkbindings/home.html
>
>
> On 04/14/2014 07:47 AM, Andy Twigg wrote:
>
>>       +1 to removing present Random Forests. Andy Twigg had provided a
>>> Spark
>>> based Streaming Random Forests impl sometime last year. Its time to
>>> restart
>>> that conversation and integrate that into the codebase if the contributor
>>> is still willing i.e.
>>>
>>
>> I'm happy to contribute this, but as it stands it's written against
>> spark, even forgetting the 'streaming' aspect. Do you have any advice
>> on how to proceed?
>>
>>
>

Reply via email to