not all things unfortunately map gracefully into algebra. But hopefully some of the whole can still be.
I am even a little bit worried that we may develop almost too much (is there such thing) of ML before we have a chance to cyrstallize data frames and perhaps dictionary discussions. these are more tools to keep abstracted. I just don't want Mahout to be yet-another mllib. I shudder every time somebody says "we want to create a Spark version of (an|the) algorithm". I know it will be creating wrong talking points for somebody anxious to draw parallels. On Sun, Apr 13, 2014 at 10:51 PM, Sebastian Schelter <s...@apache.org> wrote: > Andy, that would be awesome. Have you had a look at our new scala DSL [1]? > Does it offer enough constructs for you to rewrite your implementation with > it? > > --sebastian > > > [1] https://mahout.apache.org/users/sparkbindings/home.html > > > On 04/14/2014 07:47 AM, Andy Twigg wrote: > >> +1 to removing present Random Forests. Andy Twigg had provided a >>> Spark >>> based Streaming Random Forests impl sometime last year. Its time to >>> restart >>> that conversation and integrate that into the codebase if the contributor >>> is still willing i.e. >>> >> >> I'm happy to contribute this, but as it stands it's written against >> spark, even forgetting the 'streaming' aspect. Do you have any advice >> on how to proceed? >> >> >