Congratulations, Andrew!
- G
> On Jul 18, 2018, at 22:30, Andrew Palumbo wrote:
>
> Please join me in congratulating Andrew Musselman as the new Chair of the
> Apache Mahout Project Management Committee. I would like to thank Andrew
> for stepping up, all of us who have worked with him over
I believe -Dhadoop2 is also redundant.
mvn clean install -Dhadoop.version=2.3.0 should be sufficient
Sent from my iPhone
On Dec 16, 2014, at 12:29, Lee S sle...@gmail.com wrote:
Hi all , I have figured this out.
The command should be (mvn clean install -Dhadoop2 -Dhadoop.version=2.3.0
The problem is that seqdirectory doesn't do what you want. From the
documentation page:
The output of seqDirectory will be a Sequence file Text, Text of
all documents (/sub-directory-path/documentFileName, documentText).
Please see
Since these very names were used in the Example, I thought nothing more
was required to be done in creating sequence file.
What further steps are needed? Please suggest simple shell command.
Thanks,
Ashok Kumar Harnal
On 25 November 2014 at 14:52, Gokhan Capan gkhn...@gmail.com
Max,
Check this out: https://mahout.apache.org/developers/buildingmahout.html
Gokhan
On Fri, Nov 14, 2014 at 4:11 PM, mw m...@plista.com wrote:
Hi,
i am working on a rest api for mahout called kornakapi.
I heared that it is possible to compile the mahout trunk such that it is
compatible
for org.apache.mahout:mahout-core:jar:1.0-SNAPSHOT is
missing, no dependency information available
Was this one renamed somehow?
Best,
Max
On 11/14/2014 04:51 PM, Gokhan Capan wrote:
Hi Max,
If it is installed correctly, just adding the module you require a
dependency should work
Welcome Pat, congrats!
Gokhan
On Fri, Apr 25, 2014 at 12:08 PM, sumit sharma pro.su...@gmail.com wrote:
Congrats Pats! Like to keep seeing those revolutionary ideas in code.
On Apr 24, 2014 3:50 PM, Sebastian Schelter s...@apache.org wrote:
Hi,
this is to announce that the Project
8576 5771
M + 61 4 1463 7424
Etroung.p...@team.telstra.com
W www.telstra.com
-Original Message-
From: Gokhan Capan [mailto:gkhn...@gmail.com]
Sent: Friday, 4 April 2014 5:50 AM
To: user@mahout.apache.org
Subject: Re: Mahout v0.9 is not working with 2.2.0-cdh5.0.0-beta-1
Could you checkout the latest trunk and build the project again?
Sent from my iPhone
On Apr 2, 2014, at 9:50, Sean Owen sro...@gmail.com wrote:
Not sure what's going on there. Your build is somehow internally
inconsistent at runtime since it fails on the client side due to
mismatching
Welcome and congratulations, Andrew!
Sent from my iPhone
On Mar 7, 2014, at 20:20, Scott C. Cote scottcc...@gmail.com wrote:
I personally am looking forward to the ³advice from the newest
³recommended² committer to hadoop.
Congratulations to Mahout team for increasing and growing :)
Now
Kevin,
From trunk, can you build mahout for hadoop2 using this command:
mvn clean package -DskipTests=true -Dhadoop2.version=YOUR_HADOOP2_VERSION
Then can you verify that you have the right hadoop jars with the following
command:
find . -name hadoop*.jar
Gokhan
On Thu, Mar 6, 2014 at
I liked both of them
Great work Lucas!
Gokhan
On Wed, Mar 5, 2014 at 2:11 PM, Sebastian Schelter s...@apache.org wrote:
Hi everyone,
In our latest discussion, I argued that the lack (and errors) of
documentation on our website is one of the main pain points of Mahout atm.
To be honest,
Sent from my iPhone
On Mar 4, 2014, at 22:13, Sebastian Schelter s...@apache.org wrote:
I think we should introduce a new parameter for the recommend() method
in the Recommender interface that tells whether already known items
should be recommended or not.
+1 for that
What do you think?
Margusja,
From trunk, can you build mahout using the following command and try again:
mvn clean package -DskipTests=true -Dhadoop2.version=2.2.0
Best
Gokhan
On Tue, Mar 4, 2014 at 4:25 PM, Margusja mar...@roo.ee wrote:
Hi thanks for reply.
Here is my output:
[hduser@vm38 ~]$
Quick note, you don't need to apply a patch for building mahout for hadoop
2.2.0, it is committed to trunk.
Just get the latest mahout source, build it with:
mvn clean package -DskipTests=true -Dhadoop2.version=2.2.0
And please see MAHOUT-1310https://issues.apache.org/jira/browse/MAHOUT-1310for
I'm personally positive on this.
Could you give an example code snippet that shows how the usage is going to be?
Sent from my iPhone
On Feb 22, 2014, at 5:37, Jay Vyas jayunit...@gmail.com wrote:
Hi dead. Sure I will take a look.
On Fri, Feb 21, 2014 at 7:51 PM, Ted Dunning
If you have a chance to build mahout from source, could you try if that
works when the patch in MAHOUT-1329 applied? mvn packaging mahout with
-DskipTests=true is pretty fast
Gokhan
On Thu, Feb 20, 2014 at 11:34 PM, Suneel Marthi suneel_mar...@yahoo.comwrote:
On Thursday, February 20,
I imagine in Mahout offering an option to the users to select from
different execution engines (just like we currently do by giving M/R or
sequential options), and starting from Spark. I am not sure what changes
needed in the codebase, though. Maybe following MLI (or alike) and
implementing some
I think ClusterDumper can also export cluster results in GraphML format,
which can be used to visualize clustered points in a graph visualization
software such as Gephi.
Gokhan
On Fri, Dec 13, 2013 at 10:44 AM, David G davidgr...@gmail.com wrote:
Hi,
I see what you're saying. For more than 3
Hi Vishnu,
You may reduce the split size by setting mapred.max.split.size
configuration parameter of hadoop.
Number of map tasks then will be equal to number of splits (input
size/split size)
Best
Sent from my iPhone
On Dec 13, 2013, at 21:08, Vishnu Modi vishnu.modi...@gmail.com wrote:
I
and there is no
mapreduce API left? As far as I know, the mapreduce API needs to be
recompiled and I remember needing to recompile Mahout for CDH4 when it first
came out.
Thanks, Zoltan
On 12/10/13 10:02 PM, Gokhan Capan wrote:
I meant that you shouldn't need to modify mahout's dependencies, just mvn
, Gokhan Capan gkhn...@gmail.com
wrote:
Mahout actually should work with hadoop-2 stable without recompiling,
not with hadoop-2 alpha though.
We're, by the way, currently in the process of adding support to build
mahout with hadoop-2.
Please see mahout-1354 for the relevant issue
In the meantime, you might apply the patch in MAHOUT-1354, build mahout
using mvn package -Phadoop2 -DskipTests=true, use that mahout version and
see if that works
Gokhan
On Wed, Dec 11, 2013 at 10:09 PM, Gokhan Capan gkhn...@gmail.com wrote:
I apologize, Suneel is right, Counter breaks
the following version of hadoop:
dependency
groupIdorg.apache.hadoop/groupId
artifactIdhadoop-client/artifactId
version2.2.0/version
/dependency
That's not alpha, right?
Thanks!
On Monday, December 9, 2013 10:05 AM, Gokhan Capan gkhn...@gmail.com wrote:
Mahout
Mahout actually should work with hadoop-2 stable without recompiling,
not with hadoop-2 alpha though.
We're, by the way, currently in the process of adding support to build
mahout with hadoop-2.
Please see mahout-1354 for the relevant issue
Sent from my iPhone
On Dec 9, 2013, at 19:54, Hi
Gokhan
On Thu, Nov 28, 2013 at 3:18 AM, Ted Dunning ted.dunn...@gmail.com wrote:
On Wed, Nov 27, 2013 at 7:07 AM, Vishal Santoshi
vishal.santo...@gmail.com
Are we to assume that SGD is still a work in progress and
implementations (
Cross Fold, Online, Adaptive ) are too flawed to
the most recent user
intent.
Unfortunately we haven’t A/B tested this conclusion but the candidate for
best recommender was using most recent prefs in H and all prefs in B.
On Nov 7, 2013, at 11:36 PM, Gokhan Capan gkhn...@gmail.com wrote:
On Fri, Nov 8, 2013 at 6:24 AM, Ted Dunning
Cassio,
I am not sure if there are direct/indirect ways to to this with existing
code.
Recall that an item neighborhood based score prediction, in simplest terms,
is a weighted average of the active user's ratings on other items, where
the weights are item-to-item similarities. Applying a decay
On Fri, Nov 8, 2013 at 6:24 AM, Ted Dunning ted.dunn...@gmail.com wrote:
On Thu, Nov 7, 2013 at 12:50 AM, Gokhan Capan gkhn...@gmail.com wrote:
This particular approach is discussed, and proven to increase the
accuracy
in Collaborative filtering with Temporal Dynamics by Yehuda Koren
Cassio,
I would implement a CandidateItemsStrategy that returns products that are
available now. A neighborhood based recommender would iterate over those
products, and rank them based on the similarity measure you provide.
If the DataModel of your recommender does not contain most of your
Hi Shengjie,
I believe this means the list of the prior clusters has a zero length.
Could you check if the output of the iteration prior to the failed one is
written to the path you provided properly?
Best
Gokhan
On Fri, Nov 1, 2013 at 8:15 AM, Shengjie Min shengjie@gmail.com wrote:
)
at
org.apache.mahout.mahoutnewsrecommender2.Recommender.myRecommender(Recommender.java:99)
at org.apache.mahout.mahoutnewsrecommender2.App.main(App.java:26)
BUILD SUCCESSFUL (total time: 3 seconds)
Thanks,
Darius
2013/9/12 Gokhan Capan gkhn...@gmail.com
Although Windows is not officially supported, your
reut2-017.sgm
reut2-021.sgm
all-places-strings.lc.txt lewis.dtd
reut2-002.sgm reut2-006.sgm reut2-010.sgm reut2-014.sgm reut2-018.sgm
Anyway, I do not get any clustering. So, where is the problem?
Best,
Darius
2013/9/13 Gokhan Capan gkhn...@gmail.com
You need to have 'curl
You need to have 'curl' installed, as the error message tells.
Gokhan
On Fri, Sep 13, 2013 at 2:37 PM, Darius Miliauskas
dariui.miliaus...@gmail.com wrote:
Dear All,
I tried to run Reuters Example on my Windows machine (Windows 7), using
Cygwin, but got the following error:
Although Windows is not officially supported, your
svsf.run(new String[]{inputPath.toString(), outputPath.toString()})
should be
svsf.run(new String[]{-i,inputPath.toString(), -o,
outputPath.toString()}) anyway.
Best
Gokhan
On Thu, Sep 12, 2013 at 4:14 PM, Darius Miliauskas
Hi Parnab,
When running lda using commandline cvb utility, you may pass -o option for
the output path for topic-term distributions, and -dt option for the output
path for doc-topic distributions.
Hope that helps.
Best
Gokhan
On Wed, Sep 11, 2013 at 11:38 PM, parnab kumar
(reader.next(key, value)) {
//System.out.println(key.toString()+ belongs to cluster
+value.toString());
//}
//reader.close();
}
}
On Wed, Sep 4, 2013 at 3:29 PM, Gokhan Capan gkhn...@gmail.com wrote:
Taner,
A few questions:
Is there a specific reason
will send the command to
compare command line parameters to TFIDFConverter params.
Suneel, I had checked the seqfiles. I didn't see any problem other
generated seqfiles but I will checked and send samples from each
seqfiles.
On Sun, Sep 1, 2013 at 11:02 PM, Gokhan Capan gkhn...@gmail.com
Taner,
Could you try reducing minLLR value? (It is not a normalized measure, but
its default value is 1.0)
Best,
Gokhan
On Sun, Sep 1, 2013 at 9:24 AM, Taner Diler taner.di...@gmail.com wrote:
Hi all,
I try to run Reuters KMeans example in Java, but TFIDFComverter generates
tfidf-vectors
Suneel is right indeed. I assumed that everything performed prior to vector
generation is done correctly.
By the way, if the suggestions do not work, could you try running
seq2sparse from commandline with the same arguments and see if that works
well?
On Sun, Sep 1, 2013 at 7:23 PM, Suneel
Hi Mario,
Could you create a JIRA ticket for that, and submit your diff as a patch if
possible?
http://issues.apache.org/jira/browse/MAHOUT
Best,
Gokhan
On Sat, Aug 31, 2013 at 8:56 PM, Mario Rodriguez mario.rod...@gmail.comwrote:
Hi everyone,
It seems MAHOUT_OPTS is not getting picked up
Hi Michael,
Those are collaborative filtering examples, which would recommend a news
article i, to a user u, based on:
- A weighted average of other users' ratings on i (where weight is the
similarity of two users' rating histories)
- A weighted average of u's ratings on other items (where weight
Hi Darius,
There is no module called mahout-utils in mahout 0.8. The documentation you
referred to is prepared for mahout 0.4, it is outdated.
So if you want to use mahout 0.8, remove that dependency, mahout-core would
be sufficient.
But that wouldn't compile, since SlopeOneRecommender doesn't
A little digression: Might a Matrix implementation backed by a Solr index
and uses SolrJ for querying help at all for the Solr recommendation
approach?
It supports multiple fields of String, Text, or boolean flags.
Best
Gokhan
On Wed, Aug 7, 2013 at 9:42 PM, Pat Ferrel pat.fer...@gmail.com
Just to make sure if I understood correctly, Ted, could you please correct
me?:)
1. Using a search engine, I will treat items as documents, where each
document vector consists of other items (similar to words of documents)
with co-occurrence (LLR) weights (instead of tf-idf in a search engine
Congratulations, Sebastian!
Gokhan
On Sun, Jul 21, 2013 at 10:21 PM, Suneel Marthi suneel_mar...@yahoo.comwrote:
Congrats again Sebastian.
Sent from my iPhone
On Jul 21, 2013, at 2:22 PM, Sebastian Schelter s...@apache.org wrote:
I'm happy to anounce that a paper called Distributed
:01 PM, Robin Anil robin.a...@gmail.com wrote:
Yes! Yes! Go for it!.
On Apr 17, 2013 5:52 PM, Gokhan Capan gkhn...@gmail.com wrote:
I didn't quite get that, and assuming you tell me to implement it
Thanks
On Thu, Apr 18, 2013 at 1:44 AM, Robin Anil robin.a...@gmail.com
sparse matrices where the data existed.
Worked a charm on the app in question.
It doesn't appear that I committed it.
On Thu, Apr 18, 2013 at 7:02 AM, Gokhan Capan gkhn...@gmail.com wrote:
Ok, I'll do that.
Thank you, Jake.
On Thu, Apr 18, 2013 at 4:43 AM, Jake Mannix jake.man
Hi,
Using Mahout Matrix interface I want to represent some data where the row
vector is dense iff an instance is associated to this row index, empty
otherwise. The max possible index for rows (a.k.a. rowSize) is defined.
I never query the matrix by column.
I want to be able to add rows if the
Robin,
Aren't SparseRowMatrix rows are sparse vectors? In my use case row vectors
don't need to be sparse, they are either full or empty.
On Thu, Apr 18, 2013 at 1:32 AM, Robin Anil robin.a...@gmail.com wrote:
SparseRowMatrix?
On Apr 17, 2013 5:26 PM, Gokhan Capan gkhn...@gmail.com wrote
I didn't quite get that, and assuming you tell me to implement it
Thanks
On Thu, Apr 18, 2013 at 1:44 AM, Robin Anil robin.a...@gmail.com wrote:
Make one?
On Apr 17, 2013 5:37 PM, Gokhan Capan gkhn...@gmail.com wrote:
Robin,
Aren't SparseRowMatrix rows are sparse vectors? In my use
, having a big blob of stuff is probably fine.
On Sun, Sep 9, 2012 at 7:01 AM, Gokhan Capan gkhn...@gmail.com wrote:
On Fri, Sep 7, 2012 at 12:48 AM, Ted Dunning ted.dunn...@gmail.com
wrote:
This sounds pretty exciting. Beyond that, it is hard to say much.
Can you say a bit more
the library into
mahout. I believe in a week I will be able to add documentation and mahout
jobs for experiments and start submitting patches to JIRA.
On Thu, Sep 6, 2012 at 9:14 AM, Gokhan Capan gkhn...@gmail.com wrote:
By the way, I want to mention that my thesis is advised by Ozgur
Yilmazel,
who
Dear Mahout community,
I would like to introduce a set of tools for recommender systems those are
implemented as a part of my MSc. thesis. This is inspired by our
conversations in the user-list, and I tried to stick it to existing Taste
framework for possible contribution to Mahout.
The library
By the way, I want to mention that my thesis is advised by Ozgur Yilmazel,
who is a founding member of the Mahout project. I conducted this study and
kept the implementation integrable to Mahout with his guidance.
On Thu, Sep 6, 2012 at 6:04 PM, Gokhan Capan gkhn...@gmail.com wrote:
Dear Mahout
Hi Jake,
Today I submitted the diff. It is available at
https://issues.apache.org/jira/browse/MAHOUT-1051
Thanks for the advices
On Tue, Aug 7, 2012 at 1:06 AM, Jake Mannix jake.man...@gmail.com wrote:
Sounds great Gokhan!
On Mon, Aug 6, 2012 at 2:53 PM, Gokhan Capan gkhn...@gmail.com wrote
be
very nice! :)
On Mon, Aug 6, 2012 at 4:00 AM, Gokhan Capan gkhn...@gmail.com wrote:
Hi,
My question is about interpreting lda document-topics output.
I am using trunk.
I have a directory of documents, each of which are named by integers, and
there is no sub-directory
57 matches
Mail list logo