On Thu, Feb 21, 2013 at 11:48 AM, David LaBarbera <
davidlabarb...@localresponse.com> wrote:
> Is there a rule of thumb for determining "leveling off" of perplexity? Is
> this value controlled by the convergence delta?
>
The value of where the driver will automatically stop issuing new
iterations
Is there a rule of thumb for determining "leveling off" of perplexity? Is this
value controlled by the convergence delta?
Sorry for the table view. I reformatted it with just space.
Document Count corpus size(MB) Topic Count Perplexity
D
I really can't read your results here, the formatting of your columns is
pretty destroyed... you look like you've got results for 20 topics, as
well as for 10, with different sized corpora?
You can't compare convergence between corpora sizes - the perplexity will
vary by order of magnitude betwee
I've been running some performance test with the LDA algorithm and I'm unsure
how to gauge them. I ran 10 iterations each time and collected the perplexity
value every 2 iterations with test fraction set to 0.1. These were all run on
an AWS cluster with 10 nodes (70 mapper, 30 reducers). I'm not