Hi Lane That wasn't me I'm afraid. There was a group working on a way of batching LM requests so they could be sent to a server, but I don't the project was ever finished.
cheers - Barry On Wednesday 02 February 2011 14:14, Lane Schwartz wrote: > I recall Barry working on the LM server at the Dublin MT Marathon. What is > the current status of that? > > On Wed, Feb 2, 2011 at 3:29 AM, Miles Osborne <mi...@inf.ed.ac.uk> wrote: > > to add to Barry's excellent answer, we are currently working on a > > client-server language model. this will mean that a cluster of > > machines can be used, with a shared resource. it should also work > > with multicore > > > > but in the short-term, you are probably better off with multicore > > > > Miles > > > > On 2 February 2011 06:06, Noubours, Sandra > > > > <sandra.noubo...@fkie.fraunhofer.de> wrote: > > > Hello Barry, hello Tom, > > > > > > thank you for your answers. I think I have a better idea about > > > different > > > > approaches to MOSES efficiency issues now. > > > > > Best regards, > > > Sandra > > > > > > -----Ursprüngliche Nachricht----- > > > Von: Barry Haddow [mailto:bhad...@inf.ed.ac.uk] > > > Gesendet: Montag, 31. Januar 2011 10:52 > > > An: moses-support@mit.edu > > > Cc: Noubours, Sandra; Tom Hoar > > > Betreff: Re: [Moses-support] running moses on a cluster with sge > > > > > > Hi Sandra > > > > > > The short answer is that it really depends how big your models are. > > > > Running on > > > > > a cluster helps speed up tuning because most of the time in tuning is > > > > spent > > > > > decoding, which can be easily parallelised by splitting up the file > > > into chunks. So each of the individual machines should be capable of > > > loading > > > > your > > > > > models and running a decoder. > > > > > > The problem with using a cluster (as opposed to multicore) is that each > > > machine has to have its own ram, and if you want to load large models > > > > then > > > > > you need a lot of ram. Whereas with multicore, each thread can access > > > the same model. Sure, binarising saves a lot on ram usage, but it slows > > > you > > > > down > > > > > and puts a lot of load on the filesystem which can cause problems on > > > clusters. > > > > > > Our group's machines are a mixture of 8 and 16 core Xeon 2.67GHz, with > > > > 36-72G > > > > > ram, no sge. We also have access to the university cluster, but since > > > the most ram you can get is 16G and sge hold jobs don't work at the > > > moment we don't really use it for moses any more, > > > > > > hope that helps - regards - Barry > > > > > > On Monday 31 January 2011 07:42, Noubours, Sandra wrote: > > >> Hello, > > >> > > >> > > >> > > >> thanks for the tips! When talking about using a Sun Grid Engine I was > > >> referring tuning. Making use of a cluster is supposed to speed up the > > >> tuning process (see http://www.statmt.org/moses/?n=Moses.FAQ#ntoc10). > > > > In > > > > >> this context I wondered what hardware exactly is needed for such a > > > > cluster. > > > > >> Sandra > > >> > > >> > > >> > > >> > > >> > > >> > > >> > > >> Von: Tom Hoar [mailto:tah...@precisiontranslationtools.com] > > >> Gesendet: Freitag, 28. Januar 2011 09:01 > > >> An: Noubours, Sandra > > >> Cc: moses-support@mit.edu > > >> Betreff: Re: [Moses-support] running moses on a cluster with sge > > >> > > >> > > >> > > >> Sandra, > > >> > > >> What kind of capacity do you need to support? I just finished > > > > translating > > > > >> 21,000 pages, over 1/2 million phrases, in 22 hours on an old Intel > > >> Core2Quad, 2.4 Ghz with 4 GB RAM and a 4-disk RAID-0. Moses was > > > > configured > > > > >> with binarized phrase/reordering tables and kenlm binarized language > > > > model. > > > > >> The advances in Moses supporting efficient binarized tables/models are > > >> great! > > >> > > >> We're planning tests for a 2-socket host with two Intel Xeon 5680 > > >> 6-core 3.33 Ghz CPU's, 48 GB RAM and 4 1-TB disks as RAID0. With 12 > > >> cores (totaling 24 simultaneous threads according to Intel specs), > > >> we're expecting to boot capacity to well over 15 million phrases per > > >> day on > > > > one > > > > >> host. > > >> > > >> What's the advantage of running Moses on a grid or cluster? > > >> > > >> Tom > > >> > > >> > > >> > > >> On Fri, 28 Jan 2011 08:40:22 +0100, "Noubours, Sandra" > > >> <sandra.noubo...@fkie.fraunhofer.de> wrote: > > >> > > >> Hello, > > >> > > >> > > >> > > >> I would like to run Moses on a cluster. I am yet inexperienced > > >> in > > > > using > > > > >> Sun Grid as well as clusters in common. Could you give me any > > > > instructions > > > > >> or tips for implementing a Linux-Cluster with Sun Grid Engine for > > > > running > > > > >> Moses? > > >> > > >> a) What kind of cluster would you recommend, i.e. how many > > > > machines, > > > > >> how many cpus, what memory, etc.? > > >> > > >> b) When tuning is performed with the multicore option it > > >> does > > > > not use > > > > >> more than one cpu. Does the tuning step use more than one cpu when run > > > > on a > > > > >> cluster? > > >> > > >> c) Can Sun Grid implement a cluster virtually on one > > > > computer, so > > > > >> that jobs are spread locally to different cpus of one computer? > > >> > > >> > > >> > > >> Thank you and best regards! > > >> > > >> > > >> > > >> Sandra > > > > > > -- > > > The University of Edinburgh is a charitable body, registered in > > > Scotland, with registration number SC005336. > > > > > > > > > _______________________________________________ > > > Moses-support mailing list > > > Moses-support@mit.edu > > > http://mailman.mit.edu/mailman/listinfo/moses-support > > > > -- > > The University of Edinburgh is a charitable body, registered in > > Scotland, with registration number SC005336. > > > > _______________________________________________ > > Moses-support mailing list > > Moses-support@mit.edu > > http://mailman.mit.edu/mailman/listinfo/moses-support -- The University of Edinburgh is a charitable body, registered in Scotland, with registration number SC005336. _______________________________________________ Moses-support mailing list Moses-support@mit.edu http://mailman.mit.edu/mailman/listinfo/moses-support