On Wed, Nov 01, 2006 at 05:03:43PM -0500, Scott Adkins wrote:
> >And that's precisely what I'm trying to find out: since there are people
> >that have succeeded in running Cyrus in an enviroment that's essentially
> >the same as mine, I'd like to know how they have found their way around
> >the DB
I have to agree here, the reason we like the cluster filesystem method
is its shear simplicity. It is nothing to add extra nodes to split the
load across more servers. We don't have to partition users, we don't
have to worry about losing part of the user community when a server goes
down, only o
--On Monday, October 30, 2006 10:38 AM +0200 Janne Peltonen <[EMAIL PROTECTED]>
wrote:
IIRC there are people running
Cyrus servers that way on other systems like Tru64 or Veritas cluster.
And that's precisely what I'm trying to find out: since there are people
that have succeeded in running C
Hi!
Thanks for the good points about the benefits of a Murder setup.
On Mon, Oct 30, 2006 at 09:47:07AM -0500, Robert Banz wrote:
> On the other hand, there's some benefit to a "murder" style cluster
> as well. You've got multiple isolated systems, limiting your risk if
> "something horrible
On Mon, Oct 30, 2006 at 07:46:40AM -0900, Matthew Schumacher wrote:
> > --clip--
> > Oct 30 09:21:19 lcluster2 imap[10378]: login: localhost.localdomain
> > [127.0.0.1] cyrus plaintext User logged in
> > Oct 30 09:22:21 lcluster2 imap[10378]: DBERROR db4: PANIC: fatal region
> > error detected; run
Janne Peltonen wrote:
> But I still seem to get some weird DB errors, the same I used to: if I
> log in and out on the node on which Cyrus was started first, the imapd
> process that accepted my connection complains about DBERROR on exit:
>
> --clip--
> Oct 30 09:21:19 lcluster2 imap[10378]: logi
Michael Menge wrote:
Quoting Dave McMurtrie <[EMAIL PROTECTED]>:
I believe you'll also need to make some minor code changes. When
University of Pittsburgh implemented their Cyrus cluster, they added a
nodename config option and then used that nodename as a filename
component along with the p
Janne Peltonen wrote:
If you decide not to pursue a cluster solution, Perdition would probably
help you with this part.
Ok. Is there any gain in using Perdition instead of Murder? Is it more
stable? Less complicated? More widely used? Better suited to the system
of our size (why)?
I w
Hi,
Janne Peltonen wrote:
> People /are/ running Cyrus setups on different *nixen with different
> clustered filesystems. But those cases don't seem to be documented on
> the Web, either. That's why I'm asking here. ;)
>
If I read Ken's (historic) post correctly then the memory-mapping on the
On Mon, Oct 30, 2006 at 11:09:27AM +0200, Janne Peltonen wrote:
> solution. When my predecessor made his decision on what solution to
> pursue, he considered Murder much too complicated, and wasn't sure
> whether 2.3 was mature enough (that is, replication and friends), but
> this all might have ch
Hi,
Quoting Dave McMurtrie <[EMAIL PROTECTED]>:
I believe you'll also need to make some minor code changes. When
University of Pittsburgh implemented their Cyrus cluster, they added a
nodename config option and then used that nodename as a filename
component along with the pid for the lmtp tem
Hi, and thanks for the answer.
On Sun, Oct 29, 2006 at 09:11:53PM -0500, Wesley Craig wrote:
> On 27 Oct 2006, at 03:35, Janne Peltonen wrote:
> >Or should I just give up and start considering Murder?
> Before you decide on whether to give up on clustering, you should
> thoroughly consider Murde
On Sat, Oct 28, 2006 at 06:01:56PM -0700, Adam Kramer wrote:
> Has anyone documented running a high volume Cyrus setup on Linux with
> a clustered filesystem?
Having googled around for such a document for the last three weeks, I
can say with some conviction that if anyone has, the document isn't o
Hi,
thanks for the answers!
On Fri, Oct 27, 2006 at 07:22:43AM -0400, Dave McMurtrie wrote:
> >And if further splitting of users on more servers is needed - downtime
> >again. Moreover, it's confusing for the users to have to determine their
> >correct imap server name - we haven't really had tro
Hi, and thanks for clearing things up!
On Sat, Oct 28, 2006 at 10:19:53AM +0200, Simon Matter wrote:
[...]
> 2) You have to consider GFS volumes
> a local storage because it is usually on SAN which is also virtually local
> storage. It really has nothing to do with networked filesystems like NFS.
On 27 Oct 2006, at 03:35, Janne Peltonen wrote:
Or should I just give up and start considering Murder?
Before you decide on whether to give up on clustering, you should
thoroughly consider Murder. At a minimum, it provides the location
independence that you'd like. With replication, you h
On 10/28/06, Simon Matter <[EMAIL PROTECTED]> wrote:
Now I think you really mix things up. 1) AFAIK quota is a per user
database which is updated whenever there is a change to the users mailbox.
Cyrus only scans all mail for their size with you do a "quota -f" after
something messed with your mai
> Hello,
>
> maybe I have understood GFS wrong, but isn't it ment to stripe data of
> several servers instead of mirroring them but make it accessable from
> several servers? If one server goes down, then you can only access the
> metadata from the GFS metadata server, but not the file itself from
Hello,
maybe I have understood GFS wrong, but isn't it ment to stripe data of
several servers instead of mirroring them but make it accessable from
several servers? If one server goes down, then you can only access the
metadata from the GFS metadata server, but not the file itself from the
se
Janne Peltonen wrote:
And if further splitting of users on more servers is needed - downtime
again. Moreover, it's confusing for the users to have to determine their
correct imap server name - we haven't really had trouble with this, but
it would be nice if the users saw a unified system image.
Hi list.
Sorry for the long post. I hope someone has time to read it and shed
some light on my concerns. This all boils down to one question: those
that have succeeded in running active-active Cyrus cluster configs, how
have you done it?
So. Some background:
I inherited a university imap system
21 matches
Mail list logo