Re: live website mirroring
On Friday 26 September 2003 01:14, Sagi Bashari wrote: If your application is writen as a Java web-app then you can take advantage of clustered Servlet engines, which enable sharing of sessions in a cluster, just for this kind of situations (in addition to the scalability advantage). Does PHP have such stuff? Thats not a problem. PHP uses plain files to store sessions data, so all I need to do is to set the sessions directory to a location that is shared by both servers. Or you can store it in the same database as the rest of the site's data. -- Oded = To unsubscribe, send mail to [EMAIL PROTECTED] with the word unsubscribe in the message body, e.g., run the command echo unsubscribe | mail [EMAIL PROTECTED]
Re: live website mirroring
On Fri, Sep 26, 2003 at 12:21:12AM +0200, Sagi Bashari wrote: So going back to my original question: is there a simple way to synchronize a directory between two linux servers, like rsync does -- but in real time? A directory also has its metadata, which will be a lot harder to synchronize in run time. If you're willing to go lower than the file system, to the block device, you can take a look at DRBD (distributed replicated block device, http://www.complang.tuwien.ac.at/reisner/drbd/), which can synchronize a local and remove block device at run time. Coupled with the linux-ha.org software, you can have automatic failover, which sounds pretty much like what you want. I haven't used either, but I've been giving them some serious thought and reading the code, and both look quite reasonable to me. Cheers, Muli -- Muli Ben-Yehuda http://www.mulix.org signature.asc Description: Digital signature
Re: live website mirroring
Hi, Heartbeat can give you solutions for that.. The only real problem is indeed the storage which you want to be synchronized. High Availability clusters should provide a way for having shared storage (i.e. scsi disk / JBOD connected to two servers), while the main node is down, the other will mount the disk. (Note that there's a big risk if two nodes mount the same storage. The cluster should be connected with at least two internal cables so it could monitor each node even when main network is down) Anyway you'd better read some in www.linux-ha.org . Also note that RedHat has its own HA cluster solution (pirhana I think) which comes with their RH Advanced Server, which I have no experience with. - Oren On Thu, 2003-09-25 at 02:42, Sagi Bashari wrote: Hello We are running a website that is written in PHP and is using a MySQL database. It runs on Linux, of course. We're trying to find a method to setup a complete mirror of the server on a standby server that we can bring up easily if the main server goes down. The website is very dynamic and data changes all the time. I'm looking for a way to keep the standby server in identical state as the master server. After doing some research I found out that I could use the built-in MySQL replication for that, but we also have dynamic data directories on our server - I could setup an rsync script that runs every few minutes, but it'll probably take too many resources and won't be accurate. I would like to hear the experiences of others who had the same problem (and hopefully solved it). * Both servers are running at the same location right now, so I can takeover the main IP address easily. However, we considered having a backup server on another location (where we can't use the same ip address). Is there any way to do this? Sagi = To unsubscribe, send mail to [EMAIL PROTECTED] with the word unsubscribe in the message body, e.g., run the command echo unsubscribe | mail [EMAIL PROTECTED] = To unsubscribe, send mail to [EMAIL PROTECTED] with the word unsubscribe in the message body, e.g., run the command echo unsubscribe | mail [EMAIL PROTECTED]
Re: live website mirroring
Dan Fruehauf wrote: On Thursday 25 September 2003 02:42, Sagi Bashari wrote: Hello, I have a sollution which confronts the problem in a totally different way. Given the time and resources, what you can do is connect both servers to a NAS switch, mapping them the same storage. It's a possible solution, but: 1. The NAS is still a single point of failure. 2. They are talking about having a server at a different location, so they'll need to replicate the NAS as well. My suggestion for a general direction to investigate: 1. Check rsync more closely - as far as I remember from using it, it is actually a very efficient program in CPU, memory, and network usage. 2. What about using a content-management (CM) solution which manages your site instead of handling files manualy? Keeping track of your files through a CM software will enable it to copy over every file when it's updated, and it can also sort-of track changes and versions. I can't recall a particular software for this, maybe you can start by looking for WebDAV implementations (e.g. Apache's WebDAV module). --Amos = To unsubscribe, send mail to [EMAIL PROTECTED] with the word unsubscribe in the message body, e.g., run the command echo unsubscribe | mail [EMAIL PROTECTED]
Re: live website mirroring
On 25/09/2003 10:43, [EMAIL PROTECTED] wrote: I have a sollution which confronts the problem in a totally different way. Given the time and resources, what you can do is connect both servers to a NAS switch, mapping them the same storage. It's a possible solution, but: 1. The NAS is still a single point of failure. 2. They are talking about having a server at a different location, so they'll need to replicate the NAS as well. We would like to do this with our current hardware at this stage, we already have RAID1 on both servers so I would just like to keep the arrays identical. Right now we will stick to having both servers at the same location, my question was about changing the IP address dynamically if something happens if we ever decide to setup a backup server on another location. My suggestion for a general direction to investigate: 1. Check rsync more closely - as far as I remember from using it, it is actually a very efficient program in CPU, memory, and network usage. Yes, but I'm not sure that running it every 10 minutes is a good idea, especially if you have many files and some of them are being changed while rsync is running. I found out about a kernel module named enbd which is supposed to do RAID over network somehow, but I would like to hear about the experience of others before I start to play with it. 2. What about using a content-management (CM) solution which manages your site instead of handling files manualy? Keeping track of your files through a CM software will enable it to copy over every file when it's updated, and it can also sort-of track changes and versions. I can't recall a particular software for this, maybe you can start by looking for WebDAV implementations (e.g. Apache's WebDAV module). I'm not sure. Basically we're storing all the data in our database and only use the filesystem for files - image files, flash files, website root directories etc. All the files must be accessible by apache (it is serving the images directly, we don't read them in our application) . Sagi = To unsubscribe, send mail to [EMAIL PROTECTED] with the word unsubscribe in the message body, e.g., run the command echo unsubscribe | mail [EMAIL PROTECTED]
Re: live website mirroring
Sagi Bashari wrote: It's a possible solution, but: 1. The NAS is still a single point of failure. 2. They are talking about having a server at a different location, so they'll need to replicate the NAS as well. We would like to do this with our current hardware at this stage, we already have RAID1 on both servers so I would just like to keep the arrays identical. As far as I understood it, the meaning of NAS is that you'll have common storage accessed over the network (NAS = Network Attached Storage). What does this have to do with each server having its own RAID? BTW, If you already have RAID 1 (mirroring) and if you have more than one disk on each mirror side (i.e. at least 4 disks total in each mirror) then you better do RAID 0+1 (mirroring over stripping), to enhance your access times. (e.g. http://www.acnc.com/04_01_0p1.html) Right now we will stick to having both servers at the same location, my question was about changing the IP address dynamically if something happens if we ever decide to setup a backup server on another location. You had two questions, haven't you? As for this IP question: 1. If you want the backup server to serve also at regular times then just advertise both addresses in A records in DNS, the DNS server will alternate their order when resolving the server's address which will be a very cheap way to load-balance between them. 2. If you just want the bakcup server to be accessed only when the primary one goes down then the only way I can think off right now is with having setting the A record with a very short time to live (5-15) minutes, as much as you can stand with having an inaccessible server) so your DNS will be accessed almost every time the site is accessed, when the primary goes down (which can be discovered with various High Availability solutions) the secondary updates your DNS server and takes over control. See http://linux-ha.org/ for instance. (1) and (2) can be implemented together (advertise both, when one server goes down the other will delete its address from DNS, when the downed server goes back up it will register with the DNS). There are also mailing lists, forums and sites about Linux high availability, maybe you'll find there more knowledgeable crowde, though I for one would be curious to hear what you decided to do. My suggestion for a general direction to investigate: 1. Check rsync more closely - as far as I remember from using it, it is actually a very efficient program in CPU, memory, and network usage. Yes, but I'm not sure that running it every 10 minutes is a good idea, especially if you have many files and some of them are being changed while rsync is running. Have you conducted your own tests or found results of such tests? In general, rsync is supposed to cope well with changing files. I found out about a kernel module named enbd which is supposed to do RAID over network somehow, but I would like to hear about the experience of others before I start to play with it. Never heard of it, though I did see the Network Block Device option in the Kernel 2.4 config, which sounds related to this. I'd expect you can google for a community of this module. 2. What about using a content-management (CM) solution which manages your site instead of handling files manualy? Keeping track of your files through a CM software will enable it to copy over every file when it's updated, and it can also sort-of track changes and versions. I can't recall a particular software for this, maybe you can start by looking for WebDAV implementations (e.g. Apache's WebDAV module). I'm not sure. Basically we're storing all the data in our database and only use the filesystem for files - image files, flash files, website root directories etc. That's what content management is about. All the files must be accessible by apache (it is serving the images directly, we don't read them in our application) . What I was suggesting (never tried this myself) is that instead of just writing files directly to the disk and hope that some sync program will find it and copy it over, use a tool which writes it to the disk (so it's accessible by Apache) but in the same time can send it (or notify about the change) to the backup server. WebDAV is one standard protocol to achive the write a file to a Web server part. Also when you have such tools instead of directly writing files to disk, you can start using other tools on top of them (QA, testting, versioning, authentication, authorization, general site management) --Amos = To unsubscribe, send mail to [EMAIL PROTECTED] with the word unsubscribe in the message body, e.g., run the command echo unsubscribe | mail [EMAIL PROTECTED]
Re: live website mirroring
On Thursday 25 September 2003 11:45, [EMAIL PROTECTED] wrote: it was probably my mistake, i said NAS, but i meant SAN (Storage Area Network). I can see people are showing interest, so i'll give some more details on what were doing at work. Assuming one server is always active, and the other one is passive, we map the same disks drives to them through the SAN (while only one server mounts the storage at a given time). when a failover happens, the passive server steals the ip of the active server, and it's storage as well and mounts it. after this, the application is being executed and all should be fine. what's not too good is that sessions are lost, so if youre application is stateful, your users will feel the crash that happened, while if it's stateless the damage will be much smaller. Dan. Sagi Bashari wrote: It's a possible solution, but: 1. The NAS is still a single point of failure. 2. They are talking about having a server at a different location, so they'll need to replicate the NAS as well. We would like to do this with our current hardware at this stage, we already have RAID1 on both servers so I would just like to keep the arrays identical. As far as I understood it, the meaning of NAS is that you'll have common storage accessed over the network (NAS = Network Attached Storage). What does this have to do with each server having its own RAID? BTW, If you already have RAID 1 (mirroring) and if you have more than one disk on each mirror side (i.e. at least 4 disks total in each mirror) then you better do RAID 0+1 (mirroring over stripping), to enhance your access times. (e.g. http://www.acnc.com/04_01_0p1.html) Right now we will stick to having both servers at the same location, my question was about changing the IP address dynamically if something happens if we ever decide to setup a backup server on another location. You had two questions, haven't you? As for this IP question: 1. If you want the backup server to serve also at regular times then just advertise both addresses in A records in DNS, the DNS server will alternate their order when resolving the server's address which will be a very cheap way to load-balance between them. 2. If you just want the bakcup server to be accessed only when the primary one goes down then the only way I can think off right now is with having setting the A record with a very short time to live (5-15) minutes, as much as you can stand with having an inaccessible server) so your DNS will be accessed almost every time the site is accessed, when the primary goes down (which can be discovered with various High Availability solutions) the secondary updates your DNS server and takes over control. See http://linux-ha.org/ for instance. (1) and (2) can be implemented together (advertise both, when one server goes down the other will delete its address from DNS, when the downed server goes back up it will register with the DNS). There are also mailing lists, forums and sites about Linux high availability, maybe you'll find there more knowledgeable crowde, though I for one would be curious to hear what you decided to do. My suggestion for a general direction to investigate: 1. Check rsync more closely - as far as I remember from using it, it is actually a very efficient program in CPU, memory, and network usage. Yes, but I'm not sure that running it every 10 minutes is a good idea, especially if you have many files and some of them are being changed while rsync is running. Have you conducted your own tests or found results of such tests? In general, rsync is supposed to cope well with changing files. I found out about a kernel module named enbd which is supposed to do RAID over network somehow, but I would like to hear about the experience of others before I start to play with it. Never heard of it, though I did see the Network Block Device option in the Kernel 2.4 config, which sounds related to this. I'd expect you can google for a community of this module. 2. What about using a content-management (CM) solution which manages your site instead of handling files manualy? Keeping track of your files through a CM software will enable it to copy over every file when it's updated, and it can also sort-of track changes and versions. I can't recall a particular software for this, maybe you can start by looking for WebDAV implementations (e.g. Apache's WebDAV module). I'm not sure. Basically we're storing all the data in our database and only use the filesystem for files - image files, flash files, website root directories etc. That's what content management is about. All the files must be accessible by apache (it is serving the images directly, we don't read them in our application) . What I was suggesting (never tried this myself) is that instead of just writing files directly to the disk and hope that some sync program will find it and
Re: live website mirroring
As far as I know, the cost of SAN is very high. It also gives your more GB then you need. And if you need less GB than what given - you just pay too much extra on each GB. Also, don't forget the equipment using to work with fiber optic cables (control cards, switches and so on). At this cost I think you can build a HA solution: 1. A red rubin DNS. 2. heart bit. 3. using another computer for db/files and two servers to serve the files (sure, has lower performance). Regards, Lior Kaplan [EMAIL PROTECTED] http://www.Guides.co.il Come to write at the forums: http://www.guides.co.il/forums - Original Message - From: Dan Fruehauf [EMAIL PROTECTED] To: [EMAIL PROTECTED]; [EMAIL PROTECTED] Sent: Thursday, September 25, 2003 7:39 PM Subject: Re: live website mirroring On Thursday 25 September 2003 11:45, [EMAIL PROTECTED] wrote: it was probably my mistake, i said NAS, but i meant SAN (Storage Area Network). I can see people are showing interest, so i'll give some more details on what were doing at work. Assuming one server is always active, and the other one is passive, we map the same disks drives to them through the SAN (while only one server mounts the storage at a given time). when a failover happens, the passive server steals the ip of the active server, and it's storage as well and mounts it. after this, the application is being executed and all should be fine. what's not too good is that sessions are lost, so if youre application is stateful, your users will feel the crash that happened, while if it's stateless the damage will be much smaller. Dan. Sagi Bashari wrote: It's a possible solution, but: 1. The NAS is still a single point of failure. 2. They are talking about having a server at a different location, so they'll need to replicate the NAS as well. We would like to do this with our current hardware at this stage, we already have RAID1 on both servers so I would just like to keep the arrays identical. As far as I understood it, the meaning of NAS is that you'll have common storage accessed over the network (NAS = Network Attached Storage). What does this have to do with each server having its own RAID? BTW, If you already have RAID 1 (mirroring) and if you have more than one disk on each mirror side (i.e. at least 4 disks total in each mirror) then you better do RAID 0+1 (mirroring over stripping), to enhance your access times. (e.g. http://www.acnc.com/04_01_0p1.html) Right now we will stick to having both servers at the same location, my question was about changing the IP address dynamically if something happens if we ever decide to setup a backup server on another location. You had two questions, haven't you? As for this IP question: 1. If you want the backup server to serve also at regular times then just advertise both addresses in A records in DNS, the DNS server will alternate their order when resolving the server's address which will be a very cheap way to load-balance between them. 2. If you just want the bakcup server to be accessed only when the primary one goes down then the only way I can think off right now is with having setting the A record with a very short time to live (5-15) minutes, as much as you can stand with having an inaccessible server) so your DNS will be accessed almost every time the site is accessed, when the primary goes down (which can be discovered with various High Availability solutions) the secondary updates your DNS server and takes over control. See http://linux-ha.org/ for instance. (1) and (2) can be implemented together (advertise both, when one server goes down the other will delete its address from DNS, when the downed server goes back up it will register with the DNS). There are also mailing lists, forums and sites about Linux high availability, maybe you'll find there more knowledgeable crowde, though I for one would be curious to hear what you decided to do. My suggestion for a general direction to investigate: 1. Check rsync more closely - as far as I remember from using it, it is actually a very efficient program in CPU, memory, and network usage. Yes, but I'm not sure that running it every 10 minutes is a good idea, especially if you have many files and some of them are being changed while rsync is running. Have you conducted your own tests or found results of such tests? In general, rsync is supposed to cope well with changing files. I found out about a kernel module named enbd which is supposed to do RAID over network somehow, but I would like to hear about the experience of others before I start to play with it. Never heard of it, though I did see the Network Block Device option in the Kernel 2.4 config, which sounds related to this. I'd expect you can google for a community of this module. 2. What about using a content-management (CM
Re: live website mirroring
Dan Fruehauf wrote: On Thursday 25 September 2003 11:45, [EMAIL PROTECTED] wrote: it was probably my mistake, i said NAS, but i meant SAN (Storage Area Network). But still, wouldn't that keep the NAS as a single point of failure? Or is the NAS implemented by some HA cluster of servers? Also, the original author said, as far as I understood it, that when the backup server moves to a separate location he won't be able to use the same IP (or maybe he can channel the same IP over IP?) what's not too good is that sessions are lost, so if youre application is stateful, your users will feel the crash that happened, while if it's stateless the damage will be much smaller. If your application is writen as a Java web-app then you can take advantage of clustered Servlet engines, which enable sharing of sessions in a cluster, just for this kind of situations (in addition to the scalability advantage). Does PHP have such stuff? --Amos = To unsubscribe, send mail to [EMAIL PROTECTED] with the word unsubscribe in the message body, e.g., run the command echo unsubscribe | mail [EMAIL PROTECTED]
Re: live website mirroring
Lior Kaplan wrote: As far as I know, the cost of SAN is very high. It also gives your more GB then you need. And if you need less GB than what given - you just pay too much extra on each GB. Also, don't forget the equipment using to work with fiber optic cables (control cards, switches and so on). At this cost I think you can build a HA solution: 1. A red rubin DNS. 2. heart bit. 3. using another computer for db/files and two servers to serve the files (sure, has lower performance). What you are saying is called NAS. With SAN, you can achieve true no single point of failure. If you disks are fiber connected, and you are doing RAID over them, you can have two disks (raid-1) serving two computers (not simultaniously). Viola! With NAS, you introduce a third computer, that serves the files to both servers (single point of failure), you need some way to ballance the load (round robin DNS, in your case, which suffers serious problems when trying to cope with a failure). I'm not even sure that the fiber option is so expensive (quite honestly, I don't know). In any case, it appears as if the entire industry is shifting torwards NAS (Netapps, special Emc hardware that incorporates an NFS server, etc.). This means that most IT managers agree with you, on one level or another. Shachar -- Shachar Shemesh Open Source integration consultant Home page resume - http://www.shemesh.biz/ = To unsubscribe, send mail to [EMAIL PROTECTED] with the word unsubscribe in the message body, e.g., run the command echo unsubscribe | mail [EMAIL PROTECTED]
Re: live website mirroring
On 25/09/2003 21:53, [EMAIL PROTECTED] wrote: But still, wouldn't that keep the NAS as a single point of failure? Or is the NAS implemented by some HA cluster of servers? Also, the original author said, as far as I understood it, that when the backup server moves to a separate location he won't be able to use the same IP (or maybe he can channel the same IP over IP?) For now both servers will be on the same location. I might have confused you when I asked that question, it wasn't directly connected to the original question (how to sync the data directories and database) - I just wanted to know if such thing as possible. You confirmed my original thoughts about the DNS in your previous message. what's not too good is that sessions are lost, so if youre application is stateful, your users will feel the crash that happened, while if it's stateless the damage will be much smaller. If your application is writen as a Java web-app then you can take advantage of clustered Servlet engines, which enable sharing of sessions in a cluster, just for this kind of situations (in addition to the scalability advantage). Does PHP have such stuff? Thats not a problem. PHP uses plain files to store sessions data, so all I need to do is to set the sessions directory to a location that is shared by both servers. Sagi = To unsubscribe, send mail to [EMAIL PROTECTED] with the word unsubscribe in the message body, e.g., run the command echo unsubscribe | mail [EMAIL PROTECTED]
Re: live website mirroring
On 25/09/2003 21:58, Lior Kaplan wrote: As far as I know, the cost of SAN is very high. It also gives your more GB then you need. And if you need less GB than what given - you just pay too much extra on each GB. Exactly. Those solutions are way out of our current budget - and right now the files that we store outside the database do not even reach 1GB so we won't invest in such expensive solution. 3. using another computer for db/files and two servers to serve the files (sure, has lower performance). We don't need this, our system runs perfectly on our server. My original idea was having 1 low end PC as a backup - not as an extra server (it won't be used when the main server is up). So going back to my original question: is there a simple way to synchronize a directory between two linux servers, like rsync does -- but in real time? Sagi = To unsubscribe, send mail to [EMAIL PROTECTED] with the word unsubscribe in the message body, e.g., run the command echo unsubscribe | mail [EMAIL PROTECTED]
live website mirroring
Hello We are running a website that is written in PHP and is using a MySQL database. It runs on Linux, of course. We're trying to find a method to setup a complete mirror of the server on a standby server that we can bring up easily if the main server goes down. The website is very dynamic and data changes all the time. I'm looking for a way to keep the standby server in identical state as the master server. After doing some research I found out that I could use the built-in MySQL replication for that, but we also have dynamic data directories on our server - I could setup an rsync script that runs every few minutes, but it'll probably take too many resources and won't be accurate. I would like to hear the experiences of others who had the same problem (and hopefully solved it). * Both servers are running at the same location right now, so I can takeover the main IP address easily. However, we considered having a backup server on another location (where we can't use the same ip address). Is there any way to do this? Sagi = To unsubscribe, send mail to [EMAIL PROTECTED] with the word unsubscribe in the message body, e.g., run the command echo unsubscribe | mail [EMAIL PROTECTED]
Re: live website mirroring
On Thursday 25 September 2003 02:42, Sagi Bashari wrote: Hello, I have a sollution which confronts the problem in a totally different way. Given the time and resources, what you can do is connect both servers to a NAS switch, mapping them the same storage. Then, in case of failing over, the active computer will shut itself down and will release it's NAS mountpoints so the passive server can take it's storage and launch mysql and the webserver, with the same identical data. I know NAS is an expensive sollution, but that's what we mostly do in the enterprise. Dan. After doing some research I found out that I could use the built-in MySQL replication for that, but we also have dynamic data directories on our server - I could setup an rsync script that runs every few minutes, but it'll probably take too many resources and won't be accurate. = To unsubscribe, send mail to [EMAIL PROTECTED] with the word unsubscribe in the message body, e.g., run the command echo unsubscribe | mail [EMAIL PROTECTED]