I've had some luck putting a load balancer infront of multiple zones to get
around the multiple URL issue. You can get the LB to send POST/DELETE et al
to the primary zone, but GET requests can be distributed to multiple zones.
The only issue is the replication delay; your data may not be available on
the secondary for reading yet...

I'm pretty sure you can most LB's to check multiple zones for the object
existing when doing a GET, and redirect to the primary if replication
hasn't caught up; for what I was looking at, I didn't need this!

Dave

On Tue, Oct 7, 2014 at 1:33 AM, Craig Lewis <cle...@centraldesktop.com>
wrote:

> This sounds doable, with a few caveats.
>
>
> Currently, replication is only one direction.  You can only write to the
> primary zone, and you can read from the primary or secondary zones.  A
> cluster can have many zones on it.
>
> I'm thinking your setup would be a star topology.  Each telescope will be
> a primary zone, and replicate to a secondary zone in the main storage
> cluster.  The main cluster will have one read-only secondary zone for each
> telescope.  If you have other needs to write data to the main cluster, you
> can create another zone that only exists on the main cluster (possibly
> replicated to one of the telescopes with a good network connection).
>
> Each zone has it's own URL (primary and secondary), so you'd have a bit of
> management problem remembering to use the correct URL. The URLs can be
> whatever.  Convention follow's Amazon's naming scheme, but you'd probably
> want to create your own scheme, something like
> http://telescope_name-site.inasan.ru/ and
> http://telescope_name-campus.inasan.ru/
>
>
> You might have some problems with the replication if your VPN connections
> aren't stable.  The replication agent isn't very tolerant of cluster
> problem, so I suspect (but haven't tested) that long VPN outages will need
> a replication agent restart.  For sites that don't have permanent
> connections, just make the replication agent startup and shutdown part of
> the connection startup and shutdown process.  Replication state is
> available via a REST api, so it can be monitored.
>
>
> I have tested large backlogs in replication.  When I initially imported my
> data, I deliberately imported faster than I had bandwidth to replicate.  At
> one point, my secondary cluster was ~10 million objects, and ~10TB behind
> the primary cluster.  It eventually caught up, but the process doesn't
> handle stops and restarts well.  Restarting the replication while it was
> dealing with the backlog will start from the beginning of the backlog.
> This can be a problem if your backlog is so large that it won't finish in a
> day, because log rotation will restart the replication agent.  If that's
> something you think might be a problem, I have some strategies to deal with
> it, but they're manual and hacky.
>
>
> Does that sound feasible?
>
>
> On Mon, Oct 6, 2014 at 5:42 AM, Pavel V. Kaygorodov <pa...@inasan.ru>
> wrote:
>
>> Hi!
>>
>> Our institute now planning to deploy a set of robotic telescopes across a
>> country.
>> Most of the telescopes will have low bandwidth and high latency, or even
>> not permanent internet connectivity.
>> I think, we can set up synchronization of observational data with ceph,
>> using federated gateways:
>>
>> 1. The main big storage ceph cluster will be set up in our institute main
>> building
>> 2. The small ceph clusters will be set up near each telescope, to store
>> only the data from local telescope
>> 3. VPN tunnels will be set up from each telescope site to our institute
>> 4. Federated gateways mechanism will do all the magic to synchronize data
>>
>> Is this a realistic plan?
>> What problems we can meet with this setup?
>>
>> Thanks in advance,
>>   Pavel.
>>
>> _______________________________________________
>> ceph-users mailing list
>> ceph-users@lists.ceph.com
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to