So I just have a few more questions that are coming to mind.  Firstly, I
have OSDi whose underlying filesystems can be..  Dun dun dun!!!!  Resized!!
 If I choose to expand my allocation to ceph, I can in theory do so by
expanding the quota on the OSDi.  (I'm using ZFS)  Similarly, if the OSD is
underutilized enough, I can in theory shrink the OSD.  The question:  How
will Ceph respond to it's OSDi spontaneously changing in size?  Do I have
to make some modification to the config before restarting the OSDi, or shut
them down, resize, and restart them?  Or are they permanently lost, and my
only hope is to replace them with larger OSDi and slowly migrate data off
of them?

My other question is this:  I'm trying to limit my systems to effectively
one OSD per pool.  I have one pool per VM host, and three VM hosts.  VM
Hosts have a slow link path between them such that A can communicate in low
latency, high throughput time with B or C, but B<==>C is very slow and high
latency for obscure, probably stupid, technical reasons.  Therefore, I want
to place OSDs, (one per host) such that VMs on either B or C are mirrored
only to OSD.A, OSD.B for host B, and OSDA.A, OSD.C for host C.  Host C
should be mirrored across all three OSDi, so as to not unduly overburdon
either of the other two lower power hosts.  The tree should look something
like this:

OSD.A
OSD.B
OSD.C

Host A
    OSD.A
    OSD.B
    OSD.C

Host B
    OSD.B
    OSD.A

Host C
    OSD.C
    OSD.A

Will this work?  Or will I be forced to run additional OSDi to be able to
distribute data like this?  I can figure the config out myself, but before
testing it, I just want to be fairly certain that it will actually work as
I intend, and not give some error about OSDi appearing in multiple places
in the config.

Thank you for taking the time to read and respond to my confound questions,
everyone, and thanks for making Ceph as awesome as it is today :)
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to