Christian is correct that min_size does not affect how many need to ACK the write, it is responsible for how many copies need to be available for the PG to be accessible. This is where SSD journals for filestore and SSD DB/WAL partitions come into play. The write is considered ACK'd as soon as the journal has received the write.
Additionally please keep in mind that a write to an SSD across a network is not going to be as fast as the SSD statistics claim on their own. You are adding in network latency to the device. On Tue, Oct 10, 2017 at 7:51 PM Christian Balzer <ch...@gol.com> wrote: > > Hello, > > On Wed, 11 Oct 2017 00:05:26 +0200 Jack wrote: > > > Hi, > > > > I would like some information about the following > > > > Let say I have a running cluster, with 4 OSDs: 2 SSDs, and 2 HDDs > > My single pool has size=3, min_size=2 > > > > For a write-only pattern, I thought I would get SSDs performance level, > > because the write would be acked as soon as min_size OSDs acked > > > > But I am right ? > > > You're the 2nd person in very recent times to come up with that wrong > conclusion about min_size. > > All writes have to be ACKed, the only time where hybrid stuff helps is to > accelerate reads. > Which is something that people like me at least have very little interest > in as the writes need to be fast. > > Christian > > > (the same setup could involve some high latency OSDs, in the case of > > country-level cluster) > > _______________________________________________ > > ceph-users mailing list > > ceph-users@lists.ceph.com > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > > > > -- > Christian Balzer Network/Systems Engineer > ch...@gol.com Rakuten Communications > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com