Ceph is definitely a good choice for storing millions of files.  It sounds like 
you plan to use this like s3, so my first question would be:  Are the deletes 
done for a specific reason?  ( e.g. the files are used for a process and 
discarded  )  If its an age thing, you can set the files to expire when putting 
them in, then ceph will automatically clear them.

The more spinners you have the more performance you will end up with.  Network 
10Gb or higher?

Octopus is production stable and contains many performance enhancements.  
Depending on the OS, you may not be able to upgrade from nautilus until they 
work out that process ( e.g. centos 7/8 ).  

Delete speed is not that great but you would have to test it with your cluster 
to see how it performs for your use case.  If you have enough space present, is 
there a process that breaks if the files are not deleted?  


Regards,
-Brent

Existing Clusters:
Test: Ocotpus 15.2.5 ( all virtual on nvme )
US Production(HDD): Nautilus 14.2.11 with 11 osd servers, 3 mons, 4 gateways, 2 
iscsi gateways
UK Production(HDD): Nautilus 14.2.11 with 18 osd servers, 3 mons, 4 gateways, 2 
iscsi gateways
US Production(SSD): Nautilus 14.2.11 with 6 osd servers, 3 mons, 4 gateways, 2 
iscsi gateways
UK Production(SSD): Octopus 15.2.5 with 5 osd servers, 3 mons, 4 gateways




-----Original Message-----
From: Adrian Nicolae <adrian.nico...@rcs-rds.ro> 
Sent: Wednesday, November 11, 2020 3:42 PM
To: ceph-users <ceph-users@ceph.io>
Subject: [ceph-users] question about rgw delete speed


Hey guys,


I'm in charge of a local cloud-storage service. Our primary object storage is a 
vendor-based one and I want to replace it in the near future with Ceph with the 
following setup :

- 6 OSD servers with 36 SATA 16TB drives each and 3 big NVME per server
(1 big NVME for every 12 drives so I can reserve 300GB NVME storage for every 
SATA drive), 3 MON, 2 RGW with Epyc 7402p and 128GB RAM. So in the end we'll 
have ~ 3PB of raw data and 216 SATA drives.

Currently we have ~ 100 millions of files on the primary storage with the 
following distribution :

- ~10% = very small files ( less than 1MB - thumbnails, text&office files and 
so on)

- ~60%= small files (between 1MB and 10MB)

-  20% = medium files ( between 10MB and 1GB)

- 10% = big files (over 1GB).

My main concern is the speed of delete operations. We have around 500k-600k 
delete ops every 24 hours so quite a lot. Our current storage is not deleting 
all the files fast enough (it's always 1 week-10 days
behind) , I guess is not only a software issue and probably the delete speed 
will get better if we add more drives (we now have 108).

What do you think about Ceph delete speed ? I read on other threads that it's 
not very fast . I wonder if this hw setup can handle our current delete load 
better than our current storage. On RGW servers I want to use Swift , not S3.

And another question :   can I start deploying in production directly the 
latest Ceph version (Octopus) or is it safer to start with Nautilus until 
Octopus will be more stable ?

Any input would be greatly appreciated !


Thanks,

Adrian.




_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to 
ceph-users-le...@ceph.io
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to