We have a 19.2.3 cluster managed by cephadm with five management nodes and 21 
OSD nodes.  We have roughly 300 linux clients across several subnets doing 
cephfs kernel mounts running a wide range of applications.

Each management node has 760GB of memory.  We’re currently using a single 
active MDS daemon, but have experimented with multiple MDS daemons and 
directory pinning.

Each OSD for the cephfs data pool uses hdd (SATA) drives, with the DB and WAL 
on nvme's internal to the storage node.  The cephfs metadata pool is on nvme 
drives internal to the management nodes.

After a lot of testing in attempts to quiet down persistent 
MDS_CLIENT_LATE_RELEASE "clients failing to respond to capability release” and 
MDS_CLIENT_RECALL "clients failing to respond to cache pressure” warnings, 
we’ve ended up with the following settings.  This seems to be working well.  We 
still get periodic MDS_CLIENT_RECALL warnings, but not nearly as many as we 
were seeing and they clear relatively quickly.

I’d greatly appreciate any suggestions for further improvements, or any 
concerns anyone has with these.

Many thanks,
Devin

———
cephfs  session_timeout 120
mds     mds_cache_memory_limit  549755813888                                    
                                              mds       mds_cache_mid   
0.700000                                                                        
              mds       mds_cache_reservation   0.100000                        
                                                              mds       
mds_cache_trim_decay_rate       0.900000                                        
                                              mds       
mds_cache_trim_threshold        524288                                          
                                              mds       
mds_cap_revoke_eviction_timeout 30.000000                                       
                                              mds       
mds_health_cache_threshold      2.000000                                        
                                              mds       mds_log_max_segments    
256                                                                             
              mds       mds_max_caps_per_client 5000000                         
                                                              mds       
mds_recall_global_max_decay_threshold   1048576                                 
                                                      mds       
mds_recall_max_caps     5000                                                    
                                      mds       mds_recall_max_decay_rate       
1.500000                                                                        
              mds       mds_recall_max_decay_threshold  524288                  
                                                                      mds       
mds_recall_warning_decay_rate   120.000000                                      
                                              mds       
mds_recall_warning_threshold    262144                                          
                                                                                
                                             
———
_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]

Reply via email to