[ceph-users] Re: Tuning CephFS on NVME for HPC / IO500

2022-12-04 Thread Janne Johansson
Den lör 3 dec. 2022 kl 22:52 skrev Sebastian : > > One thing to this discussion. > I had a lot of problems with my clusters. I spent some time debugging. > What I found and what I confirmed on AMD nodes, everything starts working > like a charm when I added to kernel param iommu=pt > Plus some oth

[ceph-users] Re: Tuning CephFS on NVME for HPC / IO500

2022-12-04 Thread Manuel Holtgrewe
Dear Sebastian, Thank you for this insight. It sounds like something that is easy to try. Does this relate to the Ceph cluster? My use case is cephfs only. All my clients are Intel based and strictly separated from the Ceph servers. Everything is bare metal. Most information I found on IOMMU we

[ceph-users] Re: Tuning CephFS on NVME for HPC / IO500

2022-12-03 Thread Sebastian
One thing to this discussion. I had a lot of problems with my clusters. I spent some time debugging. What I found and what I confirmed on AMD nodes, everything starts working like a charm when I added to kernel param iommu=pt Plus some other tunings, I can’t share, all information now, but this i

[ceph-users] Re: Tuning CephFS on NVME for HPC / IO500

2022-12-02 Thread Manuel Holtgrewe
Dear Mark. Thank you very much for all of this information. I learned a lot! In particular that I need to learn more about pinning. In the end, I want to run the whole thing in production with real world workloads. My main aim in running the benchmark is to ensure that my hardware and OS is corre

[ceph-users] Re: Tuning CephFS on NVME for HPC / IO500

2022-12-01 Thread Mark Nelson
Hi Manuel, I did the IO500 runs back in 2020 and wrote the cephfs aiori backend for IOR/mdtest.  Not sure about the segfault, it's been a while since I've touched that code.  It was working the last time I used it. :D  Having said that, I don't think that's your issue.   The userland backend

[ceph-users] Re: Tuning CephFS on NVME for HPC / IO500

2022-12-01 Thread Fox, Kevin M
: Robert Sander Sent: Wednesday, November 30, 2022 11:58 PM To: ceph-users Subject: [ceph-users] Re: Tuning CephFS on NVME for HPC / IO500 Check twice before you click! This email originated from outside PNNL. Hi, On 2022-12-01 8:26, Manuel Holtgrewe wrote: > The Ceph cluster nodes have

[ceph-users] Re: Tuning CephFS on NVME for HPC / IO500

2022-11-30 Thread Robert Sander
Hi, On 2022-12-01 8:26, Manuel Holtgrewe wrote: The Ceph cluster nodes have 10x enterprise NVMEs each (all branded as "Dell enterprise disks"), 8 older nodes (last year) have "Dell Ent NVMe v2 AGN RI U.2 15.36TB" which are Samsung disks, 2 newer nodes (just delivered) have "Dell Ent NVMe CM6