Re: [ceph-users] IO to OSD with librados

2018-06-19 Thread Hervé Ballans
Le 19/06/2018 à 09:02, Dan van der Ster a écrit : The storage arrays are Nexsan E60 arrays having two active-active redundant controllers, 60 3 TB disk drives. The disk drives are organized into six 8+2 Raid 6 LUNs of 24 TB each. This is not the ideal Ceph hardware. Ceph is designed to use

Re: [ceph-users] IO to OSD with librados

2018-06-19 Thread Jialin Liu
Thanks for the advice, Dan. I'll try to reconfigure the cluster and see if the performance changes. Best, Jialin On Tue, Jun 19, 2018 at 12:02 AM Dan van der Ster wrote: > On Tue, Jun 19, 2018 at 1:04 AM Jialin Liu wrote: > > > > Hi Dan, Thanks for the follow-ups. > > > > I have just tried

Re: [ceph-users] IO to OSD with librados

2018-06-18 Thread Jialin Liu
Hi Dan, Thanks for the follow-ups. I have just tried running multiple librados MPI applications from multiple nodes, it does show increased bandwidth, with ceph -w, I observed as high as 500MB/sec (previously only 160MB/sec ), I think I can do finer tuning by coordinating more concurrent

Re: [ceph-users] IO to OSD with librados

2018-06-18 Thread Dan van der Ster
Hi, Have you tried running rados bench in parallel from several client machines? That would demonstrate the full BW capacity of the cluster. e.g. make a test pool with e.g. 256 PGs (which will average 16 per OSD on your cluster). Then from several clients at once do `rados bench -p test 60

Re: [ceph-users] IO to OSD with librados

2018-06-18 Thread Jialin Liu
Hi, To make the the problem clearer, here is the configuration of the cluster: The 'problem' I have is the low bandwidth no matter how I increase the concurrency. I have tried using MPI to launch 322 processes, each calling librados to create a handle and initialize the io context, and write one

Re: [ceph-users] IO to OSD with librados

2018-06-18 Thread Jialin Liu
Thank you Dan. I’ll try it. Best, Jialin NERSC/LBNL > On Jun 18, 2018, at 12:22 AM, Dan van der Ster wrote: > > Hi, > > One way you can see exactly what is happening when you write an object > is with --debug_ms=1. > > For example, I write a 100MB object to a test pool: rados > --debug_ms=1

Re: [ceph-users] IO to OSD with librados

2018-06-18 Thread Dan van der Ster
Hi, One way you can see exactly what is happening when you write an object is with --debug_ms=1. For example, I write a 100MB object to a test pool: rados --debug_ms=1 -p test put 100M.dat 100M.dat I pasted the output of this here: https://pastebin.com/Zg8rjaTV In this case, it first gets the

Re: [ceph-users] IO to OSD with librados

2018-06-18 Thread Jialin Liu
Sorry about the misused term 'OSS: object storage server' (a term often used in Lustre filesystem), what I meant is 4 hosts, each manages 12 OSDs. Thanks for anyone who may answer any of my questions. Best, Jialin NERSC/LBNL On Sun, Jun 17, 2018 at 11:29 AM Jialin Liu wrote: > Hello, > > I

[ceph-users] IO to OSD with librados

2018-06-17 Thread Jialin Liu
Hello, I have a couple questions regarding the IO on OSD via librados. 1. How to check which osd is receiving data? 2. Can the write operation return immediately to the application once the write to the primary OSD is done? or does it return only when the data is replicated twice? (size=3) 3.