Hello Christian. First of all, thanks for your considerations, I really appreciate it.
2017-08-23 21:34 GMT-03:00 Christian Balzer <ch...@gol.com>: > > Hello, > > On Wed, 23 Aug 2017 09:11:18 -0300 Guilherme Steinmüller wrote: > > > Hello! > > > > I recently installed INTEL SSD 400GB 750 SERIES PCIE 3.0 X4 in 3 of my > OSD > > nodes. > > > Well, you know what's coming now, don't you? > > That's a consumer device, with 70GB writes per day endurance. > unless you're essentially having a read-only cluster, you're throwing away > money. > Yes, we knew that we were going to buy a consumer device due to our limited budget and our objective of constructing a small plan of a production cloud. This model seemed acceptable. It was the top list of the consumer models on Sebastien's benchmarks We are a lab that depends on different budget sources to accquire equipments, so they can vary and most of the time we are limited by different budget ranges. > > > First of all, here's is an schema describing how my cluster is: > > > > [image: Imagem inline 1] > > > > [image: Imagem inline 2] > > > > I primarily use my ceph as a beckend for OpenStack nova, glance, swift > and > > cinder. My crushmap is configured to have rulesets for SAS disks, SATA > > disks and another ruleset that resides in HPE nodes using SATA disks too. > > > > Before installing the new journal in HPE nodes, i was using one of the > > disks that today are OSDs (osd.35, osd.34 and osd.33). After upgrading > the > > journal, i noticed that a dd command writing 1gb blocks in openstack nova > > instances doubled the throughput but the value expected was actually 400% > > or 500% since in the Dell nodes that we have another nova pool the > > throughput is around this value. > > > Apples, oranges and bananas. > You're comparing different HW (and no, I'm not going to look this up) > which may or may not have vastly different capabilities (like HW cache), > RAM and (unlikely relevant) CPU. > Indeed, we took this into account. The HP server were cheaper and have a poor configuration due that limited budget source. > Your NVMe may also be plugged into a different, insufficient PCIe slot for > all we know. > I checked this. I compared the slots identifying the slot information between the 3 dell nodes and 3 hp nodes by running: # ls -l /sys/block/nvme0n1 # lspci -vvv -s 0000:06:00.0 <- slot identifier The only difference is: Dell has a parameter called *Cache Line Size: 32 bytes* and HP doesn't have this. > You're also using very different HDDs, which definitely will be a factor. > > I thought that the backend disks would not interfer that much. For example, the ceph journal has a parameter called filestore max sync interval, which means that ceph journal will commit the transactions to the backend OSDs in a defined interval, ours is set to 35. So the client requests go first to SSD and than is commited to the OSDs. > But most importanly you're comparing 2 pools of vastly different ODS > count, no wonder a pool with 15 OSDs is faster in sequential writes than > one with 9. > > Here is a demonstration of the scenario and the difference in performance > > between Dell nodes and HPE nodes: > > > > > > > > Scenario: > > > > > > - Using pools to store instance disks for OpenStack > > > > > > - Pool nova in "ruleset SAS" placed on c4-osd201, c4-osd202 and > > c4-osd203 with 5 osds per hosts > > > SAS > > > > - Pool nova_hpedl180 in "ruleset NOVA_HPEDL180" placed on > c4-osd204, > > c4-osd205, c4-osd206 with 3 osds per hosts > > > SATA > > > > - Every OSD has one partition of 35GB in a INTEL SSD 400GB 750 > > SERIES PCIE 3.0 X4 > > > Overkill, but since your NVMe will die shortly anyway... > > With large sequential tests, the journal will have nearly NO impact on the > result, even if tuned to that effect. > > > > > - Internal link for cluster and public network of 10Gbps > > > > > > - Deployment via ceph-ansible. Same configuration define in > ansible > > for every host on cluster > > > > > > > > *Instance on pool nova in ruleset SAS:* > > > > > > # dd if=/dev/zero of=/mnt/bench bs=1G count=1 oflag=direct > > 1+0 records in > > 1+0 records out > > 1073741824 bytes (1.1 GB, 1.0 GiB) copied, 2.56255 s, 419 MB/s > > > This is a very small test for what you're trying to determine and not > going to be very representative. > If for example there _is_ a HW cache of 2GB on the Dell nodes, it would > fit nicely in there. > > Dell has PERC H730 Mini (Embedded) each with cache memory size of 1024 MB otherwise my HP uses a B140i dynamic array. Both HP and Dell doesn't use any raid level for the OSDs, just Dell for the Operating System. > > > > *Instance on pool nova in ruleset NOVA_HPEDL180:* > > > > # dd if=/dev/zero of=/mnt/bench bs=1G count=1 oflag=direct > > 1+0 records in > > 1+0 records out > > 1073741824 bytes (1.1 GB, 1.0 GiB) copied, 11.8243 s, 90.8 MB/s > > > > > > I made some FIO benchmarks as suggested by Sebastien ( > > https://www.sebastien-han.fr/blog/2014/10/10/ceph-how-to- > > test-if-your-ssd-is-suitable-as-a-journal-device/ ) and the command > with 1 > > job returned me about 180MB/s of throughput in recently installed nodes > > (HPE nodes). I made some hdparm benchmark in all SSDs and everything > seems > > normal. > > > I'd consider a 180MB/s result from a device that supposedly does 900MB/s a > fail, but then again those tests above do NOT reflect journal usage > reality but a more of a hint if something is totally broken or not. > > > > > I can't see what is causing this difference of throughput since the > network > > is not a problem and i think that cpu and memory are not crucial since i > > was monitoring the cluster with atop command and i didn't notice > saturation > > of resources. My only though is that I have less workload in > nova_hpedl180 > > pool in HPE nodes and less disks per node and this ca influence in the > > throughput of the journal. > > > How busy are your NVMe journals during that test on the Dells and HPs > respectively? > Same for the HDDs. > I can't say it now precisely, but what I can tell you for sure is that monitoring these two pools, both thoughtput and disk usage, I can see that the workload for the pool that is placed on the Dell nodes is significantly higher than the pool in the HP node. For example, the OSDs in the Dell node often keep the usage between 70% and 100%, different than HP OSDs, that vary between 10% and 40%. > > Again, run longer, larger tests to get something that will actually > register, also atop with shorter intervals. > > Christian > > > > Any clue about what is missing or what is happening? > > > > Thanks in advance. > > > -- > Christian Balzer Network/Systems Engineer > ch...@gol.com Rakuten Communications >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com