Hi Bruce,
hmm, sounds for me like the rbd cache.
Can you look, if the cache is realy disabled in the running config with

ceph --admin-daemon /var/run/ceph/ceph-osd.0.asok config show | grep cache

Udo

On 30.01.2015 21:51, Bruce McFarland wrote:
>
> I have a cluster and have created a rbd device - /dev/rbd1. It shows
> up as expected with ‘rbd –image test info’ and rbd showmapped. I have
> been looking at cluster performance with the usual Linux block device
> tools – fio and vdbench. When I look at writes and large block
> sequential reads I’m seeing what I’d expect with performance limited
> by either my cluster interconnect bandwidth or the backend device
> throughput speeds – 1 GE frontend and cluster network and 7200rpm SATA
> OSDs with 1 SSD/osd for journal. Everything looks good EXCEPT 4K
> random reads. There is caching occurring somewhere in my system that I
> haven’t been able to detect and suppress - yet.
>
>  
>
> I’ve set ‘rbd_cache=false’ in the [client] section of ceph.conf on the
> client, monitor, and storage nodes. I’ve flushed the system caches on
> the client and storage nodes before test run ie vm.drop_caches=3 and
> set the huge pages to the maximum available to consume free system
> memory so that it can’t be used for system cache . I’ve also disabled
> read-ahead on all of the HDD/OSDs.
>
>  
>
> When I run a 4k randon read workload on the client the most I could
> expect would be ~100iops/osd x number of osd’s – I’m seeing an order
> of magnitude greater than that AND running IOSTAT on the storage nodes
> show no read activity on the OSD disks.
>
>  
>
> Any ideas on what I’ve overlooked? There appears to be some read-ahead
> caching that I’ve missed.
>
>  
>
> Thanks,
>
> Bruce
>
>
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to