On 29/07/2016 5:00 AM, Julien Danjou wrote: > Best way is probably to do some bench… but I think it really depends on > the use cases here. The interest of having many small splits is that you > can parallelize the read. > > Considering the compression ratio we have, I think we should split in > smaller files. I'd pick 3600 and give it a try.
i gave this a quick try with a series of ~68k points with object size of 14400 points (uncompressed), i got: [gchung@gchung-dev ~(keystone_admin)]$ time gnocchi measures show dc51c402-67e6-4b28-aba0-9d46b35b5397 --granularity 60 &> /tmp/blah real 0m6.398s user 0m5.003s sys 0m0.071s it took ~39.45s to process into 24 different aggregated series and created 6 split objects. with object size of 3600 points (uncompressed), i got: [gchung@gchung-dev ~(keystone_admin)]$ time gnocchi measures show 301947fd-97ee-428a-b445-41a67ee62c38 --granularity 60 &> /tmp/blah real 0m6.495s user 0m4.970s sys 0m0.073s it took ~39.89s to process into 24 different aggregated series and created 21 split objects so at first glance, it doesn't really seem to affect performance much whether it's one 'larger' file or many smaller files. that said, with new proposed v3 serialisation format, a larger file has a greater requirement for additional padding which is not a good thing. cheers, -- gord __________________________________________________________________________ OpenStack Development Mailing List (not for usage questions) Unsubscribe: openstack-dev-requ...@lists.openstack.org?subject:unsubscribe http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev