Hi Ryan Thanks for helping out!!
The cluster consists of 4 exactly similar nodes - all dedicated to riak use
only - no other zones or tasks going on. We use Riak-EE 0.13. The servers is HP
servers with 4 x 146GB 10K RPM SAS disks. There is a memorycache on the RAID
controller and it is used during both read and writes but the RAID iis built
usin Solaris-10u9 ZFS in a setup as such:
pool: pool01
state: ONLINE
scrub: scrub completed after 0h0m with 0 errors on Tue Oct 26 21:25:05 2010
config:
NAME STATE READ WRITE CKSUM
pool01 ONLINE 0 0 0
mirror-0 ONLINE 0 0 0
c0t0d0s7 ONLINE 0 0 0
c0t1d0s7 ONLINE 0 0 0
mirror-1 ONLINE 0 0 0
c0t2d0 ONLINE 0 0 0
c0t3d0 ONLINE 0 0 0
errors: No known data errors
metrics during load gives 5% CPU load and about 10% IO load (iostat reports 30
iops and the disks should be able to handle 300 iops each). So basically, the
servers is unloaded....
One question remains - we use ZFS with default blocksize of 128Kb - what is the
optimal blocksize with bitcask?
But I believe, that we should look somewhere else for the challenge - the
hardware is not loaded significant, so I suspect, that we have a faulty
datamodel or usage...?
Best regards,
Karsten
On Nov 8, 2010, at 23:43 , Ryan Tilder wrote:
> Hi, Jan. Your description of the behaviour you're seeing below is frequently
> the result of slow access times to data on disk due to low spindle count for
> a given data set. Can you tell me the hardware specifications of the disks
> in each of the machines in the Riak ring? Primarily the number of disks,
> their capacity in GB or TB, their spindle speed, any on disk buffer they
> might have, and whether or not they're all the same hardware model?[1]
>
> One other question that might be relate: is there any other software of note
> running on the hardware? i.e. Riak running on multi-tenant hardware in a
> Solaris Zone while another DB is running in a separate Zone on the same
> hardware, another disk seek heavy daemon running on the same machine, etc.
>
> --Ryan
>
> 1. If you don't know the above, you can get a lot of information from the
> "prtconf -v | more" command and just search for the work "disk". The model
> number(s) for the relevant disks is enough that we can hunt down the
> technical specs quickly enough.
>
> On Mon, Nov 8, 2010 at 12:44 PM, Jan Buchholdt <[email protected]> wrote:
> Another information is that the first time I do a link walk (using curl) on a
> total idle cluster it takes 2.71 second for a person with 363 documents. If I
> repeat the request it takes 319 milliseconds. I would expect that the
> performance would be almost the same.
>
> If I run my performance test with 20 treads, that randomly pick a Person from
> 5 millions, the minimum time is 2.8s, average 6.7s, 90% 8.9s and max 12.4s.
>
> _______________________________________________
> riak-users mailing list
> [email protected]
> http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com
smime.p7s
Description: S/MIME cryptographic signature
_______________________________________________ riak-users mailing list [email protected] http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com
