> Op 3 augustus 2016 om 10:08 schreef Rob Reus <[email protected]>:
>
>
> Hi all,
>
>
> I built a CRUSH map, with the goal to distinguish between SSD and HDD storage
> machines using only 1 root. The map can be found here:
> http://pastebin.com/VQdB0CE9
>
>
> The issue I am having is this:
>
>
> root@ceph2:~/crush_files# crushtool -i crushmap --test --show-utilization
> --rule 0 --num-rep 3
> rule 0 (replicated_ruleset_ssd), x = 0..1023, numrep = 3..3
> rule 0 (replicated_ruleset_ssd) num_rep 3 result size == 0: 84/1024
> rule 0 (replicated_ruleset_ssd) num_rep 3 result size == 1: 437/1024
> rule 0 (replicated_ruleset_ssd) num_rep 3 result size == 2: 438/1024
> rule 0 (replicated_ruleset_ssd) num_rep 3 result size == 3: 65/1024
>
>
> And then the same test using num-rep 46 (the lowest possible number that
> shows full utilization):
>
>
> root@ceph2:~/crush_files# crushtool -i crushmap --test --show-utilization
> --rule 0 --num-rep 46
> rule 0 (replicated_ruleset_ssd), x = 0..1023, numrep = 46..46
> rule 0 (replicated_ruleset_ssd) num_rep 46 result size == 3: 1024/1024
>
>
> Full output of above commands can be found here http://pastebin.com/2mbBnmSM
> and here http://pastebin.com/ar6SAFnX
>
>
> The fact that amount of num-rep seems to scale with how many OSDs I am using,
> leads me to believe I am doing something wrong.
>
>
> When using 2 roots (1 dedicated to SSD and 1 to HDD), everything works
> perfectly (example: http://pastebin.com/Uthxesut).
>
>
> Would love to know what I am missing.
>
Can you tell me the reasoning behind adding the rack in between? Since there is
only one rack, why add the machines there?
In your case I wouldn't add a new type either, but I would do this:
host machineA-ssd {
}
host machineB-ssd {
}
host machineA-hdd {
}
host machineB-hdd {
}
root ssd {
item machineA-sdd
item machineB-ssd
}
root hdd {
item machineA-hdd
item machineB-hdd
}
rule replicated_ruleset_ssd {
ruleset 0
type replicated
min_size 1
max_size 10
step take ssd
step chooseleaf firstn 0 type host
step emit
}
rule replicated_ruleset_hdd {
ruleset 0
type replicated
min_size 1
max_size 10
step take hdd
step chooseleaf firstn 0 type host
step emit
}
And try again :)
Wido
>
> Thanks!
>
>
> - Rob
>
> [http://pastebin.com/i/facebook.png]<http://pastebin.com/Uthxesut>
>
> # begin crush map tunable choose_local_tries 0 tunable
> choose_local_fallback_t - Pastebin.com<http://pastebin.com/Uthxesut>
> pastebin.com
>
>
> [http://pastebin.com/i/facebook.png]<http://pastebin.com/ar6SAFnX>
>
> root@ceph2:~/crush_files# crushtool -i crushmap --test --show-utilization-all
> -- - Pastebin.com<http://pastebin.com/ar6SAFnX>
> pastebin.com
>
> [http://pastebin.com/i/facebook.png]<http://pastebin.com/2mbBnmSM>
>
> root@ceph2:~/crush_files# crushtool -i crushmap --test --show-utilization-all
> -- - Pastebin.com<http://pastebin.com/2mbBnmSM>
> pastebin.com
>
>
> [http://pastebin.com/i/facebook.png]<http://pastebin.com/VQdB0CE9>
>
> # begin crush map tunable choose_local_tries 0 tunable
> choose_local_fallback_t - Pastebin.com<http://pastebin.com/VQdB0CE9>
> pastebin.com
>
>
> _______________________________________________
> ceph-users mailing list
> [email protected]
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com