Perfect, thanks Kevin. Looking at SAR this morning, I can see that I'm sometime reaching 300tps, and spiked at 80% WIO... That will cost me 5 new additional hard drives :(
I’m not sure I will have time to install them all today, but as soon as it’s done I will give you some news. JM 2013/2/8, Kevin O'dell <kevin.od...@cloudera.com>: > JM, > > Basically, you will have to replace failed disk and rebuild RAID0 since > the other half of the data is worthless. There is not a real recommended > value, but anything under 150 - 200 would make me more comfortable. > > On Fri, Feb 8, 2013 at 10:43 AM, Jean-Marc Spaggiari < > jean-m...@spaggiari.org> wrote: > >> Hi Kevin, >> >> I think it will take time before I get a chance to have 5 drives in >> the same server, so I will see at that time to test RAID5. >> >> I'm going to add one drive per server today or tomorrow to try to >> improve that. What IOPs should I try to have? 100? Less? It will all >> be SATA3 drives and I will configure all in RAID0. >> >> It doesn't seems to me to be an issue to lose one node, since data >> will be replicated everywhere else. I will "simply" have to replace >> the failing disk and restart the node, no? >> >> JM >> >> 2013/2/8, Kevin O'dell <kevin.od...@cloudera.com>: >> > Azuryy, >> > >> > The main reason to recommend against RAID is that it is slow and it >> adds >> > redundancy that we already have in Hadoop. RAID0 is another story as >> long >> > as all of the drives are healthy and you don't mind losing the whole >> volume >> > if you lose one drive. >> > >> > JM, >> > >> > I would not even waste my time testing RAID5 or RAID6(unless it is >> > just >> > for educational purposes :) ). 200+ IOPs consistently on one SATA >> > drive >> is >> > pretty high, that would explain your high I/O wait time. If your use >> case >> > allows for you to lose the whole node, there is not a good reason for >> > you >> > to shy away from RAID0. Please let us know how this plays out with >> > your >> > environment. >> > >> > On Thu, Feb 7, 2013 at 10:23 PM, Azuryy Yu <azury...@gmail.com> wrote: >> > >> >> JM, >> >> >> >> I don't have the context, but if you are using Hadoop/Hbase, so don't >> >> do >> >> RAID on your disk. >> >> >> >> >> >> On Fri, Feb 8, 2013 at 11:15 AM, Jean-Marc Spaggiari < >> >> jean-m...@spaggiari.org> wrote: >> >> >> >> > Ok. I see. For my usecase I prefer to loose the data and have faster >> >> > process. So I will go for RAID0 and keep the replication factor to >> >> > 3... If at some point I have 5 disks in the node, I will most >> >> > probably >> >> > give a try to RAID5 and see the performances compared to the other >> >> > RAID/JBOD options. >> >> > >> >> > Is there a "rule", like, 1 HD per core? Or we can't really simplify >> >> > that >> >> > much? >> >> > >> >> > So far I have that in the sar output: >> >> > 21:35:03 tps rtps wtps bread/s bwrtn/s >> >> > 21:45:03 218,85 215,97 2,88 45441,95 308,04 >> >> > 21:55:02 209,73 206,67 3,06 43985,28 378,32 >> >> > 22:05:04 215,03 211,71 3,33 44831,00 312,95 >> >> > Average : 214,54 211,45 3,09 44753,09 333,07 >> >> > >> >> > But I'm not sure what it means. I will wait for tomorrow to get more >> >> > results, but my job will be done over night, so I'm not sure the >> >> > average will be accurate... >> >> > >> >> > JM >> >> > >> >> > >> >> > 2013/2/7, Kevin O'dell <kevin.od...@cloudera.com>: >> >> > > JM, >> >> > > >> >> > > I think you misunderstood me. I am not advocating any form of >> RAID >> >> for >> >> > > Hadoop. It is true that we already have redundancy built in with >> >> > > HDFS. >> >> > So >> >> > > unless you were going to do something silly like sacrifice speed >> >> > > to >> >> > > run >> >> > > RAID1 or RAID5 and lower your replication to 2...just don't do it >> >> > > :) >> >> > > Anyway, yes you probably should have 3 - 4 drives per node if not >> >> more. >> >> > > At that point then the you will really see the benefit of JBOD >> >> > > over >> >> > RAID0 >> >> > > >> >> > > Do you want to be able to lose a drive and keep the node up? If >> yes, >> >> > then >> >> > > JBOD is for you. Do you not care if you lose that node due to >> >> > > drive >> >> > > failure? You just need speed, then RAID0 may be the correct >> >> > > choice. >> >> Sar >> >> > > will take some time to populate. Give it about 24 hours and you >> >> > > should >> >> > be >> >> > > able to glean some interesting information. >> >> > > >> >> > > On Thu, Feb 7, 2013 at 9:50 PM, Jean-Marc Spaggiari >> >> > > <jean-m...@spaggiari.org >> >> > >> wrote: >> >> > > >> >> > >> Ok. I see with RAID0 might be better for me compare to JBOD. >> >> > >> Also, >> >> > >> why >> >> > >> do we want to use RAID1 or RAID5? We already have the redundancy >> >> > >> done >> >> > >> by hadoop, is it not going to add another non-required level of >> >> > >> redundancy? >> >> > >> >> >> > >> Should I already think to have 3 or even 4 drives in each node? >> >> > >> >> >> > >> I tried sar -A and it's only giving me 2 lines. >> >> > >> root@node7:/home/hbase# sar -A >> >> > >> Linux 3.2.0-4-amd64 (node7) 2013-02-07 _x86_64_ >> >> > >> (4 >> >> CPU) >> >> > >> >> >> > >> 21:29:54 LINUX RESTART >> >> > >> >> >> > >> It was not enabled, so I just enabled it and restart sysstat, but >> >> > >> seems that it's still not populated. >> >> > >> >> >> > >> I have the diskstats plugin installed on ganglia, so I have a LOT >> of >> >> > >> disks information, but not this specific one. >> >> > >> >> >> > >> My write_bytes_per_sec is pretty low. Average is 232K for the >> >> > >> last >> 2 >> >> > >> hours. But my erad_bytes_per_sec is avera 22.83M for the same >> >> > >> period. >> >> > >> The graph is looking like a comb. >> >> > >> >> >> > >> I just retried sar and some data is coming.. I will need to let >> >> > >> it >> >> > >> run >> >> > >> for few more minutes to get some more data ... >> >> > >> >> >> > >> JM >> >> > >> >> >> > >> >> >> > >> 2013/2/7, Kevin O'dell <kevin.od...@cloudera.com>: >> >> > >> > JM, >> >> > >> > >> >> > >> > Okay, I think I see what was happening. You currently only >> have >> >> one >> >> > >> > drive in the system that is showing High I/O wait correct? You >> >> > >> > are >> >> > >> looking >> >> > >> > at bringing in a second drive to help distribute the load? In >> >> > >> > your >> >> > >> testing >> >> > >> > with two drives you saw that RAID0 offerred superior >> >> > >> > performance >> >> > >> > vs >> >> > >> > JBOD. >> >> > >> > Typically when we see RAID vs JBOD we are dealing with about 6 >> >> > >> > - >> >> > >> > 12 >> >> > >> > drives. Here are some of the pluses and minuses: >> >> > >> > >> >> > >> > RAID0 - faster performance since the data is striped, but you >> >> > >> > are >> >> > >> > as >> >> > >> > fast >> >> > >> > as your slowest drive and one drive failure you lose the whole >> >> volume. >> >> > >> > >> >> > >> > JBOD - Better redundancy and faster than a RAID1, or a RAID5 >> >> > >> > configuration(unsure about a RAID4), but you are slower than >> RAID0 >> >> > >> > >> >> > >> > It sounds like since you only have 1 drive in the node right >> >> > >> > now, >> >> you >> >> > >> > wouldn't be gaining or losing any redundancy by going with >> >> > >> > RAID0. >> >> For >> >> > >> what >> >> > >> > it is worth, I would agree that you are I/O bound. If you run >> >> > >> > a >> >> > >> > sar >> >> > -A >> >> > >> > > >> >> > >> > /tmp/sar.out and you take a look at the drive utilization what >> >> > >> > is >> >> your >> >> > >> > TPS(IOPs) count that you are seeing? >> >> > >> > >> >> > >> > On Thu, Feb 7, 2013 at 9:00 PM, Jean-Marc Spaggiari >> >> > >> > <jean-m...@spaggiari.org >> >> > >> >> wrote: >> >> > >> > >> >> > >> >> Hi Kevin, >> >> > >> >> >> >> > >> >> I'm facing some issues on one of my nodes and I'm trying to >> >> > >> >> find >> >> > >> >> a >> >> > way >> >> > >> >> to fix that. CPU is used about 10% by user, and 80% for WIO. >> >> > >> >> So >> >> > >> >> I'm >> >> > >> >> looking for a way to improve that. The mother board can do >> >> > >> >> RAIDx >> >> and >> >> > >> >> JBOD too. It's the server I used few weeks ago to run some >> >> > >> >> disks >> >> > >> >> benchs. >> >> > >> >> >> >> > >> >> >> http://www.spaggiari.org/index.php/hbase/hard-drives-performances >> >> > >> >> >> >> > >> >> The conclusion was that RAID0 was 70% faster than JBOD. But >> >> > >> >> JBOD >> >> was >> >> > >> >> faster than RAID1. >> >> > >> >> >> >> > >> >> I have a 2TB drive in this server and was thinking about just >> >> adding >> >> > >> >> another 2TB drive. >> >> > >> >> >> >> > >> >> What are the advantages of JBOD compared to RAID0? From the >> >> > >> >> last >> >> > tests >> >> > >> >> I did, it was slower. >> >> > >> >> >> >> > >> >> Since I will have to re-format the disks anyway, I can re-run >> the >> >> > >> >> tests just in case I did not configured something properly.... >> >> > >> >> >> >> > >> >> JM >> >> > >> >> >> >> > >> >> 2013/2/7, Kevin O'dell <kevin.od...@cloudera.com>: >> >> > >> >> > Hey JM, >> >> > >> >> > >> >> > >> >> > Why RAID0? That has a lot of disadvantages to using a >> >> > >> >> > JBOD >> >> > >> >> > configuration? Wait I/O is a symptom, not a problem. Are >> >> > >> >> > you >> >> > >> actually >> >> > >> >> > experiencing a problem or are you treating for something you >> >> think >> >> > >> >> > should >> >> > >> >> > be lower? >> >> > >> >> > >> >> > >> >> > On Thu, Feb 7, 2013 at 8:19 PM, Jean-Marc Spaggiari >> >> > >> >> > <jean-m...@spaggiari.org >> >> > >> >> >> wrote: >> >> > >> >> > >> >> > >> >> >> Hi, >> >> > >> >> >> >> >> > >> >> >> What is an acceptable CPU_WIO % while running an heavy MR >> job? >> >> > >> >> >> Should >> >> > >> >> >> we also try to keep that under 10%? Or it's not realistic >> >> > >> >> >> and >> >> > >> >> >> we >> >> > >> >> >> will >> >> > >> >> >> see more about 50%? >> >> > >> >> >> >> >> > >> >> >> One of my nodes is showing 70% :( It's WAY to much. I will >> add >> >> > >> another >> >> > >> >> >> disk tomorrow and put them in RAID0, but I'm wondering how >> low >> >> > >> >> >> shoud >> >> > >> I >> >> > >> >> >> go? >> >> > >> >> >> >> >> > >> >> >> JM >> >> > >> >> >> >> >> > >> >> > >> >> > >> >> > >> >> > >> >> > >> >> > >> >> > -- >> >> > >> >> > Kevin O'Dell >> >> > >> >> > Customer Operations Engineer, Cloudera >> >> > >> >> > >> >> > >> >> >> >> > >> > >> >> > >> > >> >> > >> > >> >> > >> > -- >> >> > >> > Kevin O'Dell >> >> > >> > Customer Operations Engineer, Cloudera >> >> > >> > >> >> > >> >> >> > > >> >> > > >> >> > > >> >> > > -- >> >> > > Kevin O'Dell >> >> > > Customer Operations Engineer, Cloudera >> >> > > >> >> > >> >> >> > >> > >> > >> > -- >> > Kevin O'Dell >> > Customer Operations Engineer, Cloudera >> > >> > > > > -- > Kevin O'Dell > Customer Operations Engineer, Cloudera >