Re: RAID0 performance question
- Original Message - From: "Al Boldi" <[EMAIL PROTECTED]> To: "JaniD++" <[EMAIL PROTECTED]> Cc: Sent: Monday, December 19, 2005 12:16 PM Subject: Re: RAID0 performance question > JaniD++ wrote: > > For me, the performance bottleneck is cleanly about RAID0 layer used > > exactly as "concentrator" to join the 4x2TB to 1x8TB. > > Did you try running RAID0 over nbd directly and found it to be faster? Now i trying the NBD + RAID0 without the middle layer raid1. I wondering, the speed is much more better! If i use dd in the raid0 and the raid1 layer is active, the traffic is 350-400Mbit/s If i use dd in the raid0 and the raid1 layer is inactive, the traffic is 512-620Mbit/s! If i use parallel dd on all NBD devices, the traffic is 650-720Mbit/s. (on my system's current minimal load) I found it very interesting! The kernel never reports timing information about raids in /sys/block/mdX/stat files! I cannot understand... Cheers, Janos - To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to [EMAIL PROTECTED] More majordomo info at http://vger.kernel.org/majordomo-info.html
Re: RAID0 performance question
- Original Message - From: "Neil Brown" <[EMAIL PROTECTED]> To: "JaniD++" <[EMAIL PROTECTED]> Cc: Sent: Thursday, December 22, 2005 5:49 AM Subject: Re: RAID0 performance question > On Tuesday November 22, [EMAIL PROTECTED] wrote: > > I have already try the all available options, including readahead in all > > layer (result in earlyer mails), and chunksize. > > But with this settings, i cannot workaround this. > > And the result is incomprehensible for me! > > The raid0 performance is not equal with one component , with sum of all > > component , and not equal with the slowest component! > > This is quite perplexing. > > My next step would probably be to watch the network traffic with > tcpdump or ethereal. I would look for any differences between when it > is going quickly (without raid0) and when slowly (with raid0). > > Rather than tcpdump, it might be easier to instrument the nbd server > to print out requests and timestamps. Yes, it is good idea! I will try it, thanks! > > Sorry I cannot be more helpful, and do have a Merry Christmas anyway > :-) The same to you, and everybody on this list! :-) Cheers, Janos > > NeilBrown > - > To unsubscribe from this list: send the line "unsubscribe linux-raid" in > the body of a message to [EMAIL PROTECTED] > More majordomo info at http://vger.kernel.org/majordomo-info.html - To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to [EMAIL PROTECTED] More majordomo info at http://vger.kernel.org/majordomo-info.html
Re: RAID0 performance question
On Tuesday November 22, [EMAIL PROTECTED] wrote: > I have already try the all available options, including readahead in all > layer (result in earlyer mails), and chunksize. > But with this settings, i cannot workaround this. > And the result is incomprehensible for me! > The raid0 performance is not equal with one component , with sum of all > component , and not equal with the slowest component! This is quite perplexing. My next step would probably be to watch the network traffic with tcpdump or ethereal. I would look for any differences between when it is going quickly (without raid0) and when slowly (with raid0). Rather than tcpdump, it might be easier to instrument the nbd server to print out requests and timestamps. Sorry I cannot be more helpful, and do have a Merry Christmas anyway :-) NeilBrown - To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to [EMAIL PROTECTED] More majordomo info at http://vger.kernel.org/majordomo-info.html
Re: RAID0 performance question
- Original Message - From: "Neil Brown" <[EMAIL PROTECTED]> To: "JaniD++" <[EMAIL PROTECTED]> Cc: "Al Boldi" <[EMAIL PROTECTED]>; Sent: Wednesday, December 21, 2005 2:40 AM Subject: Re: RAID0 performance question > On Sunday December 18, [EMAIL PROTECTED] wrote: > > > > The raid (md) device why dont have scheduler in sysfs? > > And if it have scheduler, where can i tune it? > > raid0 doesn't do any scheduling. > All it does is take requests from the filesystem, decide which device > they should go do (possibly splitting them if needed) and forwarding > them on to the device. That is all. > > > The raid0 can handle multiple requests at one time? > > Yes. But raid0 doesn't exactly 'handle' requests. It 'directs' > requests for other devices to 'handle'. > > > > > For me, the performance bottleneck is cleanly about RAID0 layer used exactly > > as "concentrator" to join the 4x2TB to 1x8TB. > > But it is only a software, and i cant beleave it is unfixable, or > > tunable. > > There is really nothing to tune apart from chunksize. > > You can tune the way the filesystem/vm accesses the device by setting > readahead (readahead on component devices of a raid0 has exactly 0 > effect). First i want to sorry, about "Neil not interested" thing in previous mail... :-( I have already try the all available options, including readahead in all layer (result in earlyer mails), and chunksize. But with this settings, i cannot workaround this. And the result is incomprehensible for me! The raid0 performance is not equal with one component , with sum of all component , and not equal with the slowest component! > > You can tune the underlying devices by choosing a scheduler (for a > disk drive) or a packet size (for over-the-network devices) or > whatever. The NBD has a scheduler, and this is already tuned for really top performance, and for the components it is really great! :-) (I have planned to set the NBD to 4KB packets, but this is hard, becaused by my NICs are not supported the jumbo packets...) > > But there is nothing to tune in raid0. > > > Also, rather than doing measurements on the block devices (/dev/mdX) > do measurements on a filesystem created on that device. > I have often found that the filesystem goes faster than the block > device. I use XFS, and the two performance is almost equal, depends on kind of load. But in most often case, it is almost equal. Thanks, Janos > > > NeilBrown - To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to [EMAIL PROTECTED] More majordomo info at http://vger.kernel.org/majordomo-info.html
Re: RAID0 performance question
On Sunday December 18, [EMAIL PROTECTED] wrote: > > The raid (md) device why dont have scheduler in sysfs? > And if it have scheduler, where can i tune it? raid0 doesn't do any scheduling. All it does is take requests from the filesystem, decide which device they should go do (possibly splitting them if needed) and forwarding them on to the device. That is all. > The raid0 can handle multiple requests at one time? Yes. But raid0 doesn't exactly 'handle' requests. It 'directs' requests for other devices to 'handle'. > > For me, the performance bottleneck is cleanly about RAID0 layer used exactly > as "concentrator" to join the 4x2TB to 1x8TB. > But it is only a software, and i cant beleave it is unfixable, or > tunable. There is really nothing to tune apart from chunksize. You can tune the way the filesystem/vm accesses the device by setting readahead (readahead on component devices of a raid0 has exactly 0 effect). You can tune the underlying devices by choosing a scheduler (for a disk drive) or a packet size (for over-the-network devices) or whatever. But there is nothing to tune in raid0. Also, rather than doing measurements on the block devices (/dev/mdX) do measurements on a filesystem created on that device. I have often found that the filesystem goes faster than the block device. NeilBrown - To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to [EMAIL PROTECTED] More majordomo info at http://vger.kernel.org/majordomo-info.html
Re: RAID0 performance question
- Original Message - From: "Al Boldi" <[EMAIL PROTECTED]> To: "JaniD++" <[EMAIL PROTECTED]> Cc: Sent: Monday, December 19, 2005 12:16 PM Subject: Re: RAID0 performance question > JaniD++ wrote: > > For me, the performance bottleneck is cleanly about RAID0 layer used > > exactly as "concentrator" to join the 4x2TB to 1x8TB. > > Did you try running RAID0 over nbd directly and found it to be faster? At this time i cannot test it, because the system is loaded, and the result will fase. Anyway i will probe this > > IIRC, stacking raid modules does need a considerable amount of tuning, and > even then it does not scale linearly. > > Maybe NeilBrown can help? Maybe, but it looks like Neil is not interested. :-( And the more probles is, i plan to modify the entire system structure, and will not be spare drives to testing soon... :-/ Cheers, Janos - To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to [EMAIL PROTECTED] More majordomo info at http://vger.kernel.org/majordomo-info.html
Re: RAID0 performance question
JaniD++ wrote: > For me, the performance bottleneck is cleanly about RAID0 layer used > exactly as "concentrator" to join the 4x2TB to 1x8TB. Did you try running RAID0 over nbd directly and found it to be faster? IIRC, stacking raid modules does need a considerable amount of tuning, and even then it does not scale linearly. Maybe NeilBrown can help? -- Al - To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to [EMAIL PROTECTED] More majordomo info at http://vger.kernel.org/majordomo-info.html
Re: RAID0 performance question
- Original Message - From: "Al Boldi" <[EMAIL PROTECTED]> To: "JaniD++" <[EMAIL PROTECTED]> Cc: Sent: Friday, December 02, 2005 8:53 PM Subject: Re: RAID0 performance question > JaniD++ wrote: > > > > > > > But the cat /dev/md31 >/dev/null (RAID0, the sum of 4 nodes) > > > > > > > only makes ~450-490 Mbit/s, and i dont know why > > > > > > > > > > > > > > Somebody have an idea? :-) > > > > > > > > > > > > Try increasing the read-ahead setting on /dev/md31 using > > > > > > 'blockdev'. network block devices are likely to have latency > > > > > > issues and would benefit from large read-ahead. > > > > > > > > > > Also try larger chunk-size ~4mb. > > > > But i don't know exactly what to try. > > increase or decrease the chunksize? > > In the top layer raid (md31,raid0) or in the middle layer raids (md1-4, > > raid1) or both? > > > > What I found is that raid over nbd is highly max-chunksize dependent, due to > nbd running over TCP. But increasing chunksize does not necessarily mean > better system utilization. Much depends on your application request size. > > Tuning performance to maximize cat/dd /dev/md# throughput may only be > suitable for a synthetic indication of overall performance in system > comparisons. Yes, you have right! I already know that. ;-) But the bottleneck-effect is visible with dd/cat too. (and i am a litte bit lazy :-) Now i try the system with my spare drives, with the bigger chunk size (=4096K on RAID0 and all RAID1), and the slowness is still here. :( The problem is _exactly_ the same as previously. I think unneccessary to try smaller chunk size, because the 32k is allready small for 2,5,8MB readahead. The problem is somewhere else... :-/ I have got one (or more) question for the raid list! The raid (md) device why dont have scheduler in sysfs? And if it have scheduler, where can i tune it? The raid0 can handle multiple requests at one time? For me, the performance bottleneck is cleanly about RAID0 layer used exactly as "concentrator" to join the 4x2TB to 1x8TB. But it is only a software, and i cant beleave it is unfixable, or tunable. ;-) Cheers, Janos > > If your aim is to increase system utilization, then look for a good benchmark > specific to your application requirements which would mimic a realistic > load. > > -- > Al > > - > To unsubscribe from this list: send the line "unsubscribe linux-raid" in > the body of a message to [EMAIL PROTECTED] > More majordomo info at http://vger.kernel.org/majordomo-info.html - To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to [EMAIL PROTECTED] More majordomo info at http://vger.kernel.org/majordomo-info.html
Re: RAID0 performance question
JaniD++ wrote: > > > > > > But the cat /dev/md31 >/dev/null (RAID0, the sum of 4 nodes) > > > > > > only makes ~450-490 Mbit/s, and i dont know why > > > > > > > > > > > > Somebody have an idea? :-) > > > > > > > > > > Try increasing the read-ahead setting on /dev/md31 using > > > > > 'blockdev'. network block devices are likely to have latency > > > > > issues and would benefit from large read-ahead. > > > > > > > > Also try larger chunk-size ~4mb. > > But i don't know exactly what to try. > increase or decrease the chunksize? > In the top layer raid (md31,raid0) or in the middle layer raids (md1-4, > raid1) or both? > What I found is that raid over nbd is highly max-chunksize dependent, due to nbd running over TCP. But increasing chunksize does not necessarily mean better system utilization. Much depends on your application request size. Tuning performance to maximize cat/dd /dev/md# throughput may only be suitable for a synthetic indication of overall performance in system comparisons. If your aim is to increase system utilization, then look for a good benchmark specific to your application requirements which would mimic a realistic load. -- Al - To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to [EMAIL PROTECTED] More majordomo info at http://vger.kernel.org/majordomo-info.html
Re: RAID0 performance question
Hello, > > > > > But the cat /dev/md31 >/dev/null (RAID0, the sum of 4 nodes) only > > > > > makes ~450-490 Mbit/s, and i dont know why > > > > > > > > > > Somebody have an idea? :-) > > > > > > > > Try increasing the read-ahead setting on /dev/md31 using 'blockdev'. > > > > network block devices are likely to have latency issues and would > > > > benefit from large read-ahead. > > > > > > Also try larger chunk-size ~4mb. > > > > Ahh. > > This is what i can't do. :-( > > I dont know how to backup 8TB! ;-) > > Maybe you could use your mirror!? I have one idea! :-) I can use the spare drives in the disknodes! :-) But i don't know exactly what to try. increase or decrease the chunksize? In the top layer raid (md31,raid0) or in the middle layer raids (md1-4, raid1) or both? Can somebody help me to find the performance problem source? Thanks, Janos > > -- > Al - To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to [EMAIL PROTECTED] More majordomo info at http://vger.kernel.org/majordomo-info.html
Re: RAID0 performance question
- Original Message - From: "Al Boldi" <[EMAIL PROTECTED]> To: "JaniD++" <[EMAIL PROTECTED]> Cc: Sent: Sunday, November 27, 2005 6:40 PM Subject: Re: RAID0 performance question > JaniD++ wrote: > > Al Boldi wrote: > > > Neil Brown wrote: > > > > On Saturday November 26, [EMAIL PROTECTED] wrote: > > > > > The cat /dev/nb# >/dev/nullmakes ~ 350 Mbit/s on each nodes. > > > > > The cat /dev/nb0 + nb1 + nb2 + nb3 in one time parallel makes > > > > > ~780-800 Mbit/s. - i think this is my network bottleneck. > > > > > > How much do you get w/ nb0+1,2,3 and nb0+1+2,3 respectively? > > > > I unable to test write, because this is a productive system. > > Think this value is about 75-80 per node and ~200-250 in md31. > > How much do you get with: > cat nb# + nb# > /dev/null > cat nb# + nb# + nb# > /dev/null > respectively? md1 = 280- 291Mbit md1+md2 = 450-480 Mbit md1+md2+md3 = 615-630 Mbit md1+md2+md3+md4 = now the peak is 674 Mbit ...on a lightly used online system. (~44Mbit download + ~60Mbit upload) This time dd if=/dev/md1 of=/dev/null bs=1M count=4096 what i have used. I think, this is normal. I have try the md31 with different readahead settings: 1. nb0,1,2,3 + md1,2,3,4 readahead = 0 and md31 readahead =4096 result: 380 Mbit 2 all readahead . = 0 result: 88-98 Mbit 3. nb0,1,2,3 + md1,2,3,4 readahead = 2048 and md31 readahead = 0 result: 88-96 Mbit - I wonder! :-O 4. nb# + md# readahead = 0 and md31 readahead = 8192 result: 96-114 Mbit The winner is my default profile :-D all 2048 and md31 = 4096 result : 403-423 Mbit Neil! What do you say? :-) > > > > > > But the cat /dev/md31 >/dev/null (RAID0, the sum of 4 nodes) only > > > > > makes ~450-490 Mbit/s, and i dont know why > > > > > > > > > > Somebody have an idea? :-) > > > > > > > > Try increasing the read-ahead setting on /dev/md31 using 'blockdev'. > > > > network block devices are likely to have latency issues and would > > > > benefit from large read-ahead. > > > > > > Also try larger chunk-size ~4mb. > > > > Ahh. > > This is what i can't do. :-( > > I dont know how to backup 8TB! ;-) > > Maybe you could use your mirror!? There is no mirrors! This is only further options! To be able easy replace, repair one node Thanks Janos > > -- > Al - To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to [EMAIL PROTECTED] More majordomo info at http://vger.kernel.org/majordomo-info.html
Re: RAID0 performance question
JaniD++ wrote: > Al Boldi wrote: > > Neil Brown wrote: > > > On Saturday November 26, [EMAIL PROTECTED] wrote: > > > > The cat /dev/nb# >/dev/nullmakes ~ 350 Mbit/s on each nodes. > > > > The cat /dev/nb0 + nb1 + nb2 + nb3 in one time parallel makes > > > > ~780-800 Mbit/s. - i think this is my network bottleneck. > > > > How much do you get w/ nb0+1,2,3 and nb0+1+2,3 respectively? > > I unable to test write, because this is a productive system. > Think this value is about 75-80 per node and ~200-250 in md31. How much do you get with: cat nb# + nb# > /dev/null cat nb# + nb# + nb# > /dev/null respectively? > > > > But the cat /dev/md31 >/dev/null (RAID0, the sum of 4 nodes) only > > > > makes ~450-490 Mbit/s, and i dont know why > > > > > > > > Somebody have an idea? :-) > > > > > > Try increasing the read-ahead setting on /dev/md31 using 'blockdev'. > > > network block devices are likely to have latency issues and would > > > benefit from large read-ahead. > > > > Also try larger chunk-size ~4mb. > > Ahh. > This is what i can't do. :-( > I dont know how to backup 8TB! ;-) Maybe you could use your mirror!? -- Al - To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to [EMAIL PROTECTED] More majordomo info at http://vger.kernel.org/majordomo-info.html
Re: RAID0 performance question
Hi, - Original Message - From: "Al Boldi" <[EMAIL PROTECTED]> To: "JaniD++" <[EMAIL PROTECTED]> Cc: ; "Neil Brown" <[EMAIL PROTECTED]> Sent: Sunday, November 27, 2005 4:39 PM Subject: Re: RAID0 performance question > Neil Brown wrote: > > On Saturday November 26, [EMAIL PROTECTED] wrote: > > > > > > The cat /dev/nb# >/dev/nullmakes ~ 350 Mbit/s on each nodes. > > Why is this so slow? Yes. :-) This system tries to serv about 500-800 downloads. The final goal is the 1k! ;-) And this point why i am asking the list, because the HW performance is much more without raid0 layer. :-/ The 780-800 Mbit is "almost" enough to 1k downloaders. > Or is this the max node-HD throughput? Yes. > What's the node HW config? P4-3G -HT 12x 200Gb hdd (10 IDE+2 SATA) 2G Ram realtek gige. RAID5 inside! ;-) > > > > The cat /dev/nb0 + nb1 + nb2 + nb3 in one time parallel makes ~ 780-800 > > > Mbit/s. - i think this is my network bottleneck. > > How much do you get w/ nb0+1,2,3 and nb0+1+2,3 respectively? I unable to test write, because this is a productive system. Think this value is about 75-80 per node and ~200-250 in md31. > > > > But the cat /dev/md31 >/dev/null (RAID0, the sum of 4 nodes) only makes > > > ~450-490 Mbit/s, and i dont know why > > > > > > Somebody have an idea? :-) > > > > Try increasing the read-ahead setting on /dev/md31 using 'blockdev'. > > network block devices are likely to have latency issues and would > > benefit from large read-ahead. > > Also try larger chunk-size ~4mb. Ahh. This is what i can't do. :-( I dont know how to backup 8TB! ;-) Thanks, Janos > > -- > Al > > - > To unsubscribe from this list: send the line "unsubscribe linux-raid" in > the body of a message to [EMAIL PROTECTED] > More majordomo info at http://vger.kernel.org/majordomo-info.html - To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to [EMAIL PROTECTED] More majordomo info at http://vger.kernel.org/majordomo-info.html
Re: RAID0 performance question
Neil Brown wrote: > On Saturday November 26, [EMAIL PROTECTED] wrote: > > > > The cat /dev/nb# >/dev/nullmakes ~ 350 Mbit/s on each nodes. Why is this so slow? Or is this the max node-HD throughput? What's the node HW config? > > The cat /dev/nb0 + nb1 + nb2 + nb3 in one time parallel makes ~ 780-800 > > Mbit/s. - i think this is my network bottleneck. How much do you get w/ nb0+1,2,3 and nb0+1+2,3 respectively? > > But the cat /dev/md31 >/dev/null (RAID0, the sum of 4 nodes) only makes > > ~450-490 Mbit/s, and i dont know why > > > > Somebody have an idea? :-) > > Try increasing the read-ahead setting on /dev/md31 using 'blockdev'. > network block devices are likely to have latency issues and would > benefit from large read-ahead. Also try larger chunk-size ~4mb. -- Al - To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to [EMAIL PROTECTED] More majordomo info at http://vger.kernel.org/majordomo-info.html
Re: RAID0 performance question
Hello Neil, This is my init script's one part: blockdev --setra 2048 /dev/nb0 blockdev --setra 2048 /dev/nb1 blockdev --setra 2048 /dev/nb2 blockdev --setra 2048 /dev/nb3 blockdev --setra 2048 /dev/md1 blockdev --setra 2048 /dev/md2 blockdev --setra 2048 /dev/md3 blockdev --setra 2048 /dev/md4 blockdev --setra 4096 /dev/md31 :-) This is the "default" for me. The test is have made with this settings. The problem is somewhere else Thanks Janos - Original Message - From: "Neil Brown" <[EMAIL PROTECTED]> To: "JaniD++" <[EMAIL PROTECTED]> Cc: Sent: Sunday, November 27, 2005 12:27 AM Subject: Re: RAID0 performance question > On Saturday November 26, [EMAIL PROTECTED] wrote: > > Hello list, > > > > I have searching the bottleneck of my system, and found something what i > > cant cleanly understand. > > > > I have use NBD with 4 disk nodes. (raidtab is the bottom of mail) > > > > The cat /dev/nb# >/dev/nullmakes ~ 350 Mbit/s on each nodes. > > The cat /dev/nb0 + nb1 + nb2 + nb3 in one time parallel makes ~ 780-800 > > Mbit/s. - i think this is my network bottleneck. > > > > But the cat /dev/md31 >/dev/null (RAID0, the sum of 4 nodes) only makes > > ~450-490 Mbit/s, and i dont know why > > > > Somebody have an idea? :-) > > Try increasing the read-ahead setting on /dev/md31 using 'blockdev'. > network block devices are likely to have latency issues and would > benefit from large read-ahead. > > NeilBrown > - > To unsubscribe from this list: send the line "unsubscribe linux-raid" in > the body of a message to [EMAIL PROTECTED] > More majordomo info at http://vger.kernel.org/majordomo-info.html - To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to [EMAIL PROTECTED] More majordomo info at http://vger.kernel.org/majordomo-info.html
Re: RAID0 performance question
On Saturday November 26, [EMAIL PROTECTED] wrote: > Hello list, > > I have searching the bottleneck of my system, and found something what i > cant cleanly understand. > > I have use NBD with 4 disk nodes. (raidtab is the bottom of mail) > > The cat /dev/nb# >/dev/nullmakes ~ 350 Mbit/s on each nodes. > The cat /dev/nb0 + nb1 + nb2 + nb3 in one time parallel makes ~ 780-800 > Mbit/s. - i think this is my network bottleneck. > > But the cat /dev/md31 >/dev/null (RAID0, the sum of 4 nodes) only makes > ~450-490 Mbit/s, and i dont know why > > Somebody have an idea? :-) Try increasing the read-ahead setting on /dev/md31 using 'blockdev'. network block devices are likely to have latency issues and would benefit from large read-ahead. NeilBrown - To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to [EMAIL PROTECTED] More majordomo info at http://vger.kernel.org/majordomo-info.html
Re: RAID0 performance question
Hi, If you don't speak hungarian, forget this sentence: Beszelsz magyarul? akkor folytathatjuk ugy is. On Sat, 26 Nov 2005, JaniD++ wrote: > Intel xeon motherboard, intel e1000 x2. (64bit) > But i already write that, if i cut out the raid, and starts the 4 cat at one > time the traffic is rise to 780-800 Mbit! :-) > > This is not hardware related problem. > Only tune, or missconfiguration problem. - I think... What is in the /proc/interrupts? interruts distibuted over cpus, or all irq goes for one cpu? What about, if you switch off HT? Bye, -=Lajbi= LAJBER Zoltan Szent Istvan Egyetem, Informatika Hivatal Most of the time, if you think you are in trouble, crank that throttle! - To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to [EMAIL PROTECTED] More majordomo info at http://vger.kernel.org/majordomo-info.html
Re: RAID0 performance question
Hello, Zoltán! - Original Message - From: "Lajber Zoltan" <[EMAIL PROTECTED]> To: "JaniD++" <[EMAIL PROTECTED]> Cc: Sent: Saturday, November 26, 2005 6:11 PM Subject: Re: RAID0 performance question > On Sat, 26 Nov 2005, JaniD++ wrote: > > > Hello, Raz, > > > > Think this is not cpu usage problem. :-) > > The system is divided to 4 cpuset, and each cpuset uses only one disknode. > > (CPU0->nb0, CPU1->nb1, ...) > > Seams to be CPU problem. Which kind of NIC do you have? Intel xeon motherboard, intel e1000 x2. (64bit) But i already write that, if i cut out the raid, and starts the 4 cat at one time the traffic is rise to 780-800 Mbit! :-) This is not hardware related problem. Only tune, or missconfiguration problem. - I think... > > > CPU2 states: 2.0% user 74.0% system0.0% nice 3.0% iowait 18.0% > > idle > > CPU3 states: 10.0% user 57.0% system0.0% nice 5.0% iowait 26.0% > > Do you have 4 cpu, or 2 HT cpu? 2x HT :-) But in the previous top, it was an used system! Thanks, Janos > > Bye, > -=Lajbi= > LAJBER Zoltan Szent Istvan Egyetem, Informatika Hivatal > Most of the time, if you think you are in trouble, crank that throttle! > - > To unsubscribe from this list: send the line "unsubscribe linux-raid" in > the body of a message to [EMAIL PROTECTED] > More majordomo info at http://vger.kernel.org/majordomo-info.html - To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to [EMAIL PROTECTED] More majordomo info at http://vger.kernel.org/majordomo-info.html
Re: RAID0 performance question
On Sat, 26 Nov 2005, JaniD++ wrote: > Hello, Raz, > > Think this is not cpu usage problem. :-) > The system is divided to 4 cpuset, and each cpuset uses only one disknode. > (CPU0->nb0, CPU1->nb1, ...) Seams to be CPU problem. Which kind of NIC do you have? > CPU2 states: 2.0% user 74.0% system0.0% nice 3.0% iowait 18.0% > idle > CPU3 states: 10.0% user 57.0% system0.0% nice 5.0% iowait 26.0% Do you have 4 cpu, or 2 HT cpu? Bye, -=Lajbi= LAJBER Zoltan Szent Istvan Egyetem, Informatika Hivatal Most of the time, if you think you are in trouble, crank that throttle! - To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to [EMAIL PROTECTED] More majordomo info at http://vger.kernel.org/majordomo-info.html
Re: RAID0 performance question
Hello, Raz, Think this is not cpu usage problem. :-) The system is divided to 4 cpuset, and each cpuset uses only one disknode. (CPU0->nb0, CPU1->nb1, ...) this top is under cat /dev/md31 (raid0) Thanks, Janos 17:16:01 up 14:19, 4 users, load average: 7.74, 5.03, 4.20 305 processes: 301 sleeping, 4 running, 0 zombie, 0 stopped CPU0 states: 33.1% user 47.0% system0.0% nice 0.0% iowait 18.0% idle CPU1 states: 21.0% user 52.0% system0.0% nice 6.0% iowait 19.0% idle CPU2 states: 2.0% user 74.0% system0.0% nice 3.0% iowait 18.0% idle CPU3 states: 10.0% user 57.0% system0.0% nice 5.0% iowait 26.0% idle Mem: 4149412k av, 3961084k used, 188328k free, 0k shrd, 557032k buff 911068k active,2881680k inactive Swap: 0k av, 0k used, 0k free 2779388k cached PID USER PRI NI SIZE RSS SHARE STAT %CPU %MEM TIME CPU COMMAND 2410 root 0 -19 1584 10836 S < 48.3 0.0 21:57 3 nbd-client 16191 root 25 0 4832 820 664 R48.3 0.0 3:04 0 grep 2408 root 0 -19 1588 11236 S < 47.3 0.0 24:05 2 nbd-client 2406 root 0 -19 1584 10836 S < 40.8 0.0 22:56 1 nbd-client 18126 root 18 0 5780 1604 508 D38.0 0.0 0:12 1 dd 2404 root 0 -19 1588 11236 S < 36.2 0.0 22:56 0 nbd-client 294 root 15 0 00 0 SW7.4 0.0 3:22 1 kswapd0 2284 root 16 0 13500 5376 3040 S 7.4 0.1 8:53 2 httpd 18307 root 16 0 6320 2232 1432 S 4.6 0.0 0:00 2 sendmail 16789 root 16 0 5472 1552 952 R 3.7 0.0 0:03 3 top 2431 root 10 -5 00 0 SW< 2.7 0.0 7:32 2 md2_raid1 29076 root 17 0 4776 772 680 S 2.7 0.0 1:09 3 xfs_fsr 6955 root 15 0 1588 10836 S 2.7 0.0 0:56 2 nbd-client - Original Message - From: "Raz Ben-Jehuda(caro)" <[EMAIL PROTECTED]> To: "JaniD++" <[EMAIL PROTECTED]> Cc: Sent: Saturday, November 26, 2005 4:56 PM Subject: Re: RAID0 performance question > look at the cpu consumption. > > On 11/26/05, JaniD++ <[EMAIL PROTECTED]> wrote: > > Hello list, > > > > I have searching the bottleneck of my system, and found something what i > > cant cleanly understand. > > > > I have use NBD with 4 disk nodes. (raidtab is the bottom of mail) > > > > The cat /dev/nb# >/dev/nullmakes ~ 350 Mbit/s on each nodes. > > The cat /dev/nb0 + nb1 + nb2 + nb3 in one time parallel makes ~ 780-800 > > Mbit/s. - i think this is my network bottleneck. > > > > But the cat /dev/md31 >/dev/null (RAID0, the sum of 4 nodes) only makes > > ~450-490 Mbit/s, and i dont know why > > > > Somebody have an idea? :-) > > > > (the nb31,30,29,28 only possible mirrors) > > > > Thanks > > Janos > > > > raiddev /dev/md1 > > raid-level 1 > > nr-raid-disks 2 > > chunk-size 32 > > persistent-superblock 1 > > device /dev/nb0 > > raid-disk 0 > > device /dev/nb31 > > raid-disk 1 > > failed-disk /dev/nb31 > > > > raiddev /dev/md2 > > raid-level 1 > > nr-raid-disks 2 > > chunk-size 32 > > persistent-superblock 1 > > device /dev/nb1 > > raid-disk 0 > > device /dev/hb30 > > raid-disk 1 > > failed-disk /dev/nb30 > > > > raiddev /dev/md3 > > raid-level 1 > > nr-raid-disks 2 > > chunk-size 32 > > persistent-superblock 1 > > device /dev/nb2 > > raid-disk 0 > > device /dev/nb29 > > raid-disk 1 > > failed-disk /dev/nb29 > > > > raiddev /dev/md4 > > raid-level 1 > > nr-raid-disks 2 > > chunk-size 32 > > persistent-superblock 1 > > device /dev/nb3 > > raid-disk 0 > > device /dev/nb28 > > raid-disk 1 > > failed-disk /dev/nb28 > > > > raiddev /dev/md31 > > raid-level 0 > > nr-raid-disks 4 > > chunk-size 32 > > persistent-superblock 1 > > device /dev/md1 > > raid-disk 0 > > device /dev/md2 > > raid-disk 1 > > device /dev/md3 > > raid-disk 2 > > device /dev/md4 > > raid-disk 3 > > > > > > - > > To unsubscribe from this list: send the line "unsubscribe linux-raid" in > > the body of a message to [EMAIL PROTECTED] > > More majordomo info at http://vger.kernel.org/majordomo-info.html > > > > > -- > Raz - To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to [EMAIL PROTECTED] More majordomo info at http://vger.kernel.org/majordomo-info.html
Re: RAID0 performance question
look at the cpu consumption. On 11/26/05, JaniD++ <[EMAIL PROTECTED]> wrote: > Hello list, > > I have searching the bottleneck of my system, and found something what i > cant cleanly understand. > > I have use NBD with 4 disk nodes. (raidtab is the bottom of mail) > > The cat /dev/nb# >/dev/nullmakes ~ 350 Mbit/s on each nodes. > The cat /dev/nb0 + nb1 + nb2 + nb3 in one time parallel makes ~ 780-800 > Mbit/s. - i think this is my network bottleneck. > > But the cat /dev/md31 >/dev/null (RAID0, the sum of 4 nodes) only makes > ~450-490 Mbit/s, and i dont know why > > Somebody have an idea? :-) > > (the nb31,30,29,28 only possible mirrors) > > Thanks > Janos > > raiddev /dev/md1 > raid-level 1 > nr-raid-disks 2 > chunk-size 32 > persistent-superblock 1 > device /dev/nb0 > raid-disk 0 > device /dev/nb31 > raid-disk 1 > failed-disk /dev/nb31 > > raiddev /dev/md2 > raid-level 1 > nr-raid-disks 2 > chunk-size 32 > persistent-superblock 1 > device /dev/nb1 > raid-disk 0 > device /dev/hb30 > raid-disk 1 > failed-disk /dev/nb30 > > raiddev /dev/md3 > raid-level 1 > nr-raid-disks 2 > chunk-size 32 > persistent-superblock 1 > device /dev/nb2 > raid-disk 0 > device /dev/nb29 > raid-disk 1 > failed-disk /dev/nb29 > > raiddev /dev/md4 > raid-level 1 > nr-raid-disks 2 > chunk-size 32 > persistent-superblock 1 > device /dev/nb3 > raid-disk 0 > device /dev/nb28 > raid-disk 1 > failed-disk /dev/nb28 > > raiddev /dev/md31 > raid-level 0 > nr-raid-disks 4 > chunk-size 32 > persistent-superblock 1 > device /dev/md1 > raid-disk 0 > device /dev/md2 > raid-disk 1 > device /dev/md3 > raid-disk 2 > device /dev/md4 > raid-disk 3 > > > - > To unsubscribe from this list: send the line "unsubscribe linux-raid" in > the body of a message to [EMAIL PROTECTED] > More majordomo info at http://vger.kernel.org/majordomo-info.html > -- Raz - To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to [EMAIL PROTECTED] More majordomo info at http://vger.kernel.org/majordomo-info.html