Re: Performance degradation with FFSB between 2.6.20 and 2.6.21-rc7

2007-04-19 Thread Valerie Clement

Jens Axboe wrote:

On Thu, Apr 19 2007, Valerie Clement wrote:

Jens Axboe wrote:

On Thu, Apr 19 2007, Valerie Clement wrote:

Jens Axboe wrote:

Please tell me how you are running ffsb, and also please include a
dmessg from a booted system.


Hi,
our mails crossed! please see my response to Andrew.
You could reproduce the problem with dd command as suggested, it's more 
easy.

I'm sending you the dmesg info. For my tests I used the scsci sdc device.

Thanks, it does. Can you try one thing for me? If you run the test on
sdc, try doing:

# echo 64 > /sys/block/sdc/queue/iosched/quantum

and repeat the test.


OK, that's done.

With the change of quantum, the throughput scores are now a little bit 
better in 2.6.21 than in 2.6.20.


Wonderful, now try the patch I sent in the next mail and repeat the
test.



That's done too.
Now I've got the same results in 2.6.21 as in 2.6.20.
Thanks,

  Valérie

-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/


Re: Performance degradation with FFSB between 2.6.20 and 2.6.21-rc7

2007-04-19 Thread Jens Axboe
On Thu, Apr 19 2007, Valerie Clement wrote:
> Jens Axboe wrote:
> >On Thu, Apr 19 2007, Valerie Clement wrote:
> >>Jens Axboe wrote:
> >>>Please tell me how you are running ffsb, and also please include a
> >>>dmessg from a booted system.
> >>>
> >>Hi,
> >>our mails crossed! please see my response to Andrew.
> >>You could reproduce the problem with dd command as suggested, it's more 
> >>easy.
> >>I'm sending you the dmesg info. For my tests I used the scsci sdc device.
> >
> >Thanks, it does. Can you try one thing for me? If you run the test on
> >sdc, try doing:
> >
> ># echo 64 > /sys/block/sdc/queue/iosched/quantum
> >
> >and repeat the test.
> >
> 
> OK, that's done.
> 
> With the change of quantum, the throughput scores are now a little bit 
> better in 2.6.21 than in 2.6.20.

Wonderful, now try the patch I sent in the next mail and repeat the
test.

-- 
Jens Axboe

-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/


Re: Performance degradation with FFSB between 2.6.20 and 2.6.21-rc7

2007-04-19 Thread Valerie Clement

Jens Axboe wrote:

On Thu, Apr 19 2007, Valerie Clement wrote:

Jens Axboe wrote:

Please tell me how you are running ffsb, and also please include a
dmessg from a booted system.


Hi,
our mails crossed! please see my response to Andrew.
You could reproduce the problem with dd command as suggested, it's more 
easy.

I'm sending you the dmesg info. For my tests I used the scsci sdc device.


Thanks, it does. Can you try one thing for me? If you run the test on
sdc, try doing:

# echo 64 > /sys/block/sdc/queue/iosched/quantum

and repeat the test.



OK, that's done.

With the change of quantum, the throughput scores are now a little bit 
better in 2.6.21 than in 2.6.20.


  Valérie

-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/


Re: Performance degradation with FFSB between 2.6.20 and 2.6.21-rc7

2007-04-19 Thread Jens Axboe
On Thu, Apr 19 2007, Jens Axboe wrote:
> On Thu, Apr 19 2007, Valerie Clement wrote:
> > Jens Axboe wrote:
> > >Please tell me how you are running ffsb, and also please include a
> > >dmessg from a booted system.
> > >
> > Hi,
> > our mails crossed! please see my response to Andrew.
> > You could reproduce the problem with dd command as suggested, it's more 
> > easy.
> > I'm sending you the dmesg info. For my tests I used the scsci sdc device.
> 
> Thanks, it does. Can you try one thing for me? If you run the test on
> sdc, try doing:
> 
> # echo 64 > /sys/block/sdc/queue/iosched/quantum
> 
> and repeat the test.

And, then try this one as well (and don't tweak quantum for that
kernel):

diff --git a/block/cfq-iosched.c b/block/cfq-iosched.c
index b6491c0..9e37971 100644
--- a/block/cfq-iosched.c
+++ b/block/cfq-iosched.c
@@ -986,9 +986,9 @@ __cfq_dispatch_requests(struct cfq_data *cfqd, struct 
cfq_queue *cfqq,
 * expire an async queue immediately if it has used up its slice. idle
 * queue always expire after 1 dispatch round.
 */
-   if ((!cfq_cfqq_sync(cfqq) &&
+   if (cfqd->busy_queues > 1 && ((!cfq_cfqq_sync(cfqq) &&
cfqd->dispatch_slice >= cfq_prio_to_maxrq(cfqd, cfqq)) ||
-   cfq_class_idle(cfqq)) {
+   cfq_class_idle(cfqq))) {
cfqq->slice_end = jiffies + 1;
cfq_slice_expired(cfqd, 0, 0);
}
@@ -1051,19 +1051,21 @@ cfq_dispatch_requests(request_queue_t *q, int force)
while ((cfqq = cfq_select_queue(cfqd)) != NULL) {
int max_dispatch;
 
-   /*
-* Don't repeat dispatch from the previous queue.
-*/
-   if (prev_cfqq == cfqq)
-   break;
+   if (cfqd->busy_queues > 1) {
+   /*
+* Don't repeat dispatch from the previous queue.
+*/
+   if (prev_cfqq == cfqq)
+   break;
 
-   /*
-* So we have dispatched before in this round, if the
-* next queue has idling enabled (must be sync), don't
-* allow it service until the previous have continued.
-*/
-   if (cfqd->rq_in_driver && cfq_cfqq_idle_window(cfqq))
-   break;
+   /*
+* So we have dispatched before in this round, if the
+* next queue has idling enabled (must be sync), don't
+* allow it service until the previous have continued.
+*/
+   if (cfqd->rq_in_driver && cfq_cfqq_idle_window(cfqq))
+   break;
+   }
 
cfq_clear_cfqq_must_dispatch(cfqq);
cfq_clear_cfqq_wait_request(cfqq);
@@ -1370,7 +1372,9 @@ retry:
atomic_set(>ref, 0);
cfqq->cfqd = cfqd;
 
-   cfq_mark_cfqq_idle_window(cfqq);
+   if (key != CFQ_KEY_ASYNC)
+   cfq_mark_cfqq_idle_window(cfqq);
+
cfq_mark_cfqq_prio_changed(cfqq);
cfq_mark_cfqq_queue_new(cfqq);
cfq_init_prio_data(cfqq);

-- 
Jens Axboe

-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/


Re: Performance degradation with FFSB between 2.6.20 and 2.6.21-rc7

2007-04-19 Thread Jens Axboe
On Thu, Apr 19 2007, Valerie Clement wrote:
> Jens Axboe wrote:
> >Please tell me how you are running ffsb, and also please include a
> >dmessg from a booted system.
> >
> Hi,
> our mails crossed! please see my response to Andrew.
> You could reproduce the problem with dd command as suggested, it's more 
> easy.
> I'm sending you the dmesg info. For my tests I used the scsci sdc device.

Thanks, it does. Can you try one thing for me? If you run the test on
sdc, try doing:

# echo 64 > /sys/block/sdc/queue/iosched/quantum

and repeat the test.

-- 
Jens Axboe

-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/


Re: Performance degradation with FFSB between 2.6.20 and 2.6.21-rc7

2007-04-19 Thread Valerie Clement

Jens Axboe wrote:

Please tell me how you are running ffsb, and also please include a
dmessg from a booted system.


Hi,
our mails crossed! please see my response to Andrew.
You could reproduce the problem with dd command as suggested, it's more 
easy.

I'm sending you the dmesg info. For my tests I used the scsci sdc device.

Hope this help,

  Valérie

Linux version 2.6.21-rc7 ([EMAIL PROTECTED]) (gcc version 3.3.5 (Debian 
1:3.3.5-13)) #1 SMP Wed Apr 18 11:39:56 CEST 2007
Command line: console=ttyS0,57600n8 
BIOS-provided physical RAM map:
 BIOS-e820:  - 0009b400 (usable)
 BIOS-e820: 0009b400 - 000a (reserved)
 BIOS-e820: 000cc000 - 000d8000 (reserved)
 BIOS-e820: 000e - 0010 (reserved)
 BIOS-e820: 0010 - 7ff7 (usable)
 BIOS-e820: 7ff7 - 7ff7b000 (ACPI data)
 BIOS-e820: 7ff7b000 - 7ff8 (ACPI NVS)
 BIOS-e820: 7ff8 - 8000 (reserved)
 BIOS-e820: e000 - f000 (reserved)
 BIOS-e820: fec0 - fec1 (reserved)
 BIOS-e820: fee0 - fee01000 (reserved)
 BIOS-e820: ff80 - ffc0 (reserved)
 BIOS-e820: fc00 - 0001 (reserved)
Entering add_active_range(0, 0, 155) 0 entries of 3200 used
Entering add_active_range(0, 256, 524144) 1 entries of 3200 used
end_pfn_map = 1047552
DMI 2.3 present.
ACPI: RSDP 000F6940, 0014 (r0 PTLTD )
ACPI: RSDT 7FF766C5, 0034 (r1 PTLTDRSDT604  LTP0)
ACPI: FACP 7FF7AE94, 0074 (r1 INTEL  LINDHRST  604 PTL 3)
ACPI: DSDT 7FF766F9, 479B (r1  Intel LINDHRST  604 MSFT  10E)
ACPI: FACS 7FF7BFC0, 0040
ACPI: MCFG 7FF7AF08, 0040 (r1 PTLTD  MCFG604  LTP0)
ACPI: APIC 7FF7AF48, 0090 (r1 PTLTD  APIC604  LTP0)
ACPI: BOOT 7FF7AFD8, 0028 (r1 PTLTD  $SBFTBL$  604  LTP1)
No NUMA configuration found
Faking a node at -7ff7
Entering add_active_range(0, 0, 155) 0 entries of 3200 used
Entering add_active_range(0, 256, 524144) 1 entries of 3200 used
Bootmem setup node 0 -7ff7
Zone PFN ranges:
  DMA 0 -> 4096
  DMA324096 ->  1048576
  Normal1048576 ->  1048576
early_node_map[2] active PFN ranges
0:0 ->  155
0:  256 ->   524144
On node 0 totalpages: 524043
  DMA zone: 56 pages used for memmap
  DMA zone: 1829 pages reserved
  DMA zone: 2110 pages, LIFO batch:0
  DMA32 zone: 7110 pages used for memmap
  DMA32 zone: 512938 pages, LIFO batch:31
  Normal zone: 0 pages used for memmap
ACPI: PM-Timer IO Port: 0x1008
ACPI: Local APIC address 0xfee0
ACPI: LAPIC (acpi_id[0x00] lapic_id[0x00] enabled)
Processor #0 (Bootup-CPU)
ACPI: LAPIC (acpi_id[0x01] lapic_id[0x06] enabled)
Processor #6
ACPI: LAPIC (acpi_id[0x02] lapic_id[0x01] enabled)
Processor #1
ACPI: LAPIC (acpi_id[0x03] lapic_id[0x07] enabled)
Processor #7
ACPI: LAPIC_NMI (acpi_id[0x00] high edge lint[0x1])
ACPI: LAPIC_NMI (acpi_id[0x01] high edge lint[0x1])
ACPI: LAPIC_NMI (acpi_id[0x02] high edge lint[0x1])
ACPI: LAPIC_NMI (acpi_id[0x03] high edge lint[0x1])
ACPI: IOAPIC (id[0x02] address[0xfec0] gsi_base[0])
IOAPIC[0]: apic_id 2, address 0xfec0, GSI 0-23
ACPI: IOAPIC (id[0x03] address[0xfec1] gsi_base[24])
IOAPIC[1]: apic_id 3, address 0xfec1, GSI 24-47
ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 high edge)
ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level)
ACPI: IRQ0 used by override.
ACPI: IRQ2 used by override.
ACPI: IRQ9 used by override.
Setting APIC routing to physical flat
Using ACPI (MADT) for SMP configuration information
Nosave address range: 0009b000 - 0009c000
Nosave address range: 0009c000 - 000a
Nosave address range: 000a - 000cc000
Nosave address range: 000cc000 - 000d8000
Nosave address range: 000d8000 - 000e
Nosave address range: 000e - 0010
Allocating PCI resources starting at 8800 (gap: 8000:6000)
SMP: Allowing 4 CPUs, 0 hotplug CPUs
PERCPU: Allocating 37760 bytes of per cpu data
Built 1 zonelists.  Total pages: 515048
Kernel command line: console=ttyS0,57600n8 
Initializing CPU#0
PID hash table entries: 4096 (order: 12, 32768 bytes)
time.c: Detected 2793.059 MHz processor.
Console: colour VGA+ 80x25
Dentry cache hash table entries: 262144 (order: 9, 2097152 bytes)
Inode-cache hash table entries: 131072 (order: 8, 1048576 bytes)
Checking aperture...
Memory: 2056596k/2096576k available (3816k kernel code, 39576k reserved, 2214k 
data, 336k init)
Calibrating delay using timer specific routine.. 5590.47 BogoMIPS (lpj=11180941)
Mount-cache hash table entries: 256
CPU: Trace cache: 12K uops, L1 D cache: 16K
CPU: L2 cache: 1024K
CPU 0/0 -> Node 0
using mwait in idle threads.
CPU: Physical Processor ID: 0
CPU: Processor Core 

Re: Performance degradation with FFSB between 2.6.20 and 2.6.21-rc7

2007-04-19 Thread Valerie Clement

Andrew Morton wrote:

It could be due to I/O scheduler changes.  Which one are you using?  CFQ?

Or it could be that there has been some changed behaviour at the VFS/pagecache
layer: the VFS might be submitting little hunks of lots of files, rather than
large hunks of few files.

Or it could be a block-layer thing: perhaps some driver change has caused
us to be placing less data into the queue.  Which device driver is that machine
using?

Being a simple soul, the first thing I'll try when I get near a test box
will be

for i in $(seq 1 16)
do
time dd if=/dev/zero of=$i bs=1M count=1024 &
done



I tried first the test with dd, the results are similar to those of FFSB 
tests, about 15 percent of degradation between 2.6.20.7 and 2.6.21-rc7.


I'm using the CFQ I/O scheduler. I changed it to the "deadline" one and 
I don't have any more the problem, I've got similar throughput values 
with 2.6.20.7 and 2.6.21-rc7 kernels.

So can we conclude that it's due to the CFQ scheduler?

I also checked the device driver used, the revision number is the same 
in 2.6.20 and 2.6.21.


  Valérie

-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/


Re: Performance degradation with FFSB between 2.6.20 and 2.6.21-rc7

2007-04-19 Thread Jens Axboe
On Wed, Apr 18 2007, Valerie Clement wrote:
> 
> Running benchmark tests (FFSB) on an ext4 filesystem, I noticed a 
> performance degradation (about 15-20 percent) in sequential write tests 
> between 2.6.19-rc6 and 2.6.21-rc4 kernels.
> 
> I ran the same tests on ext3 and XFS filesystems and I saw the same 
> performance difference between the two kernel versions for these two 
> filesystems.
> 
> I have also reproduced it between 2.6.20.7 and 2.6.21-rc7.
> The FFSB tests run 16 threads, each creating 1GB files. The tests were 
> done on the same x86_64 system, with the same kernel configuration and 
> on the same scsi device. Below are the throughput values given by FFSB.
> 
>   kernel   XFSext3
> --
>  2.6.20.748 MB/sec 44 MB/sec
> 
>  2.6.21-rc7  38 MB/sec 37 MB/sec
> 
> Did anyone else run across the problem?
> Is there a known issue?

Please tell me how you are running ffsb, and also please include a
dmessg from a booted system.

-- 
Jens Axboe

-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/


Re: Performance degradation with FFSB between 2.6.20 and 2.6.21-rc7

2007-04-19 Thread Jens Axboe
On Wed, Apr 18 2007, Valerie Clement wrote:
 
 Running benchmark tests (FFSB) on an ext4 filesystem, I noticed a 
 performance degradation (about 15-20 percent) in sequential write tests 
 between 2.6.19-rc6 and 2.6.21-rc4 kernels.
 
 I ran the same tests on ext3 and XFS filesystems and I saw the same 
 performance difference between the two kernel versions for these two 
 filesystems.
 
 I have also reproduced it between 2.6.20.7 and 2.6.21-rc7.
 The FFSB tests run 16 threads, each creating 1GB files. The tests were 
 done on the same x86_64 system, with the same kernel configuration and 
 on the same scsi device. Below are the throughput values given by FFSB.
 
   kernel   XFSext3
 --
  2.6.20.748 MB/sec 44 MB/sec
 
  2.6.21-rc7  38 MB/sec 37 MB/sec
 
 Did anyone else run across the problem?
 Is there a known issue?

Please tell me how you are running ffsb, and also please include a
dmessg from a booted system.

-- 
Jens Axboe

-
To unsubscribe from this list: send the line unsubscribe linux-kernel in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/


Re: Performance degradation with FFSB between 2.6.20 and 2.6.21-rc7

2007-04-19 Thread Valerie Clement

Andrew Morton wrote:

It could be due to I/O scheduler changes.  Which one are you using?  CFQ?

Or it could be that there has been some changed behaviour at the VFS/pagecache
layer: the VFS might be submitting little hunks of lots of files, rather than
large hunks of few files.

Or it could be a block-layer thing: perhaps some driver change has caused
us to be placing less data into the queue.  Which device driver is that machine
using?

Being a simple soul, the first thing I'll try when I get near a test box
will be

for i in $(seq 1 16)
do
time dd if=/dev/zero of=$i bs=1M count=1024 
done



I tried first the test with dd, the results are similar to those of FFSB 
tests, about 15 percent of degradation between 2.6.20.7 and 2.6.21-rc7.


I'm using the CFQ I/O scheduler. I changed it to the deadline one and 
I don't have any more the problem, I've got similar throughput values 
with 2.6.20.7 and 2.6.21-rc7 kernels.

So can we conclude that it's due to the CFQ scheduler?

I also checked the device driver used, the revision number is the same 
in 2.6.20 and 2.6.21.


  Valérie

-
To unsubscribe from this list: send the line unsubscribe linux-kernel in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/


Re: Performance degradation with FFSB between 2.6.20 and 2.6.21-rc7

2007-04-19 Thread Valerie Clement

Jens Axboe wrote:

Please tell me how you are running ffsb, and also please include a
dmessg from a booted system.


Hi,
our mails crossed! please see my response to Andrew.
You could reproduce the problem with dd command as suggested, it's more 
easy.

I'm sending you the dmesg info. For my tests I used the scsci sdc device.

Hope this help,

  Valérie

Linux version 2.6.21-rc7 ([EMAIL PROTECTED]) (gcc version 3.3.5 (Debian 
1:3.3.5-13)) #1 SMP Wed Apr 18 11:39:56 CEST 2007
Command line: console=ttyS0,57600n8 
BIOS-provided physical RAM map:
 BIOS-e820:  - 0009b400 (usable)
 BIOS-e820: 0009b400 - 000a (reserved)
 BIOS-e820: 000cc000 - 000d8000 (reserved)
 BIOS-e820: 000e - 0010 (reserved)
 BIOS-e820: 0010 - 7ff7 (usable)
 BIOS-e820: 7ff7 - 7ff7b000 (ACPI data)
 BIOS-e820: 7ff7b000 - 7ff8 (ACPI NVS)
 BIOS-e820: 7ff8 - 8000 (reserved)
 BIOS-e820: e000 - f000 (reserved)
 BIOS-e820: fec0 - fec1 (reserved)
 BIOS-e820: fee0 - fee01000 (reserved)
 BIOS-e820: ff80 - ffc0 (reserved)
 BIOS-e820: fc00 - 0001 (reserved)
Entering add_active_range(0, 0, 155) 0 entries of 3200 used
Entering add_active_range(0, 256, 524144) 1 entries of 3200 used
end_pfn_map = 1047552
DMI 2.3 present.
ACPI: RSDP 000F6940, 0014 (r0 PTLTD )
ACPI: RSDT 7FF766C5, 0034 (r1 PTLTDRSDT604  LTP0)
ACPI: FACP 7FF7AE94, 0074 (r1 INTEL  LINDHRST  604 PTL 3)
ACPI: DSDT 7FF766F9, 479B (r1  Intel LINDHRST  604 MSFT  10E)
ACPI: FACS 7FF7BFC0, 0040
ACPI: MCFG 7FF7AF08, 0040 (r1 PTLTD  MCFG604  LTP0)
ACPI: APIC 7FF7AF48, 0090 (r1 PTLTD  APIC604  LTP0)
ACPI: BOOT 7FF7AFD8, 0028 (r1 PTLTD  $SBFTBL$  604  LTP1)
No NUMA configuration found
Faking a node at -7ff7
Entering add_active_range(0, 0, 155) 0 entries of 3200 used
Entering add_active_range(0, 256, 524144) 1 entries of 3200 used
Bootmem setup node 0 -7ff7
Zone PFN ranges:
  DMA 0 - 4096
  DMA324096 -  1048576
  Normal1048576 -  1048576
early_node_map[2] active PFN ranges
0:0 -  155
0:  256 -   524144
On node 0 totalpages: 524043
  DMA zone: 56 pages used for memmap
  DMA zone: 1829 pages reserved
  DMA zone: 2110 pages, LIFO batch:0
  DMA32 zone: 7110 pages used for memmap
  DMA32 zone: 512938 pages, LIFO batch:31
  Normal zone: 0 pages used for memmap
ACPI: PM-Timer IO Port: 0x1008
ACPI: Local APIC address 0xfee0
ACPI: LAPIC (acpi_id[0x00] lapic_id[0x00] enabled)
Processor #0 (Bootup-CPU)
ACPI: LAPIC (acpi_id[0x01] lapic_id[0x06] enabled)
Processor #6
ACPI: LAPIC (acpi_id[0x02] lapic_id[0x01] enabled)
Processor #1
ACPI: LAPIC (acpi_id[0x03] lapic_id[0x07] enabled)
Processor #7
ACPI: LAPIC_NMI (acpi_id[0x00] high edge lint[0x1])
ACPI: LAPIC_NMI (acpi_id[0x01] high edge lint[0x1])
ACPI: LAPIC_NMI (acpi_id[0x02] high edge lint[0x1])
ACPI: LAPIC_NMI (acpi_id[0x03] high edge lint[0x1])
ACPI: IOAPIC (id[0x02] address[0xfec0] gsi_base[0])
IOAPIC[0]: apic_id 2, address 0xfec0, GSI 0-23
ACPI: IOAPIC (id[0x03] address[0xfec1] gsi_base[24])
IOAPIC[1]: apic_id 3, address 0xfec1, GSI 24-47
ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 high edge)
ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level)
ACPI: IRQ0 used by override.
ACPI: IRQ2 used by override.
ACPI: IRQ9 used by override.
Setting APIC routing to physical flat
Using ACPI (MADT) for SMP configuration information
Nosave address range: 0009b000 - 0009c000
Nosave address range: 0009c000 - 000a
Nosave address range: 000a - 000cc000
Nosave address range: 000cc000 - 000d8000
Nosave address range: 000d8000 - 000e
Nosave address range: 000e - 0010
Allocating PCI resources starting at 8800 (gap: 8000:6000)
SMP: Allowing 4 CPUs, 0 hotplug CPUs
PERCPU: Allocating 37760 bytes of per cpu data
Built 1 zonelists.  Total pages: 515048
Kernel command line: console=ttyS0,57600n8 
Initializing CPU#0
PID hash table entries: 4096 (order: 12, 32768 bytes)
time.c: Detected 2793.059 MHz processor.
Console: colour VGA+ 80x25
Dentry cache hash table entries: 262144 (order: 9, 2097152 bytes)
Inode-cache hash table entries: 131072 (order: 8, 1048576 bytes)
Checking aperture...
Memory: 2056596k/2096576k available (3816k kernel code, 39576k reserved, 2214k 
data, 336k init)
Calibrating delay using timer specific routine.. 5590.47 BogoMIPS (lpj=11180941)
Mount-cache hash table entries: 256
CPU: Trace cache: 12K uops, L1 D cache: 16K
CPU: L2 cache: 1024K
CPU 0/0 - Node 0
using mwait in idle threads.
CPU: Physical Processor ID: 0
CPU: Processor Core ID: 0

Re: Performance degradation with FFSB between 2.6.20 and 2.6.21-rc7

2007-04-19 Thread Jens Axboe
On Thu, Apr 19 2007, Valerie Clement wrote:
 Jens Axboe wrote:
 Please tell me how you are running ffsb, and also please include a
 dmessg from a booted system.
 
 Hi,
 our mails crossed! please see my response to Andrew.
 You could reproduce the problem with dd command as suggested, it's more 
 easy.
 I'm sending you the dmesg info. For my tests I used the scsci sdc device.

Thanks, it does. Can you try one thing for me? If you run the test on
sdc, try doing:

# echo 64  /sys/block/sdc/queue/iosched/quantum

and repeat the test.

-- 
Jens Axboe

-
To unsubscribe from this list: send the line unsubscribe linux-kernel in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/


Re: Performance degradation with FFSB between 2.6.20 and 2.6.21-rc7

2007-04-19 Thread Jens Axboe
On Thu, Apr 19 2007, Jens Axboe wrote:
 On Thu, Apr 19 2007, Valerie Clement wrote:
  Jens Axboe wrote:
  Please tell me how you are running ffsb, and also please include a
  dmessg from a booted system.
  
  Hi,
  our mails crossed! please see my response to Andrew.
  You could reproduce the problem with dd command as suggested, it's more 
  easy.
  I'm sending you the dmesg info. For my tests I used the scsci sdc device.
 
 Thanks, it does. Can you try one thing for me? If you run the test on
 sdc, try doing:
 
 # echo 64  /sys/block/sdc/queue/iosched/quantum
 
 and repeat the test.

And, then try this one as well (and don't tweak quantum for that
kernel):

diff --git a/block/cfq-iosched.c b/block/cfq-iosched.c
index b6491c0..9e37971 100644
--- a/block/cfq-iosched.c
+++ b/block/cfq-iosched.c
@@ -986,9 +986,9 @@ __cfq_dispatch_requests(struct cfq_data *cfqd, struct 
cfq_queue *cfqq,
 * expire an async queue immediately if it has used up its slice. idle
 * queue always expire after 1 dispatch round.
 */
-   if ((!cfq_cfqq_sync(cfqq) 
+   if (cfqd-busy_queues  1  ((!cfq_cfqq_sync(cfqq) 
cfqd-dispatch_slice = cfq_prio_to_maxrq(cfqd, cfqq)) ||
-   cfq_class_idle(cfqq)) {
+   cfq_class_idle(cfqq))) {
cfqq-slice_end = jiffies + 1;
cfq_slice_expired(cfqd, 0, 0);
}
@@ -1051,19 +1051,21 @@ cfq_dispatch_requests(request_queue_t *q, int force)
while ((cfqq = cfq_select_queue(cfqd)) != NULL) {
int max_dispatch;
 
-   /*
-* Don't repeat dispatch from the previous queue.
-*/
-   if (prev_cfqq == cfqq)
-   break;
+   if (cfqd-busy_queues  1) {
+   /*
+* Don't repeat dispatch from the previous queue.
+*/
+   if (prev_cfqq == cfqq)
+   break;
 
-   /*
-* So we have dispatched before in this round, if the
-* next queue has idling enabled (must be sync), don't
-* allow it service until the previous have continued.
-*/
-   if (cfqd-rq_in_driver  cfq_cfqq_idle_window(cfqq))
-   break;
+   /*
+* So we have dispatched before in this round, if the
+* next queue has idling enabled (must be sync), don't
+* allow it service until the previous have continued.
+*/
+   if (cfqd-rq_in_driver  cfq_cfqq_idle_window(cfqq))
+   break;
+   }
 
cfq_clear_cfqq_must_dispatch(cfqq);
cfq_clear_cfqq_wait_request(cfqq);
@@ -1370,7 +1372,9 @@ retry:
atomic_set(cfqq-ref, 0);
cfqq-cfqd = cfqd;
 
-   cfq_mark_cfqq_idle_window(cfqq);
+   if (key != CFQ_KEY_ASYNC)
+   cfq_mark_cfqq_idle_window(cfqq);
+
cfq_mark_cfqq_prio_changed(cfqq);
cfq_mark_cfqq_queue_new(cfqq);
cfq_init_prio_data(cfqq);

-- 
Jens Axboe

-
To unsubscribe from this list: send the line unsubscribe linux-kernel in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/


Re: Performance degradation with FFSB between 2.6.20 and 2.6.21-rc7

2007-04-19 Thread Valerie Clement

Jens Axboe wrote:

On Thu, Apr 19 2007, Valerie Clement wrote:

Jens Axboe wrote:

Please tell me how you are running ffsb, and also please include a
dmessg from a booted system.


Hi,
our mails crossed! please see my response to Andrew.
You could reproduce the problem with dd command as suggested, it's more 
easy.

I'm sending you the dmesg info. For my tests I used the scsci sdc device.


Thanks, it does. Can you try one thing for me? If you run the test on
sdc, try doing:

# echo 64  /sys/block/sdc/queue/iosched/quantum

and repeat the test.



OK, that's done.

With the change of quantum, the throughput scores are now a little bit 
better in 2.6.21 than in 2.6.20.


  Valérie

-
To unsubscribe from this list: send the line unsubscribe linux-kernel in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/


Re: Performance degradation with FFSB between 2.6.20 and 2.6.21-rc7

2007-04-19 Thread Jens Axboe
On Thu, Apr 19 2007, Valerie Clement wrote:
 Jens Axboe wrote:
 On Thu, Apr 19 2007, Valerie Clement wrote:
 Jens Axboe wrote:
 Please tell me how you are running ffsb, and also please include a
 dmessg from a booted system.
 
 Hi,
 our mails crossed! please see my response to Andrew.
 You could reproduce the problem with dd command as suggested, it's more 
 easy.
 I'm sending you the dmesg info. For my tests I used the scsci sdc device.
 
 Thanks, it does. Can you try one thing for me? If you run the test on
 sdc, try doing:
 
 # echo 64  /sys/block/sdc/queue/iosched/quantum
 
 and repeat the test.
 
 
 OK, that's done.
 
 With the change of quantum, the throughput scores are now a little bit 
 better in 2.6.21 than in 2.6.20.

Wonderful, now try the patch I sent in the next mail and repeat the
test.

-- 
Jens Axboe

-
To unsubscribe from this list: send the line unsubscribe linux-kernel in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/


Re: Performance degradation with FFSB between 2.6.20 and 2.6.21-rc7

2007-04-19 Thread Valerie Clement

Jens Axboe wrote:

On Thu, Apr 19 2007, Valerie Clement wrote:

Jens Axboe wrote:

On Thu, Apr 19 2007, Valerie Clement wrote:

Jens Axboe wrote:

Please tell me how you are running ffsb, and also please include a
dmessg from a booted system.


Hi,
our mails crossed! please see my response to Andrew.
You could reproduce the problem with dd command as suggested, it's more 
easy.

I'm sending you the dmesg info. For my tests I used the scsci sdc device.

Thanks, it does. Can you try one thing for me? If you run the test on
sdc, try doing:

# echo 64  /sys/block/sdc/queue/iosched/quantum

and repeat the test.


OK, that's done.

With the change of quantum, the throughput scores are now a little bit 
better in 2.6.21 than in 2.6.20.


Wonderful, now try the patch I sent in the next mail and repeat the
test.



That's done too.
Now I've got the same results in 2.6.21 as in 2.6.20.
Thanks,

  Valérie

-
To unsubscribe from this list: send the line unsubscribe linux-kernel in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/


Re: Performance degradation with FFSB between 2.6.20 and 2.6.21-rc7

2007-04-18 Thread Andrew Morton
> On Wed, 18 Apr 2007 15:54:00 +0200 Valerie Clement <[EMAIL PROTECTED]> wrote:
> 
> Running benchmark tests (FFSB) on an ext4 filesystem, I noticed a 
> performance degradation (about 15-20 percent) in sequential write tests 
> between 2.6.19-rc6 and 2.6.21-rc4 kernels.
> 
> I ran the same tests on ext3 and XFS filesystems and I saw the same 
> performance difference between the two kernel versions for these two 
> filesystems.
> 
> I have also reproduced it between 2.6.20.7 and 2.6.21-rc7.
> The FFSB tests run 16 threads, each creating 1GB files. The tests were 
> done on the same x86_64 system, with the same kernel configuration and 
> on the same scsi device. Below are the throughput values given by FFSB.
> 
>kernel   XFSext3
> --
>   2.6.20.748 MB/sec 44 MB/sec
> 
>   2.6.21-rc7  38 MB/sec 37 MB/sec
> 
> Did anyone else run across the problem?
> Is there a known issue?
> 

That's a new discovery, thanks.

It could be due to I/O scheduler changes.  Which one are you using?  CFQ?

Or it could be that there has been some changed behaviour at the VFS/pagecache
layer: the VFS might be submitting little hunks of lots of files, rather than
large hunks of few files.

Or it could be a block-layer thing: perhaps some driver change has caused
us to be placing less data into the queue.  Which device driver is that machine
using?

Being a simple soul, the first thing I'll try when I get near a test box
will be

for i in $(seq 1 16)
do
time dd if=/dev/zero of=$i bs=1M count=1024 &
done
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/


Performance degradation with FFSB between 2.6.20 and 2.6.21-rc7

2007-04-18 Thread Valerie Clement


Running benchmark tests (FFSB) on an ext4 filesystem, I noticed a 
performance degradation (about 15-20 percent) in sequential write tests 
between 2.6.19-rc6 and 2.6.21-rc4 kernels.


I ran the same tests on ext3 and XFS filesystems and I saw the same 
performance difference between the two kernel versions for these two 
filesystems.


I have also reproduced it between 2.6.20.7 and 2.6.21-rc7.
The FFSB tests run 16 threads, each creating 1GB files. The tests were 
done on the same x86_64 system, with the same kernel configuration and 
on the same scsi device. Below are the throughput values given by FFSB.


  kernel   XFSext3
--
 2.6.20.748 MB/sec 44 MB/sec

 2.6.21-rc7  38 MB/sec 37 MB/sec

Did anyone else run across the problem?
Is there a known issue?

   Valérie


(in attachment, my kernel configuration file)
#
# Automatically generated make config: don't edit
# Linux kernel version: 2.6.21-rc7
# Wed Apr 18 11:29:53 2007
#
CONFIG_X86_64=y
CONFIG_64BIT=y
CONFIG_X86=y
CONFIG_GENERIC_TIME=y
CONFIG_GENERIC_TIME_VSYSCALL=y
CONFIG_ZONE_DMA32=y
CONFIG_LOCKDEP_SUPPORT=y
CONFIG_STACKTRACE_SUPPORT=y
CONFIG_SEMAPHORE_SLEEPERS=y
CONFIG_MMU=y
CONFIG_ZONE_DMA=y
CONFIG_RWSEM_GENERIC_SPINLOCK=y
CONFIG_GENERIC_HWEIGHT=y
CONFIG_GENERIC_CALIBRATE_DELAY=y
CONFIG_X86_CMPXCHG=y
CONFIG_EARLY_PRINTK=y
CONFIG_GENERIC_ISA_DMA=y
CONFIG_GENERIC_IOMAP=y
CONFIG_ARCH_MAY_HAVE_PC_FDC=y
CONFIG_ARCH_POPULATES_NODE_MAP=y
CONFIG_DMI=y
CONFIG_AUDIT_ARCH=y
CONFIG_GENERIC_BUG=y
# CONFIG_ARCH_HAS_ILOG2_U32 is not set
# CONFIG_ARCH_HAS_ILOG2_U64 is not set
CONFIG_DEFCONFIG_LIST="/lib/modules/$UNAME_RELEASE/.config"

#
# Code maturity level options
#
CONFIG_EXPERIMENTAL=y
CONFIG_LOCK_KERNEL=y
CONFIG_INIT_ENV_ARG_LIMIT=32

#
# General setup
#
CONFIG_LOCALVERSION=""
CONFIG_LOCALVERSION_AUTO=y
CONFIG_SWAP=y
CONFIG_SYSVIPC=y
# CONFIG_IPC_NS is not set
CONFIG_SYSVIPC_SYSCTL=y
CONFIG_POSIX_MQUEUE=y
# CONFIG_BSD_PROCESS_ACCT is not set
# CONFIG_TASKSTATS is not set
# CONFIG_UTS_NS is not set
# CONFIG_AUDIT is not set
CONFIG_IKCONFIG=y
CONFIG_IKCONFIG_PROC=y
# CONFIG_CPUSETS is not set
CONFIG_SYSFS_DEPRECATED=y
# CONFIG_RELAY is not set
CONFIG_BLK_DEV_INITRD=y
CONFIG_INITRAMFS_SOURCE=""
CONFIG_CC_OPTIMIZE_FOR_SIZE=y
CONFIG_SYSCTL=y
# CONFIG_EMBEDDED is not set
CONFIG_UID16=y
CONFIG_SYSCTL_SYSCALL=y
CONFIG_KALLSYMS=y
CONFIG_KALLSYMS_ALL=y
# CONFIG_KALLSYMS_EXTRA_PASS is not set
CONFIG_HOTPLUG=y
CONFIG_PRINTK=y
CONFIG_BUG=y
CONFIG_ELF_CORE=y
CONFIG_BASE_FULL=y
CONFIG_FUTEX=y
CONFIG_EPOLL=y
CONFIG_SHMEM=y
CONFIG_SLAB=y
CONFIG_VM_EVENT_COUNTERS=y
CONFIG_RT_MUTEXES=y
# CONFIG_TINY_SHMEM is not set
CONFIG_BASE_SMALL=0
# CONFIG_SLOB is not set

#
# Loadable module support
#
CONFIG_MODULES=y
CONFIG_MODULE_UNLOAD=y
CONFIG_MODULE_FORCE_UNLOAD=y
# CONFIG_MODVERSIONS is not set
# CONFIG_MODULE_SRCVERSION_ALL is not set
# CONFIG_KMOD is not set
CONFIG_STOP_MACHINE=y

#
# Block layer
#
CONFIG_BLOCK=y
# CONFIG_BLK_DEV_IO_TRACE is not set

#
# IO Schedulers
#
CONFIG_IOSCHED_NOOP=y
# CONFIG_IOSCHED_AS is not set
CONFIG_IOSCHED_DEADLINE=y
CONFIG_IOSCHED_CFQ=y
# CONFIG_DEFAULT_AS is not set
# CONFIG_DEFAULT_DEADLINE is not set
CONFIG_DEFAULT_CFQ=y
# CONFIG_DEFAULT_NOOP is not set
CONFIG_DEFAULT_IOSCHED="cfq"

#
# Processor type and features
#
CONFIG_X86_PC=y
# CONFIG_X86_VSMP is not set
# CONFIG_MK8 is not set
# CONFIG_MPSC is not set
# CONFIG_MCORE2 is not set
CONFIG_GENERIC_CPU=y
CONFIG_X86_L1_CACHE_BYTES=128
CONFIG_X86_L1_CACHE_SHIFT=7
CONFIG_X86_INTERNODE_CACHE_BYTES=128
CONFIG_X86_TSC=y
CONFIG_X86_GOOD_APIC=y
# CONFIG_MICROCODE is not set
CONFIG_X86_MSR=y
CONFIG_X86_CPUID=y
CONFIG_X86_HT=y
CONFIG_X86_IO_APIC=y
CONFIG_X86_LOCAL_APIC=y
CONFIG_MTRR=y
CONFIG_SMP=y
CONFIG_SCHED_SMT=y
CONFIG_SCHED_MC=y
# CONFIG_PREEMPT_NONE is not set
CONFIG_PREEMPT_VOLUNTARY=y
# CONFIG_PREEMPT is not set
CONFIG_PREEMPT_BKL=y
CONFIG_NUMA=y
CONFIG_K8_NUMA=y
CONFIG_NODES_SHIFT=6
CONFIG_X86_64_ACPI_NUMA=y
CONFIG_NUMA_EMU=y
CONFIG_ARCH_DISCONTIGMEM_ENABLE=y
CONFIG_ARCH_DISCONTIGMEM_DEFAULT=y
CONFIG_ARCH_SPARSEMEM_ENABLE=y
CONFIG_SELECT_MEMORY_MODEL=y
# CONFIG_FLATMEM_MANUAL is not set
CONFIG_DISCONTIGMEM_MANUAL=y
# CONFIG_SPARSEMEM_MANUAL is not set
CONFIG_DISCONTIGMEM=y
CONFIG_FLAT_NODE_MEM_MAP=y
CONFIG_NEED_MULTIPLE_NODES=y
# CONFIG_SPARSEMEM_STATIC is not set
CONFIG_SPLIT_PTLOCK_CPUS=4
CONFIG_MIGRATION=y
CONFIG_RESOURCES_64BIT=y
CONFIG_ZONE_DMA_FLAG=1
CONFIG_HAVE_ARCH_EARLY_PFN_TO_NID=y
CONFIG_OUT_OF_LINE_PFN_TO_PAGE=y
CONFIG_NR_CPUS=32
CONFIG_HOTPLUG_CPU=y
CONFIG_ARCH_ENABLE_MEMORY_HOTPLUG=y
CONFIG_HPET_TIMER=y
CONFIG_HPET_EMULATE_RTC=y
CONFIG_IOMMU=y
# CONFIG_CALGARY_IOMMU is not set
CONFIG_SWIOTLB=y
CONFIG_X86_MCE=y
CONFIG_X86_MCE_INTEL=y
CONFIG_X86_MCE_AMD=y
# CONFIG_KEXEC is not set
# CONFIG_CRASH_DUMP is not set
CONFIG_PHYSICAL_START=0x20
CONFIG_SECCOMP=y
# CONFIG_CC_STACKPROTECTOR is not set
# CONFIG_HZ_100 is not set
CONFIG_HZ_250=y
# CONFIG_HZ_300 is not set
# CONFIG_HZ_1000 is not set
CONFIG_HZ=250
# 

Performance degradation with FFSB between 2.6.20 and 2.6.21-rc7

2007-04-18 Thread Valerie Clement


Running benchmark tests (FFSB) on an ext4 filesystem, I noticed a 
performance degradation (about 15-20 percent) in sequential write tests 
between 2.6.19-rc6 and 2.6.21-rc4 kernels.


I ran the same tests on ext3 and XFS filesystems and I saw the same 
performance difference between the two kernel versions for these two 
filesystems.


I have also reproduced it between 2.6.20.7 and 2.6.21-rc7.
The FFSB tests run 16 threads, each creating 1GB files. The tests were 
done on the same x86_64 system, with the same kernel configuration and 
on the same scsi device. Below are the throughput values given by FFSB.


  kernel   XFSext3
--
 2.6.20.748 MB/sec 44 MB/sec

 2.6.21-rc7  38 MB/sec 37 MB/sec

Did anyone else run across the problem?
Is there a known issue?

   Valérie


(in attachment, my kernel configuration file)
#
# Automatically generated make config: don't edit
# Linux kernel version: 2.6.21-rc7
# Wed Apr 18 11:29:53 2007
#
CONFIG_X86_64=y
CONFIG_64BIT=y
CONFIG_X86=y
CONFIG_GENERIC_TIME=y
CONFIG_GENERIC_TIME_VSYSCALL=y
CONFIG_ZONE_DMA32=y
CONFIG_LOCKDEP_SUPPORT=y
CONFIG_STACKTRACE_SUPPORT=y
CONFIG_SEMAPHORE_SLEEPERS=y
CONFIG_MMU=y
CONFIG_ZONE_DMA=y
CONFIG_RWSEM_GENERIC_SPINLOCK=y
CONFIG_GENERIC_HWEIGHT=y
CONFIG_GENERIC_CALIBRATE_DELAY=y
CONFIG_X86_CMPXCHG=y
CONFIG_EARLY_PRINTK=y
CONFIG_GENERIC_ISA_DMA=y
CONFIG_GENERIC_IOMAP=y
CONFIG_ARCH_MAY_HAVE_PC_FDC=y
CONFIG_ARCH_POPULATES_NODE_MAP=y
CONFIG_DMI=y
CONFIG_AUDIT_ARCH=y
CONFIG_GENERIC_BUG=y
# CONFIG_ARCH_HAS_ILOG2_U32 is not set
# CONFIG_ARCH_HAS_ILOG2_U64 is not set
CONFIG_DEFCONFIG_LIST=/lib/modules/$UNAME_RELEASE/.config

#
# Code maturity level options
#
CONFIG_EXPERIMENTAL=y
CONFIG_LOCK_KERNEL=y
CONFIG_INIT_ENV_ARG_LIMIT=32

#
# General setup
#
CONFIG_LOCALVERSION=
CONFIG_LOCALVERSION_AUTO=y
CONFIG_SWAP=y
CONFIG_SYSVIPC=y
# CONFIG_IPC_NS is not set
CONFIG_SYSVIPC_SYSCTL=y
CONFIG_POSIX_MQUEUE=y
# CONFIG_BSD_PROCESS_ACCT is not set
# CONFIG_TASKSTATS is not set
# CONFIG_UTS_NS is not set
# CONFIG_AUDIT is not set
CONFIG_IKCONFIG=y
CONFIG_IKCONFIG_PROC=y
# CONFIG_CPUSETS is not set
CONFIG_SYSFS_DEPRECATED=y
# CONFIG_RELAY is not set
CONFIG_BLK_DEV_INITRD=y
CONFIG_INITRAMFS_SOURCE=
CONFIG_CC_OPTIMIZE_FOR_SIZE=y
CONFIG_SYSCTL=y
# CONFIG_EMBEDDED is not set
CONFIG_UID16=y
CONFIG_SYSCTL_SYSCALL=y
CONFIG_KALLSYMS=y
CONFIG_KALLSYMS_ALL=y
# CONFIG_KALLSYMS_EXTRA_PASS is not set
CONFIG_HOTPLUG=y
CONFIG_PRINTK=y
CONFIG_BUG=y
CONFIG_ELF_CORE=y
CONFIG_BASE_FULL=y
CONFIG_FUTEX=y
CONFIG_EPOLL=y
CONFIG_SHMEM=y
CONFIG_SLAB=y
CONFIG_VM_EVENT_COUNTERS=y
CONFIG_RT_MUTEXES=y
# CONFIG_TINY_SHMEM is not set
CONFIG_BASE_SMALL=0
# CONFIG_SLOB is not set

#
# Loadable module support
#
CONFIG_MODULES=y
CONFIG_MODULE_UNLOAD=y
CONFIG_MODULE_FORCE_UNLOAD=y
# CONFIG_MODVERSIONS is not set
# CONFIG_MODULE_SRCVERSION_ALL is not set
# CONFIG_KMOD is not set
CONFIG_STOP_MACHINE=y

#
# Block layer
#
CONFIG_BLOCK=y
# CONFIG_BLK_DEV_IO_TRACE is not set

#
# IO Schedulers
#
CONFIG_IOSCHED_NOOP=y
# CONFIG_IOSCHED_AS is not set
CONFIG_IOSCHED_DEADLINE=y
CONFIG_IOSCHED_CFQ=y
# CONFIG_DEFAULT_AS is not set
# CONFIG_DEFAULT_DEADLINE is not set
CONFIG_DEFAULT_CFQ=y
# CONFIG_DEFAULT_NOOP is not set
CONFIG_DEFAULT_IOSCHED=cfq

#
# Processor type and features
#
CONFIG_X86_PC=y
# CONFIG_X86_VSMP is not set
# CONFIG_MK8 is not set
# CONFIG_MPSC is not set
# CONFIG_MCORE2 is not set
CONFIG_GENERIC_CPU=y
CONFIG_X86_L1_CACHE_BYTES=128
CONFIG_X86_L1_CACHE_SHIFT=7
CONFIG_X86_INTERNODE_CACHE_BYTES=128
CONFIG_X86_TSC=y
CONFIG_X86_GOOD_APIC=y
# CONFIG_MICROCODE is not set
CONFIG_X86_MSR=y
CONFIG_X86_CPUID=y
CONFIG_X86_HT=y
CONFIG_X86_IO_APIC=y
CONFIG_X86_LOCAL_APIC=y
CONFIG_MTRR=y
CONFIG_SMP=y
CONFIG_SCHED_SMT=y
CONFIG_SCHED_MC=y
# CONFIG_PREEMPT_NONE is not set
CONFIG_PREEMPT_VOLUNTARY=y
# CONFIG_PREEMPT is not set
CONFIG_PREEMPT_BKL=y
CONFIG_NUMA=y
CONFIG_K8_NUMA=y
CONFIG_NODES_SHIFT=6
CONFIG_X86_64_ACPI_NUMA=y
CONFIG_NUMA_EMU=y
CONFIG_ARCH_DISCONTIGMEM_ENABLE=y
CONFIG_ARCH_DISCONTIGMEM_DEFAULT=y
CONFIG_ARCH_SPARSEMEM_ENABLE=y
CONFIG_SELECT_MEMORY_MODEL=y
# CONFIG_FLATMEM_MANUAL is not set
CONFIG_DISCONTIGMEM_MANUAL=y
# CONFIG_SPARSEMEM_MANUAL is not set
CONFIG_DISCONTIGMEM=y
CONFIG_FLAT_NODE_MEM_MAP=y
CONFIG_NEED_MULTIPLE_NODES=y
# CONFIG_SPARSEMEM_STATIC is not set
CONFIG_SPLIT_PTLOCK_CPUS=4
CONFIG_MIGRATION=y
CONFIG_RESOURCES_64BIT=y
CONFIG_ZONE_DMA_FLAG=1
CONFIG_HAVE_ARCH_EARLY_PFN_TO_NID=y
CONFIG_OUT_OF_LINE_PFN_TO_PAGE=y
CONFIG_NR_CPUS=32
CONFIG_HOTPLUG_CPU=y
CONFIG_ARCH_ENABLE_MEMORY_HOTPLUG=y
CONFIG_HPET_TIMER=y
CONFIG_HPET_EMULATE_RTC=y
CONFIG_IOMMU=y
# CONFIG_CALGARY_IOMMU is not set
CONFIG_SWIOTLB=y
CONFIG_X86_MCE=y
CONFIG_X86_MCE_INTEL=y
CONFIG_X86_MCE_AMD=y
# CONFIG_KEXEC is not set
# CONFIG_CRASH_DUMP is not set
CONFIG_PHYSICAL_START=0x20
CONFIG_SECCOMP=y
# CONFIG_CC_STACKPROTECTOR is not set
# CONFIG_HZ_100 is not set
CONFIG_HZ_250=y
# CONFIG_HZ_300 is not set
# CONFIG_HZ_1000 is not set
CONFIG_HZ=250
# CONFIG_REORDER is 

Re: Performance degradation with FFSB between 2.6.20 and 2.6.21-rc7

2007-04-18 Thread Andrew Morton
 On Wed, 18 Apr 2007 15:54:00 +0200 Valerie Clement [EMAIL PROTECTED] wrote:
 
 Running benchmark tests (FFSB) on an ext4 filesystem, I noticed a 
 performance degradation (about 15-20 percent) in sequential write tests 
 between 2.6.19-rc6 and 2.6.21-rc4 kernels.
 
 I ran the same tests on ext3 and XFS filesystems and I saw the same 
 performance difference between the two kernel versions for these two 
 filesystems.
 
 I have also reproduced it between 2.6.20.7 and 2.6.21-rc7.
 The FFSB tests run 16 threads, each creating 1GB files. The tests were 
 done on the same x86_64 system, with the same kernel configuration and 
 on the same scsi device. Below are the throughput values given by FFSB.
 
kernel   XFSext3
 --
   2.6.20.748 MB/sec 44 MB/sec
 
   2.6.21-rc7  38 MB/sec 37 MB/sec
 
 Did anyone else run across the problem?
 Is there a known issue?
 

That's a new discovery, thanks.

It could be due to I/O scheduler changes.  Which one are you using?  CFQ?

Or it could be that there has been some changed behaviour at the VFS/pagecache
layer: the VFS might be submitting little hunks of lots of files, rather than
large hunks of few files.

Or it could be a block-layer thing: perhaps some driver change has caused
us to be placing less data into the queue.  Which device driver is that machine
using?

Being a simple soul, the first thing I'll try when I get near a test box
will be

for i in $(seq 1 16)
do
time dd if=/dev/zero of=$i bs=1M count=1024 
done
-
To unsubscribe from this list: send the line unsubscribe linux-kernel in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/