By the way,
there are more than fifty bugs logged for marevell88sx, many of them about
problems with DMA handling and/or driver behaviour under stress.
Can it be that I'm stumbling upon something along these lines?
http://bugs.opensolaris.org/bugdatabase/view_bug.do?bug_id=6826483
Maurilio.
Richard,
thanks for the explanation.
So can we say that the problem is in the disks loosing a command now and then
under stress?
Best regards.
Maurilio.
--
This message posted from opensolaris.org
___
zfs-discuss mailing list
On Oct 4, 2009, at 11:52 PM, Maurilio Longo wrote:
Richard,
thanks for the explanation.
So can we say that the problem is in the disks loosing a command now
and then under stress?
It may be the disks or the HBA. I'll bet a steak dinner it is the HBA.
-- richard
Richard,
it is the same controller used inside Sun's thumpers; It could be a problem in
my unit (which is a couple of years old now), though.
Is there something I can do to find out if I owe you that steak? :)
Thanks.
Maurilio.
--
This message posted from opensolaris.org
Hi,
I have a pc with a MARVELL AOC-SAT2-MV8 controller and a pool made up of a six
disks in a raid-z pool with a hot spare.
pre
-bash-3.2$ /sbin/zpool status
pool: nas
stato: ONLINE
scrub: scrub in progress for 9h4m, 81,59% done, 2h2m to go
config:
NAMESTATE READ WRITE
Maurilio Longo wrote:
Hi,
I have a pc with a MARVELL AOC-SAT2-MV8 controller and a pool made up of a
six disks in a raid-z pool with a hot spare.
...
Now, the problem is that issuing an
iostat -Cmnx 10
or any other time intervall, I've seen, sometimes, a complete stall of disk
I/O due to a
Carson,
the strange thing is that this is happening on several disks (can it be that
are all failing?)
What is the controller bug you're talking about? I'm running snv_114 on this
pc, so it is fairly recent.
Best regards.
Maurilio.
--
This message posted from opensolaris.org
Maurilio Longo wrote:
the strange thing is that this is happening on several disks (can it be that
are all failing?)
Possible, but less likely. I'd suggest running some disk I/O tests, looking at
the drive error counters before/after.
What is the controller bug you're talking about? I'm
Possible, but less likely. I'd suggest running some
disk I/O tests, looking at
the drive error counters before/after.
These disks have a few months of life and are scrubbed weekly, no errors so far.
I did try to use smartmontools, but it cannot report SMART logs nor start SMART
tests, so
Maurilio Longo wrote:
I did try to use smartmontools, but it cannot report SMART logs nor start
SMART tests, so I don't know how to look at their internal state.
Really? That's odd...
You could also have a firmware bug on your disks. You might try lowering
the number of tagged commands per
Maurilio Longo wrote:
Carson,
the strange thing is that this is happening on several disks (can it be that
are all failing?)
What is the controller bug you're talking about? I'm running snv_114 on this
pc, so it is fairly recent.
Best regards.
Maurilio.
See 'iostat -En' output.
Milek,
this is it
pre
# iostat -En
c1t0d0 Soft Errors: 0 Hard Errors: 0 Transport Errors: 0
Vendor: ATA Product: ST3808110AS Revision: DSerial No:
Size: 80,03GB 80026361856 bytes
Media Error: 0 Device Not Ready: 0 No Device: 0 Recoverable: 0
Illegal Request: 91 Predictive
Errata,
they're ST31000333AS and not 340AS
Maurilio.
--
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
For the archives...
On Oct 2, 2009, at 12:41 AM, Maurilio Longo wrote:
Hi,
I have a pc with a MARVELL AOC-SAT2-MV8 controller and a pool made
up of a six disks in a raid-z pool with a hot spare.
pre
-bash-3.2$ /sbin/zpool status
pool: nas
stato: ONLINE
scrub: scrub in progress for 9h4m,
14 matches
Mail list logo