Re: Architecture usertags

2009-03-03 Thread dann frazier
On Tue, Mar 03, 2009 at 10:03:18PM +0200, Vasilios Karaklioumis wrote:
> Wouter Verhelst wrote:
>> Hi,
>>
>> I've often thought that it would be useful to have tags in the BTS so
>> that users or maintainers could mark a bug as specific to a particular
>> architecture. This way, when I have some spare time, I could go to the
>> BTS, fetch a list of bugs that are specific to an architecture I care
>> about, and see what I can do about them, or give some feedback if that
>> would help.
>>
>> Using a set of usertags under my own email address on the BTS wouldn't
>> really work for that, since it kindof defeats the whole purpose; I want
>> to use these tags to find bugs that I care about in the first place, but
>> in order for me to be able to add a usertag, I would have to know about
>> the bug before I could find it. Kind of a chicken-and-egg problem.
>>
>> So I suggested to Don Armstrong that he add a set of
>> architecture-specific regular tags; but he seemed averse to this, as the
>> current informal policy of the debbugs maintainers is to require that a
>> usertag is used for a particular purpose before they add a new regular
>> tag; this is so that no tags get created which won't be used. I guess
>> this is a worty goal.
>>
>> After a short discussion on IRC, we came up with another option: a set
>> of publically documented usertags, the definition of which would be
>> announced on debian-devel-announce and linked to from the BTS homepage,
>> so that maintainers could apply them to architecture-specific bugs when
>> necessary. The format, suggested by Steve Langasek, was to use the
>> porters mailinglist as the user, and the architecture name as the
>> usertag (e.g., 'debian-m...@lists.debian.org' as user, and 'm68k' as
>> tag).
>>
>> Before I'll fire off an email to d-d-a announcing that, does anyone have
>> any comments, objections, or suggestions to improve this proposal?
>>
>> Thanks,
>>
>>   
> Excellent news.This will come very handy.Thanks.

+1

-- 
dann frazier


-- 
To UNSUBSCRIBE, email to debian-amd64-requ...@lists.debian.org
with a subject of "unsubscribe". Trouble? Contact listmas...@lists.debian.org



[tutorial] kde 4.3 over SVN on debian sid

2009-03-03 Thread Ghost Kilah

I have maked a tutorial about how to install kde 4.3 on debian sid over svn.

Link : 
http://www.unixcod.org/component/option,com_fireboard/Itemid,83/func,view/catid,21/id,88/




Regards,
Vina Petre Jan
http://www.unixcod.org


--
To UNSUBSCRIBE, email to debian-amd64-requ...@lists.debian.org
with a subject of "unsubscribe". Trouble? Contact listmas...@lists.debian.org



Re: raid1 issue, somewhat related to recent "debian on big machines"

2009-03-03 Thread Lennart Sorensen
On Tue, Mar 03, 2009 at 10:59:50PM +0100, Francesco Pietra wrote:
> I understand that the double recommendation is fine. Though, I am
> pressed by answering the referees about a submitted paper as they
> requested additional computation. That was going on until the host
> suspended access to sda. As I find risky to go on with one disk only
> (for a many days computation), could you please explain how to
> reactivate the removed sda, or format it to try if it recovers? I made
> some proposals in previous post. Or indicated that the best is
> replacing the disk with a new one.

You can simply ask mdadm to readd it back and let it rebuild, but likely
the error will happen again and you will need to replace the disk.

If you replace the disk (with a disk at least as big as the old one),
then copy the partition table from the working drive and reread the
partition table (hdparm -z /dev/sda can do that for you) and finally
readd to the various raids again.  You do not need to create a filesystem,
since the filesystem runs on the raid, not the individual partitions,
and hence you already have your filesystems made.

-- 
Len Sorensen


-- 
To UNSUBSCRIBE, email to debian-amd64-requ...@lists.debian.org
with a subject of "unsubscribe". Trouble? Contact listmas...@lists.debian.org



Re: raid1 issue, somewhat related to recent "debian on big machines"

2009-03-03 Thread Francesco Pietra
On Tue, Mar 3, 2009 at 9:52 PM, Alex Samad  wrote:
> On Tue, Mar 03, 2009 at 12:26:27PM +0100, Goswin von Brederlow wrote:
>> Francesco Pietra  writes:
> [snip]
>
>>
>> That is a lot of raids. Have you ever thought about using LVM? The
>> different raid1 will mess up each others assumption about the head
>> positioning of the component devices. On read the linux kernel tries
>> to use the disk with the shorter seek and assumes the head is where it
>> left it on the last access. But if one of the other raids used that
>> disk the head will be way off.
>>
>> I would suggest the following scheme:
>
> this is what I would recommend as well

I understand that the double recommendation is fine. Though, I am
pressed by answering the referees about a submitted paper as they
requested additional computation. That was going on until the host
suspended access to sda. As I find risky to go on with one disk only
(for a many days computation), could you please explain how to
reactivate the removed sda, or format it to try if it recovers? I made
some proposals in previous post. Or indicated that the best is
replacing the disk with a new one.

Thanks
francesco


>>
>> sda1 / sdb1 : 100Mb raid1 for /boot (or 1GB for / + /boot)
>> sda2 / sdb2 : rest raid1 with lvm
>>
>> MfG
>>         Goswin
>>
>>
>> --
>> To UNSUBSCRIBE, email to debian-amd64-requ...@lists.debian.org
>> with a subject of "unsubscribe". Trouble? Contact listmas...@lists.debian.org
>>
>>
>
> --
> "Perhaps one way will be, if we use military force, in the post-Saddam Iraq 
> the U.N. will definitely need to have a role. And that way it can begin to 
> get its legs, legs of responsibility back."
>
>        - George W. Bush
> 03/16/2003
> the Azores, Portugal
>
> -BEGIN PGP SIGNATURE-
> Version: GnuPG v1.4.9 (GNU/Linux)
>
> iEYEARECAAYFAkmtmHwACgkQkZz88chpJ2OpcACgmoZ41ASQaImVnKgcXiovFAya
> DKwAnA4YwO7GWaL4QHnx02mSnAQdgmSM
> =SAtK
> -END PGP SIGNATURE-
>
>


--
To UNSUBSCRIBE, email to debian-amd64-requ...@lists.debian.org
with a subject of "unsubscribe". Trouble? Contact listmas...@lists.debian.org



Re: blacklist doesn't blacklist !!

2009-03-03 Thread Gilles Mocellin
On Tue, Mar 03, 2009 at 05:44:11PM -0300, macdowell@dpf.gov.br wrote:
> Hi
> 
> 
> I put ssd in my blacklist but after boot there it comes again.
> 
> m...@hp:~$ cat /etc/modprobe.d/blacklist | grep ssb
> blacklist ssb
> m...@hp:~$ lsmod | grep ssb
> ssb43140  0
> pcmcia 38680  1 ssb
> pcmcia_core41508  2 ssb,pcmcia
> m...@hp:~$
> 
> 
> Why ?

Perhaps it's loaded during boot in the initrd.

Try rebuild your initrd with :
# update-initramfs -u


signature.asc
Description: Digital signature


Re: raid1 issue, somewhat related to recent "debian on big machines"

2009-03-03 Thread Alex Samad
On Tue, Mar 03, 2009 at 12:26:27PM +0100, Goswin von Brederlow wrote:
> Francesco Pietra  writes:
[snip]

> 
> That is a lot of raids. Have you ever thought about using LVM? The
> different raid1 will mess up each others assumption about the head
> positioning of the component devices. On read the linux kernel tries
> to use the disk with the shorter seek and assumes the head is where it
> left it on the last access. But if one of the other raids used that
> disk the head will be way off.
> 
> I would suggest the following scheme:

this is what I would recommend as well
> 
> sda1 / sdb1 : 100Mb raid1 for /boot (or 1GB for / + /boot)
> sda2 / sdb2 : rest raid1 with lvm
> 
> MfG
> Goswin
> 
> 
> -- 
> To UNSUBSCRIBE, email to debian-amd64-requ...@lists.debian.org
> with a subject of "unsubscribe". Trouble? Contact listmas...@lists.debian.org
> 
> 

-- 
"Perhaps one way will be, if we use military force, in the post-Saddam Iraq the 
U.N. will definitely need to have a role. And that way it can begin to get its 
legs, legs of responsibility back."

- George W. Bush
03/16/2003
the Azores, Portugal


signature.asc
Description: Digital signature


blacklist doesn't blacklist !!

2009-03-03 Thread macdowell . smd
Hi


I put ssd in my blacklist but after boot there it comes again.

m...@hp:~$ cat /etc/modprobe.d/blacklist | grep ssb
blacklist ssb
m...@hp:~$ lsmod | grep ssb
ssb43140  0
pcmcia 38680  1 ssb
pcmcia_core41508  2 ssb,pcmcia
m...@hp:~$


Why ?

thank you for help

s macdowell




-- 
To UNSUBSCRIBE, email to debian-amd64-requ...@lists.debian.org
with a subject of "unsubscribe". Trouble? Contact listmas...@lists.debian.org



Re: Architecture usertags

2009-03-03 Thread Vasilios Karaklioumis

Wouter Verhelst wrote:

Hi,

I've often thought that it would be useful to have tags in the BTS so
that users or maintainers could mark a bug as specific to a particular
architecture. This way, when I have some spare time, I could go to the
BTS, fetch a list of bugs that are specific to an architecture I care
about, and see what I can do about them, or give some feedback if that
would help.

Using a set of usertags under my own email address on the BTS wouldn't
really work for that, since it kindof defeats the whole purpose; I want
to use these tags to find bugs that I care about in the first place, but
in order for me to be able to add a usertag, I would have to know about
the bug before I could find it. Kind of a chicken-and-egg problem.

So I suggested to Don Armstrong that he add a set of
architecture-specific regular tags; but he seemed averse to this, as the
current informal policy of the debbugs maintainers is to require that a
usertag is used for a particular purpose before they add a new regular
tag; this is so that no tags get created which won't be used. I guess
this is a worty goal.

After a short discussion on IRC, we came up with another option: a set
of publically documented usertags, the definition of which would be
announced on debian-devel-announce and linked to from the BTS homepage,
so that maintainers could apply them to architecture-specific bugs when
necessary. The format, suggested by Steve Langasek, was to use the
porters mailinglist as the user, and the architecture name as the
usertag (e.g., 'debian-m...@lists.debian.org' as user, and 'm68k' as
tag).

Before I'll fire off an email to d-d-a announcing that, does anyone have
any comments, objections, or suggestions to improve this proposal?

Thanks,

  

Excellent news.This will come very handy.Thanks.


--
To UNSUBSCRIBE, email to debian-amd64-requ...@lists.debian.org
with a subject of "unsubscribe". Trouble? Contact listmas...@lists.debian.org



Architecture usertags

2009-03-03 Thread Wouter Verhelst
Hi,

I've often thought that it would be useful to have tags in the BTS so
that users or maintainers could mark a bug as specific to a particular
architecture. This way, when I have some spare time, I could go to the
BTS, fetch a list of bugs that are specific to an architecture I care
about, and see what I can do about them, or give some feedback if that
would help.

Using a set of usertags under my own email address on the BTS wouldn't
really work for that, since it kindof defeats the whole purpose; I want
to use these tags to find bugs that I care about in the first place, but
in order for me to be able to add a usertag, I would have to know about
the bug before I could find it. Kind of a chicken-and-egg problem.

So I suggested to Don Armstrong that he add a set of
architecture-specific regular tags; but he seemed averse to this, as the
current informal policy of the debbugs maintainers is to require that a
usertag is used for a particular purpose before they add a new regular
tag; this is so that no tags get created which won't be used. I guess
this is a worty goal.

After a short discussion on IRC, we came up with another option: a set
of publically documented usertags, the definition of which would be
announced on debian-devel-announce and linked to from the BTS homepage,
so that maintainers could apply them to architecture-specific bugs when
necessary. The format, suggested by Steve Langasek, was to use the
porters mailinglist as the user, and the architecture name as the
usertag (e.g., 'debian-m...@lists.debian.org' as user, and 'm68k' as
tag).

Before I'll fire off an email to d-d-a announcing that, does anyone have
any comments, objections, or suggestions to improve this proposal?

Thanks,

-- 
 Home is where you have to wash the dishes.
  -- #debian-devel, Freenode, 2004-09-22


signature.asc
Description: Digital signature


Re: raid1 issue, somewhat related to recent "debian on big machines"

2009-03-03 Thread Francesco Pietra
It came to mind how to check the raid. Below the key tests. I am not
sure if the disk removed by the system (sda) has to be replaced with a
new one or it could be forced into place, tried, reformatted, and what
else should be better tried. If to be replaced, it would be the second
time in a less than 2000 hours service, while there is a good UPS and
the disks are mounted on Zahlman coolers with a strong fan directly on
them.

Unfortunately I have a badly planned huge raid.  If trying to reuse
sda, is it correct to command directly:

# sfdisk -d /dev/sdb > /tmp/partitions

# sfdisk --force /dev/sda < /tmp/partitions

# mdadm -add /dev/md0 /dev/sdb1

# mdadm -add /dev/md1 /dev/sdb2

# mdadm -add /dev/md2 /dev/sdb3

# mdadm -add /dev/md3 /dev/sdb5

# mdadm -add /dev/md4 /dev/sdb6

# mdadm -add /dev/md5 /dev/sdb7

# mdadm -add /dev/md6 /dev/sdb8


or should sda be first ext3 formatted?

=

cat /proc/mdstat


Personalities : [raid1]
md6 : active raid1 sdb8[1]
  102341952 blocks [2/1] [_U]

md5 : active raid1 sdb7[1]
  1951744 blocks [2/1] [_U]

md4 : active raid1 sdb6[1]
  2931712 blocks [2/1] [_U]

md3 : active raid1 sdb5[1]
  14651136 blocks [2/1] [_U]

md1 : active (auto-read-only) raid1 sdb2[1]
  6835584 blocks [2/1] [_U]

md0 : active raid1 sdb1[1]
  2931712 blocks [2/1] [_U]

md2 : active raid1 sdb3[1]
  14651200 blocks [2/1] [_U]

unused devices: 


mdadm --detail /dev/md0

/dev/md0:
Version : 00.90
  Creation Time : Wed Sep 13 16:50:11 2006
 Raid Level : raid1
 Array Size : 2931712 (2.80 GiB 3.00 GB)
  Used Dev Size : 2931712 (2.80 GiB 3.00 GB)
   Raid Devices : 2
  Total Devices : 1
Preferred Minor : 0
Persistence : Superblock is persistent

Update Time : Tue Mar  3 19:05:21 2009
  State : clean, degraded
 Active Devices : 1
Working Devices : 1
 Failed Devices : 0
  Spare Devices : 0

   UUID : 5afe70f3:fa1dcb8c:5cc73e84:1c5db3e0
 Events : 0.912

Number   Major   Minor   RaidDevice State
   0   000  removed
   1   8   171  active sync   /dev/sdb1
==
Similar situation with all other raids
==



smartctl --all /dev/sda (the disk removed):

smartctl version 5.38 [x86_64-unknown-linux-gnu] Copyright (C) 2002-8
Bruce Allen
Home page is http://smartmontools.sourceforge.net/

=== START OF INFORMATION SECTION ===
Model Family: Western Digital Raptor family
Device Model: WDC WD1500ADFD-00NLR1
Serial Number:WD-WMAP41185084
Firmware Version: 20.07P20
User Capacity:150,039,945,216 bytes
Device is:In smartctl database [for details use: -P show]
ATA Version is:   7
ATA Standard is:  ATA/ATAPI-7 published, ANSI INCITS 397-2005
Local Time is:Tue Mar  3 19:41:15 2009 CET
SMART support is: Available - device has SMART capability.
SMART support is: Enabled

=== START OF READ SMART DATA SECTION ===
SMART overall-health self-assessment test result: PASSED

General SMART Values:
Offline data collection status:  (0x84) Offline data collection activity
was suspended by an interrupting 
command from host.
Auto Offline Data Collection: Enabled.
Self-test execution status:  (   0) The previous self-test routine completed
without error or no self-test has ever
been run.
Total time to complete Offline
data collection: (4783) seconds.
Offline data collection
capabilities:(0x7b) SMART execute Offline immediate.
Auto Offline data collection on/off 
support.
Suspend Offline collection upon new
command.
Offline surface scan supported.
Self-test supported.
Conveyance Self-test supported.
Selective Self-test supported.
SMART capabilities:(0x0003) Saves SMART data before entering
power-saving mode.
Supports SMART auto save timer.
Error logging capability:(0x01) Error logging supported.
General Purpose Logging supported.
Short self-test routine
recommended polling time:(   2) minutes.
Extended self-test routine
recommended polling time:(  72) minutes.
Conveyance self-test routine
recommended polling time:(   5) minutes.
SCT capabilities:  (0x103f) SCT Status supported.
   

Life - Un salto nella tua filiera

2009-03-03 Thread Gruppo Life-City Srl | Bellaria
To view the message, please use an HTML compatible email viewer!



Fwd: raid1 issue, somewhat related to recent "debian on big machines"

2009-03-03 Thread Francesco Pietra
How to run selftest with raid1?

smartctl -t /dev/sda (or sda1, sda2, ..)

is incorrect. I don't remember what should first be done with mdadm.

I would like to run the test because

smartctl --all /dev/sd#

report errors for only sda, as follows:

SMART Error Log Version: 1
ATA Error Count: 12 (device log contains only the most recent five errors)
CR = Command Register [HEX]
FR = Features Register [HEX]
SC = Sector Count Register [HEX]
SN = Sector Number Register [HEX]
CL = Cylinder Low Register [HEX]
CH = Cylinder High Register [HEX]
DH = Device/Head Register [HEX]
DC = Device Command Register [HEX]
ER = Error register [HEX]
ST = Status register [HEX]
Powered_Up_Time is measured from power on, and printed as
DDd+hh:mm:SS.sss where DD=days, hh=hours, mm=minutes,
SS=sec, and sss=millisec. It "wraps" after 49.710 days.

Error 12 occurred at disk power-on lifetime: 1940 hours (80 days + 20 hours)
  When the command that caused the error occurred, the device was
active or idle.

  After command completion occurred, registers were:
  ER ST SC SN CL CH DH
  -- -- -- -- -- -- --
  40 51 08 4a 9d 9b ec  Error: UNC 8 sectors at LBA = 0x0c9b9d4a = 211524938

  Commands leading to the command that caused the error were:
  CR FR SC SN CL CH DH DC   Powered_Up_Time  Command/Feature_Name
  -- -- -- -- -- -- -- --    
  c8 00 08 47 9d 9b 4c 00  00:52:49.600  READ DMA
  ec 00 08 4a 9d 9b 00 00  00:52:49.600  IDENTIFY DEVICE
  c8 00 08 47 9d 9b 4c 00  00:52:49.600  READ DMA
  ec 00 08 4a 9d 9b 00 00  00:52:49.600  IDENTIFY DEVICE
  c8 00 08 47 9d 9b 4c 00  00:52:49.600  READ DMA

Error 11 occurred at disk power-on lifetime: 1940 hours (80 days + 20 hours)
  When the command that caused the error occurred, the device was
active or idle.

  After command completion occurred, registers were:
  ER ST SC SN CL CH DH
  -- -- -- -- -- -- --
  40 51 08 4a 9d 9b ec  Error: UNC 8 sectors at LBA = 0x0c9b9d4a = 211524938

  Commands leading to the command that caused the error were:
  CR FR SC SN CL CH DH DC   Powered_Up_Time  Command/Feature_Name
  -- -- -- -- -- -- -- --    
  c8 00 08 47 9d 9b 4c 00  00:52:46.400  READ DMA
  ec 00 08 4a 9d 9b 00 00  00:52:46.400  IDENTIFY DEVICE
  c8 00 08 47 9d 9b 4c 00  00:52:46.400  READ DMA
  ec 00 08 4a 9d 9b 00 00  00:52:46.400  IDENTIFY DEVICE
  c8 00 08 47 9d 9b 4c 00  00:52:46.400  READ DMA

Error 10 occurred at disk power-on lifetime: 1940 hours (80 days + 20 hours)
  When the command that caused the error occurred, the device was
active or idle.

  After command completion occurred, registers were:
  ER ST SC SN CL CH DH
  -- -- -- -- -- -- --
  40 51 08 4a 9d 9b ec  Error: UNC 8 sectors at LBA = 0x0c9b9d4a = 211524938

  Commands leading to the command that caused the error were:
  CR FR SC SN CL CH DH DC   Powered_Up_Time  Command/Feature_Name
  -- -- -- -- -- -- -- --    
  c8 00 08 47 9d 9b 4c 00  00:52:42.950  READ DMA
  ec 00 08 4a 9d 9b 00 00  00:52:42.950  IDENTIFY DEVICE
  c8 00 08 47 9d 9b 4c 00  00:52:42.950  READ DMA
  ec 00 08 4a 9d 9b 00 00  00:52:42.950  IDENTIFY DEVICE
  c8 00 08 47 9d 9b 4c 00  00:52:42.950  READ DMA

Error 9 occurred at disk power-on lifetime: 1940 hours (80 days + 20 hours)
  When the command that caused the error occurred, the device was
active or idle.

  After command completion occurred, registers were:
  ER ST SC SN CL CH DH
  -- -- -- -- -- -- --
  40 51 08 4a 9d 9b ec  Error: UNC 8 sectors at LBA = 0x0c9b9d4a = 211524938

  Commands leading to the command that caused the error were:
  CR FR SC SN CL CH DH DC   Powered_Up_Time  Command/Feature_Name
  -- -- -- -- -- -- -- --    
  c8 00 08 47 9d 9b 4c 00  00:52:39.800  READ DMA
  ec 00 08 4a 9d 9b 00 00  00:52:39.800  IDENTIFY DEVICE
  c8 00 08 47 9d 9b 4c 00  00:52:39.800  READ DMA
  ec 00 08 4a 9d 9b 00 00  00:52:39.800  IDENTIFY DEVICE
  c8 00 08 47 9d 9b 4c 00  00:52:39.800  READ DMA

Error 8 occurred at disk power-on lifetime: 1940 hours (80 days + 20 hours)
  When the command that caused the error occurred, the device was
active or idle.

  After command completion occurred, registers were:
  ER ST SC SN CL CH DH
  -- -- -- -- -- -- --
  40 51 08 4a 9d 9b ec  Error: UNC 8 sectors at LBA = 0x0c9b9d4a = 211524938

  Commands leading to the command that caused the error were:
  CR FR SC SN CL CH DH DC   Powered_Up_Time  Command/Feature_Name
  -- -- -- -- -- -- -- --    
  c8 00 08 47 9d 9b 4c 00  00:52:36.600  READ DMA
  ec 00 08 4a 9d 9b 00 00  00:52:36.600  IDENTIFY DEVICE
  c8 00 08 47 9d 9b 4c 00  00:52:36.600  READ DMA
  c8 00 08 3f 9d 9b 4c 00  00:52:36.600  READ DMA
  c8 00 08 37 9d 9b 4c 00  00:52:36.600  READ DMA

SMART Self-test l

Re: raid1 issue, somewhat related to recent "debian on big machines"

2009-03-03 Thread Goswin von Brederlow
Francesco Pietra  writes:

The important bit is:

> Disk failure on sda1, disabling device

So one of your disks failed. The kernel marked it as such and
continious on the other disk alone. This is invisible to the
application (apart from the short hickup) and no data is corrupted or
lost. That is why you use raid1 after all.

> Personalities : [raid1]
> md6 : active raid1 sda8[2](F) sdb8[1]
>   102341952 blocks [2/1] [_U]
>
> md5 : active raid1 sda7[2](F) sdb7[1]
>   1951744 blocks [2/1] [_U]
>
> md4 : active raid1 sda6[2](F) sdb6[1]
>   2931712 blocks [2/1] [_U]
>
> md3 : active raid1 sda5[2](F) sdb5[1]
>   14651136 blocks [2/1] [_U]
>
> md1 : active raid1 sda2[2](F) sdb2[1]
>   6835584 blocks [2/1] [_U]
>
> md0 : active raid1 sda1[2](F) sdb1[1]
>   2931712 blocks [2/1] [_U]
>
> md2 : active raid1 sda3[2](F) sdb3[1]
>   14651200 blocks [2/1] [_U]

That is a lot of raids. Have you ever thought about using LVM? The
different raid1 will mess up each others assumption about the head
positioning of the component devices. On read the linux kernel tries
to use the disk with the shorter seek and assumes the head is where it
left it on the last access. But if one of the other raids used that
disk the head will be way off.

I would suggest the following scheme:

sda1 / sdb1 : 100Mb raid1 for /boot (or 1GB for / + /boot)
sda2 / sdb2 : rest raid1 with lvm

MfG
Goswin


-- 
To UNSUBSCRIBE, email to debian-amd64-requ...@lists.debian.org
with a subject of "unsubscribe". Trouble? Contact listmas...@lists.debian.org



Re: raid1 issue, somewhat related to recent "debian on big machines"

2009-03-03 Thread Ron Johnson

On 03/03/2009 02:53 AM, Francesco Pietra wrote:

lupus in fabula as a follow up of my short intervention on raid1 with
my machine to the thread "Debian on big systems".

System: supermicro H8QC8 m.board, two WD Raptor SATA 150GB, Debian
amd64 lenny, raid1

While running an electronic molecular calculation - estimated to four
days time - I noticed by chance on the screen (what is not in the out
file of the calculation) that there was a disk problem. I took some
scattered notes from the scree:

RAID1 conf printout

wd: 1 rd:2

[snip]

What you are looking for should be in syslog, not your application's 
log.


--
Ron Johnson, Jr.
Jefferson LA  USA

The feeling of disgust at seeing a human female in a Relationship
with a chimp male is Homininphobia, and you should be ashamed of
yourself.


--
To UNSUBSCRIBE, email to debian-amd64-requ...@lists.debian.org
with a subject of "unsubscribe". Trouble? Contact listmas...@lists.debian.org



raid1 issue, somewhat related to recent "debian on big machines"

2009-03-03 Thread Francesco Pietra
lupus in fabula as a follow up of my short intervention on raid1 with
my machine to the thread "Debian on big systems".

System: supermicro H8QC8 m.board, two WD Raptor SATA 150GB, Debian
amd64 lenny, raid1

While running an electronic molecular calculation - estimated to four
days time - I noticed by chance on the screen (what is not in the out
file of the calculation) that there was a disk problem. I took some
scattered notes from the scree:

RAID1 conf printout

wd: 1 rd:2

disk0 wd:1 o:0 dev: sda6

disk0 wd:1 o:0 dev: sdb6

md: recovery of raid array md4

minimum guaranteed speed 1000 kB/sec/disk

using max available idle I/O bandwidth but no more than 20

..

Disk failure on sda1, disabling device

Operation continues on 1 devices.

raid sdb1: redirecting sector 262176 to another mirror

RAID1 conf printout

wd:1 rd:2
..

disk1, wd:0 0:1 dev:sdb7
=

Then, the electronic molecular calculation resumed - with all CPUs at
work, as indicated by top - and in its output file there was no trace
of the above problems.

Command:

lshw -class disk

reported:

  *-cdrom
   description: DVD writer
   product: PIONEER DVD-RW DVR-111D
   vendor: Pioneer
   physical id: 0
   bus info: i...@0.0
   logical name: /dev/hda
   version: 1.02
   capabilities: packet atapi cdrom removable nonmagnetic dma lba
iordy pm audio cd-r cd-rw dvd dvd-r
   configuration: mode=udma4 status=nodisc
  *-disk:0
   description: SCSI Disk
   physical id: 0
   bus info: s...@0:0.0.0
   logical name: /dev/sda
   size: 139GiB (150GB)
  *-disk:1
   description: ATA Disk
   product: WDC WD1500ADFD-0
   vendor: Western Digital
   physical id: 1
   bus info: s...@1:0.0.0
   logical name: /dev/sdb
   version: 20.0
   serial: WD-WMAP41173675
   size: 139GiB (150GB)
   capabilities: partitioned partitioned:dos
   configuration: ansiversion=5 signature=000b05ba

The description of disk 0 was cryptic to me.


As there have been RAM problems, I also run

lshw -class memory

all DIMMs are correctly reported. No mem problem.
===

Then I run:

/proc/mdstat

the output was:

Personalities : [raid1]
md6 : active raid1 sda8[2](F) sdb8[1]
  102341952 blocks [2/1] [_U]

md5 : active raid1 sda7[2](F) sdb7[1]
  1951744 blocks [2/1] [_U]

md4 : active raid1 sda6[2](F) sdb6[1]
  2931712 blocks [2/1] [_U]

md3 : active raid1 sda5[2](F) sdb5[1]
  14651136 blocks [2/1] [_U]

md1 : active raid1 sda2[2](F) sdb2[1]
  6835584 blocks [2/1] [_U]

md0 : active raid1 sda1[2](F) sdb1[1]
  2931712 blocks [2/1] [_U]

md2 : active raid1 sda3[2](F) sdb3[1]
  14651200 blocks [2/1] [_U]

unused devices: 
===

I would appreciate advice.
thanks

francesco pietra


-- 
To UNSUBSCRIBE, email to debian-amd64-requ...@lists.debian.org
with a subject of "unsubscribe". Trouble? Contact listmas...@lists.debian.org