tkinter-2.3.4-14.4.el4_6.1.i386.rpm

2008-03-27 Thread Aldo Saavedra

Hi,

I have trying to install gLite3.1 on my SL4.5 and it complains that tkinter
is not found. Browsing at the link I can see it is not there. What 
should I do? What other

mirror can I use?
Downloaded from somwhere and just do a yum localinstall

These are the steps I followed:

yum install glite-UI

Everything is fine until it gets to this step

Downloading header for tkinter to pack into transaction set.
http://ftp.scientificlinux.org/linux/scientific/4x/i386/errata/SL/RPMS/tkinter-2.3.4-14.4.el4_6.1.i386.rpm: 
[Errno 4] IOError: HTTP Error 404: Not Found

Trying other mirror.
ftp://ftp.scientificlinux.org/linux/scientific/4x/i386/errata/SL/RPMS/tkinter-2.3.4-14.4.el4_6.1.i386.rpm: 
[Errno 4] IOError: HTTP Error 404: Not Found

Trying other mirror.
Error: failure: tkinter-2.3.4-14.4.el4_6.1.i386.rpm from sl-errata: 
[Errno 256] No more mirrors to try.


If I do
yum info tkinter
Loading "kernel-module" plugin
Setting up repositories
Reading repository metadata in from local files
Available Packages
Name   : tkinter
Arch   : i386
Version: 2.3.4
Release: 14.4.el4_6.1
Size   : 271 k
Repo   : sl-errata
Summary: A graphical user interface for the Python scripting language.
Description:

So it expects it under sl-errata. Can I change that?

Thanks for your help,
Aldo


--
--
Aldo F. Saavedra
Phone: +61 (0) 2 9351 5970
Fax  : +61 (0) 2 9351 7726
Email: [EMAIL PROTECTED]   
Mail : High Energy Physics

  School of Physics A28
  University of Sydney NSW 2006
  Sydney, Australia


RE: fsck.ext3 on large file systems?

2008-03-27 Thread Jon Peatfield

On Thu, 27 Mar 2008, Bly, MJ (Martin) wrote:



On our hardware RAID arrays (3ware, Areca, Infortrend) with many (12/14)
SATA disks, 500/750GB each, we fsck 2TB+ ext3 filesystems (as
infrequently as possible!) and it takes ~2 hours each.  We have some
5.5TB arrays that take less than three hours.  Note that these are
created with '-T largefile4 -O dir_index' among other options.


At least one pass of the ext3 fsck involves checking every inode table 
entry so '-T largefile4' would help you since you will get smaller numbers 
of inodes.  As the inode tables are spread through the disk it will read 
the various chunks then seek off somewhere and read more...


[ one of the planned features for ext4 is a way to safely mark than an 
entire inode-table lump is unused to save things like fsck from having to 
scan all those unusd blocks.  Of course doing so safely isn't quite 
trivial and it causes problems with the current model of how to choose the 
locations for an inode for a new file... ]



I'd be very suspicious of a HW RAID controller that took 'days' to fsck
a file system unless the filesystem was already in serious trouble, and
from bitter experience, fsck on a filesystem with holes it it caused by
a bad raid controller interconnect (SCSI!) can do more damage than good.


To give you one example at least one fsck pass needs to check that every 
inode in use has the right (link-count) number of directory entries 
pointing at it.  The current ext3 fsck seems to do a good impersonation of 
a linear search though it's in-memory inode-state-table for each directory 
entry - at least for files with non-trivial link-counts.


A trivial analysis shows that such a set of checks would be O(n^2) in the 
number of files needing to be checked, not counting the performance 
problems when the 'in-memory' tables get too big for ram...


[ In case I'm slandering the ext3 fsck people - I've not actually checked 
the ext3 fsck code really does anything as simple as a linear search but 
anything more complex will need to use more memory and ... ]


Last year we were trying to fsck a ~6.8TB ext3 fs which was about 70% 
filled with hundereds of hard-link trees of home directories.  So huge 
numbers of inode entries (many/most files are small), and having a 
link-counts of say 150.  Our poor server had only 8G ram and the ext3 fsck 
wanted rather a lot more.  Obviously in such a case it will be *slow*.


Of course that was after we built a version which didn't simply go into an 
infinite loop somewhere between 3 and 4TB into scanning through the 
inode-table.


Now as you can guess any dump needs to do much the same kind of work wrt 
scanning the inodes, looking for hard-links, so you may not be shocked to 
discover that attempting a backup was rather slow too...


--
Jon Peatfield,  Computer Officer,  DAMTP,  University of Cambridge
Mail:  [EMAIL PROTECTED] Web:  http://www.damtp.cam.ac.uk/


Re: fsck.ext3 on large file systems?

2008-03-27 Thread Stephen John Smoogen
On Thu, Mar 27, 2008 at 4:40 PM, Michael Hannon <[EMAIL PROTECTED]> wrote:
> On Thu, Mar 27, 2008 at 05:06:53PM +0100, [EMAIL PROTECTED] wrote:
>  > On Thu, 27 Mar 2008, Stephen John Smoogen wrote:
>  .
>  .
>
> .
>  > Hmm, we successfully fsck'd ext3 filesystems 1.4 TB in size frequently a
>  > couple of years ago, under 2.4 (back then, it was SuSE 8.2 + a vanilla
>  > kernel). This took no more than a few hours (maybe 2,3, or 4).  It was
>  > hardware RAID, not too reliable (hence "frequently"), and not too fast (<
>  > 100 MB/s). A contemporary linux server with software RAID should complete
>  > an fsck *much* faster, or something is wrong.
>
>  Hi, Stephan.  Yea, I think I must be doing something wrong here, but I
>  haven't been able to figure out what it is.
>

Well this could be a hardware error on the wire (bad scsi wire etc).
It could also depend on how the data is laid out on the disk. The long
fsck's were tons of directories and files.. and they were read,
deleted, etc in random order (eg INN news). but then again it could be
that I had crap hardware then too :).

>
>  > And I still wonder why fsck at at all just because a broken disk was
>  > replaced in a redundant array?
>
>  The system seems to insist on it.  Again, there may be some cockpit
>  error involved.
>

Usually it will require an fsck if the disk did not shutdown clearly
or some other issue it is detecting. I would need to know more about
the hardware in place (controller, number of drives, type of drives,
how many spares etc) to make a more educated guess.



-- 
Stephen J Smoogen. -- CSIRT/Linux System Administrator
How far that little candle throws his beams! So shines a good deed
in a naughty world. = Shakespeare. "The Merchant of Venice"


Re: fsck.ext3 on large file systems?

2008-03-27 Thread Michael Hannon
On Thu, Mar 27, 2008 at 05:06:53PM +0100, [EMAIL PROTECTED] wrote:
> On Thu, 27 Mar 2008, Stephen John Smoogen wrote:
.
.
.
> Hmm, we successfully fsck'd ext3 filesystems 1.4 TB in size frequently a 
> couple of years ago, under 2.4 (back then, it was SuSE 8.2 + a vanilla 
> kernel). This took no more than a few hours (maybe 2,3, or 4).  It was 
> hardware RAID, not too reliable (hence "frequently"), and not too fast (< 
> 100 MB/s). A contemporary linux server with software RAID should complete 
> an fsck *much* faster, or something is wrong.

Hi, Stephan.  Yea, I think I must be doing something wrong here, but I
haven't been able to figure out what it is.

> And I still wonder why fsck at at all just because a broken disk was 
> replaced in a redundant array?

The system seems to insist on it.  Again, there may be some cockpit
error involved.

Thanks.

- Mike
-- 
Michael Hannonmailto:[EMAIL PROTECTED]
Dept. of Physics  530.752.4966
University of California  530.752.4717 FAX
Davis, CA 95616-8677


Re: fsck.ext3 on large file systems?

2008-03-27 Thread Michael Hannon
On Thu, Mar 27, 2008 at 06:56:17AM -0600, Stephen John Smoogen wrote:
> On Wed, Mar 26, 2008 at 7:34 PM, Michael Hannon <[EMAIL PROTECTED]> wrote:
> > Greetings.  We have a lately had a lot of trouble with relatively large
> >  (order of 1TB) file systems mounted on RAID 5 or RAID 6 volumes.  The
> >  file systems in question are based on ext3.
.
.
.
> >  After the RAID array has completed its rebuild, we run fsck on the RAID
> >  device.  When we do that, fsck seems to run forever, i.e., for days at a
> >  time, occasionally spitting out messages about files with recognizable
> >  names, but never completing satisfactorily.
> >
> 
> fsck of 1TB is going to take days  due to the linear nature of it
> checking the disk. [ I think the disks for mirrors.kernel.org take
> many weeks to fsck.] The bigger question is what kind of data are you
> writing to these disks, and is the ext3 journal large enough for those
> writes?
> 
> 
> >  The systems in question are typically running SL 4.x.  We've read that
> >  the version of fsck that is standard in SL 4 has some known bugs,
> >  especially wrt large file systems.
> >
> >  Hence, we've attempted to repeat the exercise with fsck.ext3 taken from
> >  the Fedora 8 distribution.  This gives us improved, but still not
> >  satisfactory, results.
> >
> 
> Did you recompile the binary from source, or did you use it straight?
> I am just wondering if fsck is dependant on some kernel particulars...

We just used the binary straight.  Thanks, Stephen.

- Mike
-- 
Michael Hannonmailto:[EMAIL PROTECTED]
Dept. of Physics  530.752.4966
University of California  530.752.4717 FAX
Davis, CA 95616-8677


emacs versions in SL5.1

2008-03-27 Thread Dr Andrew C Aitchison
http://ftp.scientificlinux.org/linux/scientific/51/x86_64/SL/ has 
emacs-21.4-18.1.el5.x86_64.rpm


but http://ftp.scientificlinux.org/linux/scientific/51/i386/SL/ has
emacs-21.4-20.el5.i386.rpm

... so x86_64 has an earlier emacs than i386

http://ftp.scientificlinux.org/linux/scientific/51/SRPMS/vendor/ has
emacs-21.4-18.1.el5.src.rpm

Should I be looking for emacs-21.4-20.el5.{src,x86_64}.rpm somewhere 
else ?


--
Dr. Andrew C. Aitchison Computer Officer, DPMMS, Cambridge
[EMAIL PROTECTED]   http://www.dpmms.cam.ac.uk/~werdna


RE: fsck.ext3 on large file systems?

2008-03-27 Thread Bly, MJ (Martin)
In-line...

> -Original Message-
> From: [EMAIL PROTECTED] 
> [mailto:[EMAIL PROTECTED] On 
> Behalf Of [EMAIL PROTECTED]
> Sent: 27 March 2008 16:07
> To: [EMAIL PROTECTED]
> Subject: Re: fsck.ext3 on large file systems?
> 
> On Thu, 27 Mar 2008, Stephen John Smoogen wrote:
> 
> > On Wed, Mar 26, 2008 at 7:34 PM, Michael Hannon 
> <[EMAIL PROTECTED]> wrote:
> >> Greetings.  We have a lately had a lot of trouble with 
> relatively large
> >>  (order of 1TB) file systems mounted on RAID 5 or RAID 6 
> volumes.  The
> >>  file systems in question are based on ext3.
> >>
> >>  In a typical scenario, we have a drive go bad in a RAID 
> array.  We then
> >>  remove it from the array, if it isn't already, add a new 
> hard drive
> >>  (i.e., by hand, not from a hot spare), and add it back to the RAID
> >>  array.  The RAID operations are all done using mdadm.

This is of course a software RAID...

> >>  After the RAID array has completed its rebuild, we run 
> fsck on the RAID
> >>  device.  When we do that, fsck seems to run forever, 
> i.e., for days at a
> >>  time, occasionally spitting out messages about files with 
> recognizable
> >>  names, but never completing satisfactorily.
> >>
> >
> > fsck of 1TB is going to take days  due to the linear nature of it
> 
> Hmm, we successfully fsck'd ext3 filesystems 1.4 TB in size 
> frequently a 
> couple of years ago, under 2.4 (back then, it was SuSE 8.2 + 
> a vanilla 
> kernel). This took no more than a few hours (maybe 2,3, or 
> 4).  It was 
> hardware RAID, not too reliable (hence "frequently"), and not 
> too fast (< 
> 100 MB/s). A contemporary linux server with software RAID 
> should complete 
> an fsck *much* faster, or something is wrong.

On our hardware RAID arrays (3ware, Areca, Infortrend) with many (12/14)
SATA disks, 500/750GB each, we fsck 2TB+ ext3 filesystems (as
infrequently as possible!) and it takes ~2 hours each.  We have some
5.5TB arrays that take less than three hours.  Note that these are
created with '-T largefile4 -O dir_index' among other options.

I'd be very suspicious of a HW RAID controller that took 'days' to fsck
a file system unless the filesystem was already in serious trouble, and
from bitter experience, fsck on a filesystem with holes it it caused by
a bad raid controller interconnect (SCSI!) can do more damage than good.


Maybe the controller has some missconfiguration.

Not so sure about software raid setups though.  But if you have a lot of
disks on a single controller chip, say driving 4 SATA disks, you *might*
max out the PCI(-e/X) bandwidth to the controller chip.

One thing that will kill you though - if you have a lot of remapped
blocks on your device, that can slow things down considerably.  The
device may be OK but performance would suck.  Another thing that some
harwdare RAID controllers do is turn off the device cache in some
circumstances - that bit us a while back and made a controller crawl
(BBU failed, controller went into 'safe' mode).  

> And I still wonder why fsck at at all just because a broken disk was 
> replaced in a redundant array?

Yes - should not be necessary, but I've done it for super-critical
(small) systems 'just in case' due to paranoid user groups.

> -- 
> Stephan Wiesand
>DESY - DV -
>Platanenallee 6
>15738 Zeuthen, Germany

Martin.
-- 
Martin Bly
RAL Tier1 Fabric Team 


Re: Firefox crash on SL5

2008-03-27 Thread Jan Iven

On 26/03/08 15:39, Faye Gibbins wrote:

Hi,

Lots of our users are having trouble when using firefox with SL5 
with the NVidia drivers (both legacy and newer ones).


Running firefox using /usr/bin/firefox --sync --no-xshm this error 
message was reported after mine crashed and disappeared from the desktop:


snip---

[EMAIL PROTECTED] ~ $ firefox
[NoScript] [NoScript] browserDOMWindow wrapped for external load 
interception
[NoScript] [NoScript] browserDOMWindow wrapped for external load 
interception
[NoScript] [NoScript] browserDOMWindow wrapped for external load 
interception
[NoScript] [NoScript] browserDOMWindow wrapped for external load 
interception
[NoScript] Component returned failure code: 0x80570009 
(NS_ERROR_XPC_BAD_CONVERT

_JS) [nsIXPCComponents_Utils.lookupMethod] while showing NOSCRIPT element
[NoScript] Component returned failure code: 0x80570009 
(NS_ERROR_XPC_BAD_CONVERT

_JS) [nsIXPCComponents_Utils.lookupMethod] while showing NOSCRIPT element
The program 'Gecko' received an X Window System error.
This probably reflects a bug in the program.
The error was 'BadRequest (invalid request code or no such operation)'.
  (Details: serial 22989464 error_code 1 request_code 0 minor_code 0)
  (Note to programmers: normally, X errors are reported asynchronously;
   that is, you will receive the error a while after causing it.
   To debug your program, run it with the --sync command line
   option to change this behavior. You can then get a meaningful
   backtrace from your debugger if you break on the gdk_x_error() 
function.)

[EMAIL PROTECTED] ~ $

snip---

has anyone seen this before/know how to fix it?


Haven't seen this myself. But the above would lead me to suspect a bug 
in the "NoScript" extension. If you can reproduce this at will, perhaps 
you could try first without NoScript, then retry with the 
latest&greatest (unless you have that already), then install the 
firefox-debuginfo RPM (matching your firefox RPM version+architecture), 
then run the bove under gdb (you will need to set up the environment for 
firefox to find its libraries, have a look at /usr/bin/firefox). This at 
least should provide a helpful backtrace, which might allow the NoScript 
people to guess where the problem is.


"request_code 0" indeed does not exist. This would look more like a bug 
in Firefox, it should not be making such a bogus request. You could see 
whether using the "vesa" X11 driver makes the problem go away, this 
would allow to take the proprietary NVidia drives out of the issue...


On the other hand, have a look at the changes proposed at the end of
http://www.nvnews.net/vbulletin/archive/index.php/t-96192.html

Regards
jan


SL5 and nx

2008-03-27 Thread WILLIAM J LUTTER
I use nx/nxclient connecting SL4 and SL5 nx servers to windows XP clients 
(nxclient) and to other linux PCs.To avoid the license restrsictions on the 
www.nomachine.com freeware versions, (2 connections per server), I've used 
freenx and  an earlier www.nomachine.com release (nxclient-1.5.0-141.i386.rpm). 
 This continues to work very well with SL4.  On SL5, gnome never played well 
with nx due to applet issues that I chose not to delve into.  Kde worked fine, 
so one could connect to SL5 via kde using nx and all is well - until recently.

Recently on SL5 (there must have been an update of a kde package?), keystrokes 
are not recognized when connecting with nx from linux or windows XP to the SL5 
nxserver.  If I hit a screen saver button (Print Scrn?) that will bring up a 
screen saver window during the SL5 nx session, but otherwise no keystrokes show 
in terminals.  Very unsatisfying.

I did try updating the kernel from  2.6.18-8.1.15.el5 to 2.6.18-53.1.14.el5 and 
I updated nx and freenx via yum (atrpms).  No change.

I figure I'm on my own on this as I reallly should be using the latest freeware 
from www.nomachine.com (and accepting their 2 connection limitation).  
Nevertheless, my questionss are:
1)  Is there a simple keystroke behavior fix that I have missed?
2)   I suspect it has to be yum updates that changed kde behavior.  As I don't 
have an exact day for when the nx keystroke misbehavior started, I'm not fully 
certain what software package triggered this, plus I don't see any recent kde 
updates when I take a look via yumex (ages list).  My best guess is that thie 
keystroke misbehavior started within the last month.

Bill Lutter

To compare SL4 and SL5 nx so you see how old these packages are that still 
work...
SL4.0
nx-1.5.0-0.FC4.1.i686
nxclient-1.5.0-141.i386
freenx-0.4.4-1.fdr.0.noarch

SL4.1
nx-1.5.0-4.FC4.1.i686
freenx-0.4.4-2.fdr.0.noarch

SL5
freenx-0.7.1-5.el5.i386
nx-2.1.0-22.el5.i386
nxclient-1.5.0-141.i386


Re: fsck.ext3 on large file systems?

2008-03-27 Thread Stephan . Wiesand

On Thu, 27 Mar 2008, Stephen John Smoogen wrote:


On Wed, Mar 26, 2008 at 7:34 PM, Michael Hannon <[EMAIL PROTECTED]> wrote:

Greetings.  We have a lately had a lot of trouble with relatively large
 (order of 1TB) file systems mounted on RAID 5 or RAID 6 volumes.  The
 file systems in question are based on ext3.

 In a typical scenario, we have a drive go bad in a RAID array.  We then
 remove it from the array, if it isn't already, add a new hard drive
 (i.e., by hand, not from a hot spare), and add it back to the RAID
 array.  The RAID operations are all done using mdadm.

 After the RAID array has completed its rebuild, we run fsck on the RAID
 device.  When we do that, fsck seems to run forever, i.e., for days at a
 time, occasionally spitting out messages about files with recognizable
 names, but never completing satisfactorily.



fsck of 1TB is going to take days  due to the linear nature of it


Hmm, we successfully fsck'd ext3 filesystems 1.4 TB in size frequently a 
couple of years ago, under 2.4 (back then, it was SuSE 8.2 + a vanilla 
kernel). This took no more than a few hours (maybe 2,3, or 4).  It was 
hardware RAID, not too reliable (hence "frequently"), and not too fast (< 
100 MB/s). A contemporary linux server with software RAID should complete 
an fsck *much* faster, or something is wrong.


And I still wonder why fsck at at all just because a broken disk was 
replaced in a redundant array?


--
Stephan Wiesand
  DESY - DV -
  Platanenallee 6
  15738 Zeuthen, Germany


Re: Problems with 4.5 TB disk array (LVM & ext3)

2008-03-27 Thread Jan Kundrát

[EMAIL PROTECTED] wrote:
We're successfully using devices up to 7 TB with a single XFS filesystem 
on them, under SL4/5 (but I think we started doing this with 4.3, not 
4.2). I have no hope to be able to check (xfs_repair) those should this 
ever become necessary though - from what I've read it would require more 
RAM than fits into a server today.


On this year's linux.conf.au, there was a talk about fixing the fsck.xfs 
[1]. They claim that they reduce both memory usage and required time.


[1] 
http://mirror.linux.org.au/pub/linux.conf.au/2008/slides/135-fixing_xfs_faster.pdf


Cheers,
-jkt


smime.p7s
Description: S/MIME Cryptographic Signature


Re: Problems with 4.5 TB disk array (LVM & ext3)

2008-03-27 Thread Stephan . Wiesand

On Wed, 26 Mar 2008, Jon Peatfield wrote:


On Wed, 26 Mar 2008, Jan Schulze wrote:


Hi all,

I have a disk array with about 4.5 TB and would like to use it as one large 
logical volume with an ext3 file system. When mounting the logical volume , 
I get an "Input/Output error: can't read superblock".


Do you get any interesting kernel messages in the output of dmesg (or 
/var/log/messages etc)?  Which exact kernel is this (uname -r) and what arch 
(i386/x86_64 etc; uname -m)?


And what driver/hardware?


I'm using SL 4.2 with kernel 2.6 and this is what I did so far:

- used parted to create a gpt disk label (mklabel gpt) and one large
  partition (mkpart primary ext3 0s -1s)

- used parted to enable LVM flag on device (set 1 LVM on)


I know it would be slow but can you test that you can read/write to all of 
/dev/sda1?


Using dd's "seek" parameter, this should not take too much time. But if
creating the GPT label & partition was successful, chances are the whole 
device is accessible.



- created one physical volume, one volume group and one logical volume
  (pvcreate /dev/sda1, vgcreate raid6 /dev/sda1, lvcreate -l 1189706 -n
  vol1 raid6)

- created an ext3 filesystem and explicitly specified a 4K blocksize, as
  this should allow a filesystem size of up to 16 TB (mkfs.ext3 -m 0 -b
  4096 /dev/raid6/vol1)


For some reason my EL4 notes tell me that we also specify -N (number of 
inodes), as well as -E (set RAID stride), -J size= (set journal size) and -O 
sparse_super,dir_index,filetype though most of that is probably the default 
these days...


Specifying the stripe width is also supposed to be a good idea, as is 
aligning the start of the partition to a stripe boundary (although that's 
more likely to be useful without LVM on top).


However, mounting (mount /dev/raid6/vol1 /raid) gives the superblock error, 
mentioned above.


Everything is working as expected, when using ext2 filesystem (with LVM) or 
ext3 filesystem (without LVM). Using a smaller volume (< 2 TB) is working


with ext3+LVM as well. Only the combination of > 2TB+ext3+LVM gives me 
trouble.


Any ideas or suggestions?


We found that in at least some combinations of kernel/hardware (drivers 
really I expect), that support for >2TB block devices was still rather flakey 
(at least in the early versions of EL4).


We ended up getting our RAID boxes to present as multiple LUNs each under 2TB 
which we can then set up as PVs and join back together into a single VG and 
still have an LV which was bigger than 2TB.  I'm rather conservative in such 
things so we still avoid big block devices at the moment.


[ obviously with single disk sizes growing at the rate they are it means that 
the block-devices >2TB code is going to get a LOT more testing! ]


We're successfully using devices up to 7 TB with a single XFS 
filesystem on them, under SL4/5 (but I think we started doing this with 
4.3, not 4.2). I have no hope to be able to check (xfs_repair) those 
should this ever become necessary though - from what I've read it would 
require more RAM than fits into a server today.


However, some of the tools (e.g. ext2/3 fsck) still seemed to fail at about 
3.5TB so we ended up needing to build the 'very latest' tools to be able to 
run fsck properly (the ones included in EL4 - and EL5 I think - get into an 
infinite loop at some point while scanning the inode tables).


Currently we try to avoid 'big' ext3 LVs ; the one where we discovered the 
fsck problems was originally ~6.8TB but we ended up splitting that into 
several smaller LVs since even with working tools it still took ~2 days to 
fsck... (and longer to dump/copy/restore it all!)


Some of my co-workers swear by XFS for 'big' volumes but then we do have SGI 
boxes where XFS (well CXFS) is the expected default fs.  I've not done much 
testing with XFS on SL mainly because TUV don't like XFS much...


I think it's still the best choice for large (> 2 TB) filesystems. The xfs 
available in SL4 contrib has done very well here. There are some 
interesting effects when such a filesystem runs full and you have to 
remount it with the "inode64" option in order to be able to create new 
files and you discover that quite a few applications are not ready for 
64-bit node numbers. But that aside, it has done very well. No other 
headaches. We're now beginning to deploy large (>10TB) XFS filesystems 
under SL5.


This being said, we now also have lustre OSTs (using a modified ext3) 
7.5 TB in size. No problems so far, but then none of them has run full or 
required an fsck yet.


--
Stephan Wiesand
   DESY - DV -
   Platanenallee 6
   15738 Zeuthen, Germany


Re: fsck.ext3 on large file systems?

2008-03-27 Thread Stephen John Smoogen
On Wed, Mar 26, 2008 at 7:34 PM, Michael Hannon <[EMAIL PROTECTED]> wrote:
> Greetings.  We have a lately had a lot of trouble with relatively large
>  (order of 1TB) file systems mounted on RAID 5 or RAID 6 volumes.  The
>  file systems in question are based on ext3.
>
>  In a typical scenario, we have a drive go bad in a RAID array.  We then
>  remove it from the array, if it isn't already, add a new hard drive
>  (i.e., by hand, not from a hot spare), and add it back to the RAID
>  array.  The RAID operations are all done using mdadm.
>
>  After the RAID array has completed its rebuild, we run fsck on the RAID
>  device.  When we do that, fsck seems to run forever, i.e., for days at a
>  time, occasionally spitting out messages about files with recognizable
>  names, but never completing satisfactorily.
>

fsck of 1TB is going to take days  due to the linear nature of it
checking the disk. [ I think the disks for mirrors.kernel.org take
many weeks to fsck.] The bigger question is what kind of data are you
writing to these disks, and is the ext3 journal large enough for those
writes?


>  The systems in question are typically running SL 4.x.  We've read that
>  the version of fsck that is standard in SL 4 has some known bugs,
>  especially wrt large file systems.
>
>  Hence, we've attempted to repeat the exercise with fsck.ext3 taken from
>  the Fedora 8 distribution.  This gives us improved, but still not
>  satisfactory, results.
>

Did you recompile the binary from source, or did you use it straight?
I am just wondering if fsck is dependant on some kernel particulars...

To tell you the truth, I have not done anything with Linux Raid in the
Terabyte range.. Usually I go with a hardware solution at that point
(usually for business reasons.. that much storage usually comes with a
box with hardware raid). I did run into a similar issue though trying
to help someone last week on a SuSE box with ext3. They also had a
long fsck and weird file names coming up. I think they went with the
same solution ( restore from backups).



-- 
Stephen J Smoogen. -- CSIRT/Linux System Administrator
How far that little candle throws his beams! So shines a good deed
in a naughty world. = Shakespeare. "The Merchant of Venice"


Re: Generating PDF via PHP in SL 5.1?

2008-03-27 Thread Klaus Steinberger
Hi,

> We're trying to make these applications work on the new system, and I
> have to say it's a can of worms.
Yes, pdflib is a Mega-Can of worms. It has an very bad record of incompatible 
API changes, and the license is way off from a free License.

> I suppose that if we remove  some more of the standard SL PHP packages
> we might eventually come into equilibrium here, but I keep thinking
> there must be a better way

One way is to rebuild PHP from SL's src.rpm's with included pdflib support and 
give it different release/version number. Of course you have to do that every 
time a security update comes out. 

Also you could try to build the pdflib support for PHP as a Module (you need 
to install the php-devel Package).

> If you've got any cleaner way to deal with PHP/PDF, will you please let
> me know ASAP?
The cleanest way, but maybe very hard way depending on how well written you're 
php apps are is probably to change the apps to use FPDF instead. 

 http://www.fpdf.org/

It's a PHP Class, so it doesn't a lib, just plain PHP, and its really free.

Sincerly,
Klaus


-- 
Klaus Steinberger Beschleunigerlaboratorium
Phone: (+49 89)289 14287  Am Coulombwall 6, D-85748 Garching, Germany
FAX:   (+49 89)289 14280  EMail: [EMAIL PROTECTED]
URL: http://www.physik.uni-muenchen.de/~Klaus.Steinberger/


smime.p7s
Description: S/MIME cryptographic signature