Greetings,
I found that for reisfer filesystems sometimes the option nolargeio=1 is added
to the fstab entry. At first blush this seems to be a workaround for a kernel
bug. Does anyone have any more information ? I am running reiser - lvm2 -
raid5 currently.
regards,
-Peter
-
To unsubscribe
Mitchell Laks wrote:
On Sunday 13 March 2005 10:49 am, David Greave wrote: Many Helpful remarks:
David I am grateful that you were there for me.
No probs - we've all been there!
My assessment (correct me if I am wrong) is that I have to rethink my
architecture. As I continue to work with
On 2005-03-14T15:43:52, Neil Brown [EMAIL PROTECTED] wrote:
Hi there, just a question about how the bitmap stuff works with
1++-redundancy, say RAID1 with 2 mirrors, or RAID6.
One disk fails and is replaced/reattached, and resync begins. Now
another disk fails and is replaced. Is the bitmap
Hi
I have two Linux boxes running kernel 2.4.21 having access to two
devices over fibre channel SAN.
What I'm trying to achive is host based mirroring with ability to
move the storage from one host to another.
On the firs host I created a raid1 array, put LVM on it, created a
filesystem. To move
Alex Turner [EMAIL PROTECTED] writes:
a 14 drive stripe will max out the PCI bus long before anything else,
Hopefully anyone with a 14 drive stripe is using some combination of 64 bit
PCI-X cards running at 66Mhz...
the only reason for a stripe this size is to get a total accessible
size
Arshavir Grigorian wrote:
Alex Turner wrote:
[]
Well, by putting the pg_xlog directory on a separate disk/partition, I
was able to increase this rate to about 50 or so per second (still
pretty far from your numbers). Next I am going to try putting the
pg_xlog on a RAID1+0 array and see if that
On Monday March 14, [EMAIL PROTECTED] wrote:
On 2005-03-14T21:22:57, Neil Brown [EMAIL PROTECTED] wrote:
Hi there, just a question about how the bitmap stuff works with
1++-redundancy, say RAID1 with 2 mirrors, or RAID6.
I assume you mean RAID1 with 3 drives (there isn't really one main
You said:
If your write size is smaller than chunk_size*N (N = number of data blocks
in a stripe), in order to calculate correct parity you have to read data
from the remaining drives.
Neil explained it in this message:
http://marc.theaimsgroup.com/?l=linux-raidm=108682190730593w=2
Guy
Neil Brown wrote:
On Wednesday March 9, [EMAIL PROTECTED] wrote:
avoid setting of sb-events_lo = 1 when creating a 0.90 superblock -- it
doesn't seem to be necessary and it was causing the event counters to
start at 4 billion+ (events_lo is actually the high part of the events
counter, on
All,
I have a 13 disk (250G each) software raid 5 set using 1 16 port adaptec SATA
controller.
I am very happy with the performance. The reason I went with the 13 disk raid 5
set was for the space NOT performance.
I have a single postgresql database that is over 2 TB with about 500 GB
10 matches
Mail list logo