[Lustre-discuss] Lustre User Group 2010 - registration now open and call for presentations

2009-11-30 Thread Sheila Barthel
*LUG 2010 - Eighth Annual Lustre User Group conference*

When: April 14-16, 2010
Where: Seascape Beach Resort, Monterey Bay, CA
Registration: http://LUG2010.org

Registration is now open for LUG 2010, the premier event to learn new 
technical information about Lustre, acquire best practices, and share 
knowledge about Lustre technology. LUG 2010 is a unique opportunity for 
users to get advice and answers about specific Lustre implementations 
from Lustre developers, community engineers and users. Any and all 
people using or interested in Lustre are welcome to attend.

Conference attendees will have access to experts and peers sharing their 
real-world experiences using Lustre. LUG 2010 will include updates on 
community development projects, Birds of a Feather sessions, demos and 
tutorials, and offer many opportunities to meet with the Lustre 
development team and discuss upcoming enhancements and capabilities.

LUG 2010 will be held at the beautiful Seascape beach resort and 
conference center on Monterey Bay, California (http://seascaperesort.com).

*Call for Presentations - LUG 2010*

Lustre users are invited to submit an abstract to be considered for 
presentation at LUG 2010. Presentations are generally 30 minutes in 
length, but can be 1 hour if needed. It is not necessary to submit a 
technical paper, just an abstract of your proposed talk and then slides 
for the talk itself. This is your opportunity to contribute to the 
Lustre community by sharing your knowledge and lessons learned from 
Lustre deployments. The deadline to submit presentation abstracts is 
February 15, 2010. For questions or to submit a presentation abstract, 
contact Dan Ferber (daniel.fer...@sun.com).

*Lustre Advanced User Seminar*

The Lustre Advanced User Seminar will be offered on April 14, 2010, 
before the User Group meeting. This seminar is designed for senior 
systems administrators, engineers and integrators needing more 
comprehensive knowledge of Lustre Administration and Troubleshooting 
techniques. To receive maximum value from this seminar, completion of 
Administering Lustre-based Clusters (CL-400) and/or prior experience 
administering Lustre is strongly recommended. Space is limited and 
registration fee discounts will be offered for LUG attendees.

To register for LUG 2010 and the Lustre Advanced User Seminar, visit our 
registration page at http://LUG2010.org

See you at LUG 2010!

___
Lustre-discuss mailing list
Lustre-discuss@lists.lustre.org
http://lists.lustre.org/mailman/listinfo/lustre-discuss


Re: [Lustre-discuss] 'Rename' kernel build result

2009-11-30 Thread Brian J. Murrell
On Mon, 2009-11-30 at 13:21 +0100, Frank Heckes wrote: 
> Hello Brian,

Hi Frank,

> first of all thanks for your patience and answers so far.

NP.

> No, no I like the idea a lot, but unfortunately the kernel didn't boot
> on the hardware we trying to install them (IBM pizzaboxes), even after
> rebuilding the initrd, Systemmap... to extend the (initial) driver list.

Hrm.  Does the Suse-supplied kernel boot those machines or does it have
the same problems?  I just want to isolate whether this is a problem
unique to our kernel build or whether it's a problem with the upstream
(Suse supplied) kernel.

> So I decided to built the kernel from source. Here the system boots as
> it should after installing it to one of the nodes of the (test)
> cluster.

What do you do differently than is being done by our build?  Do you
change anything in the source-tree before you build?  Probably it would
be great to have a complete transcript of what you do once you have
downloaded the kernel-lustre-source package from our download site.

> And of course there's also a small piece of (childlike) curiosity, too.

:-)

> I fully agree. I would to the ready made RPM's if it would be possible.
> (see above).

Indeed.  In fact rather than copying what could be some very copious
data to this list, can you open a bug in our bugzilla and report back
here for me (and the rest of the readers) what the ticket number is?

> Here're the Makefile's first 10 lines:

Unfortunately it was not enough.  Once you have opened the ticket as per
above, please attach the complete Makefile.

b.



signature.asc
Description: This is a digitally signed message part
___
Lustre-discuss mailing list
Lustre-discuss@lists.lustre.org
http://lists.lustre.org/mailman/listinfo/lustre-discuss


[Lustre-discuss] I/O on cluster with lustre

2009-11-30 Thread Goranka Bilalbegovic
Hello,

Recently the cluster I am using for computing has been updated to the VMware
with the Lustre file system.  Cluster uses: Oscar 6.0.3,  Sun Grid Engine
6.2u3, Nagios, Ganglia, InfiniBand 10 Gb/s. Nodes access the file system
using Ethernet via the Lustre InfiniBand/Ethernet router.

I used to run one type of jobs as:
---
#$ -N name
#$ -o namesys.out
#$ -e namesys.err
#$ -pe mpi 2
#$ -cwd
#$ -v LD_LIBRARY_PATH
mpirun -machinefile $TMPDIR/machines -np $NSLOTS /path/.../code.x << EOF
name.in
name.out
EOF
---

This is for an open source package (written in Fortran plus some C
utilities) and a such way of running was recommended by authors. It was
working on the previous version of the cluster, but it does not run on a new
lustre filesystem. It starts, but then stays in the queue forever.

Is it possible to run this type of jobs on lustre ?

Thank you.
Best wishes,
Goranka
___
Lustre-discuss mailing list
Lustre-discuss@lists.lustre.org
http://lists.lustre.org/mailman/listinfo/lustre-discuss


Re: [Lustre-discuss] Anyone built 1.8 or 1.6 on Fedora 12's 2.6.31 yet?

2009-11-30 Thread Stephen Willey
I've got this morning's b1_8 branch to compile on Fedora 12.  For those 
interested, here are the steps.

(Disclaimer:  I am not a programmer and the below changes were hacks I made 
based on far too little knowledge.  If I've done something horribly wrong, 
please reply saying so).

0) Just remembereed - I happened to symlink /usr/src/linux to 
kernels/2.6.31.5-127.fc12.x86_64 - Just me being lazy when running configure 
too many times.

1) /usr/src/kernels/2.6.31.5-127.fc12.x86_64/include/linux/posix_acl.h needs 
#include  adding to it.
I don't know why and this is probably my hackiest hack, but if not, it fails to 
compile in the configure and causes problems later on.

2) checkout the b1_8 tree and patch it with the attached.

3) sh autogen.sh
4) ./configure --disable-server

5) make rpms CFLAGS=-fno-strict-aliasing
I thought no-strict-aliasing was supposed to be picked up by default for kernel 
related builds, but I needed to add this.

Once they're built and installed, strangely, the mgc module will not load 
automatically.  I had to manually modprobe mgc before I could then successfully 
mount a Lustre FS.  If anyone knows why or if there's a fix, please let me know.

Hope it helps...




- "Andreas Dilger"  wrote:

> On 2009-11-19, at 01:48, Stephen Willey wrote:
> > https://bugzilla.lustre.org/show_bug.cgi?id=21113 seems to suggest 
> 
> > that Sun/Lustre will be building it as F12 should be the base for  
> > RH6 (please correct me if I'm wrong) but I was just wondering if  
> > anyone was way ahead and had tried this already?
> 
> Some of our other customers are also interested in porting to FC12 as 
> 
> a stepping stone to RHEL6.  We are getting closer on the client side, 
> 
> but the server still needs work.  If anyone is interested to port the 
> 
> server code, then I'd be interested to hear, because I'd prefer to  
> spend any development effort to remove the server patches entirely.  
> 
> Instead of reworking the patches yet again, this needs some effort in 
> 
> the Lustre and ldiskfs code to allow it to use newer kernel APIs and 
> 
> then (hopefully) we'll never need to patch for Lustre again.
> 
> > https://bugzilla.lustre.org/show_bug.cgi?id=21340 suggests I can use
>  
> > the 2.6.27 patches (are they rolled into 1.8.1.1??) plus the patch 
> 
> > in that thread to get it working.  Again, I'll give it a go but I  
> > just figured I'd ask before possibly redoing work that's already  
> > been done.
> 
> 
> The 2.6.27 support (both client and server) should be in 1.8.1.1  
> AFAIK, because it runs on SLES11.
> 
> Cheers, Andreas
> --
> Andreas Dilger
> Sr. Staff Engineer, Lustre Group
> Sun Microsystems of Canada, Inc.

-- 
Stephen Willey
Senior Systems Engineer
Framestore
19-23 Wells Street, London W1T 3PQ
+44 207 344 8000
www.framestore.com 
--- lustre/llite/lloop-old.c	2009-11-26 10:34:02.393003227 +
+++ lustre/llite/lloop.c	2009-11-26 10:35:03.590056455 +
@@ -500,7 +500,7 @@ static int loop_set_fd(struct lloop_devi
 lo->lo_queue->unplug_fn = loop_unplug;
 
 /* queue parameters */
-blk_queue_hardsect_size(lo->lo_queue, CFS_PAGE_SIZE);
+blk_queue_logical_block_size(lo->lo_queue, CFS_PAGE_SIZE);
 blk_queue_max_sectors(lo->lo_queue,
   LLOOP_MAX_SEGMENTS << (CFS_PAGE_SHIFT - 9));
 blk_queue_max_phys_segments(lo->lo_queue, LLOOP_MAX_SEGMENTS);
--- lustre/llite/file-old.c	2009-11-26 10:37:51.401943992 +
+++ lustre/llite/file.c	2009-11-26 10:58:28.326208286 +
@@ -47,6 +47,7 @@
 #include 
 #include "llite_internal.h"
 #include 
+#include 
 
 /* also used by llite/special.c:ll_special_open() */
 struct ll_file_data *ll_file_data_get(void)
@@ -1423,7 +1424,7 @@ static ssize_t ll_file_aio_read(struct k
 
 count = ll_file_get_iov_count(iov, &nr_segs);
 CDEBUG(D_VFSTRACE, "VFS Op:inode=%lu/%u(%p),size="LPSZ",offset=%Ld\n",
-   inode->i_ino, inode->i_generation, inode, count, *ppos);
+   inode->i_ino, inode->i_generation, inode, (unsigned int)count, *ppos);
 /* "If nbyte is 0, read() will return 0 and have no other results."
  *  -- Single Unix Spec */
 if (count == 0)
@@ -1593,7 +1594,7 @@ repeat:
 
 chunk = end - *ppos + 1;
 CDEBUG(D_INODE,"Read ino %lu, "LPSZ" bytes, offset %lld, i_size %llu\n",
-   inode->i_ino, chunk, *ppos, i_size_read(inode));
+   inode->i_ino, (unsigned int)chunk, *ppos, i_size_read(inode));
 
 /* turn off the kernel's read-ahead */
 if (ltd.lock_style != LL_LOCK_STYLE_NOLOCK) {
@@ -1773,7 +1774,7 @@ static ssize_t ll_file_aio_write(struct 
 count = ll_file_get_iov_count(iov, &nr_segs);
 
 CDEBUG(D_VFSTRACE, "VFS Op:inode=%lu/%u(%p),size="LPSZ",offset=%Ld\n",
-   inode->i_ino, inode->i_generation, inode, count, *ppos);
+   inode->i_ino, inode->i_generation, inode, (unsigned int)count, *ppos);
 
  

[Lustre-discuss] quota with vanilla kernel

2009-11-30 Thread Ralf Utermann
Dear list,

has anybody a running setup with Lustre servers 1.8 on a patched 
vanilla 2.6.22 kernel using quota? The system works fine otherwise,
just quota don't work:

I have configured a filesystem on 1.8.1.1 using the parameters ost.quota_type=ug
and mdt.quota_type=ug, run 'lfs quotacheck' on a client, but
when I run 'lfs quotaon' I get the response:
   error: quotaon failed (Device or resource busy)
and a syslog line:
kernel: [ 6243.093049] LustreError: 3766:0:(quota_ctl.c:381:client_quota_ctl()) 
ptlrpc_queue_wait failed, rc: -16
The client is a patchless 2.6.30.9/Lustre 1_8b.

TIA, Ralf
-- 
Ralf Utermann
_
Universität Augsburg, Institut für Physik   --   EDV-Betreuer
Universitätsstr.1 
D-86135 Augsburg Phone:  +49-821-598-3231
SMTP: ralf.uterm...@physik.uni-augsburg.de Fax: -3411
___
Lustre-discuss mailing list
Lustre-discuss@lists.lustre.org
http://lists.lustre.org/mailman/listinfo/lustre-discuss


Re: [Lustre-discuss] OSS server

2009-11-30 Thread Brian J. Murrell
On Mon, 2009-11-30 at 11:23 +0100, Stefano Elmopi wrote: 
> Hi,

Hi, 
> - one server with MGS/MDS
> - one server with OSS
> - the OST is on SAN
> 
> 
> If the server OSS has a hardware failure
> and I must physically change the machine.

If the machine boots from the network then the new system is a drop-in
replacement for the broken machine and you need to do nothing, right?

Or if the system boots from it's own disk you would either a) put the
disk from the broken OSS into the new hardware or b) re-image the OSS
and configure it to be identical (IP address, etc.) as the machine that
died, or c) restore the system image from a backup or d) some other
reasonable facsimile/combination of the above options that results in
the new machine being a like-replacement for the broken machine.

None of this is really all that particular to Lustre though and it just
typical syadmin practises that one would follow for replacing any
machine that dies.  When your mail server dies, what do you do to
replace it such that it's an identical (to the users) replacement?  Same
situation for an OSS.

> With the new machines, how can I retrieve data on the OST, located on
> SAN ?

Once the new machine is configured identical to the one it's replacing,
it will just access that OST as if nothing happened, no?

> I hope to be clear enough !!

I think you are.  I guess the not-really-particular-to-Lustre nature of
the question had me somewhat thinking that there must be something I was
missing.  I guess there was not.

b.




signature.asc
Description: This is a digitally signed message part
___
Lustre-discuss mailing list
Lustre-discuss@lists.lustre.org
http://lists.lustre.org/mailman/listinfo/lustre-discuss


[Lustre-discuss] MDS performance

2009-11-30 Thread Francois Chassaing
Hi list, 
I know this question has been asked zillion times, but I'll ask it again, 
because as most lustre-builder wannabe, I'm worried about MDS being a 
bottleneck. 
I'm planning to build a 60TB lustre install separated in two pools : 
- one big / slow / cheap 50 TB archive based on Dell MD3000i+MD1000 RAID 5 OST 
with two cheap OSS in front of it 
- one beefier (=smaller / faster / more expensive) 10 TB based on 2 servers 
being both OSS+OST each composed of 8 cores, 12 GB RAM, 8 NL SAS spindles RAID 
10. 
the big / slow archive is not really a problem but i'm asking myself about the 
smaller one, given the fact that this storage (if good enough in terms of 
performance) could double pretty soon by adding more OSS+OST 
I've been planning to have one redundant MDS over 2 Dell R610 sharing a common 
sas-attached MD3000 RAID 10 MDT. 
I've noticed that this array is not the best on the market for perf. but is 
rather good at $$... 
I plan on connecting all this on DDR infiniband (along with the main clients -4 
of them-). 

So now for my questions about MDS : 
- Should I do better having two lustre installs instead of a single install 
with two pools ? 
- Should I consider buying a better array for MDT ? 
- Should I be better using even beefier MDS with internal storage and DRBD ? 

and about OSS : 
- should I take "smaller" OSS+OST to improve perf ? 
- should I split my OST storage "inside" the OSSes for the 10GB storage pool ? 

I'll be sooo grateful if someone could answer those, I'll be glad to provide 
any other details one would need to help me out. 

Thanks. 
___
Lustre-discuss mailing list
Lustre-discuss@lists.lustre.org
http://lists.lustre.org/mailman/listinfo/lustre-discuss


Re: [Lustre-discuss] 'Rename' kernel build result

2009-11-30 Thread Frank Heckes
Hello Brian,

first of all thanks for your patience and answers so far.

> On Thu, 2009-11-26 at 15:37 +0100, Frank Heckes wrote:
> > Hello Brian,
>
> Hi,
>
> > Well, I downloaded two 'source' files:
>
> Hrm.  Why are you dealing with source files?  Does our already-built
> binary kernel RPMs not suit your environment for some reason?
>
No, no I like the idea a lot, but unfortunately the kernel didn't boot
on the hardware we trying to install them (IBM pizzaboxes), even after
rebuilding the initrd, Systemmap... to extend the (initial) driver list.
So I decided to built the kernel from source. Here the system boots as
it should after installing it to one of the nodes of the (test)
cluster.
And of course there's also a small piece of (childlike) curiosity, too.

> > -1- kernel-lustre-source-2.6.27.29-0.1_lustre.1.8.1.1.src.rpm
> > -2- kernel-lustre-source-2.6.27.29-0.1_lustre.1.8.1.1.x86_64.rpm
> >
> > I used the second for the compilation since the first only install stuff
> > in /usr/src/packages/SOURCE and some more complicated procedure has to
> > be used for compilation.
>
> The first is a source rpm which is built using the rpmbuild command.
>
> > Entries for /usr/src/linux-2.6.27.29-0.1_lustre.1.8.1.1
>
> Yes.
>
> > and /lib/modules-2.6.27.29-0.1_lustre.1.8.1.1
>
> Well, just the /lib/modules/2.6.27.29-0.1_lustre.1.8.1.1-default/source
> symlink so this is really quite irrelevant.  It's
> the /usr/src/linux-2.6.27.29-0.1_lustre.1.8.1.1 tree that you are
> interested in.
>
> > will be created when installing rpm '-2-'.
>
> > Running 'make rpm' leads to
> > an RPM with the specified install pathes inside the RPM.
>
> Can you tell me what the first 10 lines
> of /usr/src/linux-2.6.27.29-0.1_lustre.1.8.1.1-obj/x86_64/default/Makefile 
> say?  You might need to adjust some of those lines to make the version say 
> what you want it to say.
>
> But most definitely, things are a lot easier if you can use our binary
> RPM packages.  I'd be interested in knowing why they are not suitable
> for you.
>
I fully agree. I would to the ready made RPM's if it would be possible.
(see above).

Here're the Makefile's first 10 lines:

node006:~ # head
-10 /usr/src/linux-2.6.27.29-0.1_lustre.1.8.1.1-obj/x86_64/default/Makefile
# Automatically generated
by ../../../linux-2.6.27.29-0.1_lustre.1.8.1.1/scripts/mkmakefile: don't
edit

VERSION = 2
PATCHLEVEL = 6

lastword = $(word $(words $(1)),$(1))
makedir := $(dir $(call lastword,$(MAKEFILE_LIST)))

MAKEARGS := -C ../../../linux-2.6.27.29-0.1_lustre.1.8.1.1
MAKEARGS += O=$(if $(patsubst /%,,$(makedir)),$(CURDIR)/)$(patsubst %/,
%,$(makedir))

> > So kernel '-2-' shouldn't be used for compilation?
>
> That should be fine.
>
Okay. That's the one I used.

> b.
>




Forschungszentrum Juelich GmbH
52425 Juelich
Sitz der Gesellschaft: Juelich
Eingetragen im Handelsregister des Amtsgerichts Dueren Nr. HR B 3498
Vorsitzende des Aufsichtsrats: MinDir'in Baerbel Brumme-Bothe
Geschaeftsfuehrung: Prof. Dr. Achim Bachem (Vorsitzender),
Dr. Ulrich Krafft (stellv. Vorsitzender), Prof. Dr.-Ing. Harald Bolt,
Prof. Dr. Sebastian M. Schmidt


___
Lustre-discuss mailing list
Lustre-discuss@lists.lustre.org
http://lists.lustre.org/mailman/listinfo/lustre-discuss


[Lustre-discuss] OSS server

2009-11-30 Thread Stefano Elmopi



Hi,

I apologize for my lack of clarity, try to be clearer.
The scenario is:

- one server with MGS/MDS
- one server with OSS
- the OST is on SAN

If the server OSS has a hardware failure
and I must physically change the machine.
With the new machines, how can I retrieve data on the OST, located on  
SAN ?

I hope to be clear enough !!
Thanks



Ing. Stefano Elmopi
Gruppo Darco - Resp. ICT Sistemi
Via Ostiense 131/L Corpo B, 00154 Roma

cell. 3466147165
tel.  0657060500
email:stefano.elm...@sociale.it

Il giorno 20/nov/09, alle ore 20:00, lustre-discuss-requ...@lists.lustre.org 
 ha scritto:



Date: Fri, 20 Nov 2009 10:53:15 -0500
From: "Brian J. Murrell" 
Subject: Re: [Lustre-discuss] OSS server
To: lustre-discuss@lists.lustre.org
Message-ID: <1258732395.21860.30.ca...@pc.interlinx.bc.ca>
Content-Type: text/plain; charset="utf-8"

On Fri, 2009-11-20 at 14:37 +0100, Stefano Elmopi wrote:

If I have a simple scenario with one MGS/MDS and one OSS, without
failover,
and the server OSS crash,


I'm afraid "crash" is an ambiguous term in this context.  Do you mean
the O/S panics or some equivalent event?  The solution would be to
simply reboot the machine, no?

Or do you mean that the hardware itself has broken such that you  
need to
replace it?  Wouldn't you just move the disks in the old machine to  
the

new one and boot that so that as far as the rest of the network is
concerned, the machine was simply rebooted?


Simply replace the server root with one that has the same IP ??


Yes, I think so.  Although your scenario is not terribly clear for me.

b.


-- next part --
A non-text attachment was scrubbed...
Name: not available
Type: application/pgp-signature
Size: 197 bytes
Desc: This is a digitally signed message part
Url : 
http://lists.lustre.org/pipermail/lustre-discuss/attachments/20091120/bbdcf19b/attachment-0001.bin



___
Lustre-discuss mailing list
Lustre-discuss@lists.lustre.org
http://lists.lustre.org/mailman/listinfo/lustre-discuss