[OpenAFS] 1.6.0pre2 Filelog CB: WhoAreYou failed for host

2011-03-20 Thread Gémes Géza
Hi,

I know this topic has been discussed before, but the conclusion was that
it is caused by NAT.
This is impossible in my case, as openafs servers are firewalled from
the outside world.
The fileserver has 3 ethernet interfaces:
1: connected to the clients, two IP addresses one active (other in the
NetRestrict file)
2: connected to a SAN, no IP addresses
3: connected to other cluster memebers, IP address in the NetRestrict file
vos listaddrs gives nothing just the right IP address for the vol and
fileserver.
The FileLog is full of entries like:
CB: WhoAreYou failed for host FILESERVER
Besides that all the clients (1.6.0pre2 on linux, 1.5.78 and 1.6.0pre3
on windows) are working as expected.
Except one (1.6.0pre2 on linux) which has two interfaces (one connected
two the Fileservers network and the other in its NetRestrict file). This
has entries in its syslog like this:
Mar 21 07:31:36 ssh-gate kernel: [ 4879.645660] afs: WARM shutting down
of: CBCELLNAME afsCELLNAME BkGCELLNAME CTruncCELLNAME AFSDBCELLNAME
RxEventCELLNAME UnmaskRxkSignalsCELLNAME RxListenerCELLNAME  ALL
allocated tablesCELLNAME done
Mar 21 07:31:36 HOSTNAME kernel: [ 4880.229317] enabling dynamically
allocated vcaches
Mar 21 07:31:36 HOSTNAME kernel: [ 4880.229321] Starting AFS cache
scanCELLNAMEfound 70 non-empty cache files (2%).
Mar 21 07:32:51 HOSTNAME kernel: [ 4954.852104] afs: Lost contact with
file server FILESERVER in cell CELLNAME (all multi-homed ip addresses
down for the server)
Mar 21 07:32:51 HOSTNAME kernel: [ 4954.852115] afs: Lost contact with
file server FILESERVER in cell CELLNAME (all multi-homed ip addresses
down for the server)
Mar 21 07:32:51 HOSTNAME kernel: [ 4954.852119] RXAFS_GetCapabilities
failed with code -3
Mar 21 07:32:54 HOSTNAME kernel: [ 4958.304958] afs: file server
FILESERVER in cell CELLNAME is back up (multi-homed address; other
same-host interfaces may still be down)
Mar 21 07:32:54 HOSTNAME kernel: [ 4958.304965] afs: file server
FILESERVER in cell CELLNAME is back up (multi-homed address; other
same-host interfaces may still be down)
the first ls on /afs/CELLNAME/ always fail, and later on (after those
syslog entries, except restarting of course) it starts working.

Thanks for any idea!

Cheers

Geza

___
OpenAFS-info mailing list
OpenAFS-info@openafs.org
https://lists.openafs.org/mailman/listinfo/openafs-info


Re: [OpenAFS] Re: Crazy DAFS problem (with log)

2011-03-20 Thread Ryan C. Underwood

Well, that didn't take long, it already blew up again, in exactly the
same fashion, with the same two culprit volumes.

-- 
Ryan C. Underwood, 
___
OpenAFS-info mailing list
OpenAFS-info@openafs.org
https://lists.openafs.org/mailman/listinfo/openafs-info


Re: [OpenAFS] Re: Crazy DAFS problem (with log)

2011-03-20 Thread Ryan C. Underwood
On Sun, Mar 20, 2011 at 09:57:22PM -0500, Andrew Deason wrote:
> 
> Oh, hmm, this starts too late. I guess you don't have the BosLog before
> this?

Nope, wish it rotated more old logs out but haven't looked into this.

> > 03/20/2011 16:36:17 The volume header file /vicepa/V0536871274.vol is not 
> > associated with any actual data (deleted)
> > 03/20/2011 16:40:14 dispatching child to salvage volume 536871273...
> > 03/20/2011 16:40:14 2 nVolumesInInodeFile 56 
> > 03/20/2011 16:40:14 CHECKING CLONED VOLUME 536871274.
> > 03/20/2011 16:40:14 www_logs.readonly (536871274) updated 03/20/2011 16:39
> > 03/20/2011 16:40:14 Vnode 22: length incorrect; (is 1794950 should be 0)
> 
> Do you know if this volume was recreated between these times? I'm not
> sure if the salvager is just lying, but this would indicate it deleted
> the RO, and then it came back again and again has invalid metadata.

If by recreated you mean a remsite/addsite, the answer is no.

> > [.. ad infinitum ]
> 
> Actually, it may help to continue for an interation or two more. I don't
> think these salvages have repeated themselves yet.

Will attach at the end.

> > It's a RO clone, should I just delete it and add it back?
> 
> Yes, if it'll let you. The salvage is supposed to delete it for you, but
> if it's not it's not.

I deleted all the clones in question and manually salvaged the RW
volumes then recreated and re-released the clones.  Let's see if the
gremlin comes back.

> Based on the salsrv logs, I'd be looking at vnodes 264, 266, and 268 in
> 536870915/6, and 22, 202, and 204 in 536871273/4. That is, see if the
> 'UFS-Filename' file on disk actually has the length in the 'volinfo'
> output.

If/when it comes back I'll dig into this stuff.

> >>>Shortly after the weekly scheduled fileserver restart
> 
> You shouldn't need to have these on. But they shouldn't be causing
> salvages or whatever, either.

OK, this used to be standard practice, but I set up back in the 1.2 days
so a lot could have changed since then.  I'll remove it.

Long SalsrvLog excerpt:

@(#) OpenAFS 1.6.0~pre3-0-debian built  2011-03-18 
03/20/2011 04:00:29 Starting OpenAFS Online Salvage Server 2.4 
(/usr/lib/openafs/salvageserver)
03/20/2011 04:02:02 dispatching child to salvage volume 536870915...
03/20/2011 04:02:02 2 nVolumesInInodeFile 56 
03/20/2011 04:02:02 waiting for fileserver to finish scanning partition 
/vicepa...
03/20/2011 04:02:03 CHECKING CLONED VOLUME 536870916.
03/20/2011 04:02:03 root.cell.readonly (536870916) updated 03/20/2011 00:12
03/20/2011 04:02:03 Vnode 266: length incorrect; (is 2587609 should be 0)
03/20/2011 04:02:03 SALVAGING VOLUME 536870915.
03/20/2011 04:02:03 root.cell (536870915) updated 03/20/2011 00:12
03/20/2011 04:02:03 Vnode 266: length incorrect; changed from 2587609 to 0
03/20/2011 04:02:03 totalInodes 190
03/20/2011 04:02:03 Salvaged root.cell (536870915): 183 files, 62821 blocks
03/20/2011 04:02:05 dispatching child to salvage volume 536871273...
03/20/2011 04:02:05 2 nVolumesInInodeFile 56 
03/20/2011 04:02:05 CHECKING CLONED VOLUME 536871274.
03/20/2011 04:02:05 www_logs.readonly (536871274) updated 03/19/2011 18:35
03/20/2011 04:02:05 Vnode 22: length incorrect; (is 1683156 should be 0)
03/20/2011 04:02:05 SALVAGING VOLUME 536871273.
03/20/2011 04:02:05 www_logs (536871273) updated 03/19/2011 18:35
03/20/2011 04:02:05 Vnode 22: length incorrect; changed from 1683156 to 0
03/20/2011 04:02:05 Vnode 24: length incorrect; changed from 184572 to 0
03/20/2011 04:02:05 Vnode 202: length incorrect; changed from 232747 to 0
03/20/2011 04:02:05 totalInodes 245
03/20/2011 04:02:05 Salvaged www_logs (536871273): 238 files, 241815 blocks
03/20/2011 04:04:01 dispatching child to salvage volume 536870916...
03/20/2011 04:04:01 dispatching child to salvage volume 536870915...
03/20/2011 04:04:01 namei_ListAFSSubDirs: warning: VG 536870916 does not have a 
link table; salvager will recreate it.
03/20/2011 04:04:01 fileserver requested salvage of clone 536870916; scheduling 
salvage of volume group 536870915...
03/20/2011 04:04:01 1 nVolumesInInodeFile 28 
03/20/2011 04:04:01 SALVAGING VOLUME 536870915.
03/20/2011 04:04:01 root.cell (536870915) updated 03/20/2011 04:03
03/20/2011 04:04:01 totalInodes 187
03/20/2011 04:04:01 Salvaged root.cell (536870915): 183 files, 62824 blocks
03/20/2011 04:04:01 The volume header file /vicepa/V0536870916.vol is not 
associated with any actual data (deleted)
03/20/2011 04:04:13 dispatching child to salvage volume 536871274...
03/20/2011 04:04:13 dispatching child to salvage volume 536871273...
03/20/2011 04:04:13 namei_ListAFSSubDirs: warning: VG 536871274 does not have a 
link table; salvager will recreate it.
03/20/2011 04:04:13 fileserver requested salvage of clone 536871274; scheduling 
salvage of volume group 536871273...
03/20/2011 04:04:13 1 nVolumesInInodeFile 28 
03/20/2011 04:04:13 SALVAGING VOLUME 536871273.
03/20/2011 04:04:13 www_logs (536871273) updated 03/19/2011 18:35
03/20/2011 04:04:

[OpenAFS] Re: Crazy DAFS problem (with log)

2011-03-20 Thread Andrew Deason
On Sun, 20 Mar 2011 20:59:43 -0500
"Ryan C. Underwood"  wrote:

> Not much interesting in BosLog.old:
> Sun Mar 20 04:00:29 2011: Core limits now -1 -1

Oh, hmm, this starts too late. I guess you don't have the BosLog before
this?

> 03/20/2011 16:34:31 CHECKING CLONED VOLUME 536871274.
> 03/20/2011 16:34:31 www_logs.readonly (536871274) updated 03/20/2011 05:45
> 03/20/2011 16:34:31 Vnode 22: length incorrect; (is 1773498 should be 0)
> 03/20/2011 16:34:31 SALVAGING VOLUME 536871273.

The incorrect vnode length should result in the RO getting deleted. I'm
not sure why it's not; I'll look tomorrow. I expect that's not a dafs
problem, though; just a general salvager issue.

A vnode that keeps having the wrong length might be one explanation for
the repeated salvages, but it still seems like the headers are screwed
up, which is (probably) unrelated.

> 03/20/2011 16:36:17 The volume header file /vicepa/V0536871274.vol is not 
> associated with any actual data (deleted)
> 03/20/2011 16:40:14 dispatching child to salvage volume 536871273...
> 03/20/2011 16:40:14 2 nVolumesInInodeFile 56 
> 03/20/2011 16:40:14 CHECKING CLONED VOLUME 536871274.
> 03/20/2011 16:40:14 www_logs.readonly (536871274) updated 03/20/2011 16:39
> 03/20/2011 16:40:14 Vnode 22: length incorrect; (is 1794950 should be 0)

Do you know if this volume was recreated between these times? I'm not
sure if the salvager is just lying, but this would indicate it deleted
the RO, and then it came back again and again has invalid metadata.

> [.. ad infinitum ]

Actually, it may help to continue for an interation or two more. I don't
think these salvages have repeated themselves yet.

> > > Sun Mar 20 04:04:01 2011 VAttachVolume: Error reading diskDataHandle 
> > > header for vol 536870916; error=101
> > > Sun Mar 20 04:04:01 2011 Scheduling salvage for volume 536870916 on part 
> > > /vicepa over SALVSYNC
> > 
> > "volume metadata is corrupt for 536870916"
> 
> It's a RO clone, should I just delete it and add it back?

Yes, if it'll let you. The salvage is supposed to delete it for you, but
if it's not it's not.

> Does this look normal?  I.e., a whole lot of zero-length vnodes
> attached to the same vice file.

Those are fine; vnodes like this:

> 15040 Vnode 472.0.0 cloned: 0, length: 0 linkCount: 0 parent: 0 UFS-Filename: 
> /vicepa/AFSIDat/d=/d3++U/+/+/+

That is, vnodes like "Vnode X.0.0", are 'blank'; the vnode entry is just
zeroed out because the indices are expanded by some amount at a time in
chunks (512, or something?).

Based on the salsrv logs, I'd be looking at vnodes 264, 266, and 268 in
536870915/6, and 22, 202, and 204 in 536871273/4. That is, see if the
'UFS-Filename' file on disk actually has the length in the 'volinfo'
output.

And by the way:

>>>Shortly after the weekly scheduled fileserver restart

You shouldn't need to have these on. But they shouldn't be causing
salvages or whatever, either.

-- 
Andrew Deason
adea...@sinenomine.net
___
OpenAFS-info mailing list
OpenAFS-info@openafs.org
https://lists.openafs.org/mailman/listinfo/openafs-info


Re: [OpenAFS] Re: Crazy DAFS problem (with log)

2011-03-20 Thread Ryan C. Underwood

On Sun, Mar 20, 2011 at 07:42:12PM -0500, Andrew Deason wrote:
> 
> What's in SalsrvLog and SalsrvLog.old? It should have a bunch of stuff
> on the salvages here. Also, BosLog for this time period would be good
> for completeness.

Not much interesting in BosLog.old:
Sun Mar 20 04:00:29 2011: Core limits now -1 -1
Sun Mar 20 04:00:29 2011: Server directory access is okay
Sun Mar 20 16:32:17 2011: ptserver exited on signal 15
Sun Mar 20 16:32:17 2011: vlserver exited on signal 15
Sun Mar 20 16:32:17 2011: dafs:salsrv exited on signal 15
Sun Mar 20 16:32:17 2011: dafs:vol exited on signal 15
Sun Mar 20 16:32:18 2011: dafs:file exited with code 0
Sun Mar 20 16:32:18 2011: Shutdown of BOS server and processes in response to 
signal 15

SalsrvLog, on the other hand, shows craziness.  It just keeps salvaging
the same two volumes again and again, for hours.

03/20/2011 16:34:06 Starting OpenAFS Online Salvage Server 2.4 
(/usr/lib/openafs/salvageserver)
03/20/2011 16:34:30 dispatching child to salvage volume 536870915...
03/20/2011 16:34:30 dispatching child to salvage volume 536871273...
03/20/2011 16:34:30 2 nVolumesInInodeFile 56 
03/20/2011 16:34:30 waiting for fileserver to finish scanning partition 
/vicepa...
03/20/2011 16:34:31 CHECKING CLONED VOLUME 536871274.
03/20/2011 16:34:31 www_logs.readonly (536871274) updated 03/20/2011 05:45
03/20/2011 16:34:31 Vnode 22: length incorrect; (is 1773498 should be 0)
03/20/2011 16:34:31 SALVAGING VOLUME 536871273.
03/20/2011 16:34:31 www_logs (536871273) updated 03/20/2011 05:45
03/20/2011 16:34:31 Vnode 22: length incorrect; changed from 1773498 to 0
03/20/2011 16:34:31 Vnode 202: length incorrect; changed from 345557 to 0
03/20/2011 16:34:31 Vnode 204: length incorrect; changed from 86723 to 0
03/20/2011 16:34:31 totalInodes 245
03/20/2011 16:34:32 Salvaged www_logs (536871273): 238 files, 241815 blocks
03/20/2011 16:34:30 2 nVolumesInInodeFile 56 
03/20/2011 16:34:30 waiting for fileserver to finish scanning partition 
/vicepa...
03/20/2011 16:34:31 CHECKING CLONED VOLUME 536870916.
03/20/2011 16:34:31 root.cell.readonly (536870916) updated 03/20/2011 12:39
03/20/2011 16:34:31 Vnode 264: length incorrect; (is 15840312 should be 0)
03/20/2011 16:34:31 SALVAGING VOLUME 536870915.
03/20/2011 16:34:31 root.cell (536870915) updated 03/20/2011 12:41
03/20/2011 16:34:31 Vnode 264: length incorrect; changed from 15840312 to 0
03/20/2011 16:34:31 Vnode 266: length incorrect; changed from 2589178 to 0
03/20/2011 16:34:31 Vnode 268: length incorrect; changed from 7364175 to 0
03/20/2011 16:34:31 totalInodes 191
03/20/2011 16:34:32 Salvaged root.cell (536870915): 183 files, 37402 blocks
03/20/2011 16:35:24 dispatching child to salvage volume 536870915...
03/20/2011 16:35:24 1 nVolumesInInodeFile 28 
03/20/2011 16:35:25 SALVAGING VOLUME 536870915.
03/20/2011 16:35:25 root.cell (536870915) updated 03/20/2011 12:41
03/20/2011 16:35:25 totalInodes 187
03/20/2011 16:35:25 Salvaged root.cell (536870915): 183 files, 37402 blocks
03/20/2011 16:35:25 The volume header file /vicepa/V0536870916.vol is not 
associated with any actual data (deleted)
03/20/2011 16:36:17 dispatching child to salvage volume 536871274...
03/20/2011 16:36:17 dispatching child to salvage volume 536871273...
03/20/2011 16:36:17 namei_ListAFSSubDirs: warning: VG 536871274 does not have a 
link table; salvager will recreate it.
03/20/2011 16:36:17 fileserver requested salvage of clone 536871274; scheduling 
salvage of volume group 536871273...
03/20/2011 16:36:17 1 nVolumesInInodeFile 28 
03/20/2011 16:36:17 SALVAGING VOLUME 536871273.
03/20/2011 16:36:17 www_logs (536871273) updated 03/20/2011 16:36
03/20/2011 16:36:17 totalInodes 242
03/20/2011 16:36:17 Salvaged www_logs (536871273): 238 files, 243808 blocks
03/20/2011 16:36:17 The volume header file /vicepa/V0536871274.vol is not 
associated with any actual data (deleted)
03/20/2011 16:40:14 dispatching child to salvage volume 536871273...
03/20/2011 16:40:14 2 nVolumesInInodeFile 56 
03/20/2011 16:40:14 CHECKING CLONED VOLUME 536871274.
03/20/2011 16:40:14 www_logs.readonly (536871274) updated 03/20/2011 16:39
03/20/2011 16:40:14 Vnode 22: length incorrect; (is 1794950 should be 0)
03/20/2011 16:40:14 SALVAGING VOLUME 536871273.
03/20/2011 16:40:14 www_logs (536871273) updated 03/20/2011 16:39
[.. ad infinitum ]

> I am also assuming there's no SalvageLog* files with entries anywhere
> near this time. If that's not true, include those.

Correct

> > Sun Mar 20 04:04:01 2011 VAttachVolume: Error reading diskDataHandle header 
> > for vol 536870916; error=101
> > Sun Mar 20 04:04:01 2011 Scheduling salvage for volume 536870916 on part 
> > /vicepa over SALVSYNC
> 
> "volume metadata is corrupt for 536870916"

It's a RO clone, should I just delete it and add it back?

> > Sun Mar 20 04:14:14 2011 CopyOnWrite failed: volume 536871273 in partition 
> > /vicepa  (tried reading 8192, read 0, wrote 0, errno 0) volume needs salvage
> 
> Someone wrote to a file in 536

Re: [OpenAFS] Multiple logins

2011-03-20 Thread Jaap Winius

Quoting Jason Edgecombe :

Is this enforcing a policy decision or just preventing technical  
problems caused by multiple logins?


For my site it is strictly to prevent technical problems.

I'm wondering because we us gnome on RHEL5 with AFS home directories  
and multiple logins on different machines haven't been an issue, but  
trying to login to gnome multiple times on the same machine does  
cause error messages.


Interesting. Since Xfce is based on the GTK+ v2 toolkit (the same as  
GNOME), it may not suffer the problems that I anticipate. I'll run  
some tests later on when I have to opportunity and report the results.


Cheers,

Jaap
___
OpenAFS-info mailing list
OpenAFS-info@openafs.org
https://lists.openafs.org/mailman/listinfo/openafs-info


[OpenAFS] Re: Crazy DAFS problem (with log)

2011-03-20 Thread Andrew Deason
On Sun, 20 Mar 2011 16:50:16 -0500
"Ryan C. Underwood"  wrote:

> Shortly after the weekly scheduled fileserver restart, things blew up
> in a big way.  My RW root.cell was inaccessible in the end.  No kernel
> messages indicating filesystem or disk problems underneath.  I
> force-fscked the vice partition (ext4) and no problems were found.
> Any speculation on what in the world happened here?  This is running
> -pre3 with the patch from Andrew mentioned earlier.

What's in SalsrvLog and SalsrvLog.old? It should have a bunch of stuff
on the salvages here. Also, BosLog for this time period would be good
for completeness.

I am also assuming there's no SalvageLog* files with entries anywhere
near this time. If that's not true, include those.

> Sun Mar 20 04:00:30 2011 File Server started Sun Mar 20 04:00:30 2011
> Sun Mar 20 04:02:02 2011 Scheduling salvage for volume 536870915 on part 
> /vicepa over SALVSYNC

There are not many places that VRequestSalvage_r can get called without
logging something about why... I think the only places are when the
volume header flags say that the volume needs salvaging.

> Sun Mar 20 04:04:01 2011 VAttachVolume: Error reading diskDataHandle header 
> for vol 536870916; error=101
> Sun Mar 20 04:04:01 2011 Scheduling salvage for volume 536870916 on part 
> /vicepa over SALVSYNC

"volume metadata is corrupt for 536870916"

> Sun Mar 20 04:14:14 2011 CopyOnWrite failed: volume 536871273 in partition 
> /vicepa  (tried reading 8192, read 0, wrote 0, errno 0) volume needs salvage

Someone wrote to a file in 536871273, and we needed to CoW the file. But
we couldn't read all of the data from the file to make a new copy. Based
on those values, I'd guess the file is empty on disk, but it's supposed
to contain some data.

With a patch on the master branch, such a situation should be caught
earlier and more information logged. But that wouldn't even really help
too much if that were the case. You can find files like this by doing
'volinfo /vicepa 536871273 -filenames', and seeing the alleged length
for each vnode, and seeing if the actual file size on disk of the
reported filenames match.

However, salvaging is supposed to fix that, so it would be good to see
what the salvage logs say.

-- 
Andrew Deason
adea...@sinenomine.net

___
OpenAFS-info mailing list
OpenAFS-info@openafs.org
https://lists.openafs.org/mailman/listinfo/openafs-info


[OpenAFS] Crazy DAFS problem (with log)

2011-03-20 Thread Ryan C. Underwood

Shortly after the weekly scheduled fileserver restart, things blew up in
a big way.  My RW root.cell was inaccessible in the end.  No kernel
messages indicating filesystem or disk problems underneath.  I force-fscked
the vice partition (ext4) and no problems were found.  Any speculation
on what in the world happened here?  This is running -pre3 with the
patch from Andrew mentioned earlier.


Sun Mar 20 04:00:29 2011 File server starting (/usr/lib/openafs/dafileserver -p 
123 -pctspare 20 -L -busyat 50 -rxpck 2000 -rxbind -cb 400 -vattachpar 128 
-vlruthresh 1440 -vlrumax 8 -vhashsize 11)
Sun Mar 20 04:00:29 2011 afs_krb_get_lrealm failed, using icequake.net.
Sun Mar 20 04:00:30 2011 VLRU: starting scanner with the following 
configuration parameters:
Sun Mar 20 04:00:30 2011 VLRU:  offlining volumes after minimum of 86400 
seconds of inactivity
Sun Mar 20 04:00:30 2011 VLRU:  running VLRU soft detach pass every 120 seconds
Sun Mar 20 04:00:30 2011 VLRU:  taking up to 8 volumes offline per pass
Sun Mar 20 04:00:30 2011 VLRU:  scanning generation 0 for inactive volumes 
every 10800 seconds
Sun Mar 20 04:00:30 2011 VLRU:  scanning for promotion/demotion between 
generations 0 and 1 every 172800 seconds
Sun Mar 20 04:00:30 2011 VLRU:  scanning for promotion/demotion between 
generations 1 and 2 every 345600 seconds
Sun Mar 20 04:00:30 2011 Set thread id 3 for FSYNC_sync
Sun Mar 20 04:00:30 2011 VInitVolumePackage: beginning parallel fileserver 
startup
Sun Mar 20 04:00:30 2011 VInitVolumePackage: using 1 threads to pre-attach 
volumes on 1 partitions
Sun Mar 20 04:00:30 2011 Scanning partitions on thread 1 of 1
Sun Mar 20 04:00:30 2011 Partition /vicepa: pre-attaching volumes
Sun Mar 20 04:00:30 2011 fs_stateRestore: commencing fileserver state restore
Sun Mar 20 04:00:30 2011 Partition scan thread 1 of 1 ended
Sun Mar 20 04:00:30 2011 fs_stateRestore: host table restored
Sun Mar 20 04:00:30 2011 fs_stateRestore: FileEntry and CallBack tables restored
Sun Mar 20 04:00:30 2011 fs_stateRestore: host table indices remapped
Sun Mar 20 04:00:30 2011 fs_stateRestore: FileEntry and CallBack indices 
remapped
Sun Mar 20 04:00:30 2011 fs_stateRestore: restore phase complete
Sun Mar 20 04:00:30 2011 fs_stateRestore: beginning state verification phase
Sun Mar 20 04:00:30 2011 fs_stateRestore: fileserver state verification complete
Sun Mar 20 04:00:30 2011 fs_stateRestore: restore was successful
Sun Mar 20 04:00:30 2011 Getting FileServer name...
Sun Mar 20 04:00:30 2011 Set thread id 007E for 'FiveMinuteCheckLWP'
Sun Mar 20 04:00:30 2011 FileServer host name is 'valhalla'
Sun Mar 20 04:00:30 2011 Set thread id 0082 for 'FsyncCheckLWP'
Sun Mar 20 04:00:30 2011 Set thread id 007F for 'HostCheckLWP'
Sun Mar 20 04:00:30 2011 Getting FileServer address...
Sun Mar 20 04:00:30 2011 FileServer valhalla has address 10.0.1.232 (0xe801000a 
or 0xa0001e8 in host byte order)
Sun Mar 20 04:00:30 2011 File Server started Sun Mar 20 04:00:30 2011
Sun Mar 20 04:02:02 2011 Scheduling salvage for volume 536870915 on part 
/vicepa over SALVSYNC
Sun Mar 20 04:02:02 2011 VVGC_scan_partition: finished scanning /vicepa: 147 
volumes in 74 groups
Sun Mar 20 04:02:03 2011 fssync: breaking all call backs for volume 536870915
Sun Mar 20 04:02:05 2011 Scheduling salvage for volume 536871273 on part 
/vicepa over SALVSYNC
Sun Mar 20 04:02:05 2011 fssync: breaking all call backs for volume 536871273
Sun Mar 20 04:04:01 2011 VAttachVolume: Error reading diskDataHandle header for 
vol 536870916; error=101
Sun Mar 20 04:04:01 2011 Scheduling salvage for volume 536870916 on part 
/vicepa over SALVSYNC
Sun Mar 20 04:04:01 2011 warning: volume 536870916 recursively checked out by 
programType id 4
Sun Mar 20 04:04:13 2011 Scheduling salvage for volume 536871274 on part 
/vicepa over SALVSYNC
Sun Mar 20 04:04:13 2011 warning: volume 536871274 recursively checked out by 
programType id 4
Sun Mar 20 04:06:02 2011 fssync: breaking all call backs for volume 536870916
Sun Mar 20 04:10:02 2011 fssync: breaking all call backs for volume 536871274
Sun Mar 20 04:12:02 2011 fssync: breaking all call backs for volume 536871274
Sun Mar 20 04:14:03 2011 fssync: breaking all call backs for volume 536871274
Sun Mar 20 04:14:14 2011 CopyOnWrite failed: volume 536871273 in partition 
/vicepa  (tried reading 8192, read 0, wrote 0, errno 0) volume needs salvage
Sun Mar 20 04:14:14 2011 CopyOnWrite failed: requesting salvage
Sun Mar 20 04:14:14 2011 Scheduling salvage for volume 536871273 on part 
/vicepa over SALVSYNC
Sun Mar 20 04:14:14 2011 fssync: breaking all call backs for volume 536871273
Sun Mar 20 04:16:02 2011 Scheduling salvage for volume 536871274 on part 
/vicepa over SALVSYNC
Sun Mar 20 04:18:02 2011 fssync: breaking all call backs for volume 536871274
Sun Mar 20 04:20:02 2011 fssync: breaking all call backs for volume 536871274
Sun Mar 20 04:20:04 2011 CopyOnWrite failed: volume 536871273 in partition 
/vicepa  (tried reading 8192, read 0, wrote 

Re: [OpenAFS] Multiple logins

2011-03-20 Thread Jason Edgecombe

On 03/20/2011 08:19 AM, Coy Hile wrote:

On Sat, Mar 19, 2011 at 1:23 PM, Jaap Winius  wrote:

Quoting Dirk Heinrichs:


... Is it possible to prevent users from logging in more than once ...

No, you can't. ...

Couldn't you potentially write a PAM module to do exactly that?  At
the top of the session stack, have it store the status of the user's
session in LDAP somewhere (or potentially in some other database, and
then on logout, remove the "Joe has an active session" flag.  Then,
upon a second or subsequent attempt at login, the PAM module could
kick the user out?  I don't know the logistics of doing so,
unfortunately; potentially Russ could give a better hand-wave
solution?
Is this enforcing a policy decision or just preventing technical 
problems caused by multiple logins?


I'm wondering because we us gnome on RHEL5 with AFS home directories and 
multiple logins on different machines haven't been an issue, but trying 
to login to gnome multiple times on the same machine does cause error 
messages.


Jason
___
OpenAFS-info mailing list
OpenAFS-info@openafs.org
https://lists.openafs.org/mailman/listinfo/openafs-info


Re: [OpenAFS] Multiple logins

2011-03-20 Thread Coy Hile
On Sat, Mar 19, 2011 at 1:23 PM, Jaap Winius  wrote:
> Quoting Dirk Heinrichs :
>
>>> ... Is it possible to prevent users from logging in more than once ...
>>
>> No, you can't. ...

Couldn't you potentially write a PAM module to do exactly that?  At
the top of the session stack, have it store the status of the user's
session in LDAP somewhere (or potentially in some other database, and
then on logout, remove the "Joe has an active session" flag.  Then,
upon a second or subsequent attempt at login, the PAM module could
kick the user out?  I don't know the logistics of doing so,
unfortunately; potentially Russ could give a better hand-wave
solution?

-Coy
___
OpenAFS-info mailing list
OpenAFS-info@openafs.org
https://lists.openafs.org/mailman/listinfo/openafs-info