Re: SFS problem

2011-04-20 Thread John Hall
If I recall correctly, DIAG D4 is the one that manipulates the secondary ID,
aka Alternate ID.  (SECUSER is an old term).  Diag 88 provides the ability
to link minidisks and perform userid/password validations (if the issuer has
appropriate authority).

An interesting usage note for this is that DIAG D4 does not change the
userid associated with any existing (already active) SFS connections.  This
is because it is a CP function and manipulates the VMDBK.  Once set, future
connections (via APPC/VM Connect) will utilize the Alternate ID. ... This is
why severing all connections prior to setting the AltID will fix this type
of problem, because CMS will (re) connect and use the AltID.

If this is Nora's problem, an easy work around would be wrap the job with an
exec that uses DMSGETWU and DMSPUSWU to set a new default work unit that
contains the appropriate user, then run the job from the exec, then finally
reset with DMSPOPWU. (If I'm remembering all of this correctly)

John

--
John Hall
Safe Software, Inc.
727-608-8799
johnh...@safesoftware.com



On Tue, Apr 19, 2011 at 11:48 AM, Schuh, Richard rsc...@visa.com wrote:

  Isn't that DIAG 88, instead of SECUSER?


 Regards,
 Richard Schuh




  --
 *From:* The IBM z/VM Operating System [mailto:IBMVM@LISTSERV.UARK.EDU] *On
 Behalf Of *John Hall
 *Sent:* Tuesday, April 19, 2011 6:41 AM

 *To:* IBMVM@LISTSERV.UARK.EDU
 *Subject:* Re: SFS problem

 Nora,
 Batch jobs normally run with the privileges of the owner of the job,
 using the SECUSER facility in z/VM.  With SFS, this can lead to unexpected
 results when a prior batch job leaves the worker with a connection to the
 filepool under a different user's id.  If the job ordering/selection of
 batch workers is somewhat random, you could see the outcome that you're
 experiencing (sometimes it works, sometimes it fails).




Re: SFS problem

2011-04-20 Thread Graves Nora E
I do not believe that this is the problem. I was giving you information
on the failing job that I am most familiar with.  Other jobs that fail
are submitted to VMBatch by the ID that also owns the SFS directories.
 
FYI, VMBatch is the IBM VM Batch Facility Version 2.2,  I was not
aware until recently that there are other products also known as
VMBatch.  If this has caused confusion, I apologize.
 
Nora Graves
nora.e.gra...@irs.gov
Main IRS, Room 6531
(202) 622-6735 
Fax (202) 622-3123
SE:W:CAR:MP:D:KS:BRSI
 



From: The IBM z/VM Operating System [mailto:IBMVM@LISTSERV.UARK.EDU] On
Behalf Of John Hall
Sent: Wednesday, April 20, 2011 10:04 AM
To: IBMVM@LISTSERV.UARK.EDU
Subject: Re: SFS problem


If I recall correctly, DIAG D4 is the one that manipulates the secondary
ID, aka Alternate ID.  (SECUSER is an old term).  Diag 88 provides the
ability to link minidisks and perform userid/password validations (if
the issuer has appropriate authority).

An interesting usage note for this is that DIAG D4 does not change the
userid associated with any existing (already active) SFS connections.
This is because it is a CP function and manipulates the VMDBK.  Once
set, future connections (via APPC/VM Connect) will utilize the Alternate
ID. ... This is why severing all connections prior to setting the AltID
will fix this type of problem, because CMS will (re) connect and use
the AltID.

If this is Nora's problem, an easy work around would be wrap the job
with an exec that uses DMSGETWU and DMSPUSWU to set a new default work
unit that contains the appropriate user, then run the job from the exec,
then finally reset with DMSPOPWU. (If I'm remembering all of this
correctly)

John

--
John Hall
Safe Software, Inc.
727-608-8799
johnh...@safesoftware.com




On Tue, Apr 19, 2011 at 11:48 AM, Schuh, Richard rsc...@visa.com
wrote:


Isn't that DIAG 88, instead of SECUSER?
 

Regards, 
Richard Schuh 

 

 




From: The IBM z/VM Operating System
[mailto:IBMVM@LISTSERV.UARK.EDU] On Behalf Of John Hall
Sent: Tuesday, April 19, 2011 6:41 AM 

To: IBMVM@LISTSERV.UARK.EDU
Subject: Re: SFS problem


Nora,
Batch jobs normally run with the privileges of the
owner of the job, using the SECUSER facility in z/VM.  With SFS, this
can lead to unexpected results when a prior batch job leaves the worker
with a connection to the filepool under a different user's id.  If the
job ordering/selection of batch workers is somewhat random, you could
see the outcome that you're experiencing (sometimes it works, sometimes
it fails).






Re: SFS problem

2011-04-20 Thread Hughes, Jim
Interesting problem Nora.

 

Nothing recorded in the SFS log either.

 

Please post your POOLDEF statements and your DMSPARMS for the filepool
having the problems.

 

Also, what is the virtual storage size defined for the sfs server?

 

Regards,


Jim Hughes
Consulting Systems Programmer
Mainframe Technical Support Group
Department of Information Technology
State of New Hampshire
27 Hazen Drive
Concord, NH 03301
603-271-5586Fax 603.271.1516

Statement of Confidentiality: The contents of this message are
confidential. Any unauthorized disclosure, reproduction, use or
dissemination (either whole or in part) is prohibited. If you are not
the intended recipient of this message, please notify the sender
immediately and delete the message from your system.



From: The IBM z/VM Operating System [mailto:IBMVM@LISTSERV.UARK.EDU] On
Behalf Of Graves Nora E
Sent: Wednesday, April 20, 2011 10:17 AM
To: IBMVM@LISTSERV.UARK.EDU
Subject: Re: SFS problem

 

I do not believe that this is the problem. I was giving you information
on the failing job that I am most familiar with.  Other jobs that fail
are submitted to VMBatch by the ID that also owns the SFS directories.

 

FYI, VMBatch is the IBM VM Batch Facility Version 2.2,  I was not
aware until recently that there are other products also known as
VMBatch.  If this has caused confusion, I apologize.

 

Nora Graves

nora.e.gra...@irs.gov

Main IRS, Room 6531

(202) 622-6735 

Fax (202) 622-3123

SE:W:CAR:MP:D:KS:BRSI

 

 



From: The IBM z/VM Operating System [mailto:IBMVM@LISTSERV.UARK.EDU] On
Behalf Of John Hall
Sent: Wednesday, April 20, 2011 10:04 AM
To: IBMVM@LISTSERV.UARK.EDU
Subject: Re: SFS problem

If I recall correctly, DIAG D4 is the one that manipulates the secondary
ID, aka Alternate ID.  (SECUSER is an old term).  Diag 88 provides the
ability to link minidisks and perform userid/password validations (if
the issuer has appropriate authority).

An interesting usage note for this is that DIAG D4 does not change the
userid associated with any existing (already active) SFS connections.
This is because it is a CP function and manipulates the VMDBK.  Once
set, future connections (via APPC/VM Connect) will utilize the Alternate
ID. ... This is why severing all connections prior to setting the AltID
will fix this type of problem, because CMS will (re) connect and use
the AltID.

If this is Nora's problem, an easy work around would be wrap the job
with an exec that uses DMSGETWU and DMSPUSWU to set a new default work
unit that contains the appropriate user, then run the job from the exec,
then finally reset with DMSPOPWU. (If I'm remembering all of this
correctly)

John

--
John Hall
Safe Software, Inc.
727-608-8799
johnh...@safesoftware.com




On Tue, Apr 19, 2011 at 11:48 AM, Schuh, Richard rsc...@visa.com
wrote:

Isn't that DIAG 88, instead of SECUSER?

 

Regards, 
Richard Schuh 

 

 

 





From: The IBM z/VM Operating System
[mailto:IBMVM@LISTSERV.UARK.EDU] On Behalf Of John Hall
Sent: Tuesday, April 19, 2011 6:41 AM 


To: IBMVM@LISTSERV.UARK.EDU
Subject: Re: SFS problem

 

Nora,
Batch jobs normally run with the privileges of the owner of
the job, using the SECUSER facility in z/VM.  With SFS, this can lead to
unexpected results when a prior batch job leaves the worker with a
connection to the filepool under a different user's id.  If the job
ordering/selection of batch workers is somewhat random, you could see
the outcome that you're experiencing (sometimes it works, sometimes it
fails).

 



Re: SFS problem

2011-04-20 Thread Mike Walter
Nora,

Is there ever a case where the command (EXEC, or whatever) fails when it 
is NOT run in a VMBATCH worker?

If they never fail outside of VMBATCH workers, John probably has a good 
handle on the diagnosis.  You may have to read a little bit more in the 
VMBATCH documentation about how VMBATCH workers get initialized, end, and 
especially: how the post-job cleanup process works.  It might be as 
simple as changing the program to release the SFS directory before ending 
(and perhaps examining the connection and looping until it has been 
cleared).

BTW, the other common VMBATCH product is from CA, called VM:Batch.  Many 
moons ago there were other publically available freeware VMBATCH 
solutions, too -- IIRC, from colleges.

Mike Walter
Aon Corporation
The opinions expressed herein are mine alone, not my employer's.




Graves Nora E nora.e.gra...@irs.gov 

Sent by: The IBM z/VM Operating System IBMVM@LISTSERV.UARK.EDU
04/20/2011 09:17 AM
Please respond to
The IBM z/VM Operating System IBMVM@LISTSERV.UARK.EDU



To
IBMVM@LISTSERV.UARK.EDU
cc

Subject
Re: SFS problem






I do not believe that this is the problem. I was giving you information on 
the failing job that I am most familiar with.  Other jobs that fail are 
submitted to VMBatch by the ID that also owns the SFS directories.
 
FYI, VMBatch is the IBM VM Batch Facility Version 2.2,  I was not aware 
until recently that there are other products also known as VMBatch.  If 
this has caused confusion, I apologize.
 
Nora Graves
nora.e.gra...@irs.gov
Main IRS, Room 6531
(202) 622-6735 
Fax (202) 622-3123
SE:W:CAR:MP:D:KS:BRSI
 

From: The IBM z/VM Operating System [mailto:IBMVM@LISTSERV.UARK.EDU] On 
Behalf Of John Hall
Sent: Wednesday, April 20, 2011 10:04 AM
To: IBMVM@LISTSERV.UARK.EDU
Subject: Re: SFS problem

If I recall correctly, DIAG D4 is the one that manipulates the secondary 
ID, aka Alternate ID.  (SECUSER is an old term).  Diag 88 provides the 
ability to link minidisks and perform userid/password validations (if the 
issuer has appropriate authority).

An interesting usage note for this is that DIAG D4 does not change the 
userid associated with any existing (already active) SFS connections. This 
is because it is a CP function and manipulates the VMDBK.  Once set, 
future connections (via APPC/VM Connect) will utilize the Alternate ID. 
... This is why severing all connections prior to setting the AltID will 
fix this type of problem, because CMS will (re) connect and use the 
AltID.

If this is Nora's problem, an easy work around would be wrap the job with 
an exec that uses DMSGETWU and DMSPUSWU to set a new default work unit 
that contains the appropriate user, then run the job from the exec, then 
finally reset with DMSPOPWU. (If I'm remembering all of this correctly)

John

--
John Hall
Safe Software, Inc.
727-608-8799
johnh...@safesoftware.com



On Tue, Apr 19, 2011 at 11:48 AM, Schuh, Richard rsc...@visa.com wrote:
Isn't that DIAG 88, instead of SECUSER?
 
Regards, 
Richard Schuh 
 
 

From: The IBM z/VM Operating System [mailto:IBMVM@LISTSERV.UARK.EDU] On 
Behalf Of John Hall
Sent: Tuesday, April 19, 2011 6:41 AM 

To: IBMVM@LISTSERV.UARK.EDU
Subject: Re: SFS problem

Nora,
Batch jobs normally run with the privileges of the owner of the job, 
using the SECUSER facility in z/VM.  With SFS, this can lead to unexpected 
results when a prior batch job leaves the worker with a connection to the 
filepool under a different user's id.  If the job ordering/selection of 
batch workers is somewhat random, you could see the outcome that you're 
experiencing (sometimes it works, sometimes it fails).






The information contained in this e-mail and any accompanying documents may 
contain information that is confidential or otherwise protected from 
disclosure. If you are not the intended recipient of this message, or if this 
message has been addressed to you in error, please immediately alert the sender 
by reply e-mail and then delete this message, including any attachments. Any 
dissemination, distribution or other use of the contents of this message by 
anyone other than the intended recipient is strictly prohibited. All messages 
sent to and from this e-mail address may be monitored as permitted by 
applicable law and regulations to ensure compliance with our internal policies 
and to protect our business. E-mails are not secure and cannot be guaranteed to 
be error free as they can be intercepted, amended, lost or destroyed, or 
contain viruses. You are deemed to have accepted these risks if you communicate 
with us by e-mail. 


Re: SFS problem

2011-04-20 Thread John Hall
Thank you, yes, that is correct... Apologies for mixing my terms.  I should
have been saying Alternate ID.

John

On Wed, Apr 20, 2011 at 12:33 PM, Schuh, Richard rsc...@visa.com wrote:

  Actually, SECUSER still exists and is something quite
 different. SECUSER defines a secondary user for receiving console messages
 and entering commands and responses. The entries are done as though they
 came from, not on behalf of, the user.

 Regards,
 Richard Schuh



  --
 *From:* The IBM z/VM Operating System [mailto:IBMVM@LISTSERV.UARK.EDU] *On
 Behalf Of *John Hall
 *Sent:* Wednesday, April 20, 2011 7:04 AM

 *To:* IBMVM@LISTSERV.UARK.EDU
 *Subject:* Re: SFS problem

 If I recall correctly, DIAG D4 is the one that manipulates the secondary
 ID, aka Alternate ID.  (SECUSER is an old term).  Diag 88 provides the
 ability to link minidisks and perform userid/password validations (if the
 issuer has appropriate authority).

 An interesting usage note for this is that DIAG D4 does not change the
 userid associated with any existing (already active) SFS connections.  This
 is because it is a CP function and manipulates the VMDBK.  Once set, future
 connections (via APPC/VM Connect) will utilize the Alternate ID. ... This is
 why severing all connections prior to setting the AltID will fix this type
 of problem, because CMS will (re) connect and use the AltID.

 If this is Nora's problem, an easy work around would be wrap the job with
 an exec that uses DMSGETWU and DMSPUSWU to set a new default work unit that
 contains the appropriate user, then run the job from the exec, then finally
 reset with DMSPOPWU. (If I'm remembering all of this correctly)

 John

 --
 John Hall
 Safe Software, Inc.
 727-608-8799
 johnh...@safesoftware.com



 On Tue, Apr 19, 2011 at 11:48 AM, Schuh, Richard rsc...@visa.com wrote:

  Isn't that DIAG 88, instead of SECUSER?


 Regards,
 Richard Schuh




  --
 *From:* The IBM z/VM Operating System [mailto:IBMVM@LISTSERV.UARK.EDU] *On
 Behalf Of *John Hall
 *Sent:* Tuesday, April 19, 2011 6:41 AM

 *To:* IBMVM@LISTSERV.UARK.EDU
 *Subject:* Re: SFS problem

  Nora,
 Batch jobs normally run with the privileges of the owner of the job,
 using the SECUSER facility in z/VM.  With SFS, this can lead to
 unexpected results when a prior batch job leaves the worker with a
 connection to the filepool under a different user's id.  If the job
 ordering/selection of batch workers is somewhat random, you could see the
 outcome that you're experiencing (sometimes it works, sometimes it fails).





Re: SFS problem

2011-04-19 Thread Graves Nora E
Mike,

Yes, I checked out the message.  The thing is, the job runs nightly, and
only fails intermittently.  There is no reason to believe that the
authority to access the directory is changing during that time period;
there are only 2 people who regularly grant permissions to that user's
directory.   Neither of us has made a change.


Nora Graves
nora.e.gra...@irs.gov
Main IRS, Room 6531
(202) 622-6735 
Fax (202) 622-3123
SE:W:CAR:MP:D:KS:BRSI

-Original Message-
From: The IBM z/VM Operating System [mailto:IBMVM@LISTSERV.UARK.EDU] On
Behalf Of Mike Walter
Sent: Thursday, April 14, 2011 4:13 PM
To: IBMVM@LISTSERV.UARK.EDU
Subject: Re: SFS problem

Nora,

Did you issue: HELP DMSOPN1258E

The response from the msg indicates an SFS authorization problem.:
---snip---
(c) Copyright IBM Corporation 1990, 2008  
  
 DMS1258E  You are not authorized|Not authorized|Not permitted to
write 
(to) 
   file fn ft fm|pathname  
  
 Explanation: You attempted to write to a file for which you do not have

write 
 authority, or you attempted to create a new file in a directory for
which 
you 
 do not have write authority.  
  
 System Action: RC=12 or 28. Execution of the command is terminated.  
  
 User Response: Ensure that you specified the correct file. If so,
contact 
the 
 owner to gain proper authorization to the file or directory. If the
file  
 
 specified is an alias, you may enter the QUERY ALIAS command to
determine 
the 
 owner of the base file.  
  
 For a BFS file, enter the OPENVM LISTFILE command with the OWNER option

for 
 the file. This will tell you the owning user ID and group name for the 
file 
 you wish to use. Refer to the z/VM: OpenExtensions Command Reference or

enter 
 HELP OPENVM for more information on the OPENVM commands. 
---snip---

z/VM has a terrific HELP facility (including most messages), it's worth 
investing a bit of time to familiarize yourself with it.

Were you using the VMLINK command in every case with the (FORCERW 
option?  Regardless, check the syntax for your SFS authorizations.

And if all those bazillion intricate SFS authorizations become 
overwhelming, consider checking out SafeSFS from safesfs.com.  We're
vary 
satisfied customers, it makes authorizations much MUCH simpler, saves
huge 
amounts of effort and back time.

Mike Walter
Aon Corporation
The opinions expressed herein are mine alone, not my employer's.
 



Graves Nora E nora.e.gra...@irs.gov 

Sent by: The IBM z/VM Operating System IBMVM@LISTSERV.UARK.EDU
04/14/2011 02:45 PM
Please respond to
The IBM z/VM Operating System IBMVM@LISTSERV.UARK.EDU



To
IBMVM@LISTSERV.UARK.EDU
cc

Subject
SFS problem






We are having an intermittent problem with SFS and I'm hoping someone
may 
have some ideas of what to pursue next.
 
We have several batch jobs that run under VMBATCH overnight.  Sometimes 
they are not able to create a file in a directory, even though most
times 
it is successful.  The only differences in the executions are the file 
names; for many of these the File Type is the date.
 
In the job I am most familiar with, these are the specifics.
 
The job runs Monday-Saturday.  This year, it has failed on January 4, 
January 12, February 9, March 18, March 25, and April 13.  It has run 
successfully the other days.  Other than the QUERY statements below, it 
has not changed.
The job runs in a work machine, WORK7.
The job is submitted by the User ID of the database owner.
The SFS directories are owned by a 3rd user.  Failures occur in many of 
the subdirectories, not just one subdirectory owned by this user.  This 
user is the owner of most of the directories containing the data files
we 
create in batch, so I don't think it's significant that it's the ID that

has the problem.  However, as far as I know, it is the only ID that does

have the problem.
This job uses VMLINK to acquire a write link to SFS directory.  This 
always looks to be successful--no error is given.  (Other jobs use 
GETFMADDR and ACCESS to acquire the write link to the directory.  This 
always appears successful as well).
Once the file is ready to be copied from the Work Machine's 191 disk to 
the SFS directory, the intermittent error appears.  The vast majority of

the time, the write is successful.  However, sometimes, the job gets
this 
error message: 
DMSOPN1258E You are not authorized to write to file XX 20110413 Z1
 
The file is not large--last night's file was only 12 blocks. 
 
At the suggestion of our systems programmer, I've put in a lot of query 
statements.  I've issued QUERY LIMITS for the job submitter; it's only 
used 84% of the allocation, with several thousand blocks available. The 
SFS directory owner has only used 76% of its allocation, with several 
thousand more blocks still available.  The filepool is not full.
 
I've issued QUERY FILEPOOL CONFLICT.  There is no conflict.
 
I've issued QUERY ACCESSED.  The directory shows that is accessed R/W.
 
When the write is unsuccessful, the program

Re: SFS problem

2011-04-19 Thread Graves Nora E
Alan,

The submitter does have NEWRITE permissions.  And nothing in that job
erases the file.  The job runs only once per day, and the File Type is
the date, such as 20110419. 

However, I didn't know that authority to write the file disappeared with
the file if NEWRITE wasn't granted, so I learned something!


Nora Graves
nora.e.gra...@irs.gov
Main IRS, Room 6531
(202) 622-6735 
Fax (202) 622-3123
SE:W:CAR:MP:D:KS:BRSI

-Original Message-
From: The IBM z/VM Operating System [mailto:IBMVM@LISTSERV.UARK.EDU] On
Behalf Of Alan Altmark
Sent: Thursday, April 14, 2011 4:52 PM
To: IBMVM@LISTSERV.UARK.EDU
Subject: Re: SFS problem

On Thursday, 04/14/2011 at 03:47 EDT, Graves Nora E 
nora.e.gra...@irs.gov wrote:

 DMSOPN1258E You are  not authorized to write to file XX 20110413
Z1
:  
 I've issued QUERY  ACCESSED.  The directory shows that is accessed
R/W.
  
 When the write is  unsuccessful, the program then loops through 5
tries 
of 
 releasing the access,  reacquiring the access, and attempting to write

the file 
 again.   This has never been successful.  I've issued both a COPYFILE 
and a 
 PIPE to try to write the file; these do not work once there has been
a 
failure.
  
 We've looked at the  operator consoles to see if we can find any jobs 
running 
 at the same time.   We haven't found any that are accessing that 
directory 
 structure.
  
 There aren't any  dumps to look at--it looks perfectly successful
other 
than 
 the fact that it  won't write the file.
  
 Does anyone have any  suggestions of something to try next?

As you can see, it's an authorization error.  Does the worker/submittor 
have NEWRITE authority to the directory?  If not, and some part of the
job 
ERASEs the file from the directory and then tries to recreate it, it
will 
fail, as the authority to the file is deleted if the file itself is 
deleted.

In a FILECONTROL directory, R/W access to the directory does not imply
R/W 
access to all the files in it!

Alan Altmark

z/VM and Linux on System z Consultant
IBM System Lab Services and Training 
ibm.com/systems/services/labservices 
office: 607.429.3323
mobile; 607.321.7556
alan_altm...@us.ibm.com
IBM Endicott


Re: SFS problem

2011-04-19 Thread Graves Nora E
Kris,
 
The submitter is ABC, the directory owner is XYZ.  The submitter has
WRITE/NEWRITE authority to the directory and files.
 
Nora Graves
nora.e.gra...@irs.gov
Main IRS, Room 6531
(202) 622-6735 
Fax (202) 622-3123
SE:W:CAR:MP:D:KS:BRSI
 



From: The IBM z/VM Operating System [mailto:IBMVM@LISTSERV.UARK.EDU] On
Behalf Of Kris Buelens
Sent: Thursday, April 14, 2011 4:58 PM
To: IBMVM@LISTSERV.UARK.EDU
Subject: Re: SFS problem


Note: if you use VMBATCH, the worker machine connects to SFS with the
authority of the job submitter.
You say This user is the owner of most of the directories  You mean:
the submitter is userid ABC, the dirids are all named
fpoolid:ABC.something?


2011/4/14 Graves Nora E nora.e.gra...@irs.gov


We are having an intermittent problem with SFS and I'm hoping
someone may have some ideas of what to pursue next.
 
We have several batch jobs that run under VMBATCH overnight.
Sometimes they are not able to create a file in a directory, even though
most times it is successful.  The only differences in the executions are
the file names; for many of these the File Type is the date.
 
In the job I am most familiar with, these are the specifics.
 
The job runs Monday-Saturday.  This year, it has failed on
January 4, January 12, February 9, March 18, March 25, and April 13.  It
has run successfully the other days.  Other than the QUERY statements
below, it has not changed.
The job runs in a work machine, WORK7.
The job is submitted by the User ID of the database owner.
The SFS directories are owned by a 3rd user.  Failures occur in
many of the subdirectories, not just one subdirectory owned by this
user.  This user is the owner of most of the directories containing the
data files we create in batch, so I don't think it's significant that
it's the ID that has the problem.  However, as far as I know, it is the
only ID that does have the problem.
This job uses VMLINK to acquire a write link to SFS directory.
This always looks to be successful--no error is given.  (Other jobs use
GETFMADDR and ACCESS to acquire the write link to the directory.  This
always appears successful as well).
Once the file is ready to be copied from the Work Machine's 191
disk to the SFS directory, the intermittent error appears.  The vast
majority of the time, the write is successful.  However, sometimes, the
job gets this error message: 
DMSOPN1258E You are not authorized to write to file XX
20110413 Z1
 
The file is not large--last night's file was only 12 blocks.  
 
At the suggestion of our systems programmer, I've put in a lot
of query statements.  I've issued QUERY LIMITS for the job submitter;
it's only used 84% of the allocation, with several thousand blocks
available. The SFS directory owner has only used 76% of its allocation,
with several thousand more blocks still available.  The filepool is not
full.
 
I've issued QUERY FILEPOOL CONFLICT.  There is no conflict.
 
I've issued QUERY ACCESSED.  The directory shows that is
accessed R/W.
 
When the write is unsuccessful, the program then loops through 5
tries of releasing the access, reacquiring the access, and attempting to
write the file again.  This has never been successful.  I've issued both
a COPYFILE and a PIPE to try to write the file; these do not work once
there has been a failure.
 
We've looked at the operator consoles to see if we can find any
jobs running at the same time.  We haven't found any that are accessing
that directory structure.
 
There aren't any dumps to look at--it looks perfectly successful
other than the fact that it won't write the file.
 
Does anyone have any suggestions of something to try next?

 
Nora Graves
nora.e.gra...@irs.gov
 




-- 
Kris Buelens,
IBM Belgium, VM customer support



Re: SFS problem

2011-04-19 Thread Graves Nora E
Jim,

I didn't think to look there, I will.

Our regular backups are VMBACKUP, which is scheduled to run 90 minutes
after this job runs.  This job runs only 2 minutes maximum, so I know
that isn't the problem. But the SFS console may have some good
indications. I'll see what I can find there.  

Thanks,


Nora Graves
nora.e.gra...@irs.gov
Main IRS, Room 6531
(202) 622-6735 
Fax (202) 622-3123
SE:W:CAR:MP:D:KS:BRSI

-Original Message-
From: The IBM z/VM Operating System [mailto:IBMVM@LISTSERV.UARK.EDU] On
Behalf Of Hughes, Jim
Sent: Thursday, April 14, 2011 5:23 PM
To: IBMVM@LISTSERV.UARK.EDU
Subject: Re: SFS problem

Nora,

Have you looked at the SFS service machine's console log for any odd
messages?

Could the SFS server being doing some sort of control backup?

Jim Hughes
Consulting Systems Programmer 
Mainframe Technical Support Group
Department of Information Technology
State of New Hampshire
27 Hazen Drive
Concord, NH 03301
603-271-5586Fax 603.271.1516

Statement of Confidentiality: The contents of this message are
confidential. Any unauthorized disclosure, reproduction, use or
dissemination (either whole or in part) is prohibited. If you are not
the intended recipient of this message, please notify the sender
immediately and delete the message from your system.


Re: SFS problem

2011-04-19 Thread Graves Nora E
Jim,

There are basically no messages at all in the console for the SFS
service machine.  One of the failures was on April 13.  This is the
console for that day--no editing at all of the contents.

HCPMID6001I  TIME IS 00:00:00 EDT WEDNESDAY 04/13/11  
  
00:00:00  
  
  
HCPMID6001I  TIME IS 00:00:00 EDT THURSDAY 04/14/11   
  
00:00:00  
  
   


Nora Graves
nora.e.gra...@irs.gov
Main IRS, Room 6531
(202) 622-6735 
Fax (202) 622-3123
SE:W:CAR:MP:D:KS:BRSI

-Original Message-
From: The IBM z/VM Operating System [mailto:IBMVM@LISTSERV.UARK.EDU] On
Behalf Of Hughes, Jim
Sent: Thursday, April 14, 2011 5:23 PM
To: IBMVM@LISTSERV.UARK.EDU
Subject: Re: SFS problem

Nora,

Have you looked at the SFS service machine's console log for any odd
messages?

Could the SFS server being doing some sort of control backup?

Jim Hughes
Consulting Systems Programmer 
Mainframe Technical Support Group
Department of Information Technology
State of New Hampshire
27 Hazen Drive
Concord, NH 03301
603-271-5586Fax 603.271.1516

Statement of Confidentiality: The contents of this message are
confidential. Any unauthorized disclosure, reproduction, use or
dissemination (either whole or in part) is prohibited. If you are not
the intended recipient of this message, please notify the sender
immediately and delete the message from your system.


Re: SFS problem

2011-04-19 Thread John Hall
Nora,
Batch jobs normally run with the privileges of the owner of the job, using
the SECUSER facility in z/VM.  With SFS, this can lead to unexpected results
when a prior batch job leaves the worker with a connection to the filepool
under a different user's id.  If the job ordering/selection of batch workers
is somewhat random, you could see the outcome that you're experiencing
(sometimes it works, sometimes it fails).

For example:
If WORK7 first runs a job for JOHN and JOHN utilizes the same filepool,
WORK7 could be left with a connection to the filepool that appears to be
from the user JOHN.
If your failing job runs next, it would likely end up accessing the filepool
as JOHN, not as the actual owner, which could cause the failure.

If on another night, WORK7 is recycled before your failing job runs, then it
would establish a connection to the file pool with the owner's ID (say
OWNER) and then would run with the access rights of OWNER, and you'd see
success.

If this is your problem, there are many ways to solve it.  The easiest would
be to ensure that worker machines are recycled (logoff/xautolog or IPL CMS)
prior to the job running.  I wonder if there are VMBATCH options (system or
job) that provide the ability to do a more advanced clean up of a worker
prior to a job run.

John

--
John Hall
Safe Software, Inc.
727-608-8799
johnh...@safesoftware.com


On Thu, Apr 14, 2011 at 3:45 PM, Graves Nora E nora.e.gra...@irs.govwrote:

  We are having an intermittent problem with SFS and I'm hoping someone may
 have some ideas of what to pursue next.

 We have several batch jobs that run under VMBATCH overnight.  Sometimes
 they are not able to create a file in a directory, even though most times it
 is successful.  The only differences in the executions are the file names;
 for many of these the File Type is the date.

 In the job I am most familiar with, these are the specifics.

 The job runs Monday-Saturday.  This year, it has failed on January 4,
 January 12, February 9, March 18, March 25, and April 13.  It has run
 successfully the other days.  Other than the QUERY statements below, it has
 not changed.
 The job runs in a work machine, WORK7.
 The job is submitted by the User ID of the database owner.
 The SFS directories are owned by a 3rd user.  Failures occur in many of the
 subdirectories, not just one subdirectory owned by this user.  This user is
 the owner of most of the directories containing the data files we create in
 batch, so I don't think it's significant that it's the ID that has the
 problem.  However, as far as I know, it is the only ID that does have the
 problem.
 This job uses VMLINK to acquire a write link to SFS directory.  This always
 looks to be successful--no error is given.  (Other jobs use GETFMADDR and
 ACCESS to acquire the write link to the directory.  This always
 appears successful as well).
 Once the file is ready to be copied from the Work Machine's 191 disk to the
 SFS directory, the intermittent error appears.  The vast majority of the
 time, the write is successful.  However, sometimes, the job gets this error
 message:
 DMSOPN1258E You are not authorized to write to file XX 20110413 Z1

 The file is not large--last night's file was only 12 blocks.

 At the suggestion of our systems programmer, I've put in a lot of query
 statements.  I've issued QUERY LIMITS for the job submitter; it's only used
 84% of the allocation, with several thousand blocks available. The SFS
 directory owner has only used 76% of its allocation, with several thousand
 more blocks still available.  The filepool is not full.

 I've issued QUERY FILEPOOL CONFLICT.  There is no conflict.

 I've issued QUERY ACCESSED.  The directory shows that is accessed R/W.

 When the write is unsuccessful, the program then loops through 5 tries of
 releasing the access, reacquiring the access, and attempting to write the
 file again.  This has never been successful.  I've issued both a COPYFILE
 and a PIPE to try to write the file; these do not work once there has been a
 failure.

 We've looked at the operator consoles to see if we can find any jobs
 running at the same time.  We haven't found any that are accessing that
 directory structure.

 There aren't any dumps to look at--it looks perfectly successful other than
 the fact that it won't write the file.

 Does anyone have any suggestions of something to try next?


  Nora Graves
 nora.e.gra...@irs.gov




Re: SFS problem

2011-04-19 Thread Schuh, Richard
Isn't that DIAG 88, instead of SECUSER?


Regards,
Richard Schuh






From: The IBM z/VM Operating System [mailto:IBMVM@LISTSERV.UARK.EDU] On Behalf 
Of John Hall
Sent: Tuesday, April 19, 2011 6:41 AM
To: IBMVM@LISTSERV.UARK.EDU
Subject: Re: SFS problem

Nora,
Batch jobs normally run with the privileges of the owner of the job, using 
the SECUSER facility in z/VM.  With SFS, this can lead to unexpected results 
when a prior batch job leaves the worker with a connection to the filepool 
under a different user's id.  If the job ordering/selection of batch workers is 
somewhat random, you could see the outcome that you're experiencing (sometimes 
it works, sometimes it fails).



Re: SFS problem

2011-04-19 Thread Ethan Lanz
Just to exhaust the possibilities in this line, we have had occasions when a
job would fail for another reason, then be re-submitted by another userid
without the required authority.

On Tue, Apr 19, 2011 at 7:47 AM, Graves Nora E nora.e.gra...@irs.govwrote:

  Kris,

 The submitter is ABC, the directory owner is XYZ.  The submitter has
 WRITE/NEWRITE authority to the directory and files.

  Nora Graves
 nora.e.gra...@irs.gov
 Main IRS, Room 6531
 (202) 622-6735
 Fax (202) 622-3123
 SE:W:CAR:MP:D:KS:BRSI


  --
 *From:* The IBM z/VM Operating System [mailto:IBMVM@LISTSERV.UARK.EDU] *On
 Behalf Of *Kris Buelens
 *Sent:* Thursday, April 14, 2011 4:58 PM

 *To:* IBMVM@LISTSERV.UARK.EDU
 *Subject:* Re: SFS problem

 Note: if you use VMBATCH, the worker machine connects to SFS with the
 authority of the job submitter.
 You say This user is the owner of most of the directories  You mean: the
 submitter is userid ABC, the dirids are all named fpoolid:ABC.something?


 2011/4/14 Graves Nora E nora.e.gra...@irs.gov

  We are having an intermittent problem with SFS and I'm hoping someone
 may have some ideas of what to pursue next.

 We have several batch jobs that run under VMBATCH overnight.  Sometimes
 they are not able to create a file in a directory, even though most times it
 is successful.  The only differences in the executions are the file names;
 for many of these the File Type is the date.

 In the job I am most familiar with, these are the specifics.

 The job runs Monday-Saturday.  This year, it has failed on January 4,
 January 12, February 9, March 18, March 25, and April 13.  It has run
 successfully the other days.  Other than the QUERY statements below, it has
 not changed.
 The job runs in a work machine, WORK7.
 The job is submitted by the User ID of the database owner.
 The SFS directories are owned by a 3rd user.  Failures occur in many of
 the subdirectories, not just one subdirectory owned by this user.  This user
 is the owner of most of the directories containing the data files we create
 in batch, so I don't think it's significant that it's the ID that has the
 problem.  However, as far as I know, it is the only ID that does have the
 problem.
 This job uses VMLINK to acquire a write link to SFS directory.  This
 always looks to be successful--no error is given.  (Other jobs use GETFMADDR
 and ACCESS to acquire the write link to the directory.  This always
 appears successful as well).
 Once the file is ready to be copied from the Work Machine's 191 disk to
 the SFS directory, the intermittent error appears.  The vast majority of the
 time, the write is successful.  However, sometimes, the job gets this error
 message:
 DMSOPN1258E You are not authorized to write to file XX 20110413 Z1

 The file is not large--last night's file was only 12 blocks.

 At the suggestion of our systems programmer, I've put in a lot of query
 statements.  I've issued QUERY LIMITS for the job submitter; it's only used
 84% of the allocation, with several thousand blocks available. The SFS
 directory owner has only used 76% of its allocation, with several thousand
 more blocks still available.  The filepool is not full.

 I've issued QUERY FILEPOOL CONFLICT.  There is no conflict.

 I've issued QUERY ACCESSED.  The directory shows that is accessed R/W.

 When the write is unsuccessful, the program then loops through 5 tries of
 releasing the access, reacquiring the access, and attempting to write the
 file again.  This has never been successful.  I've issued both a COPYFILE
 and a PIPE to try to write the file; these do not work once there has been a
 failure.

 We've looked at the operator consoles to see if we can find any jobs
 running at the same time.  We haven't found any that are accessing that
 directory structure.

 There aren't any dumps to look at--it looks perfectly successful other
 than the fact that it won't write the file.

 Does anyone have any suggestions of something to try next?


  Nora Graves
 nora.e.gra...@irs.gov





 --
 Kris Buelens,
 IBM Belgium, VM customer support



SFS problem

2011-04-14 Thread Graves Nora E
We are having an intermittent problem with SFS and I'm hoping someone
may have some ideas of what to pursue next.
 
We have several batch jobs that run under VMBATCH overnight.  Sometimes
they are not able to create a file in a directory, even though most
times it is successful.  The only differences in the executions are the
file names; for many of these the File Type is the date.
 
In the job I am most familiar with, these are the specifics.
 
The job runs Monday-Saturday.  This year, it has failed on January 4,
January 12, February 9, March 18, March 25, and April 13.  It has run
successfully the other days.  Other than the QUERY statements below, it
has not changed.
The job runs in a work machine, WORK7.
The job is submitted by the User ID of the database owner.
The SFS directories are owned by a 3rd user.  Failures occur in many of
the subdirectories, not just one subdirectory owned by this user.  This
user is the owner of most of the directories containing the data files
we create in batch, so I don't think it's significant that it's the ID
that has the problem.  However, as far as I know, it is the only ID that
does have the problem.
This job uses VMLINK to acquire a write link to SFS directory.  This
always looks to be successful--no error is given.  (Other jobs use
GETFMADDR and ACCESS to acquire the write link to the directory.  This
always appears successful as well).
Once the file is ready to be copied from the Work Machine's 191 disk to
the SFS directory, the intermittent error appears.  The vast majority of
the time, the write is successful.  However, sometimes, the job gets
this error message: 
DMSOPN1258E You are not authorized to write to file XX 20110413 Z1
 
The file is not large--last night's file was only 12 blocks.  
 
At the suggestion of our systems programmer, I've put in a lot of query
statements.  I've issued QUERY LIMITS for the job submitter; it's only
used 84% of the allocation, with several thousand blocks available. The
SFS directory owner has only used 76% of its allocation, with several
thousand more blocks still available.  The filepool is not full.
 
I've issued QUERY FILEPOOL CONFLICT.  There is no conflict.
 
I've issued QUERY ACCESSED.  The directory shows that is accessed R/W.
 
When the write is unsuccessful, the program then loops through 5 tries
of releasing the access, reacquiring the access, and attempting to write
the file again.  This has never been successful.  I've issued both a
COPYFILE and a PIPE to try to write the file; these do not work once
there has been a failure.
 
We've looked at the operator consoles to see if we can find any jobs
running at the same time.  We haven't found any that are accessing that
directory structure.
 
There aren't any dumps to look at--it looks perfectly successful other
than the fact that it won't write the file.
 
Does anyone have any suggestions of something to try next?

 
Nora Graves
nora.e.gra...@irs.gov
 


Re: SFS problem

2011-04-14 Thread Peter . Webb
Would issuing a QUERY AUTHORITY at the time of the error give anything
useful, I wonder. Maybe try for both the directory and the file?

 

Peter

 

-Original Message-
From: The IBM z/VM Operating System [mailto:IBMVM@LISTSERV.UARK.EDU] On
Behalf Of Graves Nora E
Sent: April 14, 2011 15:46
To: IBMVM@LISTSERV.UARK.EDU
Subject: SFS problem

 

We are having an intermittent problem with SFS and I'm hoping someone
may have some ideas of what to pursue next.

 

We have several batch jobs that run under VMBATCH overnight.  Sometimes
they are not able to create a file in a directory, even though most
times it is successful.  The only differences in the executions are the
file names; for many of these the File Type is the date.

 

In the job I am most familiar with, these are the specifics.

 

The job runs Monday-Saturday.  This year, it has failed on January 4,
January 12, February 9, March 18, March 25, and April 13.  It has run
successfully the other days.  Other than the QUERY statements below, it
has not changed.

The job runs in a work machine, WORK7.

The job is submitted by the User ID of the database owner.

The SFS directories are owned by a 3rd user.  Failures occur in many of
the subdirectories, not just one subdirectory owned by this user.  This
user is the owner of most of the directories containing the data files
we create in batch, so I don't think it's significant that it's the ID
that has the problem.  However, as far as I know, it is the only ID that
does have the problem.

This job uses VMLINK to acquire a write link to SFS directory.  This
always looks to be successful--no error is given.  (Other jobs use
GETFMADDR and ACCESS to acquire the write link to the directory.  This
always appears successful as well).

Once the file is ready to be copied from the Work Machine's 191 disk to
the SFS directory, the intermittent error appears.  The vast majority of
the time, the write is successful.  However, sometimes, the job gets
this error message: 

DMSOPN1258E You are not authorized to write to file XX 20110413 Z1

 

The file is not large--last night's file was only 12 blocks.  

 

At the suggestion of our systems programmer, I've put in a lot of query
statements.  I've issued QUERY LIMITS for the job submitter; it's only
used 84% of the allocation, with several thousand blocks available. The
SFS directory owner has only used 76% of its allocation, with several
thousand more blocks still available.  The filepool is not full.

 

I've issued QUERY FILEPOOL CONFLICT.  There is no conflict.

 

I've issued QUERY ACCESSED.  The directory shows that is accessed R/W.

 

When the write is unsuccessful, the program then loops through 5 tries
of releasing the access, reacquiring the access, and attempting to write
the file again.  This has never been successful.  I've issued both a
COPYFILE and a PIPE to try to write the file; these do not work once
there has been a failure.

 

We've looked at the operator consoles to see if we can find any jobs
running at the same time.  We haven't found any that are accessing that
directory structure.

 

There aren't any dumps to look at--it looks perfectly successful other
than the fact that it won't write the file.

 

Does anyone have any suggestions of something to try next?

 

 

Nora Graves

nora.e.gra...@irs.gov

 



The information transmitted is intended only for the person or entity to which 
it is addressed and may contain confidential and/or privileged material.  Any 
review retransmission dissemination or other use of or taking any action in 
reliance upon this information by persons or entities other than the intended 
recipient or delegate is strictly prohibited.  If you received this in error 
please contact the sender and delete the material from any computer.  The 
integrity and security of this message cannot be guaranteed on the Internet.  
The sender accepts no liability for the content of this e-mail or for the 
consequences of any actions taken on the basis of information provided.  The 
recipient should check this e-mail and any attachments for the presence of 
viruses.  The sender accepts no liability for any damage caused by any virus 
transmitted by this e-mail.  This disclaimer is property of the TTC and must 
not be altered or circumvented in any manner.


Re: SFS problem

2011-04-14 Thread Mike Walter
Nora,

Did you issue: HELP DMSOPN1258E

The response from the msg indicates an SFS authorization problem.:
---snip---
(c) Copyright IBM Corporation 1990, 2008  
  
 DMS1258E  You are not authorized|Not authorized|Not permitted to write 
(to) 
   file fn ft fm|pathname  
  
 Explanation: You attempted to write to a file for which you do not have 
write 
 authority, or you attempted to create a new file in a directory for which 
you 
 do not have write authority.  
  
 System Action: RC=12 or 28. Execution of the command is terminated.  
  
 User Response: Ensure that you specified the correct file. If so, contact 
the 
 owner to gain proper authorization to the file or directory. If the file  
 
 specified is an alias, you may enter the QUERY ALIAS command to determine 
the 
 owner of the base file.  
  
 For a BFS file, enter the OPENVM LISTFILE command with the OWNER option 
for 
 the file. This will tell you the owning user ID and group name for the 
file 
 you wish to use. Refer to the z/VM: OpenExtensions Command Reference or 
enter 
 HELP OPENVM for more information on the OPENVM commands. 
---snip---

z/VM has a terrific HELP facility (including most messages), it's worth 
investing a bit of time to familiarize yourself with it.

Were you using the VMLINK command in every case with the (FORCERW 
option?  Regardless, check the syntax for your SFS authorizations.

And if all those bazillion intricate SFS authorizations become 
overwhelming, consider checking out SafeSFS from safesfs.com.  We're vary 
satisfied customers, it makes authorizations much MUCH simpler, saves huge 
amounts of effort and back time.

Mike Walter
Aon Corporation
The opinions expressed herein are mine alone, not my employer's.
 



Graves Nora E nora.e.gra...@irs.gov 

Sent by: The IBM z/VM Operating System IBMVM@LISTSERV.UARK.EDU
04/14/2011 02:45 PM
Please respond to
The IBM z/VM Operating System IBMVM@LISTSERV.UARK.EDU



To
IBMVM@LISTSERV.UARK.EDU
cc

Subject
SFS problem






We are having an intermittent problem with SFS and I'm hoping someone may 
have some ideas of what to pursue next.
 
We have several batch jobs that run under VMBATCH overnight.  Sometimes 
they are not able to create a file in a directory, even though most times 
it is successful.  The only differences in the executions are the file 
names; for many of these the File Type is the date.
 
In the job I am most familiar with, these are the specifics.
 
The job runs Monday-Saturday.  This year, it has failed on January 4, 
January 12, February 9, March 18, March 25, and April 13.  It has run 
successfully the other days.  Other than the QUERY statements below, it 
has not changed.
The job runs in a work machine, WORK7.
The job is submitted by the User ID of the database owner.
The SFS directories are owned by a 3rd user.  Failures occur in many of 
the subdirectories, not just one subdirectory owned by this user.  This 
user is the owner of most of the directories containing the data files we 
create in batch, so I don't think it's significant that it's the ID that 
has the problem.  However, as far as I know, it is the only ID that does 
have the problem.
This job uses VMLINK to acquire a write link to SFS directory.  This 
always looks to be successful--no error is given.  (Other jobs use 
GETFMADDR and ACCESS to acquire the write link to the directory.  This 
always appears successful as well).
Once the file is ready to be copied from the Work Machine's 191 disk to 
the SFS directory, the intermittent error appears.  The vast majority of 
the time, the write is successful.  However, sometimes, the job gets this 
error message: 
DMSOPN1258E You are not authorized to write to file XX 20110413 Z1
 
The file is not large--last night's file was only 12 blocks. 
 
At the suggestion of our systems programmer, I've put in a lot of query 
statements.  I've issued QUERY LIMITS for the job submitter; it's only 
used 84% of the allocation, with several thousand blocks available. The 
SFS directory owner has only used 76% of its allocation, with several 
thousand more blocks still available.  The filepool is not full.
 
I've issued QUERY FILEPOOL CONFLICT.  There is no conflict.
 
I've issued QUERY ACCESSED.  The directory shows that is accessed R/W.
 
When the write is unsuccessful, the program then loops through 5 tries of 
releasing the access, reacquiring the access, and attempting to write the 
file again.  This has never been successful.  I've issued both a COPYFILE 
and a PIPE to try to write the file; these do not work once there has been 
a failure.
 
We've looked at the operator consoles to see if we can find any jobs 
running at the same time.  We haven't found any that are accessing that 
directory structure.
 
There aren't any dumps to look at--it looks perfectly successful other 
than the fact that it won't write the file.
 
Does anyone have any suggestions of something to try next?

 
Nora Graves
nora.e.gra...@irs.gov
 




The information

Re: SFS problem

2011-04-14 Thread Alan Altmark
On Thursday, 04/14/2011 at 03:47 EDT, Graves Nora E 
nora.e.gra...@irs.gov wrote:

 DMSOPN1258E You are  not authorized to write to file XX 20110413 Z1
:  
 I've issued QUERY  ACCESSED.  The directory shows that is accessed R/W.
  
 When the write is  unsuccessful, the program then loops through 5 tries 
of 
 releasing the access,  reacquiring the access, and attempting to write 
the file 
 again.   This has never been successful.  I've issued both a COPYFILE 
and a 
 PIPE to try to write the file; these do not work once there has been  a 
failure.
  
 We've looked at the  operator consoles to see if we can find any jobs 
running 
 at the same time.   We haven't found any that are accessing that 
directory 
 structure.
  
 There aren't any  dumps to look at--it looks perfectly successful other 
than 
 the fact that it  won't write the file.
  
 Does anyone have any  suggestions of something to try next?

As you can see, it's an authorization error.  Does the worker/submittor 
have NEWRITE authority to the directory?  If not, and some part of the job 
ERASEs the file from the directory and then tries to recreate it, it will 
fail, as the authority to the file is deleted if the file itself is 
deleted.

In a FILECONTROL directory, R/W access to the directory does not imply R/W 
access to all the files in it!

Alan Altmark

z/VM and Linux on System z Consultant
IBM System Lab Services and Training 
ibm.com/systems/services/labservices 
office: 607.429.3323
mobile; 607.321.7556
alan_altm...@us.ibm.com
IBM Endicott


Re: SFS problem

2011-04-14 Thread Kris Buelens
Note: if you use VMBATCH, the worker machine connects to SFS with the
authority of the job submitter.
You say This user is the owner of most of the directories  You mean: the
submitter is userid ABC, the dirids are all named fpoolid:ABC.something?

2011/4/14 Graves Nora E nora.e.gra...@irs.gov

  We are having an intermittent problem with SFS and I'm hoping someone may
 have some ideas of what to pursue next.

 We have several batch jobs that run under VMBATCH overnight.  Sometimes
 they are not able to create a file in a directory, even though most times it
 is successful.  The only differences in the executions are the file names;
 for many of these the File Type is the date.

 In the job I am most familiar with, these are the specifics.

 The job runs Monday-Saturday.  This year, it has failed on January 4,
 January 12, February 9, March 18, March 25, and April 13.  It has run
 successfully the other days.  Other than the QUERY statements below, it has
 not changed.
 The job runs in a work machine, WORK7.
 The job is submitted by the User ID of the database owner.
 The SFS directories are owned by a 3rd user.  Failures occur in many of the
 subdirectories, not just one subdirectory owned by this user.  This user is
 the owner of most of the directories containing the data files we create in
 batch, so I don't think it's significant that it's the ID that has the
 problem.  However, as far as I know, it is the only ID that does have the
 problem.
 This job uses VMLINK to acquire a write link to SFS directory.  This always
 looks to be successful--no error is given.  (Other jobs use GETFMADDR and
 ACCESS to acquire the write link to the directory.  This always
 appears successful as well).
 Once the file is ready to be copied from the Work Machine's 191 disk to the
 SFS directory, the intermittent error appears.  The vast majority of the
 time, the write is successful.  However, sometimes, the job gets this error
 message:
 DMSOPN1258E You are not authorized to write to file XX 20110413 Z1

 The file is not large--last night's file was only 12 blocks.

 At the suggestion of our systems programmer, I've put in a lot of query
 statements.  I've issued QUERY LIMITS for the job submitter; it's only used
 84% of the allocation, with several thousand blocks available. The SFS
 directory owner has only used 76% of its allocation, with several thousand
 more blocks still available.  The filepool is not full.

 I've issued QUERY FILEPOOL CONFLICT.  There is no conflict.

 I've issued QUERY ACCESSED.  The directory shows that is accessed R/W.

 When the write is unsuccessful, the program then loops through 5 tries of
 releasing the access, reacquiring the access, and attempting to write the
 file again.  This has never been successful.  I've issued both a COPYFILE
 and a PIPE to try to write the file; these do not work once there has been a
 failure.

 We've looked at the operator consoles to see if we can find any jobs
 running at the same time.  We haven't found any that are accessing that
 directory structure.

 There aren't any dumps to look at--it looks perfectly successful other than
 the fact that it won't write the file.

 Does anyone have any suggestions of something to try next?


  Nora Graves
 nora.e.gra...@irs.gov





-- 
Kris Buelens,
IBM Belgium, VM customer support


Re: SFS problem

2011-04-14 Thread Hughes, Jim
Nora,

Have you looked at the SFS service machine's console log for any odd
messages?

Could the SFS server being doing some sort of control backup?

Jim Hughes
Consulting Systems Programmer 
Mainframe Technical Support Group
Department of Information Technology
State of New Hampshire
27 Hazen Drive
Concord, NH 03301
603-271-5586Fax 603.271.1516

Statement of Confidentiality: The contents of this message are
confidential. Any unauthorized disclosure, reproduction, use or
dissemination (either whole or in part) is prohibited. If you are not
the intended recipient of this message, please notify the sender
immediately and delete the message from your system.