Thank you all for your insightful user experiences in response to my question.
If I may summarize the points that apply to my particular situation:
* No real issues with respect to backups, particularly for
consolidations, e.g.
of volume pools -- in fact we may even see a slight
.
Ron
-Original Message-
From: IBM Mainframe Discussion List [mailto:IBM-MAIN@LISTSERV.UA.EDU] On
Behalf Of Don Williams
Sent: Friday, February 22, 2013 2:36 PM
To: IBM-MAIN@LISTSERV.UA.EDU
Subject: Re: [IBM-MAIN] Mod-9 vs. Mod-27 vs. mixed
I think DASD vendors need to create a simple
: [IBM-MAIN] Reducing Backup Time (Was: Mod-9 vs. Mod-27 vs. mixed)
On 2/23/2013 9:33 AM, Paul Gilmartin wrote:
Most of my department's large UNIX filesystems are NFS mounted from
ZFS (not zFS) on Solaris servers. Our daily backups are ZFS
snapshots, almost negligible latency, followed
On 24/02/2013 4:45, Ed Jaffe wrote:
My question is about whether a DFS/SMB ZFS should be backed up at all
given its size and the existence of a daily TSM backup of its contents.
In theory, a TSM backup of the contents should be adequate. In practice,
there are a few things that I would want
Discussion List [mailto:IBM-MAIN@LISTSERV.UA.EDU]
On Behalf Of Ron Hawkins
Sent: Sunday, February 24, 2013 5:07 PM
To: IBM-MAIN@LISTSERV.UA.EDU
Subject: Re: Mod-9 vs. Mod-27 vs. mixed
Don,
I'm can't speak for the EMC and IBM iterations, but with HDS HDP setting
up
mirrored configurations
On 2/22/2013 1:08 PM, Skip Robinson wrote:
The biggest advantage of large volumes is reduction in UCB count. One
disadvantage is that larger volumes take longer to back up and to restore.
We have some astonishingly large 'volumes' in the open systems world that
are supported by only modestly
On Sat, 23 Feb 2013 08:59:39 -0800, Ed Jaffe wrote:
One area of focus is our corporate file server, which utilizes the
DFS/SMB server on z/OS. We have large, single-volume ZFS data sets on
large volumes. (They used to be multivolume ZFS on smaller volumes). One
small update causes the entire ZFS
On 2/23/2013 9:33 AM, Paul Gilmartin wrote:
Most of my department's large UNIX filesystems are NFS mounted
from ZFS (not zFS) on Solaris servers. Our daily backups are
ZFS snapshots, almost negligible latency, followed by background
dumps to tape.
Right. FLASHCOPY of a large volume on our
Space wasted for small volumes e.g. XCF couple datasets. Just a talking
point. With Hyper-PAV, etc. most of the other points are just hot air.
The convenience of not having to support multiple geometries.
snip
A client with DS8000 DASD configured as a mix of 3390 Mod9 and Mod27s is
jo.skip.robin...@sce.com
From: Staller, Allan allan.stal...@kbmg.com
To: IBM-MAIN@LISTSERV.UA.EDU,
Date: 02/22/2013 10:04 AM
Subject:Re: Mod-9 vs. Mod-27 vs. mixed
Sent by:IBM Mainframe Discussion List IBM-MAIN@LISTSERV.UA.EDU
Space wasted for small volumes e.g. XCF
Discussion List [mailto:IBM-MAIN@LISTSERV.UA.EDU] On Behalf
Of Skip Robinson
Sent: Friday, February 22, 2013 1:40 PM
To: IBM-MAIN@LISTSERV.UA.EDU
Subject: Re: Mod-9 vs. Mod-27 vs. mixed [ External ]
In the interest of frugality, I asked my storage guys some time ago to allocate
some tiny volumes
Program Co-Manager
626-302-7535 Office
323-715-0595 Mobile
jo.skip.robin...@sce.com
From: Lester, Bob bles...@oppenheimerfunds.com
To: IBM-MAIN@LISTSERV.UA.EDU,
Date: 02/22/2013 12:54 PM
Subject:Re: Mod-9 vs. Mod-27 vs. mixed
Sent by:IBM Mainframe Discussion List IBM
We are looking at converting about 12TB of 21TB of disk from Mod 9 to
27 (32,760). We are doing those storage groups that have 250GB of
data in them so they use 10 M27s and adding another M27 adds 10% or
less. Two storage groups have 1,600GB so adding 9 or 29.4 GB are very
minor. We have some
On Friday 22 February 2013 15:08:34 Skip Robinson wrote:
Someone once observed that 'rarely used' is not the same the thing as
'lightly used'. Whenever a file is accessed for read or write, IOS goes
after it with a heavy boot. Then there's the problem of backups: if a file
is worth keeping
If I remember correctly, there is also an issue of the number of paths
available; fewer, larger volumes = fewer paths = more I/O contention. (This
may be an old issue that is no longer relevant.)
Pretty well with PAV HIPERPAV.
-
Ted MacNEIL
eamacn...@yahoo.ca
Twitter: @TedMacNEIL
15 matches
Mail list logo