Re: [zfs-discuss] aclmode gone in S10u10?

2011-09-14 Thread Ian Collins

 On 09/14/11 08:49 PM, Sami Ketola wrote:

On Sep 13, 2011, at 3:21 , Peter Tribble wrote:

On Tue, Sep 13, 2011 at 1:50 AM, Paul B. Hensonhen...@acm.org  wrote:

I recently saw a message posted to the sunmanagers list complaining
about installing a kernel patch and suddenly having his ACL's disappear
completely whenever a chmod occurred. I replied and asked him to check
if the aclmode attribute was gone, as it sounded like the default
discard that was (questionably) implemented in OpenSolaris/Solaris 11.
He confirmed it was, so it looks like the removal of aclmode was
backported to Solaris 10? I don't know exactly what kernel patch he
installed; it doesn't look like update 10 is out yet.

Update 10 has been out for about 3 weeks.

No it has not. As of now Solaris 10 u10 has not been released.


It slipped out the back door for a couple of days a few weeks ago.

--
Ian.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] booting from ashift=12 pool..

2011-09-14 Thread Hans Rosenfeld
Hi,

On Mon, Sep 05, 2011 at 02:18:48AM -0400, Daniel Carosone wrote:
 I see via the issue tracker that there have been several updates
 since, and an integration back into the main Illumos tree.   How do I
 go about getting hold of current boot blocks?

The OpenIndiana release that was announced earlier today has the fixed
boot blocks.


Hans


-- 
%SYSTEM-F-ANARCHISM, The operating system has been overthrown

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] zfs destroy snapshot runs out of memory bug

2011-09-14 Thread Paul Kraus
I know there was (is ?) a bug where a zfs destroy of a large
snapshot would run a system out of kernel memory, but searching the
list archives and on defects.opensolaris.org I cannot find it. Could
someone here explain the failure mechanism in language a Sys Admin (I
am NOT a developer) could understand. I am running Solaris 10 with
zpool 22 and I am looking for both understanding of the underlying
problem and a way to estimate the amount of kernel memory necessary to
destroy a given snapshot (based on information gathered from zfs, zdb,
and any other necessary commands).

Thanks in advance, and sorry to bring this up again. I am almost
certain I saw mention here that this bug is fixed in Solaris 11
Express and Nexenta (Oracle Support is telling me the bug is fixed in
zpool 26 which is included with Solaris 10U10, but because of our use
of ACLs I don't think I can go there, and upgrading the zpool won't
help with legacy snapshots).

-- 
{1-2-3-4-5-6-7-}
Paul Kraus
- Senior Systems Architect, Garnet River ( http://www.garnetriver.com/ )
- Sound Designer: Frankenstein, A New Musical
(http://www.facebook.com/event.php?eid=123170297765140)
- Sound Coordinator, Schenectady Light Opera Company (
http://www.sloctheater.org/ )
- Technical Advisor, RPI Players
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs destroy snapshot runs out of memory bug

2011-09-14 Thread Richard Elling
On Sep 14, 2011, at 9:50 AM, Paul Kraus wrote:

I know there was (is ?) a bug where a zfs destroy of a large
 snapshot would run a system out of kernel memory, but searching the
 list archives and on defects.opensolaris.org I cannot find it. Could
 someone here explain the failure mechanism in language a Sys Admin (I
 am NOT a developer) could understand. I am running Solaris 10 with
 zpool 22 and I am looking for both understanding of the underlying
 problem and a way to estimate the amount of kernel memory necessary to
 destroy a given snapshot (based on information gathered from zfs, zdb,
 and any other necessary commands).

I don't recall a bug with that description. However, there are several bugs that
relate to how the internals work that were fixed last summer and led to the
on-disk format change to version 26 (Improved snapshot deletion performance).

Look for details in 
http://src.illumos.org/source/history/illumos-gate/usr/src/uts/common/fs/zfs/
during the May-July 2010 timeframe. Methinks the most important change was
6948890 snapshot deletion can induce pathologically long spa_sync() times
spa_sync() is called when the transaction group is sync'ed to permanent storage.
-- richard

 
 Thanks in advance, and sorry to bring this up again. I am almost
 certain I saw mention here that this bug is fixed in Solaris 11
 Express and Nexenta (Oracle Support is telling me the bug is fixed in
 zpool 26 which is included with Solaris 10U10, but because of our use
 of ACLs I don't think I can go there, and upgrading the zpool won't
 help with legacy snapshots).

Sorry, I haven't run Solaris 10 in the past 6 years :-) can't help you there.
But I can say that NexentaStor has this bug fix in 3.0.5. For NexentaStor 3.1+
releases, zpool version is 28.
 -- richard

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs destroy snapshot runs out of memory bug

2011-09-14 Thread Richard Elling
Question below…

On Sep 14, 2011, at 12:07 PM, Paul Kraus wrote:

 On Wed, Sep 14, 2011 at 2:30 PM, Richard Elling
 richard.ell...@gmail.com wrote:
 
 I don't recall a bug with that description. However, there are several bugs 
 that
 relate to how the internals work that were fixed last summer and led to the
 on-disk format change to version 26 (Improved snapshot deletion performance).
 Look for details in 
 http://src.illumos.org/source/history/illumos-gate/usr/src/uts/common/fs/zfs/
 during the May-July 2010 timeframe. Methinks the most important change was
 6948890 snapshot deletion can induce pathologically long spa_sync() times
 spa_sync() is called when the transaction group is sync'ed to permanent 
 storage.
 
 I looked through that list, and found the following that looked applicable:
 6948911 snapshot deletion can induce unsatisfiable allocations in txg sync
 6948890 snapshot deletion can induce pathologically long spa_sync() times
 
 But all I get at bugs.opensolaris.org is a Service Temporarily
 Unavailable message (and have for at least the past few weeks). The
 MOS lookup of the 6948890 bug yields the title and not much else, no
 details. I can't even find the 6948911 bug in MOS.
 
 MOS == My Oracle Support
 
 Thanks for the pointers, I just wish I could find more data that will
 lead me to either:
 A) a mechanism to estimate the RAM needed to destroy a pre-26 snapshot
 -or-
 B) indication that there is no way to do A.
 
 From watching the system try to import this pool, it looks like it is
 still building a kernel structure in RAM when the system runs out of
 RAM. It has not committed anything to disk.

Did you experience a severe memory shortfall? 
(Do you know how to determine that condition?)
 -- richard

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Is there any implementation of VSS for a ZFS iSCSI snapshot on Solaris?

2011-09-14 Thread S Joshi




I am using a Solaris + ZFS environment to export a iSCSI
block layer device and use the snapshot facility to take a snapshot of the ZFS
volume. Is there an existing Volume Shadow Copy (VSS) implementation on
Windows for this environment?



Thanks

 

S Joshi

  ___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs destroy snapshot runs out of memory bug

2011-09-14 Thread Paul Kraus
On Wed, Sep 14, 2011 at 4:31 PM, Richard Elling
richard.ell...@gmail.com wrote:

 From watching the system try to import this pool, it looks like it is
 still building a kernel structure in RAM when the system runs out of
 RAM. It has not committed anything to disk.

 Did you experience a severe memory shortfall?
 (Do you know how to determine that condition?)

T2000 with 32 GB RAM

zpool that hangs the machine by running it out of kernel memory when
trying to import the zpool

zpool has an incomplete snapshot from a zfs recv that it is trying
to destroy on import

I *can* import the zpool readonly

So the answer is yes to the severe memory shortfall. One of the
many things I did to instrument this system was as simple as running
vmstat 10 on the console :-) The last instance before the system hung
showed a scan rate of 900,000 ! In one case I watched as it hung (it
has done this many times as I have troubleshot with Oracle Support)
and did not see *any* user level processes that would account for the
memory shortfall. I have logs of system freemem showing the memory
exhaustion. Oracle Support has confirmed (from a core dump) that it is
some combination of the two bugs you mentioned (plus they created a
new Bug ID for this specific problem).

I have asked multiple times if the incomplete snapshot could be
corrupt in a way that would cause this (early on then led us to
believe the incomplete snapshot was 7 TB when it should be about 2.5
TB), but have not gotten anything substantive back (just a one line,
The snapshot is not corrupt.).

What I am looking for is a way to estimate the kernel memory
necessary to destroy a given snapshot so that I can see if any of the
snapshots on my production server (M4000 with 16 GB) will run the
machine out of memory.

-- 
{1-2-3-4-5-6-7-}
Paul Kraus
- Senior Systems Architect, Garnet River ( http://www.garnetriver.com/ )
- Sound Designer: Frankenstein, A New Musical
(http://www.facebook.com/event.php?eid=123170297765140)
- Sound Coordinator, Schenectady Light Opera Company (
http://www.sloctheater.org/ )
- Technical Advisor, RPI Players
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] booting from ashift=12 pool..

2011-09-14 Thread Daniel Carosone
On Wed, Sep 14, 2011 at 04:08:19PM +0200, Hans Rosenfeld wrote:
 On Mon, Sep 05, 2011 at 02:18:48AM -0400, Daniel Carosone wrote:
  I see via the issue tracker that there have been several updates
  since, and an integration back into the main Illumos tree.   How do I
  go about getting hold of current boot blocks?
 
 The OpenIndiana release that was announced earlier today has the fixed
 boot blocks.

Yep, saw that and have it here ready to boot and install grub.  I hope
the fact that the pool itself is v31 for zfs crypto will not be a
problem.. 

If it should be the case that the pool version is an issue running
from the OI CD, can I take the updated stage* files and use them with
the installgrub from solaris express b151?

I guess I'll find out in due course. 

--
Dan.


pgph7UOuW8sMw.pgp
Description: PGP signature
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs destroy snapshot runs out of memory bug

2011-09-14 Thread Daniel Carosone
On Wed, Sep 14, 2011 at 05:36:53PM -0400, Paul Kraus wrote:
 T2000 with 32 GB RAM
 
 zpool that hangs the machine by running it out of kernel memory when
 trying to import the zpool
 
 zpool has an incomplete snapshot from a zfs recv that it is trying
 to destroy on import
 
 I *can* import the zpool readonly

Can you import it booting from a newer kernel (say liveDVD), and allow
that to complete the deletion? Or does this not help until the pool is
upgraded past the on-disk format in question, for which it must first
be imported writable?  

If you can import it read-only, would it be faster to just send it
somewhere else?  Is there a new-enough snapshot near the current data?

--
Dan.



pgpSAJqkjQOd1.pgp
Description: PGP signature
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss