Re: [zfs-discuss] problem with zfs receive -i

2006-08-19 Thread Matthew Ahrens
On Sat, Aug 19, 2006 at 07:21:52PM -0700, Frank Cusack wrote:
> On August 19, 2006 7:06:06 PM -0700 Matthew Ahrens <[EMAIL PROTECTED]> 
> wrote:
> >My guess is that the filesystem is not mounted.  It should be remounted
> >after the 'zfs recv', but perhaps that is not happening correctly.  You
> >can see if it's mounted by running 'df' or 'zfs list -o name,mounted'.
> 
> You are right, it's not mounted.
> 
> >Did the 'zfs recv' print any error messages?
> 
> nope.
> 
> > Are you able to reproduce this behavior?
> 
> easily.

Hmm, I think there must be something special about your filesystems or
configuration; I'm not able to reproduce it.  One possible cause for
trouble is if you are doing the 'zfs receive' into a filesystem which
has descendent filesystems (eg, you are doing 'zfs recv pool/[EMAIL PROTECTED]'
and pool/fs/child exists).  This isn't handled correctly now, but you
should get an error message in that case.  (This will be fixed by some
changes Noel is going to putback next week.)

Could you send me the output of 'truss zfs recv ...', and 'zfs list' and
'zfs get -r all ' on both the source and destination systems?

> ah ok.  Note that if I do zfs send; zfs send -i on the "local side", then
> do zfs list; zfs mount -a on the "remote side", I still show space used
> in the @7.1 snapshot, even though I didn't touch anything.  I guess mounting
> accesses the mount point and updates the atime.

Hmm, maybe.  I'm not sure if that's exactly what's happening, because
mounting and unmounting a filesystem doesn't seem to update the atime
for me.  Does the @7.1 snapshot show used space before you do the 'zfs
mount -a'?

> On the "local" side, how come after I take the 7.1 snapshot and then 'ls',
> the 7.1 snapshot doesn't start using up space?  Shouldn't my ls of the
> mountpoint update the atime also?

I believe what's happening here is that although we update the in-core
atime, we sometimes defer pushing it to disk.  You can force the atime
to be pushed to disk by unmounting the filesystem.

--matt
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] problem with zfs receive -i

2006-08-19 Thread Frank Cusack

On August 19, 2006 7:06:06 PM -0700 Matthew Ahrens <[EMAIL PROTECTED]> wrote:

On Sat, Aug 19, 2006 at 06:31:47PM -0700, Frank Cusack wrote:

But when I login to zone smb and cd to /share/tmp/.zfs I get 'no such
file or directory'.  This does exist for other filesystems like
/zone/eng/.zfs.


My guess is that the filesystem is not mounted.  It should be remounted
after the 'zfs recv', but perhaps that is not happening correctly.  You
can see if it's mounted by running 'df' or 'zfs list -o name,mounted'.


You are right, it's not mounted.


Did the 'zfs recv' print any error messages?


nope.


 Are you able to reproduce this behavior?


easily.


1. Is the need to reboot a bug?  Certainly having the other snapshots
  go missing seems like a bug.


Yes, it is a bug for the filesystem to not be remounted after the 'zfs
recv'.  FYI, you should be able to get it mounted again by running 'zfs
mount -a', you don't have to reboot the entire zone.


yay, that works.


2. Why does the 7.1 snapshot have the extra space?


If the filesystem (export/zone/smb/share/tmp) is modified, then some of
the data that it shared with the snapshot (@7.1) will not be shared any
longer, so it will become unique to the snapshot and show up as "used"
space.  Even if you didn't explicitly change anything in the filesystem,
with the default settings, simply reading files in the filesystem will
cause their atimes to be modified.


ah ok.  Note that if I do zfs send; zfs send -i on the "local side", then
do zfs list; zfs mount -a on the "remote side", I still show space used
in the @7.1 snapshot, even though I didn't touch anything.  I guess mounting
accesses the mount point and updates the atime.

On the "local" side, how come after I take the 7.1 snapshot and then 'ls',
the 7.1 snapshot doesn't start using up space?  Shouldn't my ls of the
mountpoint update the atime also?

-frank
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] problem with zfs receive -i

2006-08-19 Thread Frank Cusack

I followed the formula in zfs(1M) to replicate a zfs filesytem remotely.
The initial transfer works as I expected (and as documented).  The
filesystems were created on the remote side and the data transferred.
The man page is really difficult to interpret as to whether the data
will be in the filesystem or only in the snapshot, but anyway it did
what I wanted.

Then I added a file to one of the filesystems and took a second snapshot,
and sent it with 'zfs send -i'.  No errors from the 'zfs recv' but the data
does not appear.  There *is* a new snapshot on the remote side, but it's
not accessible.  Also, the first snapshot is now no longer accessible,
in fact all of .zfs is no longer accessible.

This is a zoned filesystem.

Here you can see they contain the same amount of data:

 [EMAIL PROTECTED]:~]# zfs list -r export/zone/smb/share/tmp
 NAME   USED  AVAIL  REFER  MOUNTPOINT
 export/zone/smb/share/tmp48K   225G  25.5K  /share/tmp
 export/zone/smb/share/[EMAIL PROTECTED]  22.5K  -  24.5K  -
 export/zone/smb/share/[EMAIL PROTECTED]  0  -  25.5K  -
 [EMAIL PROTECTED]:~]#

 [EMAIL PROTECTED]:~]# zfs list -r export/zone/smb/share/tmp
 NAME   USED  AVAIL  REFER  MOUNTPOINT
 export/zone/smb/share/tmp48K   225G  25.5K  /share/tmp
 export/zone/smb/share/[EMAIL PROTECTED]  22.5K  -  24.5K  -
 export/zone/smb/share/[EMAIL PROTECTED]  0  -  25.5K  -
 [EMAIL PROTECTED]:~]#

But when I login to zone smb and cd to /share/tmp/.zfs I get 'no such
file or directory'.  This does exist for other filesystems like
/zone/eng/.zfs.

hmm ok if I reboot the zone .zfs shows up and the filesystem itself
(not just the snapshot) contains the new data.  Curiously the 7.1
snapshot contains data now.

 NAME   USED  AVAIL  REFER  MOUNTPOINT
 export/zone/smb/share/tmp  69.5K   225G  25.5K  /share/tmp
 export/zone/smb/share/[EMAIL PROTECTED]  22.5K  -  24.5K  -
 export/zone/smb/share/[EMAIL PROTECTED]  21.5K  -  25.5K  -
 [EMAIL PROTECTED]:~]#

So, questions:

1. Is the need to reboot a bug?  Certainly having the other snapshots
  go missing seems like a bug.
2. Why does the 7.1 snapshot have the extra space?

This is S10 06/06 on x86.

thanks
-frank
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS Boot Disk

2006-08-19 Thread Torrey McMahon

Tabriz Leman wrote:


Torrey McMahon wrote:


Lori Alt wrote:



No, zfs boot will be supported on both x86 and sparc.  Sparc's
OBP, and various x86 BIOS's both have restrictions on the devices
that can be accessed at boot time, so we need to limit the
devices in a root pool on both architectures.



Hi Lori.

Can you expand a bit on the above? What sort of limitations are you 
referring too? (Boot time? Topology?)


I think what Lori is referring to here is that we need to limit the 
rootpool to BIOS/OBP visible devices; not all devices are visible from 
the BIOS/OBP (fibre channel devices, for example). 



Actually, OBP and BIOS should have access to fabric devices. I know OBP 
does and I've seen docs that mention BIOS does. (Look in the x86 fabric 
boot docs for example)


I think the problem is more around, as Lori mentioned in her email, the 
ability to traverse a ZFS pool when the ZFS drivers haven't been loaded.



___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] SPEC SFS97 benchmark of ZFS,UFS,VxFS

2006-08-19 Thread Frank Cusack

On August 19, 2006 10:53:55 AM -0700 Eric Redmond <[EMAIL PROTECTED]> wrote:

  Sun Cluster 3.2: New Features


wow, this makes 3.1 sound like dog food.

-frank
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] SPEC SFS97 benchmark of ZFS,UFS,VxFS

2006-08-19 Thread Eric Redmond




George Wilson wrote On 08/18/06 14:08,:
Frank,
  
  
The SC 3.2 beta maybe closed but I'm forwarding your request to Eric
Redmond.
  

The Sun Cluster 3.2 Beta program has been extended.  You can apply for
the Beta via this URL:
https://feedbackprograms.sun.com/callout/default.html?callid={11B4E37C-D608-433B-AF69-07F6CD714AA1}

Sun Cluster 3.2: New Features
Ease of Use
New Sun Cluster Object Oriented Command Set
The
new SC command line interface includes one command per cluster object
type and consistent use of sub-command names and option letters. It
also supports short and long command names. The commands output have
been greatly improved with better help and error messages as well as
more readable status and configuration reporting. In addition some
commands include export and import options with use of portable
XML-based configuration files allowing replication of portion of, or
entire cluster configurations.
This
new interface is easier to learn and easier to use, thereby limiting
human errors during administration of clusters. It also speeds up
partial or full configuration cloning.
Oracle RAC 10g improved integration and administration
Sun
Cluster RAC packages installation as well as configuration are now
integrated in the Sun Cluster procedures. New RAC specific resource
types and properties can be used for finer grained control.
Oracle
RAC extended manageability leads to easier set-up of Oracle RAC within
Sun Cluster as well as improved diagnosability and availability.
Agent configuration wizards
New
GUI-based wizard provides simplified configuration for popular
applications via on-line help, automatic discovery of parameter choices
and immediate validation. Supported applications include Oracle RAC and
HA, NFS, Apache and SAP.
The configuration of agents is easier and less error-prone
enabling faster set-up of popular solutions.
Flexible IP address scheme
Sun
Cluster now allows a reduced range of IP addresses for its private
interconnect. In addition it becomes also possible to customize the IP
base address and its range during or after installation.
These changes facilitate integration of Sun Cluster
environments in existing networks with limited or regulated address
spaces.


Availability
Cluster support for SMF services
Sun
Cluster now integrates tightly with Solaris 10 Service Management
Facility (SMF) and enables the encapsulation of SMF controlled
applications in the Sun Cluster resource management model. Local
service-level life-cycle management continues to be operated by SMF
while whole resource level cluster-wide failure (node, storage, ...)
handling operations are carried out by Sun Cluster.
Moving
applications from a single node Solaris 10 environment to multi-node
Sun Cluster environment enables increased availability while requiring
limited-to-no effort.  
Extended flexibility for fencing protocol
This new functionality allows the customization of the
default fencing protocol: choices include  
SCSI 3, SCSI 2 or per-device discovery.
This
flexibility enables the default usage of SCSI 3, a more recent
protocol, for better support for multi-pathing, easier integration with
non-Sun storage and shorter recovery times on newer storage while still
supporting the SC 3.0/3.1 behavior and SCSI 2 for older devices.
Quorum Server
A
new quorum device option is now available in Sun Cluster: instead of
using a shared disk and SCSI reservation protocols, it is now possible
to use a Solaris server outside of the cluster to run a quorum server
module supporting an atomic reservation protocol over TCP/IP. 

This
enables faster failover time but also lowers deployment costs: it
removes the need of a shared quorum disk for any scenario where quorum
is required (2-node) or desired.
Disk path failure handling
Sun Cluster can now be configured to automatically reboot a
node if all its path to shared disk have failed.
Faster reaction in case of severe disk path failure enables
improved availability.
HA Storage plus availability improvements
HA
Storage plus mount points are created automatically in case of mount
failure to eliminate failure-to-failover cases thus improving
availability of the environment  


Flexibility
Solaris Container expanded support


Any
application of scalable or failover type and their associated Sun
Cluster agents can now run unmodified within Solaris Containers (except
Oracle RAC).  

This
allows the combination of the benefits of application containment
offered by Solaris containers and the increased availability provided
by Sun Cluster.

Note: Currently only the following Sun Cluster Agents are supported in
Solaris Containers


  JES Application
Server
  JES Web Server
  JES MQ Server
  DNS
  Apache
  Kerberos
  HA-Oracle

HA ZFS
ZFS is supported as a failover file system in Sun Cluster.
ZFS
and Sun Cluster offers a best class file system solution combining high
availability, data integrity, performance and scalability covering the
needs of the most demanding environments.
H

Re: [zfs-discuss] in-kernel gzip compression

2006-08-19 Thread Bill Moore
On Sat, Aug 19, 2006 at 01:25:21PM +0200, michael schuster wrote:
> maybe a stupid question: what do we use for compressing dump data on the 
> dump device?

We use a variant of Lempel-Ziv called lzjb (the jb is for Jeff Bonwick).
The algorithm was designed for very small code/memory footprint and to
be very fast (single CPU can compress data at 150MB/s).  When running a
compression algorithm in panic context, you have a lot of constraints.


--Bill
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] in-kernel gzip compression

2006-08-19 Thread michael schuster

Matthew Ahrens wrote:

On Thu, Aug 17, 2006 at 02:53:09PM +0200, Robert Milkowski wrote:

Hello zfs-discuss,

  Is someone actually working on it? Or any other algorithms?
  Any dates?


Not that I know of.  Any volunteers? :-)

(Actually, I think that a RLE compression algorithm for metadata is a
higher priority, but if someone from the community wants to step up, we
won't turn your code away!)


maybe a stupid question: what do we use for compressing dump data on the 
dump device?


Michael
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss