Re: [zfs-discuss] zfs diff cannot stat shares

2010-10-13 Thread Tim Haley

On 10/13/10 10:20 AM, dirk schelfhout wrote:

Wanted to test the zfs diff command and ran into this.
I turned off all windows sharing.
the rpool has normal permissions for .zfs/shares

how do I fix this ?
Dirk

r...@osolpro:/data/.zfs# zfs diff d...@10aug2010 d...@13oct2010
Cannot stat /data/.zfs/shares/: unable to generate diffs

pwd
/data/.zfs
r...@osolpro:/data/.zfs# ls -alt
ls: cannot access shares: Operation not supported
total 4
dr-xr-xr-x  2 rootroot   2 2010-10-13 17:44 snapshot
drwxr-xr-x 18 piet staff 24 2010-10-13 17:44 ..
dr-xr-xr-x  4 rootroot   4 2009-01-28 23:08 .


What OS bits are you running?

-tim

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Making ZFS better: zfshistory

2010-04-21 Thread Tim Haley

On 04/21/10 03:24 AM, Darren J Moffat wrote:

On 21/04/2010 05:09, Edward Ned Harvey wrote:

From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss-
boun...@opensolaris.org] On Behalf Of Nicolas Williams

The .zfs/snapshot directory is most certainly available over NFS.


I'm not sure you've been following this thread. Nobody said .zfs/snapshot
wasn't available over NFS. It was said that all the snapshot
subdirectories
.zfs/snapshot/frequent-blah and .zfs/snapshot/hourly-foo and so on
...

Over NFS there's no way to know the time those snapshots were taken.
There
is a convention of writing the time of the snapshot into the name of the
snapshot, but if you can't rely on that, then the NFS client doesn't know
the order of snapshots.


Correct and think this is because POSIX has not such thing as a file
creation date (ctime,mtime,atime is all it has) - ZFS actually does have
a creation date but there isn't a way to expose that over NFS that I'm
aware of and ls/stat can't see it because stat(2) doesn't have a way to
expose it.


You can see it with ls:
# ls -ld -% all /net/server/export/ws/timh/nvc
drwxr-xr-x   9 timh staff 13 Apr 21 01:25 
/net/server/export/ws/timh/nvc/

 timestamp: atime Apr 21 07:54:05 2010
 timestamp: ctime Apr 21 01:25:48 2010
 timestamp: mtime Apr 21 01:25:48 2010
 timestamp: crtime Jun 22 08:27:23 2009

and I believe you can retrieve it with fgetattr() from the read/write
view.

-tim


The snapshot dir is just another directory and over NFS you are looking
at it with NFS so it shows the ctime,mtime,atime of the top level
directory of the ZFS dataset at the time the snapshot was created.

This RFE if it were to be implemented could give a possible way to get
this information from a remote filesystem client:

6527390 want to read zfs properties over nfs (eg via .zfs/props)



___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] [shell-discuss] getconf test for case insensitive ZFS?

2010-04-15 Thread Tim Haley

On 04/14/10 11:48 AM, Glenn Fowler wrote:


On Wed, 14 Apr 2010 17:54:02 +0200 =?KOI8-R?B?z8zYx8Egy9LZ1sHOz9fTy8HR?= wrote:

Can I use getconf to test if a ZFS file system is mounted in case
insensitive mode?


we would have to put in the zfs query (hopefull more generic that just for zfs)
the only current working case-insensitive checks are for uwin

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


There is a pathconf(2) check described in the
documentation for the case-insensitive behavior ARC case

http://arc.opensolaris.org/caselog/PSARC/2007/244/spec.txt

Not sure why it's not in the man page...

-tim
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] How to Catch ZFS error with syslog ?

2010-04-12 Thread Tim Haley

On 04/12/10 09:05 AM, J James wrote:

I have a simple mirror pool with 2 disks. I pulled out one disk to simulate a 
failed drive. zpool status shows that the pool is in DEGRADED state.

I want syslog to log these type of ZFS errors. I have syslog running and 
logging all sorts of error to a log server. But this failed disk in ZFS pool 
did not generate any syslog messages.

ZFS diagnosists engine are online as seen bleow.

hrs1zgpprd1# fmadm config | grep zfs
zfs-diagnosis 1.0 active ZFS Diagnosis Engine
zfs-retire 1.0 active ZFS Retire Agent

So, why is it not generating any syslog messages?


Try explicitly enabling fmd to send to syslog in
/usr/lib/fm/fmd/plugins/syslog-msgs.conf

-tim


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] how can I remove files when the fiile system is full?

2010-04-02 Thread Tim Haley

On 04/ 1/10 01:46 PM, Eiji Ota wrote:

During the IPS upgrade, the file system got full, then I cannot do anything to 
recover it.

# df -kl
Filesystem   1K-blocks  Used Available Use% Mounted on
rpool/ROOT/opensolaris
4976642   4976642 0 100% /
swap  14217564   244  14217320   1% /etc/svc/volatile
/usr/lib/libc/libc_hwcap2.so.1
4976642   4976642 0 100% /lib/libc.so.1
swap  14217460   140  14217320   1% /tmp
swap  1421734424  14217320   1% /var/run

# cd /var/adm
# rm messages.?
rm: cannot remove `messages.0': No space left on device
rm: cannot remove `messages.1': No space left on device

Likely a similar issue was reported a few years ago like:
http://opensolaris.org/jive/thread.jspa?messageID=241580#241580

However, my system is on snv_133.

Is there any way to work around the situation?
This is really critical since after the IPS gets the file system full, 
customers seem not able to recover.

Thanks,

-Eiji
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


On opensolaris?  Did you try deleting any old BEs?

-tim

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs diff

2010-03-29 Thread Tim Haley

On 3/29/10 8:02 PM, Daniel Carosone wrote:

On Tue, Mar 30, 2010 at 12:37:15PM +1100, Daniel Carosone wrote:


There will also need to be clear rules on output ordering, with
respect to renames, where multiple changes have happened to renamed
files.


Separately, but relevant in particular to the above due to the
potential for races:  what is the defined behaviour when diffing
against a live filesystem (rather than a snapshot)?

is there an implied snapshot (ie, diff based on content frozen at
txg_id when started) or is thhe comparison done against a moving
target?

It's not just a question of implementation if it can affect the
output, especially if it can make it internally inconsistent.

--
Dan.



Yes, a snapshot is taken and removed once the compare is performed.

-tim

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Recover rpool

2010-03-09 Thread Tim Haley

On 03/ 9/10 10:53 AM, Cindy Swearingen wrote:

Hi D,

Is this a 32-bit system?

We were looking at your panic messages and they seem to indicate a
problem with memory and not necessarily a problem with the pool or
the disk. Your previous zpool status output also indicates that the
disk is okay.

To perhaps clarify, you're panicking trying to grab a mutex, which hints 
that

something has stomped on the memory containing that mutex.   The reason
for the 32-bit question is that sometimes a deep stack can overrun on a 
32-bit

box.  That's probably not what happened here, but we ask anyway.

-tim

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] On collecting data from hangs

2009-12-20 Thread Tim Haley
There seems to be a rash of posts lately where people are resetting or 
rebooting without getting any data, so I thought I'd post a quick 
overview on collecting crash dumps.  If you think you've got a hang 
problem with ZFS and you want to gather data for someone to look at, 
then here are a few steps you should take.


If you already know all about gathering crash dumps on Solaris, feel 
free to delete now.


1) Make sure crash dumps are enabled

Enable saving of crash dumps by executing as root or with pfexec 
'dumpadm -y'.


The most reasonable trade-off of information vs size in the crash dump 
is probably 'dumpadm -c curproc'


If you're running Opensolaris you likely already have a dedicated zvol 
as a dump device.  If you're running SXCE you may need to dedicate a 
disk or slice for dump purposes.  Change or dedicate the dump device 
with 'dumpadm -d device'


See dumpadm(1M) for more info.

2) There are at least a couple of ways to capture crash dumps.

As root or with pfexec run 'savecore -L'.  This is a point-in-time 
capture of what's happening on the live system.  The system continues

to run during the capture so the results can be slightly inconsistent,
but the machine won't reboot.  Good if you think whatever is hung is
still making progress.

If you really don't mind rebooting, then 'reboot -nd' will most often 
get a dump without the dump also hanging up and forcing you to hard 
reset anyway.


---

Number 1 is best done now, before you have a hang.  It won't hurt 
anything to have crash dumps enabled - and if you ever get a panic 
you'll have the data needed for someone to analyze the issue.


If the crash dump saving works, the default location for the dumps to be 
stored is the directory /var/crash/hostname.


-tim

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Quick dedup question

2009-11-07 Thread Tim Haley

Rich Teer wrote:

Congrats for integrating dedup!  Quick question: in what build
of Nevada will dedep first be found?  b126 is the current one
presently.

Cheers,



128

-tim

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] dedupe question

2009-11-07 Thread Tim Haley

Dennis Clarke wrote:

Does the dedupe functionality happen at the file level or a lower block
level?


block level, but remember that block size may vary from file to file.


I am writing a large number of files that have the fol structure :

-- file begins
1024 lines of random ASCII chars 64 chars long
some tilde chars .. about 1000 of then
some text ( english ) for 2K
more text ( english ) for 700 bytes or so
--

Each file has the same tilde chars and then english text at the end of 64K
of random character data.

Before writing the data I see :

# zpool get size,capacity,version,dedupratio,free,allocated zp_dd
NAME   PROPERTYVALUESOURCE
zp_dd  size67.5G-
zp_dd  capacity6%   -
zp_dd  version 21   default
zp_dd  dedupratio  1.16x-
zp_dd  free63.3G-
zp_dd  allocated   4.19G-

After I see this :

# zpool get size,capacity,version,dedupratio,free,allocated zp_dd
NAME   PROPERTYVALUESOURCE
zp_dd  size67.5G-
zp_dd  capacity6%   -
zp_dd  version 21   default
zp_dd  dedupratio  1.11x-
zp_dd  free63.1G-
zp_dd  allocated   4.36G-


Note the drop in dedup ratio from 1.16x to 1.11x which seems to indicate
that dedupe does not detect the english text is identical in every file.





Theory:  Your files may end up being in one large 128K block or maybe a 
couple of 64K blocks where there isn't much redundancy to de-dup.


-tim

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS + fsck

2009-11-05 Thread Tim Haley

Orvar Korvar wrote:

Does this putback mean that I have to upgrade my zpool, or is it a zfs tool? If 
I missed upgrading my zpool I am smoked?


The putback did not bump zpool or zfs versions.  You shouldn't have to upgrade 
your pool.


-tim

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS + fsck

2009-11-05 Thread Tim Haley

Robert Milkowski wrote:

Miles Nordin wrote:

csb == Craig S Bell cb...@standard.com writes:
   csb Two: If you lost data with another filesystem, you may have
   csb overlooked it and blamed the OS or the application,

yeah, but with ZFS you often lose the whole pool in certain 
classes of

repeatable real-world failures, like hotswap disks with flakey power
or SAN's without NVRAM where the target reboots and the initiator 
does
not.  Losing the whole pool is relevantly different to corrupting 
the
insides of a few files. 
I think that most people including ZFS developers agree with you that 
losing an access to entire pool is not acceptable. And this has been 
fixed in snv_126 so now in those rare circumstances you should be able 
to import a pool. And generally you will end-up in a much better 
situation than with legacy filesystems + fsck.



Just a slight correction.  The current build in-process is 128 and that's the 
build into which the changes were pushed.


-tim

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS + fsck

2009-11-04 Thread Tim Haley

Robert Milkowski wrote:

Kevin Walker wrote:

Hi all,

Just subscribed to the list after a debate on our helpdesk lead me to 
the posting about  ZFS corruption and the need for a fsck repair tool 
of some kind...


Has there been any update on this?

  


I guess the discussion started after someone read an article on OSNEWS.

The way zfs works is that basically you get an fsck equivalent while 
using a pool.
ZFS checks checksums for all metadata and user data while reading it. 
Then all metadata are using ditto blocks to provide 2 or three copies of 
it (totally independent from any pool redundancy) depends on type of 
metadata. If it is corrupted a second (or third) copy will be used so 
correct data is returned and a corrupted block is automatically 
repaired. The ability to repair a block containing a user data depends 
on if you have a pool configured with or without redundancy. But even if 
pool is non-redundant (lets say a single disk drive) zfs still will be 
able to detect corruption and will be able to tell you what files are 
affected while metadata will be correct in most cases (unless corruption 
is so large and not localized so it affected all copies of a block in a 
pool). You will be able to read all other files and other parts of the 
file.


So fsck actually happens while you are accessing your data and it is 
even better than fsck on most other filesystems as thanks to 
checksumming of all data and metadata zfs knows exactly when/if 
something is wrong and in most cases is even able to fix it on the fly. 
If you want to scan entire pool including all redundant copies and get 
them fix if something doesn't checksum then you can schedule the pool 
scrubbing (while your applications are still using the pool!). This will 
force zfs to read all blocks from all copies to be read, their checksum 
checked and if needed data corrected if possible and the fact reported 
to user. Legacy fsck is not even close to it.



I think that the perceived need for fsck for ZFS probably comes from 
lack of understanding how ZFS works and from some frustrated users where 
under a very unlikely and rare circumstances due to data corruption a 
user might be in a position of not being able to import the pool 
therefore not being able to access any data at all while a corruption 
might have affected only a relatively small amount of data. Most other 
filesystem will allow you to access most of the data after fsck in such 
a situation (probably with some data loss) while zfs left user with no 
access to data at all. In such a case the problem lies with zfs 
uberblock and the remedy is to revert a pool to its previous uberblock 
version (or even an earlier one). In almost all the cases this will 
render a pool importable and then the mechanisms described in the first 
paragraph above will kick-in. The problem is (was) that the procedure to 
revert a pool to one of its previous uberblock is not documented nor is 
automatic and is definitely far from being sys-admin friendly. But 
thanks to some community members (most notably mr. Victor I think) some 
users affected by the issue were given a hand and were able to recover 
most/all their data. Others were probably assisted by Sun's support 
service I guess.


Fortunately a much more user-friendly mechanism has been finally 
implemented and inegrated into Open Solaris build 126 which allows a 
user to import a pool and force it to on of the previous versions of its 
uberblock if necessary. See 
http://c0t0d0s0.org/archives/6067-PSARC-2009479-zpool-recovery-support.html 
for more details.


There is another CR (don't have its number at hand) which is about 
implementing a delayed re-use on just freed blocks which should allow 
for more data to be recovered in such a case as above. Although I'm not 
sure if it has been implemented yet.


IMHO with the above CR implemented, in most cases ZFS currently provides 
*much* better solution to random data corruption than any other 
filesystem+fsck in the market.


The code for the putback of 2009/479 allows reverting to an earlier uberblock 
AND defers the re-use of blocks for a short time to make this rewind safer.


-tim
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] *Almost* empty ZFS filesystem - 14GB?

2009-08-16 Thread Tim Haley

Chris Murray wrote:
Accidentally posted the below earlier against ZFS Code, rather than ZFS Discuss. 


My ESXi box now uses ZFS filesystems which have been shared over NFS. Spotted 
something odd this afternoon - a filesystem which I thought didn't have any 
files in it, weighs in at 14GB. Before I start deleting the empty folders to 
see what happens, any ideas what's happened here?

# zfs list | grep temp
zp/nfs/esx_temp 14.0G 225G 14.0G /zp/nfs/esx_temp
# ls -la /zp/nfs/esx_temp
total 20
drwxr-xr-x 5 root root 5 Aug 13 12:54 .
drwxr-xr-x 7 root root 7 Aug 13 12:40 ..
drwxr-xr-x 2 root root 2 Aug 13 12:53 iguana
drwxr-xr-x 2 root root 2 Aug 13 12:54 meerkat
drwxr-xr-x 2 root root 2 Aug 16 19:39 panda
# ls -la /zp/nfs/esx_temp/iguana/
total 8
drwxr-xr-x 2 root root 2 Aug 13 12:53 .
drwxr-xr-x 5 root root 5 Aug 13 12:54 ..
# ls -la /zp/nfs/esx_temp/meerkat/
total 8
drwxr-xr-x 2 root root 2 Aug 13 12:54 .
drwxr-xr-x 5 root root 5 Aug 13 12:54 ..
# ls -la /zp/nfs/esx_temp/panda/
total 8
drwxr-xr-x 2 root root 2 Aug 16 19:39 .
drwxr-xr-x 5 root root 5 Aug 13 12:54 ..
#

Could there be something super-hidden, which I can't see here?

There don't appear to be any snapshots relating to zp/nfs/esx_temp.

On a suggestion, I have ran the following:

# zfs list -r zp/nfs/esx_temp
NAME USED AVAIL REFER MOUNTPOINT
zp/nfs/esx_temp 14.0G 225G 14.0G /zp/nfs/esx_temp
# du -sh /zp/nfs/esx_temp
8K /zp/nfs/esx_temp
#



Does

zfs list -t snapshot -r zp/nfs/esx_temp

show anything?

What about

zfs get refquota,refreservation,quota,reservation zp/fs/esx_tmp

-tim


Thanks,
Chris


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs-discuss Digest, Vol 46, Issue 50

2009-08-08 Thread Tim Haley

Allen Eastwood wrote:

Does DNLC even play a part in ZFS, or are the Docs out of date?

Defines the number of entries in the directory name look-up cache 
(DNLC). This parameter is used by UFS and NFS to cache elements of path 
names that have been resolved.


No mention of ZFS.  Noticed that when discussing that with a customer of 
mine.



The best workload is one that doesn't read from disk to begin with :-)
For workloads with millions of files (eg large-scale mail servers) you
will need to increase the size of the Directory Name Lookup Cache
(DNLC). By default, it is way too small for such workloads. If the
directory names are in cache, then they do not have to be read from
disk -- a big win.

You can see how well the DNLC is working by looking at the output of
vmstat -s and look for the total name lookups. You can size DNLC
by tuning the ncsize parameter, but it requires a reboot.  See the
Solaris Tunable Parameters Guide for details.
http://docs.sun.com/app/docs/doc/817-0404/chapter2-35?a=view





Yes, ZFS uses the DNLC as well.

-tim
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] APPLE: ZFS need bug corrections instead of new func! Or?

2009-06-19 Thread Tim Haley

Miles Nordin wrote:

bmm == Bogdan M Maryniuk bogdan.maryn...@gmail.com writes:


   bmm OK, so what is the status of your bugreport about this?

That's a good question if it's meant genuinely, and not to be
obstructionist.  It's hard to report one bug with clear information
because the problem isn't well-isolated yet.

In my notes:  6565042, 6749630


The first of which is marked as fixed in snv_77, 19 months ago.

The second is marked as a duplicate of 6784395, fixed in snv_107, 20 weeks ago.

-tim


but as I said before, I've found the information on the mailing list
more useful w.r.t. this particular problem.

You can see how those bugs are about specific,
methodically-reproduceable problems.  Bugs are not ``I have been
losing more zpools than I lost UFS/vxfs filesystems on top of the same
storage platform.''  


It may take a while.




___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS snapshot send/recv hangs X4540 servers

2009-06-10 Thread Tim Haley

Brent Jones wrote:

On Mon, Jun 8, 2009 at 9:38 PM, Richard Lowerichl...@richlowe.net wrote:

Brent Jones br...@servuhome.net writes:




I've had similar issues with similar traces.  I think you're waiting on
a transaction that's never going to come.

I thought at the time that I was hitting:
  CR 6367701 hang because tx_state_t is inconsistent

But given the rash of reports here, it seems perhaps this is something
different.

I, like you, hit it when sending snapshots, it seems (in my case) to be
specific to incremental streams, rather than full streams, I can send
seemingly any number of full streams, but incremental sends via send -i,
or send -R of datasets with multiple snapshots, will get into a state
like that above.

-- Rich



For now, back to snv_106 (the most stable build that I've seen, like it a lot)
I'll open a case in the morning, and see what they suggest.


After examining the dump we got from you (thanks again), we're relatively sure 
you are hitting


6826836 Deadlock possible in dmu_object_reclaim()

This was introduced in nv_111 and fixed in nv_113.

Sorry for the trouble.

-tim

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS snapshot send/recv hangs X4540 servers

2009-06-05 Thread Tim Haley

Brent Jones wrote:

Hello all,
I had been running snv_106 for about 3 or 4 months on a pair of X4540's.
I would ship snapshots from the primary server to the secondary server
nightly, which was working really well.

However, I have upgraded to 2009.06, and my replication scripts appear
to hang when performing zfs send/recv.
When one zfs send/recv process hangs, you cannot send any other
snapshots from any other filesystem to the remote host.
I have about 20 file systems I snapshots and replicate nightly.

The script I use to perform the snapshots is here:
http://www.brentrjones.com/wp-content/uploads/2009/03/replicate.ksh

On the remote side, I end up with many hung processes, like this:

  bjones 11676 11661   0 01:30:03 ?   0:00 /sbin/zfs recv -vFd pdxfilu02
  bjones 11673 11660   0 01:30:03 ?   0:00 /sbin/zfs recv -vFd pdxfilu02
  bjones 11664 11653   0 01:30:03 ?   0:00 /sbin/zfs recv -vFd pdxfilu02
  bjones 13727 13722   0 14:21:20 ?   0:00 /sbin/zfs recv -vFd pdxfilu02

And so on, one for each file system.

On the receiving end, 'zfs list' shows one filesystem attempting to
receive a snapshot, but I cannot stop it:

$ zfs list
NAME   USED  AVAIL  REFER  MOUNTPOINT
pdxfilu02/data/fs01/%20090605-00:30:00  1.74G  27.2T   208G
/pdxfilu02/data/fs01/%20090605-00:30:00



On the sending side, I CAN kill the ZFS send process, but the remote
side leaves its processes going, and I CANNOT kill -9 them. I also
cannot reboot the receiving system, at init 6, the system will just
hang trying to unmount the file systems.
I have to physically cut power to the server, but a couple days later,
this issue will occur again.


A crash dump from the receiving server with the stuck receives would be highly 
useful, if you can get it.  Reboot -d would be best, but it might just hang. 
You can try savecore -L.


-tim


I'f I boot to my snv_106 BE, everything works fine, this issue has
never occurred on that version.

Any thoughts?



___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs related google summer of code ideas - your vote

2009-03-06 Thread Tim Haley

Dave wrote:

C. Bergström wrote:

Bob Friesenhahn wrote:
I don't know if anyone has noticed that the topic is google summer 
of code.  There is only so much that a starving college student can 
accomplish from a dead-start in 1-1/2 months.  The ZFS equivalent of 
eliminating world hunger is not among the tasks which may be 
reasonably accomplished, yet tasks at this level of effort is all 
that I have seen mentioned here.
May I interject a bit.. I'm silently collecting this task list and 
even outside of gsoc may help try to arrange it from a community 
perspective.  Of course this will be volunteer based unless /we/ get a 
sponsor or sun beats /us/ to it.  So all the crazy ideas welcome..




I would really like to see a feature like 'zfs diff f...@snap1 
f...@othersnap' that would report the paths of files that have either been 
added, deleted, or changed between snapshots. If this could be done at 
the ZFS level instead of the application level it would be very cool.


--
Dave
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


this is actually in the works.  There is a functioning prototype.

-tim



___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] possibly a stupid question, why can I not set sharenfs=sec=krb5, rw?

2009-02-27 Thread Tim Haley

Alastair Neil wrote:

x86 snv 108

I have a pool with around 5300 file systems called home.  I can do:

zfs set sharenfs=on home

however

zfs set sharenfs=sec=krb5,rw home

complains:

cannot set property for 'home': 'sharenfs' cannot be set to invalid options

I feel I must be overlooking something elementary.



Thanks, Alastair




___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Does a by-hand share succeed?

I.e, share -F nfs -o sec=krb5,rw /home

?

-tim

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS upgrade mangled my share

2009-01-23 Thread Tim Haley
Colin Johnson wrote:
 I was having CIFs problems on my Mac so I upgrade to build 105.
 After getting all my shares populated with data I ran zpool scrub on  
 the raidz array and it told me the version was out of date so I  
 upgraded.
 
 One of my shares is now inaccessible and I cannot even delete it :(
 
 r...@bitchko:/nas1/backups# zfs list -r /nas1/backups/
 NAME   USED  AVAIL  REFER  MOUNTPOINT
 nas1/backups  33.3K  1.04T  33.3K  /nas1/backups
 
 r...@bitchko:/# zfs destroy /nas1/backups/
 cannot open '/nas1/backups/': invalid dataset name
 

There's a typo there, you would have to do

zfs destroy nas1/backups

Unfortunately, you can't use the mountpoint, you have to name the dataset.

-tim

 r...@bitchko:/nas1# rm -Rf backups/
 rm: cannot remove directory `backups/': Device busy

 
 
 This is pretty scary for something that's storing valuable data.
 
 Any ideas how I can destroy the dataset and recreate it?
 
 Thanks
 
 Colin
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS encryption?? - [Fwd: [osol-announce] SXCE Build 105 available

2009-01-23 Thread Tim Haley
Jerry K wrote:
 It was rumored that Nevada build 105 would have ZFS encrypted file 
 systems integrated into the main source.
 
 In reviewing the Change logs (URL's below) I did not see anything 
 mentioned that this had come to pass.  Its going to be another week 
 before I have a chance to play with b105.
 
 Does anyone know specifically if b105 has ZFS encryption?
 
It does not.

-tim

 Thanks,
 
 Jerry
 
 
  Original Message 
 Subject: [osol-announce] SXCE Build 105 available
 Date: Fri, 09 Jan 2009 08:58:40 -0800
 
 
 Please find the links to SXCE Build 105 at:
 
  http://www.opensolaris.org/os/downloads/
 
 This is still a DVD only release.
 
 -
 wget work around:
 
   http://wikis.sun.com/pages/viewpage.action?pageId=28448383
 
 ---
 Changelogs:
 
 ON (The kernel, drivers, and core utilities):
 
   http://dlc.sun.com/osol/on/downloads/b105/on-changelog-b105.html
 
 X Window System:
 
 http://opensolaris.org/os/community/x_win/changelogs/changelogs-nv_100/
 
 - Derek
 ___
 opensolaris-announce mailing list
 opensolaris-annou...@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/opensolaris-announce
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Changing casesensitivity for existing filesystems?

2008-12-11 Thread Tim Haley
Nicolas Williams wrote:
 On Wed, Dec 10, 2008 at 11:13:21AM -0800, Jay Anderson wrote:
 The casesensitivity option is just like utf8only and normalization, it
 can only be set at creation time. The result from attempting to change
 it on an existing filesystem:

 # zfs set casesensitivity=mixed pool0/data1
 cannot set property for 'pool0/data1': 'casesensitivity' is readonly

 From the ZFS Administration Guide 
 (http://docs.sun.com/app/docs/doc/817-2271/6mhupg6lc?a=view#indexterm-319):
 
 The manpage I looked at didn't mention that, though it did for
 normalization and utf8only, so I foolishly stopped there.  Sounds like a
 manpage bug.  My apologies for the confusion.
 

The manpage reads:

  The following three properties cannot be changed  after  the
  file  system  is  created, and therefore, should be set when
  the file system is created. If the properties  are  not  set
  with the zfs create or zpool create commands, these pro-
  perties are inherited from the parent dataset. If the parent
  dataset  lacks  these  properties due to having been created
  prior to these features being supported, the new file system
  will have the default values for these properties.

  casesensitivity = sensitive | insensitive | mixed

...

  normalization =none | formD | formKCf

...

  utf8only =on | off

...

We reiterate that the property cannot be changed in the paragraphs 
describing the second two, perhaps we should do so for the first as well.

 I would think only the casesensitivity=mixed should have to be set at
 creation time, that casesensitivity=insensitive could be set at any
 time.  Hmmm.
 

We don't allow this for a couple of reasons.  If the file system was 
case-sensitive or mixed and you suddenly make it insensitive, you 
potentially create name conflicts:

% ls
Readme README readme

if we make this case-insensitive all of sudden you've got two files that 
can no longer be looked up.

The other reason is performance.  Knowing beforehand that we need to 
track or disallow case conflicts helps us to optimize to keep lookups fast.

-tim

 Nico

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] cp: Operation not supported

2008-12-11 Thread Tim Haley
Kristof Van Damme wrote:
 Hi Tim,
 Thanks for having a look.
 The 'utf8only' setting is set to off.
 
 Important bit of additional information:
 We only seem to have this problem when copying to a zfs filesystem with the 
 casesensitivity=mixed property. We need this because the filesystem will 
 eventually be shared over NFS and CIFS.
 
 I attached a file with the relevant zfs properties of the filesystem that has 
 the problem.
 See zfs-prop.txt. Notice that utf8only is set to off.
 
 My locale is set to en_US.ISO8859-15 when doing the copy, but that doesn't 
 seem to change anything. I tried with en_US.UTF-8 as well with the same 
 result.
 
 A truss on the failed can be found in truss-ENOTSUP.txt.
 Notice the \xf6 in the filename of the stat64, which gives an ENOTSUP.
 
 When we take an the UTF-8 name, the stat64 succeeds just fine. A truss of 
 this can be found in truss-OK.txt
 
 Cheers,
 Kristof/
 
 
Kristof,

Your detailed information was very helpful.  I have reproduced the 
problem on my own machine.  I will need to consult with our i18n folks 
on whether the problem is within zfs or one of their routines we are 
calling.

-tim
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] cp: Operation not supported

2008-12-11 Thread Tim Haley
Kristof Van Damme wrote:
 Hi Tim,
 That's splendid!
 
 In case other people want to reproduce the issue themselves, here is how.
 In attach is a tar which contains the 2 files (UTF8 and ISO8859) like the 
 ones I used in my first post to demonstrate the problem. Here are the 
 instructions to reproduce the issue:
 
 Create a zfs filesystem with casesensitivity set to mixed and create one 
 without this setting.
 Unpack the tar outside the pool and try to copy both the files from the tar 
 on to the zfs fileystem. One (UTF8) will work fine, the other will not.
 # zfs create -o casesensitivity=mixed mypool/mixed
 # zfs create mypool/normal
 # cp UTF8-Köln.txt /mypool/mixed/
 # cp ISO8859-K?ln.txt /mypool/mixed/
 cp: cannot stat `/mypool/mixed/ISO8859-K\366ln.txt': Operation not supported
 # cp UTF8-Köln.txt /mypool/normal/
 # cp ISO8859-K?ln.txt /mypool/normal/
 # 
 
 
 Kristof/
 
Thanks, Kristof

I filed

6783995 file creation sometimes failing within a zfs where 
casesensitivity=mixed

a short while ago.  I believe I already have a fix:

chinook:~/bugs/nonutf[1016] ls
ISO8859-K?ln.txt*  UTF8-Köln.txt koeln.tar
chinook:~/bugs/nonutf[1016] zfs get casesensitivity tp/timh/mix
NAME PROPERTY VALUESOURCE
tp/timh/mix  casesensitivity  mixed-
chinook:~/bugs/nonutf[1017] cp UTF8-K\303\266ln.txt /tp/timh/mix
chinook:~/bugs/nonutf[1019] cp ISO8859-K\366ln.txt /tp/timh/mix
chinook:~/bugs/nonutf[1021] ls /tp/timh/mix
ISO8859-K?ln.txt*  UTF8-Köln.txt

Still will need some folks to look at the change, though.

-tim

 
 
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] cp: Operation not supported

2008-12-10 Thread Tim Haley
Kristof Van Damme wrote:
 Hi All,
 We have set up a zpool on OpenSolaris 2008.11, but have difficulties copying 
 files with special chars in the name when the name is encoded in ISO8859-15. 
 When the name is in UTF8 we don't have this problem. We get Operation not 
 supported.
 We want to copy the files with their name in ISO8859-15 encoding and do not 
 want to convert them to UTF8, because we think this will cause it problems 
 further down the road: The filesystem will be used as a network share (CIFS 
 and NFS) for windows and linux clients. The clients may not all properly 
 support UTF8 at the moment. So we'd like to avoid having to convert to UTF8 
 at this moment. 
 
 How can we copy files with an ISO8859-encoded name?
 
 To demonstrate the problem:
 # /usr/bin/cp ISO8859-K?ln.url /datapool/appl
 cp: cannot create /datapool/appl/ISO8859-K?ln.url: Operation not supported
 # /usr/bin/cp UTF8-Köln.txt /datapool/appl/
 #
 
 Kristof/

What is the output from:

zfs get utf8only datapool/appl

?

thanks,
-tim

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs is a co-dependent parent and won't let children leave home

2008-12-09 Thread Tim Haley
Ross wrote:
 While it's good that this is at least possible, that looks horribly 
 complicated to me.  
 Does anybody know if there's any work being done on making it easy to remove 
 obsolete 
 boot environments?

If the clones were promoted at the time of their creation the BEs would 
stay independent and individually deletable.  Promotes can fail, though, 
if there is not enough space.

I was told a little while back when I ran into this myself on an Nevada 
build where ludelete failed, that beadm *did* promote clones.  This 
thread appears to be evidence to the contrary.  I think it's a bug, we 
should either promote immediately on creation, or perhaps beadm destroy 
could do the promotion behind the covers.

-tim

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs is a co-dependent parent and won't let children leave home

2008-12-09 Thread Tim Haley
Kyle McDonald wrote:
 Tim Haley wrote:
 Ross wrote:
   
 While it's good that this is at least possible, that looks horribly 
 complicated to me.  
 Does anybody know if there's any work being done on making it easy to 
 remove obsolete 
 boot environments?
 
 If the clones were promoted at the time of their creation the BEs would 
 stay independent and individually deletable.  Promotes can fail, though, 
 if there is not enough space.

 I was told a little while back when I ran into this myself on an Nevada 
 build where ludelete failed, that beadm *did* promote clones.  This 
 thread appears to be evidence to the contrary.  I think it's a bug, we 
 should either promote immediately on creation, or perhaps beadm destroy 
 could do the promotion behind the covers.
   
 If I understand this right, the latter option looks better to me. Why 
 consume the disk space before you have to?
 What does LU do?
 

ludelete doesn't handle this any better than beadm destroy does, it 
fails for the same reasons. lucreate does not promote the clone it 
creates when a new BE is spawned, either.

-tim

   -Kyle
 
 -tim

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
   
 
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] problems accessing ZFS snapshots

2008-07-30 Thread Tim Haley
Thomas Nau wrote:
 Dear all.
 I stumbled over an issue triggered by Samba while accessing ZFS snapshots. 
 As soon as a Windows client tries to open the .zfs/snapshot folder it 
 issues the Microsoft equivalent of ls dir, dir *. It get's translates 
 by Samba all the way down into stat64(/pool/.zfs/snapshot*). The stat() 
 call fails with error EILSEQ (Illegal byte sequence). It can easily 
 reproduced by
   mybox# /bin/ls /pool/.zfs/snapshot/\*
   *: Illegal byte sequence
 
 Doing the same on the original ZFS the snapshot was taken from everything 
 works fine.
 
 BTW, I'm running snv_93.
 
 Any hints on that?
 Thomas
 
 -
 GPG fingerprint: B1 EE D2 39 2C 82 26 DA  A5 4D E0 50 35 75 9E ED
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Interesting.  What are the normalization and utf8only property values for the 
  ZFS file system?  I.e., what's the output from

'zfs get normalization,utf8only pool/fs'

-tim

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] problems accessing ZFS snapshots

2008-07-30 Thread Tim Haley
Thomas Nau wrote:
 Dear all.
 I stumbled over an issue triggered by Samba while accessing ZFS snapshots. 
 As soon as a Windows client tries to open the .zfs/snapshot folder it 
 issues the Microsoft equivalent of ls dir, dir *. It get's translates 
 by Samba all the way down into stat64(/pool/.zfs/snapshot*). The stat() 
 call fails with error EILSEQ (Illegal byte sequence). It can easily 
 reproduced by
   mybox# /bin/ls /pool/.zfs/snapshot/\*
   *: Illegal byte sequence
 
 Doing the same on the original ZFS the snapshot was taken from everything 
 works fine.
 
 BTW, I'm running snv_93.
 
 Any hints on that?
 Thomas
 
 -
 GPG fingerprint: B1 EE D2 39 2C 82 26 DA  A5 4D E0 50 35 75 9E ED
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Ah, ignore my previous question.  We believe we found the problem, and filed:

6731778  'ls *' in empty zfs snapshot directory returns EILSEQ vs. ENOENT we 
get in other empty directories

Fix will likely go back today or tomorrow and be present in snv_96.

Thanks for finding this.

-tim
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Creating root FS with mixed case sensitivity?

2008-06-28 Thread Tim Haley
Aaron Moore wrote:
 Hello,
 
 I am trying to create a file server that will be used across multiple OSs and 
 I wanted to create the root pool/zfs system with mixed case. I am not sure 
 how to do this since the root zfs fs gets automatically created without this 
 flag when I create the pool.
 
 Is there a way to do this?
 
 Thanks
  
  
 This message posted from opensolaris.org
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Not just yet, unfortunately, but we are working on it.  This is bug 6647661.

http://www.opensolaris.org/isearch.jspa;jsessionid=F5B8A30AE9981256C1DE4F37EC7B72E1?query=6647661Submit=Search

-tim



___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Solaris 10 x86 + ZFS / NFS server cp problem with AIX

2008-03-18 Thread Tim Haley
Sachin Palav wrote:
 Friends,
 I have recently built a file server on x2200 with solaris x86 having zfs 
 (version4) and running NFS version2  samba.
 
 the AIX 5.2  AIX 5.2 client give error while running command cp -R 
 zfs_nfs_mount_source zfs_nfs_mount_desticantion as below:
 cp: 0653-440 directory/1: name too long.
 cp: 0653-438 cannot read directory directory/1.
  and the cp core dumps in AIX.
 
 Can someone please help
 
This is almost certainly an instance of 6538383 in our bug database. 
scandir() is indeed broken on AIX, this is not a ZFS bug.  If you have 
IBM support, you need to file something called a PMR with them to get 
the ball rolling on a fix.  It's possible someone at Sun opened one 
already, I'm not certain.

The workaround is to use a program that doesn't rely on scandir() to do 
the copying, tar or cpio as previously mentioned.  Adding the -p option 
to cp appeared to be a workaround at one point, but that was a red 
herring, the failure with cp can be intermittent as it involves 
scandir() scribbling somewhat randomly over memory.

-tim

 Thanks
 Sachin Palav
  
  
 This message posted from opensolaris.org
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] path-name encodings

2008-02-27 Thread Tim Haley
Roland Mainz wrote:
 Bart Smaalders wrote:
 Marcus Sundman wrote:
 I'm unable to find more info about this. E.g., what does reject file
 names mean in practice? E.g., if a program tries to create a file
 using an utf8-incompatible filename, what happens? Does the fopen()
 fail? Would this normally be a problem? E.g., do tar and similar
 programs convert utf8-incompatible filenames to utf8 upon extraction if
 my locale (or wherever the fs encoding is taken from) is set to use
 utf-8? If they don't, then what happens with archives containing
 utf8-incompatible filenames?
 Note that the normal ZFS behavior is exactly what you'd expect: you
 get the filenames you wanted; the same ones back you put in.
 
 Does ZFS convert the strings to UTF-8 in this case or will it just store
 the multibyte sequence unmodified ?
 
ZFS doesn't muck with names it is sent when storing them on-disk.  The 
on-disk name is exactly the sequence of bytes provided to the open(), 
creat(), etc.  If normalization options are chosen, it may do some 
manipulation of the byte strings *when comparing* names, but the on-disk 
name should be untouched from what the user requested.

-tim

 
 
 Bye,
 Roland
 

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Can't remove corrupt file

2006-07-19 Thread Tim Haley

On Wed, 19 Jul 2006, Eric Lowe wrote:



(Also BTW that page has a typo, you might want to get the typo fixed, I 
didn't know where the doc bugs should go for those messages)


- Eric


Product: event_registry
Category: events
Sub-Category: msg

-tim
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss