Michael, please keep CCs. Its usual to keep them on kernel related lists. 
Thanks.


Am Sonntag, 28. Oktober 2012 schrieb Michael Kjörling:
> On 27 Oct 2012 23:38 +0100, from h...@carfax.org.uk (Hugo Mills):
> >>>            Data: RAID 0   System: RAID 1   Unused
> >>> 
> >>> /dev/vdb     307.25 MB                -        2.23 GB
> >>> /dev/vdc     307.25 MB             8 MB        2.69 GB
> >>> /dev/vdd     307.25 MB             8 MB        2.24 GB
> >>> 
> >>>            ============   ==============   ============
> >>> 
> >>> TOTAL        921.75 MB            16 MB        7.16 GB
> >> 
> >> It would scale better with the number of drives and there is a good
> >> way to place the totals.
> >> 
> >    Note that this could get arbitrarily wide in the presence of the
> > 
> > (planned) per-object replication config. Otherwise, it works. The
> > width is probably likely to grow more slowly than the length, though,
> > so this way round is probably the better option. IMO. Eggshell blue
> > is good enough. :)
> 
> Of course, but the suggestion in the mail I replied to can get equally
> arbitrarily wide in the presence of a large number of _drives_.
> 
> In my experience, many times it's better to put something together
> that works with the current status of the project and start using it,
> than trying to shoehorn every "we'd like to do this some day" feature
> into the original design. _Particularly_ when it's UI one is talking
> about. I can think of a few ways it might be possible to restrict the
> growth of the width of a table like this even in the face of separate
> per-object replication settings, the most obvious probably being to
> keep a tally on disk for each of the replication types, and have
> columns for each replication configuration (so you might get one
> column for RAID 0 data, one for RAID 1 data, one for SINGLE data, and
> so on, but you'll _never_ get more "data" columns than the filesystem
> itself supports replication methods for "data" data; the tally simply
> being an optimization so you don't have to scan the whole file system
> for a simple "df"), but by the time that feature gets implemented,
> maybe someone can think of a better presentation.

Good idea.

Maybe its also possible to hide to subsum empty trees as "Empty areas" or 
"Unused" or so.

> After all, UI aspects tend to be the easiest to fiddle with.

I agree. Output is not set in stone and I still think, scripts shall not 
parse it. If output is needed for scripts, provide a switch for csv or 
json like output. fio - the flexible I/O tester - has got JSON support 
recently so some code is already there. Or provide direct API by libbtrfs 
or so.

> Organizing the drives in rows also has the advantage that you don't
> _have_ to read everything before you can start printing the results,
> if you can live with the constraint of supporting only one data and
> metadata replication strategy. Whether to implement it that way is
> another matter. With large storage systems and multi-CPU/multi-core
> systems, while a multithreaded approach might not provide consistent
> device ordering between executions depending on the exact thread
> execution order, it could provide a fair performance enhancement. And
> forget KISS; don't we all _love_ a chance to do a little multithreaded
> programming before coffee if it saves the poor sysadmin a few dozen
> milliseconds per "df"? ;-)

Well if tabular I am all for having drives in rows.

Aside from all of this, I wonder how ZFS tools are doing it?

Anyone who has access to ZFS and can provide some outputs?

There are claims for it being better in that regard¹. I bet the new quota 
stuff provides for used space per subvolume, but still they claim they can 
now how much space is free exactly. If different things can have different 
replication strategies I think they can´t.

Still I would like to see the output of something ZFS commands to show 
disk-usage. Maybe that gives some additional ideas.

Hmmm, I got something:

merkaba:/> zpool create -m /mnt/zeit zeit /dev/merkaba/zeit


merkaba:/> zfs create zeit/test1
merkaba:/> zfs create zeit/test2
merkaba:/> zfs create zeit/test3


merkaba:/> mount | grep zfs
kstat on /zfs-kstat type fuse 
(rw,nosuid,nodev,relatime,user_id=0,group_id=0,allow_other)
zeit on /mnt/zeit type fuse.zfs 
(rw,relatime,user_id=0,group_id=0,default_permissions,allow_other)
zeit/test1 on /mnt/zeit/test1 type fuse.zfs 
(rw,relatime,user_id=0,group_id=0,default_permissions,allow_other)
zeit/test2 on /mnt/zeit/test2 type fuse.zfs 
(rw,relatime,user_id=0,group_id=0,default_permissions,allow_other)
zeit/test3 on /mnt/zeit/test3 type fuse.zfs 
(rw,relatime,user_id=0,group_id=0,default_permissions,allow_other)


merkaba:/> zfs list
NAME         USED  AVAIL  REFER  MOUNTPOINT
zeit         168K  19,6G    24K  /mnt/zeit
zeit/test1    21K  19,6G    21K  /mnt/zeit/test1
zeit/test2    21K  19,6G    21K  /mnt/zeit/test2
zeit/test3    21K  19,6G    21K  /mnt/zeit/test3
merkaba:/> dd if=/dev/zero of=/mnt/zeit/test2/schlumpf bs=1M count=10
10+0 Datensätze ein
10+0 Datensätze aus
10485760 Bytes (10 MB) kopiert, 0,022653 s, 463 MB/s
merkaba:/> zfs list                                                  
NAME         USED  AVAIL  REFER  MOUNTPOINT
zeit         169K  19,6G    25K  /mnt/zeit
zeit/test1    21K  19,6G    21K  /mnt/zeit/test1
zeit/test2    21K  19,6G    21K  /mnt/zeit/test2
zeit/test3    21K  19,6G    21K  /mnt/zeit/test3
merkaba:/> sync
merkaba:/> zfs list
NAME         USED  AVAIL  REFER  MOUNTPOINT
zeit         169K  19,6G    25K  /mnt/zeit
zeit/test1    21K  19,6G    21K  /mnt/zeit/test1
zeit/test2    21K  19,6G    21K  /mnt/zeit/test2
zeit/test3    21K  19,6G    21K  /mnt/zeit/test3
merkaba:/> dd if=/dev/urandom of=/mnt/zeit/test2/schlumpf bs=1M count=10
10+0 Datensätze ein
10+0 Datensätze aus
10485760 Bytes (10 MB) kopiert, 0,929913 s, 11,3 MB/s
merkaba:/> zfs list                                                     
NAME         USED  AVAIL  REFER  MOUNTPOINT
zeit        10,2M  19,6G    25K  /mnt/zeit
zeit/test1    21K  19,6G    21K  /mnt/zeit/test1
zeit/test2  10,0M  19,6G  10,0M  /mnt/zeit/test2
zeit/test3    21K  19,6G    21K  /mnt/zeit/test3

Hmmm. They just tell used/free space, but also for "subvolumes".

Which we can only do by estimation in some cases.


[1] http://rudd-o.com/linux-and-free-software/ways-in-which-zfs-is-better-
than-btrfs under "ZFS tracks used space per file system"

Thanks,
-- 
Martin 'Helios' Steigerwald - http://www.Lichtvoll.de
GPG: 03B0 0D6C 0040 0710 4AFA  B82F 991B EAAC A599 84C7
--
To unsubscribe from this list: send the line "unsubscribe linux-btrfs" in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html

Reply via email to