While not 'quick dirty', I ran across a nifty utility I refer back to every once in a
while called the 3D FileSystem Profiler. It will take the specified directory,
and create a 3D NAVIGATABLE chart of of directories/files showing file sizes/number of files.
I have a C program which produces a histogram of
files versus age and size.
It's a bit rough but if anyone wants it I can
beat it into releaseable state.
Matt
--
SLUG - Sydney Linux User's Group Mailing List - http://slug.org.au/
Subscription info and FAQs:
On Fri, 15 Apr 2005, Voytek wrote:
using Tony's method, my largest file is:
find / -ls -type f 2/dev/null | awk '{print $7 $11 }' | sort -nr |
head -10
# cd /proc
Hi,
A handy way to avoid find going into other filesystems (useful when
network mounts are around), is to use the -mount option.
using Tony's method, my largest file is:
find / -ls -type f 2/dev/null | awk '{print $7 $11 }' | sort -nr |
head -10
# cd /proc
# ls -l kc*
-r1 root root 948887552 Apr 15 18:41 kcore
what is it and do I need it ..?
--
Voytek
--
SLUG - Sydney Linux User's Group Mailing
Ben Donohue [EMAIL PROTECTED] writes:
Further to this (and this is not an answer to the question above) but
I'm buggered if i can find the largest files on the hard disk and list
them in order.
I've tried various arguements but can't seem to crack it.
like find / -S -r (or -s) -name
the
information in it.
-Original Message-
From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On
Behalf Of Julio Cesar Ody
Sent: Friday, 15 April 2005 12:29 PM
To: James Ballantine
Cc: slug@slug.org.au
Subject: Re: [SLUG] finding a file
A little bit lazy to figure how to get full paths, but
clean
Voytek wrote:
using Tony's method, my largest file is:
find / -ls -type f 2/dev/null | awk '{print $7 $11 }' | sort -nr |
head -10
# cd /proc
# ls -l kc*
-r1 root root 948887552 Apr 15 18:41 kcore
what is it and do I need it ..?
It's a core dump of memory. Do:
size core
and
# cd /proc
^^^
# ls -l kc*
-r1 root root 948887552 Apr 15 18:41 kcore
what is it and do I need it ..?
It's a core dump of memory. Do: size core
and it will tell you what program dumped it, and then think back to what
you were doing with that proggy at
On Fri, 15 Apr 2005 11:49:38 +1000
Tony Green [EMAIL PROTECTED] wrote:
This will give you the 10 largest files on your hard drive (that can be
seen by the user executing the code)
find / -ls -type f 2/dev/null | awk '{print $7 $11 }' | sort -nr |
head -10
Not that it really matters in
Gottfried Szing wrote:
# cd /proc
^^^
# ls -l kc*
-r1 root root 948887552 Apr 15 18:41 kcore
what is it and do I need it ..?
It's a core dump of memory. Do: size core
and it will tell you what program dumped it, and then think back to what
you were doing with
I'm trying to find a specific file withing a web tree, what the way to do it:
I tried this with no luck
# locate /home/domain.org.au localconf.php
only to get
find: localconf.php: No such file or directory
--
Voytek
--
SLUG - Sydney Linux User's Group Mailing List - http://slug.org.au/
On Fri, 2005-04-15 at 00:23 +1000, Voytek wrote:
I'm trying to find a specific file withing a web tree, what the way to do it:
I tried this with no luck
# locate /home/domain.org.au localconf.php
only to get
find: localconf.php: No such file or directory
Locate uses a database of file
Voytek wrote:
I'm trying to find a specific file withing a web tree, what the way to do it:
I tried this with no luck
# locate /home/domain.org.au localconf.php
only to get
find: localconf.php: No such file or directory
I have similar requirement all the time and I use,
$ find
You could run
$ slocate -U /home/domain
with this command you create a database of file locations, which later
can be used through
$ slocate localconf.php
The database needs to be updated once in a while, in case you move
files around often. Maybe you could put the first command above in a
On 4/15/05, Julio Cesar Ody [EMAIL PROTECTED] wrote:
You could run
$ slocate -U /home/domain
with this command you create a database of file locations, which later
can be used through
$ slocate localconf.php
The database needs to be updated once in a while, in case you move
files
Voytek wrote:
I'm trying to find a specific file withing a web tree, what the way to do it:
I tried this with no luck
# locate /home/domain.org.au localconf.php
only to get
find: localconf.php: No such file or directory
Further to this (and this is not an answer to the question above) but
I'm
Ben Donohue wrote:
Further to this (and this is not an answer to the question above) but
I'm buggered if i can find the largest files on the hard disk and list
them in order.
I've tried various arguements but can't seem to crack it.
like find / -S -r (or -s) -name xxx|more
Any ideas out there?
On 15/04/2005, at 11:44, Ben Donohue wrote:
Further to this (and this is not an answer to the question above) but
I'm buggered if i can find the largest files on the hard disk and list
them in order.
I've tried various arguements but can't seem to crack it.
like find / -S -r (or -s) -name
Not quite what you wanted, but to get the largest files or directories
in the current directory in order, you can use:
du -cks * |sort -nr |head -n15
This came from one of the O'Reilly UNIX books if I recall correctly.
They suggested you alias it to 'ducks' for ease of typing.
/james
Ben
A little bit lazy to figure how to get full paths, but
clean and simple:
$ ls -RShl
just the size and filename:
$ ls -Shl | awk '{print $5 $8}'
no directories, just the size and filename:
$ ls -RShl | grep -v '^d' | awk '{print $5 $8}'
there's probably easier ways to do it, but
20 matches
Mail list logo