Evren Yurtesen wrote:
> Jason Hughes wrote:
>> That drive should be more than adequate. Mine is a 5400rpm 2mb
>> buffer clunker. Works fine.
>> Are you running anything else on the backup server, besides
>> BackupPC? What OS? What filesystem? How many files total?
>
> FreeBSD, UFS2+softupdates, noatime.
>
> There are 4 hosts that have been backed up, for a total of:
>
> * 16 full backups of total size 72.16GB (prior to pooling and
> compression),
> * 24 incr backups of total size 13.45GB (prior to pooling and
> compression).
>
> # Pool is 17.08GB comprising 760528 files and 4369 directories (as of
> 3/27 05:54),
> # Pool hashing gives 38 repeated files with longest chain 6,
> # Nightly cleanup removed 10725 files of size 0.40GB (around 3/27 05:54),
> # Pool file system was recently at 10% (3/27 07:16), today's max is
> 10% (3/27 01:00) and yesterday's max was 10%.
>
> Host User #Full Full Age (days) Full Size
> (GB) Speed (MB/s) #Incr Incr Age (days) Last
> Backup (days) State Last attempt
> host1 4 5.4 3.88 0.22 6 0.4
> 0.4 idle idle
> host2 4 5.4 2.10 0.06 6 0.4
> 0.4 idle idle
> host3 4 5.4 7.57 0.14 6 0.4
> 0.4 idle idle
> host4 4 5.4 5.56 0.10 6 0.4
> 0.4 idle idle
>
Hmm. This is a tiny backup setup, even smaller than mine. However, it
appears that the average size of your file is only 22KB, which is quite
small. For comparison sake, this is from my own server:
Pool is 172.91GB comprising 217311 files and 4369 directories (as of
3/26 01:08),
The fact that you have tons of little files will probably give
significantly higher overhead when doing file-oriented work, simply
because the inodes must be fetched for each file before seeking to the
file itself. If we assume no files are shared between hosts (very
conservative), and you have an 8ms access time, you will have 190132
files per host and two seeks per file, neglecting actual i/o time, gives
you 50 minutes. Just to seek them all. If you have a high degree of
sharing, it can be up to 4x worse. Realize, the same number of seeks
must be made on the server as well as the client.
Are you sure you need to be backing up everything that you're putting
across the network? Maybe excluding some useless directories, maybe
temp files or logs that haven't been cleaned up? Perhaps you can
archive big chunks of it with a cron job?
I'd start looking for ways to cut down the number of files, because the
overhead of per-file accesses are probably eating you alive. I'm also
no expert on UFS2 or FreeBSD, so it may be worthwhile to research its
behavior with hard links and small files.
JH
-------------------------------------------------------------------------
Take Surveys. Earn Cash. Influence the Future of IT
Join SourceForge.net's Techsay panel and you'll get the chance to share your
opinions on IT & business topics through brief surveys-and earn cash
http://www.techsay.com/default.php?page=join.php&p=sourceforge&CID=DEVDEV
_______________________________________________
BackupPC-users mailing list
[email protected]
https://lists.sourceforge.net/lists/listinfo/backuppc-users
http://backuppc.sourceforge.net/