Re: [BackupPC-users] Backup fails after three days, possibly millions of files

2019-07-19 Thread David Koski



On 7/16/19 4:27 PM, Adam Goryachev wrote:

On 17/7/19 4:22 am, David Koski wrote:


Regards,
David Koski
dko...@sutinen.com

On 7/8/19 6:16 PM, Adam Goryachev wrote:

On 9/7/19 10:23 am, David Koski wrote:
I am trying to back up about 24TB of data that has millions of 
files.  It takes a day or to before it starts backing up and then 
stops with an error.  I did a CLI dump and trapped the output and 
can see the error message:


Can't write 32780 bytes to socket
Read EOF: Connection reset by peer
Tried again: got 0 bytes
finish: removing in-process file 
Shares/Archives//COR_2630.png

Child is aborting
Done: 589666 files, 1667429241846 bytes
Got fatal error during xfer (aborted by signal=PIPE)
Backup aborted by user signal
Not saving this as a partial backup since it has fewer files than 
the prior one (got 589666 and 589666 files versus 4225016)

dump failed: aborted by signal=PIPE

This backup is doing rsync over ssh.  I enabled SSH keepalive but 
it does not appear to be due to an idle network.  It does not 
appear to be a random network interruption because the time it 
takes to fail is pretty consistent, about three days. I'm stumped. 



Did you check:

$Conf{ClientTimeout} = 72000;

Also, what version of rsync on the client, what version of BackupPC 
on the server, etc?


I think BPC v4 handles this scenario significantly better, in fact a 
server I used to have trouble with on BPC3.x all the time has since 
been combined with 4 other server (so 4 x the number of files and 
total size of data) and BPC4 handles it easily.





Thank you all for your input.  More information:

rsync version on client: 3.0.8 (Windows)
rsync version on server: 3.1.2 (Debian)
BackupPC version: 3.3.1
$(Config{ClientTimeout} = 604800

I just compared the output of two verbose BackupPC_dump runs and it 
looks like the files are reported to be backed up even though they 
are not.  For example, this appears in logs of both backup runs:


create   644  4616/545  1085243184 /3412.zip

I checked and the file time stamp is year 2018.  The log files are 
full of these.  I checked the real time clock on both systems and 
they are correct.  There are also files that have been backed up that 
are not in the logs.


I suspect there are over ten million files but I don't have a good 
way of telling now.  Oddly, there are about 500,000 files backed 
according to the log captured from BackupPC_dump and almost the same 
number actually backed up and found in pc//0, but they are 
different subsets of files.  I have been tracking memory and swap 
usage on the server and see no issues.


Is this a possible bug in BackupPC 3.3.1?


Please don't top-post if you can avoid it, at least not on mailing lists.

I just realised:

Read EOF: Connection reset by peer

This is a networking issue, not BackupPC. In other words, something 
has broken the network connection (in the middle of transferring a 
file, so I would presume it isn't due to some idle timeout, dropped 
NAT entry, etc). BackupPC has been told by the operating system that 
the connection is no longer valid, and so it has "cleaned up" by 
removing the in-progress file (partial).


I just completed another backup cycle that failed in the same manner but 
this time with a continuous ping with captured output.  It didn't miss a 
beat.




It takes a day to start (presumably reading ALL the files on the 
client takes this long, you could improve disk performance, or 
increase RAM on the client to improve this).


You might be right.  But it's not a show stopper.



"and then stops with an error" - is that on the first file, or are 
some files successfully transferred? Is that the first large file? 
Does it always fail on the same file (seems not, since it previously 
got many more).


Good points.  Confirmed: Not the first file (over 600,000 files 
transferred first), not a large file (less than 20Meg), does not always 
fail on the same file or directory.




I'm thinking you need to check and/or improve network reliability, 
make sure both client and server are not running out of RAM/etc 
(mainly the backuppc client, the OOM might kill the rsync process), 
etc. Check your system logs on both client and server, and/or watch 
top output on both systems during the backup.


The network did not miss a beat and generally appears responsive. It has 
been checked.  The client and server RAM usage are tracked in Zabbix and 
not close to running out.  Only curious thing is swap is running out on 
the client (Windows Server 2016) even with 10GB RAM available, but still 
has about 2GB before crash.  Server system logs (kern.log, syslog) show 
no signs of issues.




Try backing up other systems, try backing up a smaller subset (exclude 
some large directories, and then add them back in if you complete a 
backup successfully).


That is a good idea.  I'll try adding incrementally to the data backed up.



Overall, I would advise to upgrade to BPC v4.x, it handles backups o

Re: [BackupPC-users] Backup fails after three days, possibly millions of files

2019-07-16 Thread David Koski

Thank you all for your input.  More information:

rsync version on client: 3.0.8 (Windows)
rsync version on server: 3.1.2 (Debian)
BackupPC version: 3.3.1
$(Config{ClientTimeout} = 604800

I just compared the output of two verbose BackupPC_dump runs and it 
looks like the files are reported to be backed up even though they are 
not.  For example, this appears in logs of both backup runs:


create   644  4616/545  1085243184 /3412.zip

I checked and the file time stamp is year 2018.  The log files are full 
of these.  I checked the real time clock on both systems and they are 
correct.  There are also files that have been backed up that are not in 
the logs.


I suspect there are over ten million files but I don't have a good way 
of telling now.  Oddly, there are about 500,000 files backed according 
to the log captured from BackupPC_dump and almost the same number 
actually backed up and found in pc//0, but they are different 
subsets of files.  I have been tracking memory and swap usage on the 
server and see no issues.


Is this a possible bug in BackupPC 3.3.1?

P.S.: I notice on a new run the files are populating into pc/new folder 
but not merging to pc/0, presumably because of the crash.


Regards,
David Koski
dko...@sutinen.com

On 7/8/19 6:16 PM, Adam Goryachev wrote:

On 9/7/19 10:23 am, David Koski wrote:
I am trying to back up about 24TB of data that has millions of 
files.  It takes a day or to before it starts backing up and then 
stops with an error.  I did a CLI dump and trapped the output and can 
see the error message:


Can't write 32780 bytes to socket
Read EOF: Connection reset by peer
Tried again: got 0 bytes
finish: removing in-process file 
Shares/Archives//COR_2630.png

Child is aborting
Done: 589666 files, 1667429241846 bytes
Got fatal error during xfer (aborted by signal=PIPE)
Backup aborted by user signal
Not saving this as a partial backup since it has fewer files than the 
prior one (got 589666 and 589666 files versus 4225016)

dump failed: aborted by signal=PIPE

This backup is doing rsync over ssh.  I enabled SSH keepalive but it 
does not appear to be due to an idle network.  It does not appear to 
be a random network interruption because the time it takes to fail is 
pretty consistent, about three days.  I'm stumped. 



Did you check:

$Conf{ClientTimeout} = 72000;

Also, what version of rsync on the client, what version of BackupPC on 
the server, etc?


I think BPC v4 handles this scenario significantly better, in fact a 
server I used to have trouble with on BPC3.x all the time has since 
been combined with 4 other server (so 4 x the number of files and 
total size of data) and BPC4 handles it easily.


Regards,
Adam






___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
List:https://lists.sourceforge.net/lists/listinfo/backuppc-users
Wiki:http://backuppc.wiki.sourceforge.net
Project: http://backuppc.sourceforge.net/


[BackupPC-users] Backup fails after three days, possibly millions of files

2019-07-08 Thread David Koski
I am trying to back up about 24TB of data that has millions of files.  
It takes a day or to before it starts backing up and then stops with an 
error.  I did a CLI dump and trapped the output and can see the error 
message:


Can't write 32780 bytes to socket
Read EOF: Connection reset by peer
Tried again: got 0 bytes
finish: removing in-process file Shares/Archives//COR_2630.png
Child is aborting
Done: 589666 files, 1667429241846 bytes
Got fatal error during xfer (aborted by signal=PIPE)
Backup aborted by user signal
Not saving this as a partial backup since it has fewer files than the 
prior one (got 589666 and 589666 files versus 4225016)

dump failed: aborted by signal=PIPE

This backup is doing rsync over ssh.  I enabled SSH keepalive but it 
does not appear to be due to an idle network.  It does not appear to be 
a random network interruption because the time it takes to fail is 
pretty consistent, about three days.  I'm stumped.


David Koski



___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
List:https://lists.sourceforge.net/lists/listinfo/backuppc-users
Wiki:http://backuppc.wiki.sourceforge.net
Project: http://backuppc.sourceforge.net/


[BackupPC-users] backup hangs, CPU usage high

2008-09-03 Thread David Koski
Lately, one machine has started to hang on backup.  It was a Linux machine, 
now it is a Windows 2003 Server machine that hangs.  The state of the machine 
stays at backup in progress but there is no network traffic and the CPU 
usage is at 100 percent, with two BackupPC_dump processes, one using slightly 
more CPU than the other but with a total of 100 percent.  I have changed the 
RsyncArgs and RsyncRestoreArgs options for --devices to -D.

Backuppc Version: 2.1.2-6
Backuppc host OS: Debian 4.0 (Etch)
kernel: 2.6.12.6-xen
All packages up to date.

Today I increased XferLogLevel to 6 and ran the dump manually and got the 
following error message in a repeating loop:

  create 0 /   0
attribSet(dir=fswsoft, file=)
attribSet(dir=fswsoft, file=)
makeSpecial(/var/lib/backuppc/pc/fp2003/new//fswsoft/, 9, )
Can't open /var/lib/backuppc/pc/fp2003/new//fswsoft/ for empty output\n

Oddly, the directory /var/lib/backuppc/pc/fp2003/new did not exist but I was 
able to create it.  I do have free inodes and I forced fsck.ext3 on the 
filesystems and they were clean.

Any suggestions appreciated.

Regards,
David Koski
[EMAIL PROTECTED]

-
This SF.Net email is sponsored by the Moblin Your Move Developer's challenge
Build the coolest Linux based applications with Moblin SDK  win great prizes
Grand prize is a trip for two to an Open Source event anywhere in the world
http://moblin-contest.org/redirect.php?banner_id=100url=/
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
List:https://lists.sourceforge.net/lists/listinfo/backuppc-users
Wiki:http://backuppc.wiki.sourceforge.net
Project: http://backuppc.sourceforge.net/


Re: [BackupPC-users] backup keeps failing in exactly the same place

2008-04-21 Thread David Koski
On Sunday 20 April 2008 14:39:33 Benjamin Staffin wrote:
 I've run into a frustrating problem that affects only one of my backup
 clients.  This particular client is configured as though it is two
 hosts to backuppc, such that part of its contents go to a compressed
 pool and the rest goes to an uncompressed pool.  Both of the backups
 have started failing for reasons that are compeltely eluding me.
 Here's a slightly-trimmed backup log from one of the failures: (kindly
 ignore the fact that I'm backing up lame anime videos here...)

snip

 exiting after signal ALRM

Maybe it is timing out because the file is too big.  Try increasing the 
timeout value in config.pl (the name escapes me) but I had to do that when I 
had a similar problem.

Regards,
David Koski
[EMAIL PROTECTED]

-
This SF.net email is sponsored by the 2008 JavaOne(SM) Conference 
Don't miss this year's exciting event. There's still time to save $100. 
Use priority code J8TL2D2. 
http://ad.doubleclick.net/clk;198757673;13503038;p?http://java.sun.com/javaone
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
List:https://lists.sourceforge.net/lists/listinfo/backuppc-users
Wiki:http://backuppc.wiki.sourceforge.net
Project: http://backuppc.sourceforge.net/


Re: [BackupPC-users] files already in pool are downloaded

2007-09-11 Thread David Koski
I should have mentioned that I do use rsync and have since discovered 
some pool files do not appear to be downloaded during backup.

Thanks,
David

On Monday 10 September 2007 09:00, Rob Owens wrote:
 My understanding is that with tar and smb, all files are downloaded (and
 then discarded if they're already in the pool).  Rsync is smart enough,
 though, not to download files already in the pool.

 -Rob

 David Koski wrote:
  I have been trying to get a good backup with backuppc (2.1.1) but it has
  been taking days.  I ran a dump on the command line so I could see what
  is going on and I see the files that are in the pool are being
  downloaded.  For example:
 
pool 700   511/1008039 home/daler/My
  Documents/DRAWINGS/Lakeport/Pics/C_03.tif
 
  This is a large file and at 750kb/s takes a while.  Is this expected?  I
  thought if they are in the pool they do not need to be downloaded.
 
  Thanks in advance,
  David Koski
  [EMAIL PROTECTED]
 
  -
  This SF.net email is sponsored by: Microsoft
  Defy all challenges. Microsoft(R) Visual Studio 2005.
  http://clk.atdmt.com/MRT/go/vse012070mrt/direct/01/
  ___
  BackupPC-users mailing list
  BackupPC-users@lists.sourceforge.net
  https://lists.sourceforge.net/lists/listinfo/backuppc-users
  http://backuppc.sourceforge.net/

 -
 This SF.net email is sponsored by: Microsoft
 Defy all challenges. Microsoft(R) Visual Studio 2005.
 http://clk.atdmt.com/MRT/go/vse012070mrt/direct/01/
 ___
 BackupPC-users mailing list
 BackupPC-users@lists.sourceforge.net
 https://lists.sourceforge.net/lists/listinfo/backuppc-users
 http://backuppc.sourceforge.net/

-
This SF.net email is sponsored by: Microsoft
Defy all challenges. Microsoft(R) Visual Studio 2005.
http://clk.atdmt.com/MRT/go/vse012070mrt/direct/01/
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/backuppc-users
http://backuppc.sourceforge.net/


Re: [BackupPC-users] files already in pool are downloaded, Can't link..

2007-09-11 Thread David Koski
Another wrinkle: Many of these same pool files get an error:

Can't link /var/lib/backuppc/pc/bki/new/f%2f/fhome/path-and-file 
to /var/lib/backuppc/cpool/d/1/9/d19f21440531ec9046070a9ad79190c5

Yet, the pool file does not appear to have many links:

-rw-r-  9 backuppc backuppc 38 2007-01-12 
19:25 /var/lib/backuppc/cpool/d/1/9/d19f21440531ec9046070a9ad79190c5

Regards,
David

On Monday 10 September 2007 09:00, Rob Owens wrote:
 My understanding is that with tar and smb, all files are downloaded (and
 then discarded if they're already in the pool).  Rsync is smart enough,
 though, not to download files already in the pool.

 -Rob

 David Koski wrote:
  I have been trying to get a good backup with backuppc (2.1.1) but it has
  been taking days.  I ran a dump on the command line so I could see what
  is going on and I see the files that are in the pool are being
  downloaded.  For example:
 
pool 700   511/1008039 home/daler/My
  Documents/DRAWINGS/Lakeport/Pics/C_03.tif
 
  This is a large file and at 750kb/s takes a while.  Is this expected?  I
  thought if they are in the pool they do not need to be downloaded.
 
  Thanks in advance,
  David Koski
  [EMAIL PROTECTED]
 
  -
  This SF.net email is sponsored by: Microsoft
  Defy all challenges. Microsoft(R) Visual Studio 2005.
  http://clk.atdmt.com/MRT/go/vse012070mrt/direct/01/
  ___
  BackupPC-users mailing list
  BackupPC-users@lists.sourceforge.net
  https://lists.sourceforge.net/lists/listinfo/backuppc-users
  http://backuppc.sourceforge.net/

 -
 This SF.net email is sponsored by: Microsoft
 Defy all challenges. Microsoft(R) Visual Studio 2005.
 http://clk.atdmt.com/MRT/go/vse012070mrt/direct/01/
 ___
 BackupPC-users mailing list
 BackupPC-users@lists.sourceforge.net
 https://lists.sourceforge.net/lists/listinfo/backuppc-users
 http://backuppc.sourceforge.net/

-
This SF.net email is sponsored by: Microsoft
Defy all challenges. Microsoft(R) Visual Studio 2005.
http://clk.atdmt.com/MRT/go/vse012070mrt/direct/01/
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/backuppc-users
http://backuppc.sourceforge.net/


[BackupPC-users] files already in pool are downloaded

2007-09-10 Thread David Koski
I have been trying to get a good backup with backuppc (2.1.1) but it has been
taking days.  I ran a dump on the command line so I could see what is going
on and I see the files that are in the pool are being downloaded.  For example:

  pool 700   511/1008039 home/daler/My 
Documents/DRAWINGS/Lakeport/Pics/C_03.tif

This is a large file and at 750kb/s takes a while.  Is this expected?  I 
thought if
they are in the pool they do not need to be downloaded.

Thanks in advance,
David Koski
[EMAIL PROTECTED]

-
This SF.net email is sponsored by: Microsoft
Defy all challenges. Microsoft(R) Visual Studio 2005.
http://clk.atdmt.com/MRT/go/vse012070mrt/direct/01/
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/backuppc-users
http://backuppc.sourceforge.net/


[BackupPC-users] BackupPC_fixupBackupSummary: No info for backup number 207

2007-03-31 Thread David Koski
I installed version 3 to be able to access old archives that had no
entries in file backups. But when I run the recovery tool I get:

Doing host localhost
*** No info for backup number 207
*** No info for backup number 266
*** No info for backup number 322
*** No info for backup number 407
*** No info for backup number 419
*** No info for backup number 416
*** No info for backup number 350
*** No info for backup number 39
*** No info for backup number 418
*** No info for backup number 417
*** No info for backup number 393
*** No info for backup number 379
*** No info for backup number 414
*** No info for backup number 151
*** No info for backup number 415
No changes for host localhost

Log files exist:

# ls -l LOG*
-rw-r- 1 backuppc backuppc 2599 2006-08-13 01:06 LOG
-rw-r- 1 backuppc backuppc  799 2006-07-30 01:06 LOG.0.z
-rw-r- 1 backuppc backuppc  994 2005-08-31 01:03 LOG.10.z
-rw-r- 1 backuppc backuppc 1013 2005-07-31 01:04 LOG.11.z
-rw-r- 1 backuppc backuppc 1145 2006-05-31 01:10 LOG.1.z
-rw-r- 1 backuppc backuppc 1008 2006-04-30 01:07 LOG.2.z
-rw-r- 1 backuppc backuppc 1094 2006-03-31 01:04 LOG.3.z
-rw-r- 1 backuppc backuppc  989 2006-02-28 01:05 LOG.4.z
-rw-r- 1 backuppc backuppc 1121 2006-01-31 01:20 LOG.5.z
-rw-r- 1 backuppc backuppc 1022 2005-12-31 01:13 LOG.6.z
-rw-r- 1 backuppc backuppc  994 2005-11-30 01:09 LOG.7.z
-rw-r- 1 backuppc backuppc 1030 2005-10-31 01:06 LOG.8.z
-rw-r- 1 backuppc backuppc 1170 2005-09-30 01:08 LOG.9.z

The host localhost has the following backup summary:

0   fullyes 0   6/21 02:00  1.5 283.4   
/var/lib/backuppc/pc/localhost/0 
6   fullyes 0   8/5 01:08   2.0 238.4   
/var/lib/backuppc/pc/localhost/6 
8   incrno  1   8/7 01:06   0.3 236.4   
/var/lib/backuppc/pc/localhost/8 
9   incrno  1   8/8 01:08   0.2 235.4   
/var/lib/backuppc/pc/localhost/9 
10  incrno  1   8/9 01:11   0.4 234.4   
/var/lib/backuppc/pc/localhost/10 
11  incrno  1   8/10 01:07  0.3 233.4   
/var/lib/backuppc/pc/localhost/11 
12  incrno  1   8/11 01:08  0.3 232.4   
/var/lib/backuppc/pc/localhost/12 
13  fullyes 0   8/12 01:07  0.9 231.4   
/var/lib/backuppc/pc/localhost/13 
14  incrno  1   8/13 01:06  0.3 230.4   
/var/lib/backuppc/pc/localhost/14

Regards,
David Koski
david [EMAIL PROTECTED]



-
Take Surveys. Earn Cash. Influence the Future of IT
Join SourceForge.net's Techsay panel and you'll get the chance to share your
opinions on IT  business topics through brief surveys-and earn cash
http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDEV
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/backuppc-users
http://backuppc.sourceforge.net/


Re: [BackupPC-users] recovering status.pl

2007-01-11 Thread David Koski
On Thursday 11 January 2007 09:26, you wrote:
 David writes:
   It's ok to delete status.pl.  Over time the status will get refreshed.
  
   It sounds like you deleted the archives file from the pc/HOST
   directories.  There isn't a way to restore those.  But they
   aren't used by BackupPC - they are only there for logging
   purposes.
 
  As it turns out, I had to copy pc/HOST/backups.backup to pc/HOST/backups
  without the first two lines that corresponded to the two archives that
  I had deleted. Until I did that I could see no archives in the backuppc
  application. Of course I had to delete the empty status.pl file too.

 Sorry, I misunderstood your terminology - I thought you were referring to
 BackupPC's archives (not the backups).  In BackupPC 3.0.0 you can recover
 a missing or corrupted backups file using the BackupPC_fixupBackupSummary
 script.

 That script uses meta data that is saved in each backup's directory.
 That meta data is stored automatically with each new backup, and
 pre-3.0.0 backups have the meta data added on each Saturday night
 if you are already running 3.0.0beta.

 If you have just upgraded to 3.0.0 and run BackupPC_fixupBackupSummary,
 you will need to use the -l option which tries to get the necessary
 date by parsing the log files.

Thank you Craig. I'll know next time.

David Koski
[EMAIL PROTECTED]



-
Take Surveys. Earn Cash. Influence the Future of IT
Join SourceForge.net's Techsay panel and you'll get the chance to share your
opinions on IT  business topics through brief surveys - and earn cash
http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDEV
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/backuppc-users
http://backuppc.sourceforge.net/


[BackupPC-users] recovering status.pl

2007-01-10 Thread David Koski
I made the mistake of deleting some archives manually and now my status.pl
file is zero size and backuppc will not start unless I delete status.pl
but none of my archives are visible in backuppc. How can I recover from
this?

Thank you,
David Koski
[EMAIL PROTECTED]



-
Take Surveys. Earn Cash. Influence the Future of IT
Join SourceForge.net's Techsay panel and you'll get the chance to share your
opinions on IT  business topics through brief surveys - and earn cash
http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDEV
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/backuppc-users
http://backuppc.sourceforge.net/


[BackupPC-users] 555 errors: Unable to read 65536 bytes from

2006-12-08 Thread David Koski
Two days in  a row I got 555 lines of error that are all the same for
both days:

Unable to read 65536 bytes from /home/backuppc/pc/mail/new//f%2f/RStmp got=0, 
seekPosn=402653184 (0,4,24581,401604608,402727821)

The curious thing is that the size is exactly 64k. And every line
of error is the same for two days. I looked at the system logs
and there were no errors.  Where is this error comming from? Is
it a media error? Is it on the client? The server? Is it a bug
in backuppc?

Thank you,
David Koski




-
Take Surveys. Earn Cash. Influence the Future of IT
Join SourceForge.net's Techsay panel and you'll get the chance to share your
opinions on IT  business topics through brief surveys - and earn cash
http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDEV
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/backuppc-users
http://backuppc.sourceforge.net/


Re: [BackupPC-users] Unable to read 2757426 bytes from /home/backuppc/pc/mail/new

2006-09-18 Thread David Koski
On Wednesday 06 September 2006 10:19, Les Mikesell wrote:
 On Wed, 2006-09-06 at 10:01 -0700, David Koski wrote:
  Last nights backup looks about the same with same errors. There are about
  64 of these:
 
  Unable to read 2757426 bytes from /home/backuppc/pc/mail/new//f%2f/RStmp
  got=2287922, seekPosn=78452430 (4608,339,10001,80740352,81347343)

The errors have returned. I am getting about 20 exactly like this:

Unable to read 7389184 bytes from /home/backuppc/pc/mail/new//f%2f/RStmp 
got=6832128, seekPosn=364363776 (6656,451,22691,371195904,371762304)

..and about 80 with other numeric parameters. Any idea what might cause this?

 I still don't understand those, but maybe they are a symptom of:
  There are about 225 of these:
 
  Remote[2]: file has vanished: /var/CommuniGate/Queue/27732874.msg

 These are normal when backing up active file systems.  It just means
 that the file went away between reading the directory to get the
 name and opening the file.   I'm not sure what backuppc does with
 growing/shrinking files - tar takes only the length known when
 the header is written and pads with nulls if the file has shrunk.

I have excluded the Queue directory to eliminate these.

David Koski
[EMAIL PROTECTED]

-
Using Tomcat but need to do more? Need to support web services, security?
Get stuff done quickly with pre-integrated technology to make your job easier
Download IBM WebSphere Application Server v.1.0.1 based on Apache Geronimo
http://sel.as-us.falkag.net/sel?cmd=lnkkid=120709bid=263057dat=121642
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/backuppc-users
http://backuppc.sourceforge.net/


Re: [BackupPC-users] Unable to read 2757426 bytes from /home/backuppc/pc/mail/new

2006-09-18 Thread David Koski
On Monday 18 September 2006 09:18, Les Mikesell wrote:
 On Mon, 2006-09-18 at 09:10 -0700, David Koski wrote:
   On Wed, 2006-09-06 at 10:01 -0700, David Koski wrote:
Last nights backup looks about the same with same errors. There are
about 64 of these:
   
Unable to read 2757426 bytes from
/home/backuppc/pc/mail/new//f%2f/RStmp got=2287922, seekPosn=78452430
(4608,339,10001,80740352,81347343)
 
  The errors have returned. I am getting about 20 exactly like this:
 
  Unable to read 7389184 bytes from /home/backuppc/pc/mail/new//f%2f/RStmp
  got=6832128, seekPosn=364363776 (6656,451,22691,371195904,371762304)
 
  ..and about 80 with other numeric parameters. Any idea what might cause
  this?
 
   I still don't understand those, but maybe they are a symptom of:
There are about 225 of these:
   
Remote[2]: file has vanished: /var/CommuniGate/Queue/27732874.msg
  
   These are normal when backing up active file systems.  It just means
   that the file went away between reading the directory to get the
   name and opening the file.   I'm not sure what backuppc does with
   growing/shrinking files - tar takes only the length known when
   the header is written and pads with nulls if the file has shrunk.
 
  I have excluded the Queue directory to eliminate these.

Just to clarify, I am not getting file has vanished errors since I excluded 
the Queue directory from backup.

 Same problem, different place. The files are disappearing between
 the time the directory is read to find the name and opening the
 file for copying.  That looks like a maildir directory where
 frequent changes would be expected.

The Unable to read problem indicates the same seekPosn and other 
parameters every day for the past several days. Do you still think it is a 
transient file issue?

Regards,
David

-
Using Tomcat but need to do more? Need to support web services, security?
Get stuff done quickly with pre-integrated technology to make your job easier
Download IBM WebSphere Application Server v.1.0.1 based on Apache Geronimo
http://sel.as-us.falkag.net/sel?cmd=lnkkid=120709bid=263057dat=121642
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/backuppc-users
http://backuppc.sourceforge.net/


Re: [BackupPC-users] Unable to read 2757426 bytes from /home/backuppc/pc/mail/new

2006-09-18 Thread David Koski
On Monday 18 September 2006 09:48, Les Mikesell wrote:
 On Mon, 2006-09-18 at 09:27 -0700, David Koski wrote:
  Just to clarify, I am not getting file has vanished errors since I
  excluded the Queue directory from backup.
 
   Same problem, different place. The files are disappearing between
   the time the directory is read to find the name and opening the
   file for copying.  That looks like a maildir directory where
   frequent changes would be expected.
 
  The Unable to read problem indicates the same seekPosn and other
  parameters every day for the past several days. Do you still think it is
  a transient file issue?

 That's 'not supposed' to happen on filesystems with unix semantics.
 That is, if a process gets a file open, it is supposed to continue
 to have access to it's data even if some other process removes
 or renames it.  It could happen if another process truncates
 it - which doesn't seem likely for programs using maildir format.
 It also might happen over an NFS mount, although NFS goes through
 some contortions to avoid it.  In any case, it is somehow related
 to filesystem activity and probably doesn't hurt much as long
 as you can restore the other files.

But this error occurs in the same position day after day, increasing in number 
of times it is found in the log.

Now for a dumb question: Does that error indicate a problem on the backuppc 
server or the server being backed up? I have been assuming the whole time it 
was the backuppc server.

Thanks,
David

-
Using Tomcat but need to do more? Need to support web services, security?
Get stuff done quickly with pre-integrated technology to make your job easier
Download IBM WebSphere Application Server v.1.0.1 based on Apache Geronimo
http://sel.as-us.falkag.net/sel?cmd=lnkkid=120709bid=263057dat=121642
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/backuppc-users
http://backuppc.sourceforge.net/


Re: [BackupPC-users] Unable to read 2757426 byte s from /home/backuppc/pc/mail/new

2006-09-18 Thread David Koski
On Monday 18 September 2006 11:19, Les Mikesell wrote:
 On Mon, 2006-09-18 at 09:59 -0700, David Koski wrote:
  But this error occurs in the same position day after day, increasing in
  number of times it is found in the log.
 
  Now for a dumb question: Does that error indicate a problem on the
  backuppc server or the server being backed up? I have been assuming the
  whole time it was the backuppc server.

 It's being reported on the backuppc copy.  I was assuming it got that
 way because it encountered the error when reading the remote file
 after starting the copy but I could easily be wrong.

Thank you.  I have removed a drive from the mirror of the backuppc host, 
zeroed the superblock and readded it to get a fresh resynchronization. I'll 
see what happens tonight with the next backup.

Regards,
David Koski
[EMAIL PROTECTED]

-
Take Surveys. Earn Cash. Influence the Future of IT
Join SourceForge.net's Techsay panel and you'll get the chance to share your
opinions on IT  business topics through brief surveys -- and earn cash
http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDEV
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/backuppc-users
http://backuppc.sourceforge.net/


Re: [BackupPC-users] Unable to read 2757426 bytes from /home/backuppc/pc/mail/new

2006-09-07 Thread David Koski
On Wednesday 06 September 2006 10:19, Les Mikesell wrote:
 On Wed, 2006-09-06 at 10:01 -0700, David Koski wrote:
  Last nights backup looks about the same with same errors. There are about
  64 of these:
 
  Unable to read 2757426 bytes from /home/backuppc/pc/mail/new//f%2f/RStmp
  got=2287922, seekPosn=78452430 (4608,339,10001,80740352,81347343)

 I still don't understand those, but maybe they are a symptom of:
  There are about 225 of these:
 
  Remote[2]: file has vanished: /var/CommuniGate/Queue/27732874.msg

 These are normal when backing up active file systems.  It just means
 that the file went away between reading the directory to get the
 name and opening the file.   I'm not sure what backuppc does with
 growing/shrinking files - tar takes only the length known when
 the header is written and pads with nulls if the file has shrunk.

I did a backup excluting the queue directory and the problem went away.

Thank you,
David

-
Using Tomcat but need to do more? Need to support web services, security?
Get stuff done quickly with pre-integrated technology to make your job easier
Download IBM WebSphere Application Server v.1.0.1 based on Apache Geronimo
http://sel.as-us.falkag.net/sel?cmd=lnkkid=120709bid=263057dat=121642
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/backuppc-users
http://backuppc.sourceforge.net/


Re: [BackupPC-users] Unable to read 2757426 bytes from /home/backuppc/pc/mail/new

2006-09-05 Thread David Koski
On Monday 04 September 2006 12:55, Les Mikesell wrote:
 On Mon, 2006-09-04 at 13:25, David Koski wrote:
  I got several of these errors:
 
  Unable to read 2757426 bytes from /home/backuppc/pc/mail/new//f%2f/RStmp
  got=2287922, seekPosn=78452430 (4608,339,10001,80740352,81347343)
 
  I also see this in the kernel log:
 
  Sep  4 01:35:00 backup kernel: EXT3-fs error (device md(9,0)):
  ext3_free_blocks: bit already cleared for block 5871769
 
  What do they mean?
  Sep  4 01:35:00 backup kernel: EXT3-fs error (device md(9,0)):
  ext3_free_blocks: bit already cleared for block 5871769

 Sounds like filesystem corruption to me.  Can you unmount the
 partition and run fsck on it?

I did shutdown with the -F option. It reported no errors but refused to boot. 
I entered maintenance mode and did fsck again (fsck.ext3 -f -vy) with no 
errors reported. I did shutdown with -F again and it rebooted after fsck.

Regards,
David

-
Using Tomcat but need to do more? Need to support web services, security?
Get stuff done quickly with pre-integrated technology to make your job easier
Download IBM WebSphere Application Server v.1.0.1 based on Apache Geronimo
http://sel.as-us.falkag.net/sel?cmd=lnkkid=120709bid=263057dat=121642
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/backuppc-users
http://backuppc.sourceforge.net/


Re: [BackupPC-users] Unable to read 2757426 bytes fr om /home/backuppc/pc/mail/new

2006-09-05 Thread David Koski
On Tuesday 05 September 2006 11:49, Les Mikesell wrote:
 On Tue, 2006-09-05 at 10:26 -0700, David Koski wrote:
I got several of these errors:
   
Unable to read 2757426 bytes from
/home/backuppc/pc/mail/new//f%2f/RStmp got=2287922, seekPosn=78452430
(4608,339,10001,80740352,81347343)
   
I also see this in the kernel log:
   
Sep  4 01:35:00 backup kernel: EXT3-fs error (device md(9,0)):
ext3_free_blocks: bit already cleared for block 5871769
   
What do they mean?
Sep  4 01:35:00 backup kernel: EXT3-fs error (device md(9,0)):
ext3_free_blocks: bit already cleared for block 5871769
  
   Sounds like filesystem corruption to me.  Can you unmount the
   partition and run fsck on it?
 
  I did shutdown with the -F option. It reported no errors but refused to
  boot. I entered maintenance mode and did fsck again (fsck.ext3 -f -vy)
  with no errors reported. I did shutdown with -F again and it rebooted
  after fsck.

 Do you still have trouble reading that file?  What kind of md device
 is this?  On raid1 mirrors, I think you only need to read one of the
 underlying partitions so there is a possibility of the mirrors getting
 out of sync and appearing correct if you read one of them but not
 the other.

I will watch the log but another backup is not scheduled until tonight.

It is RAID1 but /proc/mdstat shows they are in sync.

David

-
Using Tomcat but need to do more? Need to support web services, security?
Get stuff done quickly with pre-integrated technology to make your job easier
Download IBM WebSphere Application Server v.1.0.1 based on Apache Geronimo
http://sel.as-us.falkag.net/sel?cmd=lnkkid=120709bid=263057dat=121642
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/backuppc-users
http://backuppc.sourceforge.net/


[BackupPC-users] Unable to read 2757426 bytes from /home/backuppc/pc/mail/new

2006-09-04 Thread David Koski
I got several of these errors:

Unable to read 2757426 bytes from /home/backuppc/pc/mail/new//f%2f/RStmp 
got=2287922, seekPosn=78452430 (4608,339,10001,80740352,81347343)

I also see this in the kernel log:

Sep  4 01:35:00 backup kernel: EXT3-fs error (device md(9,0)): 
ext3_free_blocks: bit already cleared for block 5871769

What do they mean?
Sep  4 01:35:00 backup kernel: EXT3-fs error (device md(9,0)): 
ext3_free_blocks: bit already cleared for block 5871769

David Koski
[EMAIL PROTECTED]

-
Using Tomcat but need to do more? Need to support web services, security?
Get stuff done quickly with pre-integrated technology to make your job easier
Download IBM WebSphere Application Server v.1.0.1 based on Apache Geronimo
http://sel.as-us.falkag.net/sel?cmd=lnkkid=120709bid=263057dat=121642
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/backuppc-users
http://backuppc.sourceforge.net/


Re: [BackupPC-users] backup recovery

2006-08-20 Thread David Koski
On Saturday 19 August 2006 23:15, Craig Barratt wrote:

snip

 In 3.0.0 there is a script to recreate a missing or damaged
 backups file.  It relies on some additional information saved
 in 3.0.0.  However, it also has a mode (using the -l option)
 that allows it to do the best it can to recover pre-3.0.0
 backups too.

 So, if you are running something prior to 2.1.2 then this
 is a known bug (fixed since 2.1.2), and I recommend upgrading
 to 3.0.0 and running:

 su backuppc
 BackupPC_fixupBackupSummary -l HOST

I just upgraded from 2.1.1-2sarge2 to 2.1.2-5. I will wait for 3.0 to become 
available for Debian.

Thank you much.

David Koski
[EMAIL PROTECTED]

-
Using Tomcat but need to do more? Need to support web services, security?
Get stuff done quickly with pre-integrated technology to make your job easier
Download IBM WebSphere Application Server v.1.0.1 based on Apache Geronimo
http://sel.as-us.falkag.net/sel?cmd=lnkkid=120709bid=263057dat=121642
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/backuppc-users
http://backuppc.sourceforge.net/


[BackupPC-users] backup recovery

2006-08-13 Thread David Koski
I made some changes to config.pl to reduce the number of backups:

$Conf{FullKeepCnt} = [2,0,0,2,0,0,2];

..whereas before I had [2,2,2,2,2,2,2];

Now I need to recover files from a year ago and they are actually still there 
but do not show up in the GUI. The GUI only goes back to 7/29 but for some 
reason the files still exist on the backuppc server. When I try to execute:

/usr/share/backuppc/bin/BackupPC_tarCreate -t -n 126 -h 
tiikeri /home/users/dkoski/src

(all on one line) I get the error bad backup number 126 for host tiikeri. 
How do I restore the files?

Thank you,
David Koski
[EMAIL PROTECTED]

-
Using Tomcat but need to do more? Need to support web services, security?
Get stuff done quickly with pre-integrated technology to make your job easier
Download IBM WebSphere Application Server v.1.0.1 based on Apache Geronimo
http://sel.as-us.falkag.net/sel?cmd=lnkkid=120709bid=263057dat=121642
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/backuppc-users
http://backuppc.sourceforge.net/


Re: [BackupPC-users] backup recovery [UPDATE]

2006-08-13 Thread David Koski
On Sunday 13 August 2006 20:44, David Koski wrote:
 I made some changes to config.pl to reduce the number of backups:

 $Conf{FullKeepCnt} = [2,0,0,2,0,0,2];

 ..whereas before I had [2,2,2,2,2,2,2];

 Now I need to recover files from a year ago and they are actually still
 there but do not show up in the GUI. The GUI only goes back to 7/29 but for
 some reason the files still exist on the backuppc server. When I try to
 execute:

 /usr/share/backuppc/bin/BackupPC_tarCreate -t -n 126 -h
 tiikeri /home/users/dkoski/src

 (all on one line) I get the error bad backup number 126 for host tiikeri.
 How do I restore the files?

I also reduced $Conf{FullAgeMax} to 30 and noticed the RestoreInfo.xx files 
in /var/lib/backuppc from a year ago are lost. Is it still possible to 
restore? Is it still possible to get the archive from a year ago to be 
visible in the user interface?

Regards,
David Koski
[EMAIL PROTECTED]

-
Using Tomcat but need to do more? Need to support web services, security?
Get stuff done quickly with pre-integrated technology to make your job easier
Download IBM WebSphere Application Server v.1.0.1 based on Apache Geronimo
http://sel.as-us.falkag.net/sel?cmd=lnkkid=120709bid=263057dat=121642
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/backuppc-users
http://backuppc.sourceforge.net/


[BackupPC-users] Strange error messages in log

2005-11-27 Thread David Koski
I am getting the following error message in the error log:

snip log
Unable to read 621621 bytes from /home/backuppc/pc/mail/new//f%2f/RStmp
got=208181, seekPosn=23909067 (1536,253,10003,24117248,24576026)Unable to read 
621621 bytes from /home/backuppc/pc/mail/new//f%2f/RStmp
snip log

The second line is repeated about 125 times. I thought it might be a disk read
error so I did dd if=/dev/hda of=/dev/null bs=1M and the same for hdc. (It
is using software raid with /dev/hda and dev/hdc). It produced no errors and
the system logs reveal nothing. Any clues as to what it might be?

Thank you,
David Koski
[EMAIL PROTECTED]


---
This SF.net email is sponsored by: Splunk Inc. Do you grep through log files
for problems?  Stop!  Download the new AJAX search engine that makes
searching your log files as easy as surfing the  web.  DOWNLOAD SPLUNK!
http://ads.osdn.com/?ad_id=7637alloc_id=16865op=click
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/backuppc-users
http://backuppc.sourceforge.net/


Re: [BackupPC-users] rsync? tar? rsyncd?

2005-10-29 Thread David Koski
On Saturday 29 October 2005 04:11 pm, Rich Duzenbury wrote:
 The boxes are
 currently being backed up via ssh/rsync, but I wonder if ssh encryption
 is limiting the transfer rates.

You might want to set ssh to use the blowfish cipher if you are on a secure 
network to see if you pick up some speed.

David


---
This SF.Net email is sponsored by the JBoss Inc.
Get Certified Today * Register for a JBoss Training Course
Free Certification Exam for All Training Attendees Through End of 2005
Visit http://www.jboss.com/services/certification for more information
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/backuppc-users
http://backuppc.sourceforge.net/


[BackupPC-users] cwrsync: unable to bind any inbound sockets on port 873

2005-10-25 Thread David Koski
When I start the service it just goes back down in a few seconds (maybe 5). 
Any clues?

OS: XP-Pro

Regards,
David Koski
[EMAIL PROTECTED]


---
This SF.Net email is sponsored by the JBoss Inc.
Get Certified Today * Register for a JBoss Training Course
Free Certification Exam for All Training Attendees Through End of 2005
Visit http://www.jboss.com/services/certification for more information
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/backuppc-users
http://backuppc.sourceforge.net/