list,

I am re-iterating my problem, due to no response.  As a side note the
storage server and director are running on the same server.  This happens
on a copy job from one disk to another disk on the same system.
It is a definite problem as I am loosing backup data!

bacula 9.0.3
mariadb 10.2.8
centos 6.9

I upgraded bacula from 5.2.13 which worked very well, to version 9.0.3.
Basically installed new version from source and then upgraded the database
structure.  The source was compiled with the following:

./configure --sbindir=/usr/bacula/bin --sysconfdir=/usr/bacula/bin
--with-pid-dir=/var/run/bacula --with-subsys-dir=/var/run/bacula/working
--enable-smartalloc --with-mysql --with-working-dir=/usr/bacula/bin/working
--with-dump-email=u...@domain.com --with-job-email=u...@domain.com
--with-smtp-host=smtp.googlemail.com --enable-bat

The problem started with my offsite backups. I will get the following error:

13-Nov 01:18 distress JobId 33429: Fatal error: Socket error on Data
received command: ERR=No data available
13-Nov 01:18 distress JobId 33429: Fatal error: fd_cmds.c:157 Read
data not accepted

And the backup fails. Most of the time it is on a backup that spans
multiple disks.  So, I chatted with the ATTO raid support folks and they
suggested that I use a different hotswap raid enclosure due to the one I
was using was not very reliable in their opinion.  Although this enclosure
had worked very reliably for well into 10 years without a problem!  So, I
moved the system to a completely new system ( Supermicro with ATTO raid ).
The problem still persists!  I have rebuilt the raid disk structure and
changed the working of the backups. To no avail!
My backups worked flawlessly before the upgrade!  Once going to v 9 I can
not count how many offsite backups the have failed to complete with this
type of error.  I also get

13-Nov 01:23 distress JobId 33430: Warning: mount.c:210 Open of File
device "BottomSwap" (/BottomSwap) Volume "dcBS-104" failed:
ERR=file_dev.c:190 Could not
open(/BottomSwap/dcBS-104,OPEN_READ_WRITE,0640): ERR=No such file or
directory
and
15-Nov 17:20 kilchis JobId 35825: Error: bsock.c:849 Read error from
Storage daemon:kilchis:9103: ERR=Connection reset by peer
15-Nov 17:20 kilchis JobId 35825: Fatal error: append.c:271 Network
error reading from FD. ERR=Connection reset by peer

All of this happens on one storage server, well actually two storage
servers but they

service two different subnets/domains. It all started with the upgrade!

please, tell me that you have fixed this with a new version!

thanks


On Mon, Nov 27, 2017 at 7:45 AM, Jerry Lowry <michaiah2...@gmail.com> wrote:

> list,
>
> bacula 9.0.3
> mariadb 10.2.8
> centos 6.9
>
> I upgraded bacula from 5.2.13 which worked very well, to version 9.0.3.
> Basically installed new version from source and then upgraded the database
> structure.  The source was compiled with the following:
>
> ./configure --sbindir=/usr/bacula/bin --sysconfdir=/usr/bacula/bin
> --with-pid-dir=/var/run/bacula --with-subsys-dir=/var/run/bacula/working
> --enable-smartalloc --with-mysql --with-working-dir=/usr/bacula/bin/working
> --with-dump-email=u...@domain.com --with-job-email=u...@domain.com
> --with-smtp-host=smtp.googlemail.com --enable-bat
>
> The problem started with my offsite backups. I will get the following
> error:
>
> 13-Nov 01:18 distress JobId 33429: Fatal error: Socket error on Data received 
> command: ERR=No data available
> 13-Nov 01:18 distress JobId 33429: Fatal error: fd_cmds.c:157 Read data not 
> accepted
>
> And the backup fails. Most of the time it is on a backup that spans
> multiple disks.  So, I chatted with the ATTO raid support folks and they
> suggested that I use a different hotswap raid enclosure due to the one I
> was using was not very reliable in their opinion.  Although this enclosure
> had worked very reliably for well into 10 years without a problem!  So, I
> moved the system to a completely new system ( Supermicro with ATTO raid ).
> The problem still persists!  I have rebuilt the raid disk structure and
> changed the working of the backups. To no avail!
> My backups worked flawlessly before the upgrade!  Once going to v 9 I can
> not count how many offsite backups the have failed to complete with this
> type of error.  I also get
>
> 13-Nov 01:23 distress JobId 33430: Warning: mount.c:210 Open of File device 
> "BottomSwap" (/BottomSwap) Volume "dcBS-104" failed: ERR=file_dev.c:190 Could 
> not open(/BottomSwap/dcBS-104,OPEN_READ_WRITE,0640): ERR=No such file or 
> directory
> and
> 15-Nov 17:20 kilchis JobId 35825: Error: bsock.c:849 Read error from Storage 
> daemon:kilchis:9103: ERR=Connection reset by peer
> 15-Nov 17:20 kilchis JobId 35825: Fatal error: append.c:271 Network error 
> reading from FD. ERR=Connection reset by peer
>
> All of this happens on one storage server, well actually two storage servers 
> but they
>
> service two different subnets/domains. It all started with the upgrade!
>
> please, tell me that you have fixed this with a new version!
>
> thanks
>
>
------------------------------------------------------------------------------
Check out the vibrant tech community on one of the world's most
engaging tech sites, Slashdot.org! http://sdm.link/slashdot
_______________________________________________
Bacula-users mailing list
Bacula-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/bacula-users

Reply via email to