Re: [Bacula-users] Backup from client with deduplication

2020-10-06 Thread Radosław Korzeniewski
Hello,

śr., 7 paź 2020 o 06:37 Рукавцов Дмитрий Геннадьевич 
napisał(a):

> So, there is no solution? Bacula can't backup such clients?
>
I know nothing about windows filesystem deduplication capabilities, but I
assume to properly read contents of these files a standard win32
BackupRead() dedicated to backup function is not sufficient and requires
special handling.
I do not know if this special handling is available in Bacula, but as this
is not my area of interests in Bacula my answer could be incomplete.

best regards
-- 
Radosław Korzeniewski
rados...@korzeniewski.net
___
Bacula-users mailing list
Bacula-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/bacula-users


Re: [Bacula-users] Discrepancy between backup/restore size and files

2020-10-06 Thread Radosław Korzeniewski
Hello,

śr., 30 wrz 2020 o 12:33 Ben Laurie  napisał(a):

>
>
> On Wed, 30 Sep 2020 at 09:42, Radosław Korzeniewski <
> rados...@korzeniewski.net> wrote:
>
>> Hello,
>>
>> pon., 28 wrz 2020 o 14:34 Ben Laurie  napisał(a):
>>
>>>
>>>
>>> On Sun, 13 Sep 2020 at 18:58, Josip Deanovic 
>>> wrote:
>>>
 On Sunday 2020-09-13 15:43:31 Ben Laurie wrote:
 > > Not sure if IgnoreCase option could caused it somehow.
 >
 > According to the documentation it just makes it ignore case for
 pattern
 > matching, so seems unlikely?

 I can't say because I didn't use Bacula with Windows a lot.
 It worked fine back in 2010-2012.

 > Yes, funny I somehow had not noticed that before - I've added it. I
 > presume it won't fix this problem, tho.

 No, but it will make the difference for future backups.

>>>
>>> I tried using accurate but it just caused errors (I forget what now, but
>>> some problem with database searches?).
>>>
>>> In other news, I did a full restore as a test recently and I got quite a
>>> few errors along the lines of:
>>>
>>> 25-Sep 14:46 xxx-fd JobId 35: Error: File /tmp/bacula-restores/k/xxx
>>> already exists and could not be replaced. ERR=Permission denied.
>>>
>>
>> First of all to properly restore Windows files on any Linux or Unix
>> machine you have to set a Portable flag for your backup.
>>
>
> According to
> https://www.bacula.org/5.1.x-manuals/fr/main/main/Windows_Version_Bacula.html
> that is no longer true...
>
> "Note: with Bacula versions 1.39.x and later, non-portable Windows data
> can be restore to any machine."
>

Not exactly. You can restore some data because non windows file daemon
accept windows stream. And this is what the above statement means. But this
stream is what win32 BackupRead() generates. It could be the same bit2bit
what standard read() can generate but there is no guarantee. The Portable
flag replaces win32 BackupRead() for a simple read().

best regards
-- 
Radosław Korzeniewski
rados...@korzeniewski.net
___
Bacula-users mailing list
Bacula-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/bacula-users


Re: [Bacula-users] Backup from client with deduplication

2020-10-06 Thread Рукавцов Дмитрий Геннадьевич

So, there is no solution? Bacula can't backup such clients?

Maybe i explained something wrong. I have bacula server on centos 
(server), i don't have any deduplication on the bacula server and i 
don't need it there. I have windows 2019 server (client), it has it's 
own deduplication on it's own volume which is enabled in windows 
settings. From this windows volume i need to backup files.


Full backup size before i turn on deduplication (which is windows 
deduplication) on windows server (client) was like 200GB, full backup 
after i turn on deduplication with same files on the same server was 
6.7GB, files count same, when i restoring it, most files is 
unreadable. I'm thinking cause bacula makes snapshot of the files when 
making backup and it's not working with deduplication on windows 
server (client), but maybe i'm wrong.


I just want normal (recoverable) backup from this windows server 
(which is 200gb), i don't need any deduplication on my backups.



This is FileSet

FileSet {
    Name = "winsrv"
    Enable VSS = no
    Include {
    Options {

        compression = GZIP
            signature = MD5
    aclsupport = yes
   }
        File = D:/FS/
                  }
}

this is log from backup:

Sep 29 19:20:36 bacula bacula-dir:  Build OS: 
x86_64-redhat-linux-gnu-bacula redhat (Core)

Sep 29 19:20:36 bacula bacula-dir:  JobId:  3814
Sep 29 19:20:36 bacula bacula-dir:  Job: gc-main.2020-09-29_10.58.43_03
Sep 29 19:20:36 bacula bacula-dir:  Backup Level:   Full
Sep 29 19:20:36 bacula bacula-dir:  Client: "winsrv" 9.6.5 (11Jun20) 
Microsoft Standard Edition (build 9200), 64-bit,Cross-compile,Win64

Sep 29 19:20:36 bacula bacula-dir:  FileSet: "winsrv" 2020-09-15 22:00:00
Sep 29 19:20:36 bacula bacula-dir:  Pool: "main-pool" (From Job resource)
Sep 29 19:20:36 bacula bacula-dir:  Catalog: "default" (From Client 
resource)
Sep 29 19:20:36 bacula bacula-dir:  Storage: "backup-sd" (From Job 
resource)

Sep 29 19:20:36 bacula bacula-dir:  Scheduled time: 29-сен-2020 19:00:00
Sep 29 19:20:36 bacula bacula-dir:  Start time: 29-сен-2020 19:00:02
Sep 29 19:20:36 bacula bacula-dir:  End time: 29-сен-2020 19:20:36
Sep 29 19:20:36 bacula bacula-dir:  Elapsed time:   20 mins 34 
secs

Sep 29 19:20:36 bacula bacula-dir:  Priority:   1
Sep 29 19:20:36 bacula bacula-dir:  FD Files Written: 905,999
Sep 29 19:20:36 bacula bacula-dir:  SD Files Written: 905,999
Sep 29 19:20:36 bacula bacula-dir:  FD Bytes Written: 6,500,269,762 
(6.500 GB)
Sep 29 19:20:36 bacula bacula-dir:  SD Bytes Written: 6,708,284,663 
(6.708 GB)

Sep 29 19:20:36 bacula bacula-dir:  Rate: 5267.6 KB/s
Sep 29 19:20:36 bacula bacula-dir:  Software Compression: 15.7% 1.2:1
Sep 29 19:20:36 bacula bacula-dir:  Comm Line Compression:  0.6% 1.0:1
Sep 29 19:20:36 bacula bacula-dir:  Snapshot/VSS:   no
Sep 29 19:20:36 bacula bacula-dir:  Encryption: no
Sep 29 19:20:36 bacula bacula-dir:  Accurate:   no
Sep 29 19:20:36 bacula bacula-dir:  Volume name(s): All0121
Sep 29 19:20:36 bacula bacula-dir:  Volume Session Id:  1
Sep 29 19:20:36 bacula bacula-dir:  Volume Session Time: 1601366166
Sep 29 19:20:36 bacula bacula-dir:  Last Volume Bytes: 6,743,816,269 
(6.743 GB)

Sep 29 19:20:36 bacula bacula-dir:  Non-fatal FD errors:    2
Sep 29 19:20:36 bacula bacula-dir:  SD Errors:  0
Sep 29 19:20:36 bacula bacula-dir:  FD termination status:  OK
Sep 29 19:20:36 bacula bacula-dir:  SD termination status:  OK
Sep 29 19:20:36 bacula bacula-dir:  Termination: Backup OK -- with 
warnings



30.09.2020 13:22, Radosław Korzeniewski пишет:

Hello,

wt., 29 wrz 2020 o 09:50 Рукавцов Дмитрий Геннадьевич 
mailto:m...@santel-navi.ru>> napisał(a):


Hello, i have bacula server 9.6.5 on Centos and i need to backup
folders
from Windows 2019 server client with deduplication on (optimized
files).
Problem if i'm making Full backup, the archive is incomplete


Could you show the job terminate log?

(files
count match, but size is wrong)


I do not understand how you judge that a size is wrong?

Is there any way to tell Bacula server
to ignore deduplication (optimized files) on client or any other way?


I do not understand your statement. Could you elaborate what you want 
to achieve and what configuration you use?


Deduplication in Bacula Enterprise Edition is called GED (Global 
Endpoint Deduplication) and it is performed on client side (endpoint) 
and on storage daemon. You can enable/disable deduplication for a 
single FileSet/Include resource.
A deduplication optimized volumes functionality (aligned plugin) 
available on Bacula Community requires an external deduplication 
solution, hardware or software level to perform the required 
computation and management. This solution does not involve a backup 
client in any way.



Tried to set "Enable VSS = no" at FileSet but it gives nothing.


VSS is a Microsoft Windo

[Bacula-users] Baculum 9.6.6.1

2020-10-06 Thread Marcin Haba
Hello,

We are pleased to announce a new Baculum 9.6.6.1 version. This is a
minor new feature and bug fix release. It introduces new functions to
the Security page which help creating user accounts with restricted
Bacula resources access. Binary packages for popular distributions are
available already.

New changes and fixes since version 9.6.6:

 - New features and improvements to multi-user interface and restricted access
 - Add default days and limit values to status schedules API endpoint
 - Update Portuguese translations
 - Fix displaying multitextbox and multicombobox directive controls
 - Fix date filter on status schedule page
 - Fix #2570 fileset excludes settings with PHP 7.4

Sources:
 
https://www.bacula.org/downloads/baculum/releases/9.6.6.1/baculum-9.6.6.1.tar.gz
 
https://www.bacula.org/downloads/baculum/releases/9.6.6.1/baculum-9.6.6.1.tar.gz.sig

Useful links:
 Documentation:
https://www.bacula.org/9.6.x-manuals/en/console/Baculum_API_Web_GUI_Tools.html
 New features 9.6.6.1 video: https://www.youtube.com/watch?v=cvKtvZCvjLQ
 Online Demo: https://baculum.app

Thank you to everybody involved in preparing this release.

Best regards,
Marcin Haba (gani)

-- 
"Greater love hath no man than this, that a man lay down his life for
his friends." Jesus Christ

"Większej miłości nikt nie ma nad tę, jak gdy kto życie swoje kładzie
za przyjaciół swoich." Jezus Chrystus


___
Bacula-users mailing list
Bacula-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/bacula-users


Re: [Bacula-users] performance&design&configuration challenges

2020-10-06 Thread Žiga Žvan

Hi,
I have done some testing:
a) testing storage with dd command (eg: dd if=/dev/zero 
of=/storage/test1.img bs=1G count=1 oflag=dsync). The results are:

-writing to IBM storage (with cloud enabled) shows 300 MB/sec
-writing to local SSD storage shows 600 MB/sec.
I guess storage is not a bottleneck.
b) testing file copy from linux centos 6 server to bacula server with 
rsync (eg. rsync --info=progress2 source destination)

-writing to local storage: 82 MB/sec
-writing to IBM storage: 85 MB/sec
I guess this is ok for a 1 GB network link.
c) using bacula:
-linux centos 6 file server: 13 MB/sec on IBM storage, 16 MB/sec on 
local SSD storage (version of client 5.2.13).
-windows file server:  around 18 MB/sec - there could be some additional 
problem, because I perform a backup from a deduplicated drive (version 
of client 9.6.5)
d) I have tried to manipulate encryption/compression settings, but I 
believe there is no significant difference


I think that  bacula rate (15 MB/sec) in quite slow comparing to file 
copy results (85 MB/sec) from the same client/server. It should be 
better... Do you agree?


I have implemented autochanger in order to perform backup from both 
servers at the same time. We shall see the results tomorrow.
I have not changed the version of the client on linux server yet. My 
windows server uses new client version, so that was not my first idea... 
Will try this tomorrow if needed.


What about retention?
I would like to:
- create incremental daily backup (retention 1 week)
- create weekly full backup (retention 1 month)
- create monthly full backup (retention 1 year)

At the moment I use different job/schedule for monthly backup, but that 
triggers full backup also on Monday after monthly backup (I would like 
to run incremental then). Is there a better way? Relevant parts of conf 
below...


Regards,
Ziga

JobDefs {
Name = "bazar2-job"
Schedule = "WeeklyCycle"
...
}

Job {
  Name = "bazar2-backup"
  JobDefs = "bazar2-job"
  Full Backup Pool = bazar2-weekly-pool
  Incremental Backup Pool = bazar2-daily-pool
}

Job {
  Name = "bazar2-monthly-backup"
  Level = Full
  JobDefs = "bazar2-job"
  Pool = bazar2-monthly-pool
  Schedule = "MonthlyFull"  #schedule : see in bacula-dir.conf (monthly 
pool with longer retention)

}





Example output:

06-Oct 12:19 bacula-dir JobId 714: Bacula bacula-dir 9.6.5 (11Jun20):
  Build OS:   x86_64-redhat-linux-gnu-bacula redhat (Core)
  JobId:  714
  Job:bazar2-monthly-backup.2020-10-06_09.33.25_03
  Backup Level:   Full
  Client: "bazar2.kranj.cetrtapot.si-fd" 5.2.13 (19Jan13) 
x86_64-redhat-linux-gnu,redhat,(Core)
  FileSet:"bazar2-fileset" 2020-09-30 15:40:26
  Pool:   "bazar2-monthly-pool" (From Job resource)
  Catalog:"MyCatalog" (From Client resource)
  Storage:"FSTestBackup" (From Job resource)
  Scheduled time: 06-Oct-2020 09:33:15
  Start time: 06-Oct-2020 09:33:28
  End time:   06-Oct-2020 12:19:19
  Elapsed time:   2 hours 45 mins 51 secs
  Priority:   10
  FD Files Written:   53,682
  SD Files Written:   53,682
  FD Bytes Written:   168,149,175,433 (168.1 GB)
  SD Bytes Written:   168,158,044,149 (168.1 GB)
  Rate:   16897.7 KB/s
  Software Compression:   36.6% 1.6:1
  Comm Line Compression:  None
  Snapshot/VSS:   no
  Encryption: no
  Accurate:   no
  Volume name(s): bazar2-monthly-vol-0300
  Volume Session Id:  11
  Volume Session Time:1601893281
  Last Volume Bytes:  337,370,601,852 (337.3 GB)
  Non-fatal FD errors:0
  SD Errors:  0
  FD termination status:  OK
  SD termination status:  OK
  Termination:Backup OK


On 06.10.2020 14:28, Josh Fisher wrote:


On 10/6/20 3:45 AM, Žiga Žvan wrote:
I believe that I have my spooling attributes set correctly on jobdefs 
(see bellow). Spool attributes = yes; Spool data defaults to no. Any 
other idea for performance problems?

Regard,
Ziga



The client version is very old. First try updating the client to 9.6.x.

For testing purposes, create another storage device on local disk and 
write a full backup to that. If it is much faster to local disk 
storage than it is to the s3 driver, then there may be an issue with 
how the s3 driver is compiled, version of s3 driver, etc.


Otherwise, with attribute spooling enabled, the status of the job as 
given by the status dir command in bconsole will change to "despooling 
attributes" or something like that when the client has finished 
sending data. That is the period at the end of the job when the 
spooled attributes are being written to the catalog database. If 
despooling is taking a long time, then database performance might be 
the bottleneck.





___
Bacula-users mailing list
Bacula-users@lists.sourceforge.n

Re: [Bacula-users] performance&design&configuration challenges

2020-10-06 Thread Josh Fisher


On 10/6/20 3:45 AM, Žiga Žvan wrote:
I believe that I have my spooling attributes set correctly on jobdefs 
(see bellow). Spool attributes = yes; Spool data defaults to no. Any 
other idea for performance problems?

Regard,
Ziga



The client version is very old. First try updating the client to 9.6.x.

For testing purposes, create another storage device on local disk and 
write a full backup to that. If it is much faster to local disk storage 
than it is to the s3 driver, then there may be an issue with how the s3 
driver is compiled, version of s3 driver, etc.


Otherwise, with attribute spooling enabled, the status of the job as 
given by the status dir command in bconsole will change to "despooling 
attributes" or something like that when the client has finished sending 
data. That is the period at the end of the job when the spooled 
attributes are being written to the catalog database. If despooling is 
taking a long time, then database performance might be the bottleneck.





___
Bacula-users mailing list
Bacula-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/bacula-users


Re: [Bacula-users] performance&design&configuration challenges

2020-10-06 Thread Heitor Faria
Hello Ziga,

Your client is probably too old for the 9.2.x Director.
Even CentOS 6 is old, most likely in the end of life.
Other than that you can try some tuning: 
http://www.bacula.lat/tuning-better-performance-and-treatment-of-backup-bottlenecks/?lang=en

Rgds.
--
MSc Heitor Faria
CEO Bacula LatAm
mobile1: + 1 909 655-8971
mobile2: + 55 61 98268-4220

América Latina
[ http://bacula.lat/]

 Original Message 
From: Žiga Žvan 
Sent: Tuesday, October 6, 2020 03:11 AM
To: bacula-users@lists.sourceforge.net
Subject: [Bacula-users] performance&design&configuration challenges

>Hi,
>
>I'm having some performance challenges. I would appreciate some educated 
>guess from an experienced bacula user.
>
>I'm changing old backup sw that writes to tape drive with bacula 
>writing  to disk. The results are:
>a) windows file server backup from a deduplicated drive (1.700.000 
>files, 900 GB data, deduplicated space used 600 GB). *Bacula: 12 hours, 
>old software: 2.5 hours*
>b) linux file server backup (50.000 files, 166 GB data).*Bacula 3.5 
>hours, old software: 1 hour*.
>
>I have tried to:
>a) turn off compression&encryption. The result is the same: backup speed 
>around 13 MB/sec.
>b) change destination storage (from a new ibm storage attached over nfs, 
>to a local SSD disk attached on bacula server virtual machine). It took 
>2 hours 50 minutes to backup linux file server (instead of 3.5 hours). 
>Sequential write test tested with linux dd command shows write speed 300 
>MB/sec for IBM storage and 600 MB/sec for local SSD storage (far better 
>than actual throughput).
>
>The network bandwidth is 1 GB (1 GB on client, 10 GB on server) so I 
>guess this is not a problem; however I have noticed that bacula-fd on 
>client side uses 100% of CPU.
>
>I'm using:
>-bacula server version 9.6.5
>-bacula client version 5.2.13 (original from centos 6 repo).
>
>Any idea what is wrong and/or what performance should I expect?
>I would also appreciate some answers on the questions bellow.
>
>Kind regards,
>Ziga Zvan
>
>
>
>
>On 05.08.2020 10:52, Žiga Žvan wrote:
>>
>> Dear all,
>> I have tested bacula sw (9.6.5) and I must say I'm quite happy with 
>> the results (eg. compression, encryption, configureability). However I 
>> have some configuration/design questions I hope, you can help me with.
>>
>> Regarding job schedule, I would like to:
>> - create incremental daily backup (retention 1 week)
>> - create weekly full backup (retention 1 month)
>> - create monthly full backup (retention 1 year)
>>
>> I am using dummy cloud driver that writes to local file storage.  
>> Volume is a directory with fileparts. I would like to have seperate 
>> volumes/pools for each client. I would like to delete the data on disk 
>> after retention period expires. If possible, I would like to delete 
>> just the fileparts with expired backup.
>>
>> Questions:
>> a) At the moment, I'm using two backup job definitions per client and 
>> central schedule definition for all my clients. I have noticed that my 
>> incremental job gets promoted to full after monthly backup ("No prior 
>> Full backup Job record found"; because monthly backup is a seperate 
>> job, but bacula searches for full backups inside the same job). Could 
>> you please suggest a better configuration. If possible, I would like 
>> to keep central schedule definition (If I manipulate pools in a 
>> schedule resource, I would need to define them per client).
>>
>> b) I would like to delete expired backups on disk (and in the catalog 
>> as well). At the moment I'm using one volume in a daily/weekly/monthly 
>> pool per client. In a volume, there are fileparts belonging to expired 
>> backups (eg. part1-23 in the output bellow). I have tried to solve 
>> this with purge/prune scripts in my BackupCatalog job (as suggested in 
>> the whitepapers) but the data does not get deleted. Is there any way 
>> to delete fileparts? Should I create separate volumes after retention 
>> period? Please suggest a better configuration.
>>
>> c) Do I need a restore job for each client? I would just like to 
>> restore backup on the same client, default to /restore folder... When 
>> I use bconsole restore all command, the wizard asks me all the 
>> questions (eg. 5- last backup for a client, which client,fileset...) 
>> but at the end it asks for a restore job which changes all previously 
>> defined things (eg. client).
>>
>> d) At the moment, I have not implemented autochanger functionality. 
>> Clients compress/encrypt the data and send them to bacula server, 
>> which writes them on one central storage system. Jobs are processed in 
>> sequential order (one at a time). Do you expect any significant 
>> performance gain if i implement autochanger in order to have jobs run 
>> simultaneously?
>>
>> Relevant part of configuration attached bellow.
>>
>> Looking forward to move in the production...
>> Kind regards,
>> Ziga Zvan
>>
>>
>> *Volume example *(fileparts 1-23 should be deleted)*:*

Re: [Bacula-users] Bacula 9.6.5 TLS issue - solved in 9.6.6

2020-10-06 Thread djosip+news

On 2020-10-06 02:07, Phil Stracchino wrote:

On 9/28/20 12:33 PM, Phil Stracchino wrote:
test phase 1:  All clients and Storage on 9.6.6, Director still on 
9.6.3

No hung jobs so far.  I plan to leave it this way for at least a week
before upgrading the Director to 9.6.6. as well.


OK, a week of no issues and monthly full backups just ran.  I am now
updating the Director from 9.6.3 to 9.6.6.  No other changes.

Fingers crossed.



I am eager to hear about your conclusions. Good luck.


Regards!

--
Josip Deanovic


___
Bacula-users mailing list
Bacula-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/bacula-users


Re: [Bacula-users] performance&design&configuration challenges

2020-10-06 Thread Žiga Žvan
I believe that I have my spooling attributes set correctly on jobdefs 
(see bellow). Spool attributes = yes; Spool data defaults to no. Any 
other idea for performance problems?

Regard,
Ziga



JobDefs {
  Name = "bazar2-job"
  Type = Backup
  Level = Incremental
  Client = bazar2.kranj.cetrtapot.si-fd #Client names: will be match on 
bacula-fd.conf on client side

  FileSet = "bazar2-fileset"
  Schedule = "WeeklyCycle" #schedule : see in bacula-dir.conf
  Storage = FSTestBackup
#  Storage = FSOciCloudStandard
  Messages = Standard
  Pool = bazar2-daily-pool
  Spool Attributes = yes   # Better for backup to disk
  Max Full Interval = 15 days # Ensure that full backup exist
  Priority = 10
  Write Bootstrap = "/opt/bacula/working/%c.bsr"
}

status dir not showing files transfered:

 JobId  Type Level Files Bytes  Name  Status
==
   714  Back Full  0 0  bazar2-monthly-backup is running


On 06.10.2020 09:14, Žiga Žvan wrote:


Thanks Josh for your reply and sorry for my previous duplicate email.

I will try to disable data spooling and report back the results.

What about manipulating retention? Currently I have different jobs for 
weekly full and monthly full backup (see bellow), but that triggers 
full backup instead of incremental on Monday (because I use different 
job resource). Is there a better way to have monthly backup with 
longer retention?


Kind regards,
Ziga

#For all clients

Schedule {
  Name = "MonthlyFull"
  Run = Full 1st fri at 23:05
}

# This schedule does the catalog. It starts after the WeeklyCycle
Schedule {
  Name = "WeeklyCycleAfterBackup"
  Run = Full sun-sat at 23:10
}

#Job for each client

Job {
  Name = "oradev02-backup"
  JobDefs = "oradev02-job"
  Full Backup Pool = oradev02-weekly-pool
  Incremental Backup Pool = oradev02-daily-pool
}

Job {
  Name = "oradev02-monthly-backup"
  JobDefs = "oradev02-job"
  Pool = oradev02-monthly-pool
  Schedule = "MonthlyFull"  #schedule : see in bacula-dir.conf 
(monthly pool with longer retention)

}



On 05.10.2020 16:30, Josh Fisher wrote:



On 10/5/20 9:20 AM, Žiga Žvan wrote:


Hi,
I'm having some performance challenges. I would appreciate some 
educated guess from an experienced bacula user.


I'm changing old backup sw that writes to tape drive with bacula 
writing  to disk. The results are:
a) windows file server backup from a deduplicated drive (1.700.000 
files, 900 GB data, deduplicated space used 600 GB). *Bacula: 12 
hours, old software: 2.5 hours*
b) linux file server backup (50.000 files, 166 GB data).*Bacula 3.5 
hours, old software: 1 hour*.


I have tried to:
a) turn off compression&encryption. The result is the same: backup 
speed around 13 MB/sec.
b) change destination storage (from a new ibm storage attached over 
nfs, to a local SSD disk attached on bacula server virtual machine). 
It took 2 hours 50 minutes to backup linux file server (instead of 
3.5 hours). Sequential write test tested with linux dd command shows 
write speed 300 MB/sec for IBM storage and 600 MB/sec for local SSD 
storage (far better than actual throughput).




There are directives to enable/disable spooling of both data and the 
attributes (metadata) being written to the catalog database. When 
using disk volumes, you probably want to disable data spooling and 
enable attribute spooling. The attribute spooling will prevent a 
database write after each file backed up and instead do the database 
writes as a batch at the end of the job. Data spooling would rarely 
if ever be needed when writing to dick media.


With attribute spooling enabled, you can make a rough guess as to 
whether DB performance is the problem by judging how long the job is 
in the 'attribute despooling' state, The status dir command in 
bconsole shows the job state.



The network bandwidth is 1 GB (1 GB on client, 10 GB on server) so I 
guess this is not a problem; however I have noticed that bacula-fd 
on client side uses 100% of CPU.


I'm using:
-bacula server version 9.6.5
-bacula client version 5.2.13 (original from centos 6 repo).

Any idea what is wrong and/or what performance should I expect?
I would also appreciate some answers on the questions bellow (I 
think this email went unanswered).


Kind regards,
Ziga Zvan




On 05.08.2020 10:52, Žiga Žvan wrote:


Dear all,
I have tested bacula sw (9.6.5) and I must say I'm quite happy with 
the results (eg. compression, encryption, configureability). 
However I have some configuration/design questions I hope, you can 
help me with.


Regarding job schedule, I would like to:
- create incremental daily backup (retention 1 week)
- create weekly full backup (retention 1 month)
- create monthly full backup (retention 1 year)

I am using dummy cloud driver that writes to local file storage.  
Volume is a directory with fileparts. I would like to have seperate 
volumes/pools for each client. I wou

[Bacula-users] performance&design&configuration challenges

2020-10-06 Thread Žiga Žvan

Hi,

I'm having some performance challenges. I would appreciate some educated 
guess from an experienced bacula user.


I'm changing old backup sw that writes to tape drive with bacula 
writing  to disk. The results are:
a) windows file server backup from a deduplicated drive (1.700.000 
files, 900 GB data, deduplicated space used 600 GB). *Bacula: 12 hours, 
old software: 2.5 hours*
b) linux file server backup (50.000 files, 166 GB data).*Bacula 3.5 
hours, old software: 1 hour*.


I have tried to:
a) turn off compression&encryption. The result is the same: backup speed 
around 13 MB/sec.
b) change destination storage (from a new ibm storage attached over nfs, 
to a local SSD disk attached on bacula server virtual machine). It took 
2 hours 50 minutes to backup linux file server (instead of 3.5 hours). 
Sequential write test tested with linux dd command shows write speed 300 
MB/sec for IBM storage and 600 MB/sec for local SSD storage (far better 
than actual throughput).


The network bandwidth is 1 GB (1 GB on client, 10 GB on server) so I 
guess this is not a problem; however I have noticed that bacula-fd on 
client side uses 100% of CPU.


I'm using:
-bacula server version 9.6.5
-bacula client version 5.2.13 (original from centos 6 repo).

Any idea what is wrong and/or what performance should I expect?
I would also appreciate some answers on the questions bellow.

Kind regards,
Ziga Zvan




On 05.08.2020 10:52, Žiga Žvan wrote:


Dear all,
I have tested bacula sw (9.6.5) and I must say I'm quite happy with 
the results (eg. compression, encryption, configureability). However I 
have some configuration/design questions I hope, you can help me with.


Regarding job schedule, I would like to:
- create incremental daily backup (retention 1 week)
- create weekly full backup (retention 1 month)
- create monthly full backup (retention 1 year)

I am using dummy cloud driver that writes to local file storage.  
Volume is a directory with fileparts. I would like to have seperate 
volumes/pools for each client. I would like to delete the data on disk 
after retention period expires. If possible, I would like to delete 
just the fileparts with expired backup.


Questions:
a) At the moment, I'm using two backup job definitions per client and 
central schedule definition for all my clients. I have noticed that my 
incremental job gets promoted to full after monthly backup ("No prior 
Full backup Job record found"; because monthly backup is a seperate 
job, but bacula searches for full backups inside the same job). Could 
you please suggest a better configuration. If possible, I would like 
to keep central schedule definition (If I manipulate pools in a 
schedule resource, I would need to define them per client).


b) I would like to delete expired backups on disk (and in the catalog 
as well). At the moment I'm using one volume in a daily/weekly/monthly 
pool per client. In a volume, there are fileparts belonging to expired 
backups (eg. part1-23 in the output bellow). I have tried to solve 
this with purge/prune scripts in my BackupCatalog job (as suggested in 
the whitepapers) but the data does not get deleted. Is there any way 
to delete fileparts? Should I create separate volumes after retention 
period? Please suggest a better configuration.


c) Do I need a restore job for each client? I would just like to 
restore backup on the same client, default to /restore folder... When 
I use bconsole restore all command, the wizard asks me all the 
questions (eg. 5- last backup for a client, which client,fileset...) 
but at the end it asks for a restore job which changes all previously 
defined things (eg. client).


d) At the moment, I have not implemented autochanger functionality. 
Clients compress/encrypt the data and send them to bacula server, 
which writes them on one central storage system. Jobs are processed in 
sequential order (one at a time). Do you expect any significant 
performance gain if i implement autochanger in order to have jobs run 
simultaneously?


Relevant part of configuration attached bellow.

Looking forward to move in the production...
Kind regards,
Ziga Zvan


*Volume example *(fileparts 1-23 should be deleted)*:*
[root@bacula cetrtapot-daily-vol-0022]# ls -ltr
total 0
-rw-r--r--. 1 bacula disk   262 Jul 28 23:05 part.1
-rw-r--r--. 1 bacula disk 35988 Jul 28 23:06 part.2
-rw-r--r--. 1 bacula disk 35992 Jul 28 23:07 part.3
-rw-r--r--. 1 bacula disk 36000 Jul 28 23:08 part.4
-rw-r--r--. 1 bacula disk 35981 Jul 28 23:09 part.5
-rw-r--r--. 1 bacula disk 328795126 Jul 28 23:10 part.6
-rw-r--r--. 1 bacula disk 35988 Jul 29 23:09 part.7
-rw-r--r--. 1 bacula disk 35995 Jul 29 23:10 part.8
-rw-r--r--. 1 bacula disk 35981 Jul 29 23:11 part.9
-rw-r--r--. 1 bacula disk 35992 Jul 29 23:12 part.10
-rw-r--r--. 1 bacula disk 453070890 Jul 29 23:12 part.11
-rw-r--r--. 1 bacula disk 35995 Jul 30 23:09 part.12
-rw-r--r--. 1 bacula disk 35993 Jul 30 23:10 part.13