>>>>> On Tue, 04 Jun 2019 18:08:16 -0500, Larry Rosenman said:
>
> On 06/04/2019 6:05 pm, Chandler wrote:
> > Larry Rosenman wrote on 6/4/19 15:46:
> >> I was just wondering if it would make sense for Bacula to become
> >> smarter about this,
> >> and be helpful and not duplicate the files.
> >
> > Well bacula will do what you tell it, so it's up to you to figure out
> > if there are duplicated entries in your Fileset.
> >
> >
> >
>
> I was kind of hoping it would sort the list into the include/exclude
> lists, and especially
> if OneFS = yes is set, only hit a path ONCE.
>
> But I guess not.
>
> Oh well, was worth asking.
OneFS = yes will prevent bacula from descending into mounted filesystems, so
will prevent duplicates of those. There is no detection of duplicates within
a single filesystem though.
You can check for duplicate files in the catalog using this query (works on
PostgreSQL at least):
select dup.jobid, path.path, filename.name, countof
from (select jobid, pathid, filenameid, count(*) as countof
from file
group by jobid, pathid, filenameid having count(*)>1) as dup
inner join path on dup.pathid = path.pathid
inner join filename on dup.filenameid = filename.filenameid
where not filename.name = ''
limit 100;
Remove the limit 100 if you want to see them all.
__Martin
_______________________________________________
Bacula-users mailing list
[email protected]
https://lists.sourceforge.net/lists/listinfo/bacula-users