Well it try (for good reason, configuration) to also consolidate the full, 
so it failed when it tries to access data stored on tape 

>19-Jul 13:27 schorsch-sd JobId 63962: Fatal error: stored/acquire.cc:214
>No suitable device found to read Volume "B00006L8"

AI work with 2 distinct storage, but having the same media type.
so in your case, full and inc would be on disk, and then tape can be used 
to get a Virtual Full consolidated (like every week, month or so)
as described in documentation.

On Monday 22 July 2024 at 09:48:59 UTC+2 Thomas Kempf wrote:

> Hello Bruno,
> i thougt that i tried to do what you proposed...
> Write the full consolidation job, including the first full from tape to 
> the Consolidation pool and them migrate all back to tape. So i thought 
> there would be no need to write on the tape during the consolidated full.
> As it works on incremental consolidation, data is read of type (file) 
> and written to type (filec) too. At least, that's how i interpret the log.
> > 19-Jul 13:27 hueper-dir JobId 63962: Using Device "Bareos-AII0001"
> > to read.
> > 19-Jul 13:27 hueper-dir JobId 63962: Using Device "Bareos-AIC0001"
> > to write.
> Or am i missing something here?
>
> Am 22.07.2024 um 09:33 schrieb Bruno Friedmann (bruno-at-bareos):
> > This setup will not work, if your first full is on Tape, the 
> > consolidation won't be able to have a read and write storage of the same 
> > media type.
> > 
> > The tape should be used to create a VF copy of consolidated jobs for 
> > example.
> > 
> > On Friday 19 July 2024 at 14:50:17 UTC+2 Thomas Kempf wrote:
> > 
> > Hello
> > I'm reviving this old thread, because i think i have a setup like Brock
> > describes, but still have some error in the setup or found a bug in
> > bareos. I'd be really glad if you could help me...
> > 
> > There is One LTO-8 Drive with an associated Pool "Fulltape" (mediatype
> > "lto-8") and the Disk-Storage with Pools "AI-Incremental" (mediatype
> > "file") and "AI-Consolidated" (mediatype "filec").
> > At the beginning i did a Full in the "Fulltape" Pool on LTO-8.
> > After that daily AI Incrementals on Disk in the AI-Incremental Pool.
> > Additionaly daily Consolidation jobs to AI-Consolidated.
> > This ran errorfree and smooth for 8 months. Then "Always Incremental
> > max
> > Full Age" was reached and i expected a full backup In AI-Consolidated,
> > which i wanted to migrate back to tape again afterwards.
> > Alas this Full Consolidation throws an error and fails to change the
> > read device to the tape.
> > Is this a bug in bareos ?
> > 
> > here is the error:
> > 
> > 19-Jul 13:27 hueper-dir JobId 63962: Start Virtual Backup JobId 63962,
> > Job=BETTERONE-AI-MULTIPOLSTER.2024-07-19_13.27.23_47
> > 19-Jul 13:27 hueper-dir JobId 63962: Bootstrap records written to
> > /var/lib/bareos/hueper-dir.restore.8.bsr
> > 19-Jul 13:27 hueper-dir JobId 63962: Consolidating JobIds
> > 59638,63851,62915,62939,62963 containing 215949 files
> > 19-Jul 13:27 hueper-dir JobId 63962: Connected Storage daemon at
> > schorsch-sd.ad.hueper.de:9103
> > <http://schorsch-sd.ad.hueper.de:9103>, encryption:
> > TLS_CHACHA20_POLY1305_SHA256
> > TLSv1.3
> > 19-Jul 13:27 hueper-dir JobId 63962: Encryption:
> > TLS_CHACHA20_POLY1305_SHA256 TLSv1.3
> > 19-Jul 13:27 hueper-dir JobId 63962: Using Device "Bareos-AII0001"
> > to read.
> > 19-Jul 13:27 hueper-dir JobId 63962: Using Device "Bareos-AIC0001"
> > to write.
> > 19-Jul 13:27 schorsch-sd JobId 63962: Volume "BO-AI-Consolidated-31509"
> > previously written, moving to end of data.
> > 19-Jul 13:27 schorsch-sd JobId 63962: Ready to append to end of Volume
> > "BO-AI-Consolidated-31509" size=238
> > 19-Jul 13:27 schorsch-sd JobId 63962: stored/acquire.cc:157 Changing
> > read device. Want Media Type="LTO-8" have="File"
> > device="Bareos-AII0001" (/bareos-data/BO-AI-Incremental)
> > 19-Jul 13:27 schorsch-sd JobId 63962: Releasing device "Bareos-AII0001"
> > (/bareos-data/BO-AI-Incremental).
> > 19-Jul 13:27 schorsch-sd JobId 63962: Fatal error:
> > stored/acquire.cc:214
> > No suitable device found to read Volume "B00006L8"
> > 19-Jul 13:27 schorsch-sd JobId 63962: Releasing device "Bareos-AIC0001"
> > (/bareos-data/BO-AI-Consolidated).
> > 19-Jul 13:27 schorsch-sd JobId 63962: Releasing device "Bareos-AII0001"
> > (/bareos-data/BO-AI-Incremental).
> > 19-Jul 13:27 hueper-dir JobId 63962: Replicating deleted files from
> > jobids 59638,63851,62915,62939,62963 to jobid 63962
> > 19-Jul 13:27 hueper-dir JobId 63962: Error: Bareos hueper-dir
> > 23.0.3~pre135.a9e3d95ca (28May24):
> > Build OS: Debian GNU/Linux 11 (bullseye)
> > JobId: 63962
> > Job: BETTERONE-AI-MULTIPOLSTER.2024-07-19_13.27.23_47
> > Backup Level: Virtual Full
> > Client: "betterone-fd" 23.0.3~pre95.0aeaf0d6d
> > (15Apr24) 13.2-RELEASE,freebsd
> > FileSet: "betterone-ai-multipolster" 2021-02-12 12:02:55
> > Pool: "BO-AI-Consolidated" (From Job Pool's
> > NextPool resource)
> > Catalog: "MyCatalog" (From Client resource)
> > Storage: "Bareos-AIC" (From Storage from Pool's
> > NextPool resource)
> > Scheduled time: 19-Jul-2024 13:27:23
> > Start time: 23-Mai-2024 19:06:26
> > End time: 23-Mai-2024 19:23:02
> > Elapsed time: 16 mins 36 secs
> > Priority: 25
> > Allow Mixed Priority: yes
> > SD Files Written: 0
> > SD Bytes Written: 0 (0 B)
> > Rate: 0.0 KB/s
> > Volume name(s):
> > Volume Session Id: 9
> > Volume Session Time: 1721383710
> > Last Volume Bytes: 0 (0 B)
> > SD Errors: 1
> > SD termination status: Fatal Error
> > Accurate: yes
> > Bareos binary info: Bareos community build (UNSUPPORTED): Get
> > professional support from https://www.bareos.com
> > <https://www.bareos.com>
> > Job triggered by: User
> > Termination: *** Backup Error ***
> > 
> > Kind Regards
> > Tom
> > 
> > 
> > 
> > 
> > Am 26.12.2023 um 14:05 schrieb Brock Palen:
> > > Doesn’t work,
> > >
> > > During the consolidation of a full, you have to read back from
> > teh full pool and write to the full pool. So you always need to have
> > 2 working devices, one to read, one to write to the AI-Consolidate
> > pool.
> > >
> > > That’s why in my setup tape is really a pool to ‘migrate to make
> > space’ so I can read back from it (Bareos often correctly switches
> > to read from that pool). But the AI-Consolidate pool is on disk and
> > where all the shuffling happens.
> > >
> > > There is no way to do AI without two devices, and enough disk to
> > at least create one full backup for whatever you are backing up.
> > >
> > >
> > > Brock Palen
> > > bro...@mlds-networks.com
> > > www.mlds-networks.com <http://www.mlds-networks.com>
> > > Websites, Linux, Hosting, Joomla, Consulting
> > >
> > >
> > >
> > >> On Dec 24, 2023, at 12:54 AM, Russell Harmon
> > <eatnu...@gmail.com> wrote:
> > >>
> > >> On Sat, Dec 23, 2023 at 17:21 Brock Palen
> > <bro...@mlds-networks.com> wrote:
> > >> Correct. Because when you run your consolidate with a full it
> > has to read the old full likely from your tape drive. So it has to
> > write to disk.
> > >>
> > >> What if I flip things around: use disk for incrementals and tape
> > for full? Would I then just need to make sure I run a consolidate
> > job before I run out of disk?
> > >
> > 
> > -- 
> > You received this message because you are subscribed to the Google 
> > Groups "bareos-users" group.
> > To unsubscribe from this group and stop receiving emails from it, send 
> > an email to bareos-users...@googlegroups.com 
> > <mailto:bareos-users...@googlegroups.com>.
> > To view this discussion on the web visit 
> > 
> https://groups.google.com/d/msgid/bareos-users/46fbea15-db58-4afd-893c-2099fea15ec4n%40googlegroups.com
>  
> <
> https://groups.google.com/d/msgid/bareos-users/46fbea15-db58-4afd-893c-2099fea15ec4n%40googlegroups.com?utm_medium=email&utm_source=footer
> >.
>
>

-- 
You received this message because you are subscribed to the Google Groups 
"bareos-users" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to bareos-users+unsubscr...@googlegroups.com.
To view this discussion on the web visit 
https://groups.google.com/d/msgid/bareos-users/f5be2a6c-3e6b-4b55-9cc7-b1396986b4c4n%40googlegroups.com.

Reply via email to