ok, what i don't understand is the necessity of having the same media
type for read access. I just read from tape
Wouldn't it be possible to read full from one device (media type),
consolidate with incrementals from other device (media type) and write
the consolidated full data to consolidated pool.
Like that You could avoid having to keep a full of all AI jobs on disk
at the same time...
Am 22.07.2024 um 14:50 schrieb Bruno Friedmann (bruno-at-bareos):> Well
it try (for good reason, configuration) to also consolidate the
full, so it failed when it tries to access data stored on tape
>19-Jul 13:27 schorsch-sd JobId 63962: Fatal error: stored/acquire.cc:214
>No suitable device found to read Volume "B00006L8"
AI work with 2 distinct storage, but having the same media type.
so in your case, full and inc would be on disk, and then tape can be
used to get a Virtual Full consolidated (like every week, month or so)
as described in documentation.
On Monday 22 July 2024 at 09:48:59 UTC+2 Thomas Kempf wrote:
Hello Bruno,
i thougt that i tried to do what you proposed...
Write the full consolidation job, including the first full from tape to
the Consolidation pool and them migrate all back to tape. So i thought
there would be no need to write on the tape during the consolidated
full.
As it works on incremental consolidation, data is read of type (file)
and written to type (filec) too. At least, that's how i interpret
the log.
> 19-Jul 13:27 hueper-dir JobId 63962: Using Device "Bareos-AII0001"
> to read.
> 19-Jul 13:27 hueper-dir JobId 63962: Using Device "Bareos-AIC0001"
> to write.
Or am i missing something here?
Am 22.07.2024 um 09:33 schrieb Bruno Friedmann (bruno-at-bareos):
> This setup will not work, if your first full is on Tape, the
> consolidation won't be able to have a read and write storage of
the same
> media type.
>
> The tape should be used to create a VF copy of consolidated jobs for
> example.
>
> On Friday 19 July 2024 at 14:50:17 UTC+2 Thomas Kempf wrote:
>
> Hello
> I'm reviving this old thread, because i think i have a setup like
Brock
> describes, but still have some error in the setup or found a bug in
> bareos. I'd be really glad if you could help me...
>
> There is One LTO-8 Drive with an associated Pool "Fulltape"
(mediatype
> "lto-8") and the Disk-Storage with Pools "AI-Incremental" (mediatype
> "file") and "AI-Consolidated" (mediatype "filec").
> At the beginning i did a Full in the "Fulltape" Pool on LTO-8.
> After that daily AI Incrementals on Disk in the AI-Incremental Pool.
> Additionaly daily Consolidation jobs to AI-Consolidated.
> This ran errorfree and smooth for 8 months. Then "Always Incremental
> max
> Full Age" was reached and i expected a full backup In
AI-Consolidated,
> which i wanted to migrate back to tape again afterwards.
> Alas this Full Consolidation throws an error and fails to change the
> read device to the tape.
> Is this a bug in bareos ?
>
> here is the error:
>
> 19-Jul 13:27 hueper-dir JobId 63962: Start Virtual Backup JobId
63962,
> Job=BETTERONE-AI-MULTIPOLSTER.2024-07-19_13.27.23_47
> 19-Jul 13:27 hueper-dir JobId 63962: Bootstrap records written to
> /var/lib/bareos/hueper-dir.restore.8.bsr
> 19-Jul 13:27 hueper-dir JobId 63962: Consolidating JobIds
> 59638,63851,62915,62939,62963 containing 215949 files
> 19-Jul 13:27 hueper-dir JobId 63962: Connected Storage daemon at
> schorsch-sd.ad.hueper.de:9103 <http://schorsch-sd.ad.hueper.de:9103>
> <http://schorsch-sd.ad.hueper.de:9103
<http://schorsch-sd.ad.hueper.de:9103>>, encryption:
> TLS_CHACHA20_POLY1305_SHA256
> TLSv1.3
> 19-Jul 13:27 hueper-dir JobId 63962: Encryption:
> TLS_CHACHA20_POLY1305_SHA256 TLSv1.3
> 19-Jul 13:27 hueper-dir JobId 63962: Using Device "Bareos-AII0001"
> to read.
> 19-Jul 13:27 hueper-dir JobId 63962: Using Device "Bareos-AIC0001"
> to write.
> 19-Jul 13:27 schorsch-sd JobId 63962: Volume
"BO-AI-Consolidated-31509"
> previously written, moving to end of data.
> 19-Jul 13:27 schorsch-sd JobId 63962: Ready to append to end of
Volume
> "BO-AI-Consolidated-31509" size=238
> 19-Jul 13:27 schorsch-sd JobId 63962: stored/acquire.cc:157 Changing
> read device. Want Media Type="LTO-8" have="File"
> device="Bareos-AII0001" (/bareos-data/BO-AI-Incremental)
> 19-Jul 13:27 schorsch-sd JobId 63962: Releasing device
"Bareos-AII0001"
> (/bareos-data/BO-AI-Incremental).
> 19-Jul 13:27 schorsch-sd JobId 63962: Fatal error:
> stored/acquire.cc:214
> No suitable device found to read Volume "B00006L8"
> 19-Jul 13:27 schorsch-sd JobId 63962: Releasing device
"Bareos-AIC0001"
> (/bareos-data/BO-AI-Consolidated).
> 19-Jul 13:27 schorsch-sd JobId 63962: Releasing device
"Bareos-AII0001"
> (/bareos-data/BO-AI-Incremental).
> 19-Jul 13:27 hueper-dir JobId 63962: Replicating deleted files from
> jobids 59638,63851,62915,62939,62963 to jobid 63962
> 19-Jul 13:27 hueper-dir JobId 63962: Error: Bareos hueper-dir
> 23.0.3~pre135.a9e3d95ca (28May24):
> Build OS: Debian GNU/Linux 11 (bullseye)
> JobId: 63962
> Job: BETTERONE-AI-MULTIPOLSTER.2024-07-19_13.27.23_47
> Backup Level: Virtual Full
> Client: "betterone-fd" 23.0.3~pre95.0aeaf0d6d
> (15Apr24) 13.2-RELEASE,freebsd
> FileSet: "betterone-ai-multipolster" 2021-02-12 12:02:55
> Pool: "BO-AI-Consolidated" (From Job Pool's
> NextPool resource)
> Catalog: "MyCatalog" (From Client resource)
> Storage: "Bareos-AIC" (From Storage from Pool's
> NextPool resource)
> Scheduled time: 19-Jul-2024 13:27:23
> Start time: 23-Mai-2024 19:06:26
> End time: 23-Mai-2024 19:23:02
> Elapsed time: 16 mins 36 secs
> Priority: 25
> Allow Mixed Priority: yes
> SD Files Written: 0
> SD Bytes Written: 0 (0 B)
> Rate: 0.0 KB/s
> Volume name(s):
> Volume Session Id: 9
> Volume Session Time: 1721383710
> Last Volume Bytes: 0 (0 B)
> SD Errors: 1
> SD termination status: Fatal Error
> Accurate: yes
> Bareos binary info: Bareos community build (UNSUPPORTED): Get
> professional support from https://www.bareos.com
<https://www.bareos.com>
> <https://www.bareos.com <https://www.bareos.com>>
> Job triggered by: User
> Termination: *** Backup Error ***
>
> Kind Regards
> Tom
>
>
>
>
> Am 26.12.2023 um 14:05 schrieb Brock Palen:
> > Doesn’t work,
> >
> > During the consolidation of a full, you have to read back from
> teh full pool and write to the full pool. So you always need to have
> 2 working devices, one to read, one to write to the AI-Consolidate
> pool.
> >
> > That’s why in my setup tape is really a pool to ‘migrate to make
> space’ so I can read back from it (Bareos often correctly switches
> to read from that pool). But the AI-Consolidate pool is on disk and
> where all the shuffling happens.
> >
> > There is no way to do AI without two devices, and enough disk to
> at least create one full backup for whatever you are backing up.
> >
> >
> > Brock Palen
> > [email protected]
> > www.mlds-networks.com <http://www.mlds-networks.com>
<http://www.mlds-networks.com <http://www.mlds-networks.com>>
> > Websites, Linux, Hosting, Joomla, Consulting
> >
> >
> >
> >> On Dec 24, 2023, at 12:54 AM, Russell Harmon
> <[email protected]> wrote:
> >>
> >> On Sat, Dec 23, 2023 at 17:21 Brock Palen
> <[email protected]> wrote:
> >> Correct. Because when you run your consolidate with a full it
> has to read the old full likely from your tape drive. So it has to
> write to disk.
> >>
> >> What if I flip things around: use disk for incrementals and tape
> for full? Would I then just need to make sure I run a consolidate
> job before I run out of disk?
> >
>
> --
> You received this message because you are subscribed to the Google
> Groups "bareos-users" group.
> To unsubscribe from this group and stop receiving emails from it,
send
> an email to [email protected]
> <mailto:[email protected]>.
> To view this discussion on the web visit
>
https://groups.google.com/d/msgid/bareos-users/46fbea15-db58-4afd-893c-2099fea15ec4n%40googlegroups.com
<https://groups.google.com/d/msgid/bareos-users/46fbea15-db58-4afd-893c-2099fea15ec4n%40googlegroups.com>
<https://groups.google.com/d/msgid/bareos-users/46fbea15-db58-4afd-893c-2099fea15ec4n%40googlegroups.com?utm_medium=email&utm_source=footer
<https://groups.google.com/d/msgid/bareos-users/46fbea15-db58-4afd-893c-2099fea15ec4n%40googlegroups.com?utm_medium=email&utm_source=footer>>.
--
You received this message because you are subscribed to the Google
Groups "bareos-users" group.
To unsubscribe from this group and stop receiving emails from it, send
an email to [email protected]
<mailto:[email protected]>.
To view this discussion on the web visit
https://groups.google.com/d/msgid/bareos-users/f5be2a6c-3e6b-4b55-9cc7-b1396986b4c4n%40googlegroups.com <https://groups.google.com/d/msgid/bareos-users/f5be2a6c-3e6b-4b55-9cc7-b1396986b4c4n%40googlegroups.com?utm_medium=email&utm_source=footer>.
--
You received this message because you are subscribed to the Google Groups
"bareos-users" group.
To unsubscribe from this group and stop receiving emails from it, send an email
to [email protected].
To view this discussion on the web visit
https://groups.google.com/d/msgid/bareos-users/c163d4fd-024b-42bf-9ff3-7c9847641a05%40hueper.de.