Hello Daniele,
Am 26.01.23 um 19:18 schrieb Daniele Teneggi:
Hello Philipp,
this is the version that did the backups (and is actually running):
Version: 22.0.2~pre (02 January 2023) Debian GNU/Linux 11 (bullseye)
I see that there's a new version available on your community repo (22.0.2~pre10.81f8d3b2b-15), i can
install it tomorrow if this can help resolve our problem.
Yes, please always use the newest version and check if the problem still exists. Of course you need
to make new backups as the existing ones will not be altered.
Best regards,
Philipp
Relative to the size of the volumes, we are slowly moving toward a more manageable sizing (around
100-500 GB max based on the client data size)
Thank you and have a nice day
Daniele
From: Philipp Storz <[email protected]>
To: Daniele Teneggi <[email protected]>, [email protected]
Date: Thu, 26 Jan 2023 18:25:33 +0100
Subject: Re: [bareos-users] Very slow restore on 22.0.2
Hello Daniele,
thank you very much for your problem description.
What exact version have you used for the backup, and is your current
version the same or different?
Please specify the full version info, like e.g. 22.0.2~pre10.81f8d3b2b
BTW it is usually a good idea to limit the volume size so that the volumes
do not get too big.
Best regards,
Philipp
Am 26.01.23 um 13:22 schrieb Daniele Teneggi:
> Hello,
> following an upgrade of the OS on our Bareos server (Debian 8 -> 11), we
upgraded Bareos to
version
> 22.0.2 from 17.x. using the community packages from the site; the
upgrade went well and system is
> running correctly from the old configuration, needing only minor changes
to disable the TLS
support.
> Two days ago i had to restore a file from the last full backup of our
fileserver, the restore
went
> well but took around 7 hours to complete (for a single file); with the
old version the time spent
> was always around 2-3 minutes average. Our Bareos installation saves
data on file volumes, not
> tapes, and doing some tests we verified that the restore took 7 hours
because it started
reading the
> volume from the beginning of the job until it found the file requested;
the size of full job is
> around 5.5 TB in a single Volume and the file was near the end, so
Bareos had to do almost a full
> read of the volume before finding it.
> Digging a little bit further we found that the old version, during a
backup, wrote a sort of
"index"
> in the jobmedia table, so it can seek near the position of the requested
file and make the
restore a
> lot faster, the backup from the version 22 has only one row indicating
the starting and ending
> position of the job in the volume:
> [backup done with the old version of Bareos, full job on a single
volume, around 19 GB]:
>
+--------+-------+----------+----------------+---------------------+---------------+
> | jobid | level | jobfiles | jobbytes | starttime |
volumename |
>
+--------+-------+----------+----------------+---------------------+---------------+
> | 148366 | F | 53,780 | 19,403,934,091 | 2023-01-02 04:35:38 |
win-infopub-1 |
>
> bareos=# select * from jobmedia where jobid=148366;
> jobmediaid | jobid | mediaid | firstindex | lastindex | startfile |
endfile | startblock |
endblock | volindex | jobbytes
>
------------+--------+---------+------------+-----------+-----------+---------+------------+------------+----------+----------
> 8079653 | 148366 | 452 | 46067 | 53780 | 5 |
5 | 2845841927 |
3277427288 | 20 | 0
> 8079652 | 148366 | 452 | 30131 | 46067 | 5 |
5 | 1845905960 |
2845841926 | 19 | 0
> 8079645 | 148366 | 452 | 14749 | 30131 | 5 |
5 | 845970039 |
1845905959 | 18 | 0
> 8079640 | 148366 | 452 | 7721 | 14749 | 4 |
5 | 4141001410 |
845970038 | 17 | 0
> 8079636 | 148366 | 452 | 7512 | 7721 | 4 |
4 | 3141065414 |
4141001409 | 16 | 0
> 8079634 | 148366 | 452 | 7512 | 7512 | 4 |
4 | 2141129415 |
3141065413 | 15 | 0
> 8079633 | 148366 | 452 | 7473 | 7512 | 4 |
4 | 1141193420 |
2141129414 | 14 | 0
> 8079631 | 148366 | 452 | 7407 | 7473 | 4 |
4 | 141257426 |
1141193419 | 13 | 0
> 8079630 | 148366 | 452 | 7369 | 7407 | 3 |
4 | 3436288733 |
141257425 | 12 | 0
> 8079629 | 148366 | 452 | 7328 | 7369 | 3 |
3 | 2436352734 |
3436288732 | 11 | 0
> 8079627 | 148366 | 452 | 7298 | 7328 | 3 |
3 | 1436416734 |
2436352733 | 10 | 0
> 8079626 | 148366 | 452 | 7274 | 7298 | 3 |
3 | 436480745 |
1436416733 | 9 | 0
> 8079624 | 148366 | 452 | 6226 | 7274 | 2 |
3 | 3731512054 |
436480744 | 8 | 0
> 8079620 | 148366 | 452 | 6058 | 6226 | 2 |
2 | 2731576066 |
3731512053 | 7 | 0
> 8079617 | 148366 | 452 | 3325 | 6058 | 2 |
2 | 1731640078 |
2731576065 | 6 | 0
> 8079613 | 148366 | 452 | 3317 | 3325 | 2 |
2 | 731704078 |
1731640077 | 5 | 0
> 8079607 | 148366 | 452 | 3317 | 3317 | 1 |
2 | 4026735374 |
731704077 | 4 | 0
> 8079598 | 148366 | 452 | 1789 | 3317 | 1 |
1 | 3026799387 |
4026735373 | 3 | 0
> 8079590 | 148366 | 452 | 244 | 1789 | 1 |
1 | 2026863387 |
3026799386 | 2 | 0
> 8079582 | 148366 | 452 | 1 | 244 | 1 |
1 | 1026927388 |
2026863386 | 1 | 0
>
> [backup done with version 22, full job on a single volume, around 5,5
TB]:
>
+--------+-------+-----------+-------------------+---------------------+---------------+
> | jobid | level | jobfiles | jobbytes | starttime |
volumename |
>
+--------+-------+-----------+-------------------+---------------------+---------------+
> | 149513 | F | 2,886,927 | 5,637,588,281,412 | 2023-01-21 00:30:02 |
fileserver-12 |
> bareos=# select * from jobmedia where jobid=149513;
> jobmediaid | jobid | mediaid | firstindex | lastindex | startfile |
endfile | startblock |
> endblock | volindex | jobbytes
>
------------+--------+---------+------------+-----------+-----------+---------+------------+-----------+----------+----------
> 8086618 | 149513 | 2146 | 1 | 2886927 | 6 | 1320 | 1489561298 |
465483062 | 1 | 0
> (the jobs are from 2 different clients because the first job was one of
the few done with the old
> version remaining in the catalog)
> There's someting that has to be changed in the configuration to make it
work as before? I
tried to
> search the documentation but cannot find anything about it.. I tried to
activate the checkpoint
> function but i cannot see any visible difference regarding this
behaviour.
> Thank you
> Daniele
>
> --
> You received this message because you are subscribed to the Google Groups
"bareos-users" group.
> To unsubscribe from this group and stop receiving emails from it, send
an email to
> [email protected]
<mailto:bareos-users%2Bunsubscribe%40googlegroups.com>
<mailto:[email protected]
<mailto:bareos-users%2Bunsubscribe%40googlegroups.com>>.
> To view this discussion on the web visit
>
https://groups.google.com/d/msgid/bareos-users/WC20230126122232.490768%40zensistemi.com
<https://groups.google.com/d/msgid/bareos-users/WC20230126122232.490768%40zensistemi.com>
>
<https://groups.google.com/d/msgid/bareos-users/WC20230126122232.490768%40zensistemi.com?utm_medium=email&utm_source=footer
<https://groups.google.com/d/msgid/bareos-users/WC20230126122232.490768%40zensistemi.com?utm_medium=email&utm_source=footer>>.
--
Mit freundlichen Grüßen
Philipp Storz [email protected]
<mailto:philipp.storz%40bareos.com>
Bareos GmbH & Co. KG Phone: +49 221 63 06 93-92
http://www.bareos.com <http://www.bareos.com> Fax:
+49 221 63 06 93-10
Sitz der Gesellschaft: Köln | Amtsgericht Köln: HRA 29646
Geschäftsführer: Stephan Dühr, M. Außendorf,
J. Steffens, P. Storz
--
Mit freundlichen Grüßen
Philipp Storz [email protected]
Bareos GmbH & Co. KG Phone: +49 221 63 06 93-92
http://www.bareos.com Fax: +49 221 63 06 93-10
Sitz der Gesellschaft: Köln | Amtsgericht Köln: HRA 29646
Geschäftsführer: Stephan Dühr, M. Außendorf,
J. Steffens, P. Storz
--
You received this message because you are subscribed to the Google Groups
"bareos-users" group.
To unsubscribe from this group and stop receiving emails from it, send an email
to [email protected].
To view this discussion on the web visit
https://groups.google.com/d/msgid/bareos-users/74983379-4b3d-6795-e600-7c5cbd0cd4b8%40bareos.com.