I did it, but it still doesn't show the files for the restore.
El jueves, 30 de enero de 2025 a la(s) 11:33:57 a.m. UTC-3, Bruno Friedmann 
(bruno-at-bareos) escribió:

> You may do this in bconsole, so there's no timeout when recreating the 
> bvfs cache.
>
>
> On Thursday, 30 January 2025 at 15:31:22 UTC+1 Rodrigo Yesi wrote:
>
>> Hello, I increased the timeout time and it completed, but it still shows 
>> blank in the web ui. 
>>
>> El jueves, 30 de enero de 2025 a la(s) 11:12:00 a.m. UTC-3, Rodrigo Yesi 
>> escribió:
>>
>>> Jan 30 10:54 bareos-dir JobId 0: Fatal error: cats/bvfs.cc:247 
>>> cats/bvfs.cc:247 query INSERT INTO PathVisibility (PathId, JobId) SELECT 
>>> a.PathId,27 FROM (SELECT DISTINCT h . PPathId AS PathId FROM PathHierarchy 
>>> AS h JOIN PathVisibili ty AS p ON (h.PathId=p.PathId) WHERE p.JobId=27) AS 
>>> LEFT JOIN PathVisibility AS b ON (b.JobId=27 AND a.PathId = b .PathId) 
>>> WHERE b.PathId IS NULL failed : ERROR: canceling the sentence because it 
>>> was done? the waiting time for sentences Jan 30 10:54 bareos-dir JobId 0: 
>>> Error: cats/bvfs.cc:251 cats/bvfs.cc:251 update UPDATE Job SET HasCache=1 
>>> WHERE Job Id=27 failed: ERROR: transaction aborted, orders will be ignored 
>>> until the end of the transaction block
>>>
>>> El jueves, 30 de enero de 2025 a la(s) 6:58:53 a.m. UTC-3, Bruno 
>>> Friedmann (bruno-at-bareos) escribió:
>>>
>>>> You might want to cleanup your bvfs_cache and try to regenerate it 
>>>>
>>>> bconsole <<< ".bvfs_clear_cache yes"
>>>>
>>>> then recreate it completely
>>>>
>>>> bconsole <<< ".bvfs_update" 
>>>>
>>>> On Thursday, 30 January 2025 at 04:01:54 UTC+1 Rodrigo Yesi wrote:
>>>>
>>>>> this is to day, its ok, but dont see in web
>>>>>
>>>>>
>>>>> 29-ene 21:00 bareos-dir JobId 38: Version: 24.0.1~pre27.250812184 (24 
>>>>> January 2025) Red Hat Enterprise Linux release 9.5 (Plow) 29-ene 21:00 
>>>>> bareos-dir JobId 38: Start Backup JobId 38, 
>>>>> Job=Backup-AVY.2025-01-29_21.00.00_03
>>>>>
>>>>> 29-ene 21:00 bareos-dir JobId 38: Connected Storage daemon at 
>>>>> 192.168.12.27:9103, encryption: TLS_CHACHA20_POLY1305_SHA256 TLSv1.3 
>>>>> 29-ene 21:00 bareos-dir JobId 38:  Encryption: 
>>>>> TLS_CHACHA20_POLY1305_SHA256 
>>>>> TLSv1.3 29-ene 21:00 bareos-dir JobId 38: Probing client protocol... 
>>>>> (result will be saved until config reload) 29-ene 21:00 bareos-dir JobId 
>>>>> 38: Connected Client: AMB-SVR-AVY-fd at 192.168.6.91:9102, 
>>>>> encryption: TLS_CHACHA20_POLY1305_SHA256 TLSv1.3
>>>>>
>>>>> 29-ene 21:00 bareos-dir JobId 38:    Handshake: Immediate TLS 
>>>>>
>>>>> 29-ene 21:00 bareos-dir JobId 38:  Encryption: 
>>>>> TLS_CHACHA20_POLY1305_SHA256 TLSv1.3 29-ene 21:00 bareos-sd JobId 38: 
>>>>> Using 
>>>>> just in time reservation for job 38 29-ene 21:00 bareos-dir JobId 38: 
>>>>> Using 
>>>>> Device "JustInTime Device" to write.
>>>>>
>>>>> 29-ene 20:59 amb-svr-avy-fd JobId 38: Created 24 wildcard excludes 
>>>>> from FilesNotToBackup Registry key 29-ene 20:59 amb-svr-avy-fd JobId 38: 
>>>>> Connected Storage daemon at 192.168.12.27:9103, encryption: 
>>>>> TLS_CHACHA20_POLY1305_SHA256 TLSv1.3 29-ene 20:59 amb-svr-avy-fd JobId 
>>>>> 38:  
>>>>> Encryption: TLS_CHACHA20_POLY1305_SHA256 TLSv1.3 29-ene 21:00 bareos-sd 
>>>>> JobId 38: Version: 24.0.1~pre27.250812184 (24 January 2025) Red Hat 
>>>>> Enterprise Linux release 9.5 (Plow) 29-ene 21:00 amb-svr-avy-fd JobId 38: 
>>>>> Generate VSS snapshots. Driver="Win64 VSS"
>>>>>
>>>>> 29-ene 21:00 amb-svr-avy-fd JobId 38: VolumeMountpoints are not 
>>>>> processed as onefs = yes.
>>>>>
>>>>> 29-ene 21:00 amb-svr-avy-fd JobId 38: 
>>>>> (C:\)\\?\Volume{2077a78b-2dee-4fbe-b9a2-ab119663189b}\ -> 
>>>>> \\?\GLOBALROOT\Device\HarddiskVolumeShadowCopy5
>>>>>
>>>>> 29-ene 21:00 amb-svr-avy-fd JobId 38: Version: 24.0.1~pre27.250812184 
>>>>> (24 January 2025) Microsoft Windows Server 2012 Standard Edition (build 
>>>>> 9200), 64-bit 29-ene 21:00 bareos-sd JobId 38: JustInTime Reservation: 
>>>>> Finding drive to reserve.
>>>>>
>>>>> 29-ene 21:00 bareos-dir JobId 38: Created new Volume "VolAVY-0041" in 
>>>>> catalog.
>>>>>
>>>>> 29-ene 21:00 bareos-sd JobId 38: Using Device "MyDevice2" (/Storage2) 
>>>>> to write.
>>>>>
>>>>> 29-ene 21:00 bareos-sd JobId 38: Labeled new Volume "VolAVY-0041" on 
>>>>> device "MyDevice2" (/Storage2).
>>>>>
>>>>> 29-ene 21:00 bareos-sd JobId 38: Moving to end of data on volume 
>>>>> "VolAVY-0041"
>>>>>
>>>>> 29-ene 21:00 bareos-sd JobId 38: Ready to append to end of Volume 
>>>>> "VolAVY-0041" size=221 29-ene 21:00 bareos-dir JobId 38: Max Volume 
>>>>> jobs=1 
>>>>> exceeded. Marking Volume "VolAVY-0041" as Used.
>>>>>
>>>>> 29-ene 21:52 amb-svr-avy-fd JobId 38: VSS Writer (BackupComplete): 
>>>>> "Task Scheduler Writer", State: 0x1 (VSS_WS_STABLE) 29-ene 21:52 
>>>>> amb-svr-avy-fd JobId 38: VSS Writer (BackupComplete): "VSS Metadata Store 
>>>>> Writer", State: 0x1 (VSS_WS_STABLE) 29-ene 21:52 amb-svr-avy-fd JobId 38: 
>>>>> VSS Writer (BackupComplete): "Performance Counters Writer", State: 0x1 
>>>>> (VSS_WS_STABLE) 29-ene 21:52 amb-svr-avy-fd JobId 38: VSS Writer 
>>>>> (BackupComplete): "System Writer", State: 0x1 (VSS_WS_STABLE) 29-ene 
>>>>> 21:52 
>>>>> amb-svr-avy-fd JobId 38: VSS Writer (BackupComplete): "ASR Writer", 
>>>>> State: 
>>>>> 0x1 (VSS_WS_STABLE) 29-ene 21:52 amb-svr-avy-fd JobId 38: VSS Writer 
>>>>> (BackupComplete): "BITS Writer", State: 0x1 (VSS_WS_STABLE) 29-ene 21:52 
>>>>> amb-svr-avy-fd JobId 38: VSS Writer (BackupComplete): "WMI Writer", 
>>>>> State: 
>>>>> 0x1 (VSS_WS_STABLE) 29-ene 21:52 amb-svr-avy-fd JobId 38: VSS Writer 
>>>>> (BackupComplete): "Registry Writer", State: 0x1 (VSS_WS_STABLE) 29-ene 
>>>>> 21:52 amb-svr-avy-fd JobId 38: VSS Writer (BackupComplete): "Shadow Copy 
>>>>> Optimization Writer", State: 0x1 (VSS_WS_STABLE) 29-ene 21:52 
>>>>> amb-svr-avy-fd JobId 38: VSS Writer (BackupComplete): "COM+ REGDB 
>>>>> Writer", 
>>>>> State: 0x1 (VSS_WS_STABLE) 29-ene 21:52 bareos-sd JobId 38: Releasing 
>>>>> device "MyDevice2" (/Storage2).
>>>>>
>>>>> 29-ene 21:52 bareos-sd JobId 38: Elapsed time=00:52:25, Transfer 
>>>>> rate=2.321 M Bytes/second 29-ene 21:52 bareos-dir JobId 38: Insert of 
>>>>> attributes batch table with 70834 entries start 29-ene 21:52 bareos-dir 
>>>>> JobId 38: Insert of attributes batch table done 29-ene 21:52 bareos-dir 
>>>>> JobId 38: Bareos bareos-dir 24.0.1~pre27.250812184 (24Jan25):
>>>>>
>>>>>   Build OS:               Red Hat Enterprise Linux release 9.5 (Plow)
>>>>>
>>>>>   JobId:                  38
>>>>>
>>>>>   Job:                    Backup-AVY.2025-01-29_21.00.00_03
>>>>>
>>>>>   Backup Level:           Full
>>>>>
>>>>>   Client:                 "AMB-SVR-AVY-fd" 24.0.1~pre27.250812184 
>>>>> (24Jan25) Microsoft Windows Server 2012 Standard Edition (build 9200), 
>>>>> 64-bit,Cross-compile
>>>>>
>>>>>   FileSet:                "MyFileSetAVY" 2025-01-24 09:03:38
>>>>>
>>>>>   Pool:                   "Pool-AVY" (From Run Pool override)
>>>>>
>>>>>   Catalog:                "MyCatalog" (From Client resource)
>>>>>
>>>>>   Storage:                "MyStorage2" (From run override)
>>>>>
>>>>>   Scheduled time:         29-ene-2025 21:00:00
>>>>>
>>>>>   Start time:             29-ene-2025 21:00:00
>>>>>
>>>>>   End time:               29-ene-2025 21:52:25
>>>>>
>>>>>   Elapsed time:           52 mins 25 secs
>>>>>
>>>>>   Priority:               10
>>>>>
>>>>>   Allow Mixed Priority:   no
>>>>>
>>>>>   FD Files Written:       70,835
>>>>>
>>>>>   SD Files Written:       70,835
>>>>>
>>>>>   FD Bytes Written:       7,284,102,133 (7.284 GB)
>>>>>
>>>>>   SD Bytes Written:       7,300,803,636 (7.300 GB)
>>>>>
>>>>>   Rate:                   2316.1 KB/s
>>>>>
>>>>>   Software Compression:   None
>>>>>
>>>>>   VSS:                    yes
>>>>>
>>>>>   Encryption:             no
>>>>>
>>>>>   Accurate:               no
>>>>>
>>>>>   Volume name(s):         VolAVY-0041
>>>>>
>>>>>   Volume Session Id:      2
>>>>>
>>>>>   Volume Session Time:    1738150563
>>>>>
>>>>>   Last Volume Bytes:      7,303,834,619 (7.303 GB)
>>>>>
>>>>>   Non-fatal FD errors:    0
>>>>>
>>>>>   SD Errors:              0
>>>>>
>>>>>   FD termination status:  OK
>>>>>
>>>>>   SD termination status:  OK
>>>>>
>>>>>   Bareos binary info:     Bareos community build (UNSUPPORTED): Get 
>>>>> professional support from https://www.bareos.com
>>>>>
>>>>>   Job triggered by:       Scheduler
>>>>>
>>>>>   Termination:            Backup OK
>>>>>
>>>>>  
>>>>>
>>>>> 29-ene 21:52 bareos-dir JobId 38: Begin pruning Jobs older than 3 
>>>>> months .
>>>>>
>>>>> 29-ene 21:52 bareos-dir JobId 38: No jobids found to be purged 29-ene 
>>>>> 21:52 bareos-dir JobId 38: Begin pruning Files.
>>>>>
>>>>> 29-ene 21:52 bareos-dir JobId 38: No Files found to prune.
>>>>>
>>>>> 29-ene 21:52 bareos-dir JobId 38: End auto prune.
>>>>>
>>>>>  
>>>>> El miércoles, 29 de enero de 2025 a la(s) 8:27:58 a.m. UTC-3, Rodrigo 
>>>>> Yesi escribió:
>>>>>
>>>>>> Hello good morning. As you can see in the images, certain backups 
>>>>>> work. The webui marks them as correct for the backup, but when I want to 
>>>>>> do 
>>>>>> a restore. In some clients it does not show them to me, as in the case 
>>>>>> of 
>>>>>> MFS. It has 2 backups made and both are left blank. in the case of 
>>>>>> MANTIS, 
>>>>>> if you see the image. The back from the 28th works, but the one from the 
>>>>>> 29th doesn't! I already gave more resources to postgresql-14 and still 
>>>>>> nothing.
>>>>>
>>>>>

-- 
You received this message because you are subscribed to the Google Groups 
"bareos-users" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To view this discussion visit 
https://groups.google.com/d/msgid/bareos-users/4255807e-2a27-4176-ace2-206ddb9ff4f6n%40googlegroups.com.

Reply via email to