Kenny,

First of all, are you using the "run" command to submit a prior configured
Restore Job? I think this is not advisable, since there are several restore
variables that only the "restore" command can fetch.
Did you try to restore the same file with the restore command?

Regards,


On Mon, Sep 8, 2014 at 1:28 PM, Kenny Noe <knoe...@gmail.com> wrote:

> Dan,
>
> Thanks for the reply.   I tried this this morning and still failed the
> restore.   I see during the Status Storage - Storage_Bluewhale where the
> "Running Jobs section shows Files=0, Bytes=0 Bytes/sec=0.  However in the
> Device Status section, Device "File_bluewhale" is mounted and the Total
> Bytes Read and Block Read go up....   Now with the simplefied config It
> seems to have lost it's Pool.
>
> I'm confused...  I changed the where to restore to
> /nas/users/admin/backups and changed in the client config to remove the
> fifo headache...  But it still is trying to use it...
>
> Here is the error from the log :
> 08-Sep 11:51 BS01-DIR1 JobId 12897: Start Restore Job
> Restore_mail_bluewhale.2014-09-08_11.51.55_03
> 08-Sep 11:51 BS01-DIR1 JobId 12897: Using Device "File_bluewhale"
> 08-Sep 11:51 BS01-SD1 JobId 12897: Ready to read from volume "mail-0386"
> on device "File_bluewhale" (/nas/bacula/bluewhale).
> 08-Sep 11:51 BS01-SD1 JobId 12897: Forward spacing Volume "mail-0386" to
> file:block 0:219.
> 08-Sep 12:14 BS01-SD1 JobId 12897: End of Volume at file 28 on device
> "File_bluewhale" (/nas/bacula/bluewhale), Volume "mail-0386"
> 08-Sep 12:14 BS01-SD1 JobId 12897: End of all volumes.
> 08-Sep 11:52 BS01-FD1 JobId 12897: Error: create_file.c:292 Could not open
> /nas/users/admin/backups/data/backups/mail/fifo/mail.tar: ERR=Interrupted
> system call
> 08-Sep 12:14 BS01-DIR1 JobId 12897: Bacula BS01-DIR1 5.2.2 (26Nov11):
>   Build OS:               x86_64-unknown-linux-gnu ubuntu 11.10
>   JobId:                  12897
>   Job:                    Restore_mail_bluewhale.2014-09-08_11.51.55_03
>   Restore Client:         besc-bs01
>   Start time:             08-Sep-2014 11:51:57
>   End time:               08-Sep-2014 12:14:09
>   Files Expected:         1
>   Files Restored:         0
>   Bytes Restored:         0
>   Rate:                   0.0 KB/s
>   FD Errors:              0
>   FD termination status:  OK
>   SD termination status:  OK
>   Termination:            Restore OK -- warning file count mismatch
>
> 08-Sep 12:14 BS01-DIR1 JobId 12897: Begin pruning Jobs older than 15 days .
> 08-Sep 12:14 BS01-DIR1 JobId 12897: No Jobs found to prune.
> 08-Sep 12:14 BS01-DIR1 JobId 12897: Begin pruning Files.
> 08-Sep 12:14 BS01-DIR1 JobId 12897: No Files found to prune.
> 08-Sep 12:14 BS01-DIR1 JobId 12897: End auto prune.
>
>
> What is a file count mismatch?
>
> Here is the status during a restore :
> Connecting to Storage daemon Storage_bluewhale at 10.10.10.199:9103
>
> BS01-SD1 Version: 5.2.2 (26 November 2011) x86_64-unknown-linux-gnu ubuntu
> 11.10
> Daemon started 08-Sep-14 11:48. Jobs: run=0, running=0.
>  Heap: heap=598,016 smbytes=386,922 max_bytes=405,712 bufs=947 max_bufs=949
>  Sizes: boffset_t=8 size_t=8 int32_t=4 int64_t=8 mode=0,0
>
> Running Jobs:
> Reading: Full Restore job Restore_mail_bluewhale JobId=12897
> Volume="mail-0386"
>     pool="Pool_mail_bluewhale" device="File_bluewhale"
> (/nas/bacula/bluewhale)
>     Files=0 Bytes=0 Bytes/sec=0
>     FDReadSeqNo=6 in_msg=6 out_msg=84529 fd=6
> ====
>
> Jobs waiting to reserve a drive:
> ====
>
> Terminated Jobs:
>  JobId  Level    Files      Bytes   Status   Finished        Name
> ===================================================================
>  12889  Incr         31    67.94 M  OK       08-Sep-14 00:01
> Backup_os_besc-unixmgr01
>  12891  Full          4    501.0 M  OK       08-Sep-14 00:05 Backup_app_dev
>  12888  Incr        437    1.158 G  OK       08-Sep-14 00:06
> Backup_os_besc-bs01
>  12890  Incr          0         0   Other    08-Sep-14 00:30
> Backup_os_bluewhale
>  12893  Full          0         0   Other    08-Sep-14 01:30
> Backup_mail_bluewhale
>  12884  Full   2,361,101    154.6 G  OK       08-Sep-14 04:46
> Backup_os_mako
>  12892  Full          4    54.40 G  OK       08-Sep-14 05:56
> Backup_app_mako
>  12894                0         0   OK       08-Sep-14 08:53
> Restore_mail_bluewhale
>  12895                0         0   OK       08-Sep-14 09:28
> Restore_mail_bluewhale
>  12896                0         0   OK       08-Sep-14 10:10
> Restore_mail_bluewhale
> ====
>
> Device status:
> Device "File_asterisk" (/nas/bacula/asterisk) is not open.
> Device "File_besc-4dvapp" (/nas/bacula/besc-4dvapp) is not open.
> Device "File_besc-bs01" (/nas/bacula/besc-bs01) is not open.
> Device "File_besc-unixmgr01" (/nas/bacula/besc-unixmgr01) is not open.
> Device "File_bluewhale" (/nas/bacula/bluewhale) is mounted with:
>     Volume:      mail-0386
>     Pool:        *unknown*
>     Media type:  NAS_bluewhale
>     Total Bytes Read=1,121,412,096 Blocks Read=17,383 Bytes/block=64,512
>     Positioned at File=0 Block=1,121,412,275
> Device "File_demo" (/nas/bacula/demo) is not open.
> Device "File_dev" (/nas/bacula/dev) is not open.
> Device "File_mako" (/nas/bacula/mako) is not open.
> Device "File_qa" (/nas/bacula/qa) is not open.
> Device "File_qa2" (/nas/bacula/qa2) is not open.
> Device "File_smart" (/nas/bacula/smart) is not open.
> ====
>
> Used Volume status:
> mail-0386 on device "File_bluewhale" (/nas/bacula/bluewhale)
>     Reader=1 writers=0 devres=0 volinuse=1
> mail-0386 read volume JobId=12897
> ====
>
> ====
>
>
>
>
> Here is my simplified client config:
>
>
> #********************************************************************************
> # bluewhale
>
> #********************************************************************************
>    Client {
>       Name                   = bluewhale
>       Address                = bluewhale.bnesystems.com
>       Catalog                = BS01-Catalog
>       Password               = "xxxxxxxxx"
>       FileRetention          = 365 days
>       JobRetention           = 365 days
>       AutoPrune              = yes
>       MaximumConcurrentJobs  = 1
>    }
>    Job {
>       Name                   = Restore_mail_bluewhale
>       FileSet                = Full_mail_bluewhale
>       Type                   = Restore
>       Pool                   = Pool_mail_bluewhale
>       Client                 = bluewhale
>       Messages               = Standard
>    }
>    Pool {
>       Name                   = Pool_mail_bluewhale
>       PoolType               = Backup
>       Storage                = Storage_bluewhale
>       MaximumVolumeJobs      = 1
>       CatalogFiles           = yes
>       AutoPrune              = yes
>       VolumeRetention        = 365 days
>       Recycle                = yes
>       LabelFormat            = "mail-"
>    }
>    Storage {
>       Name                   = Storage_bluewhale
>       Address                = 10.10.10.199
>       SDPort                 = 9103
>       Password               = "xxxxxxx"
>       Device                 = File_bluewhale
>       MediaType              = NAS_bluewhale
>       MaximumConcurrentJobs  = 1
>    }
>    FileSet {
>       Name = Full_mail_bluewhale
>       Include {
>          Options {
>             signature=SHA1
> #            readfifo=yes
>          }
>          File="/mail.tar"
>       }
>    }
>
>
>
> Thanks for the help.   I appreciate all the input.
>
> --Kenny
>
>
> On Sun, Sep 7, 2014 at 8:22 AM, Dan Langille <d...@langille.org> wrote:
>
>> I suggest removing the before & after scripts.
>>
>> --
>> Dan Langille
>> http://langille.org/
>>
>>
>> > On Sep 6, 2014, at 8:38 PM, Kenny Noe <knoe...@gmail.com> wrote:
>> >
>> > Dan,
>> >
>> > Appreciate the reply....   Yes this is exactly what I want to do.
>> > However when I try to just do a "simple" restore, the job finishes
>> > with the error previously given.
>> >
>> > Any suggestions to do this would be appreciated.
>> >
>> > Thanks    --Kenny
>> >
>> >> On Sat, Sep 6, 2014 at 5:51 PM, Dan Langille <d...@langille.org> wrote:
>> >>
>> >> On Sep 5, 2014, at 5:48 PM, Kenny Noe <knoe...@gmail.com> wrote:
>> >>
>> >> Birre,
>> >>
>> >> Thanks for the reply.   I guess this is where I get lost...
>> >>
>> >>
>> >>
>> >> The fifo is reading a file that was created in the pre-process called
>> >> mail.tar.  The mail.tar is made from the following directories
>> /opt/zimbra
>> >> and /var/mail/zimbra.  This is where the Zimbra files and mailstore
>> were
>> >> kept.
>> >>
>> >> This pre-process is a script that has this :
>> >>
>> >> MailBackup.bash
>> >> #!/bin/bash
>> >>
>> >> exec >/dev/null
>> >>
>> >> MKDIR="/bin/mkdir"
>> >> MKFIFO="/usr/bin/mkfifo"
>> >> RM="/bin/rm"
>> >> TAR="/bin/tar"
>> >>
>> >> DEFCODE=0
>> >> DUMPBASE="/data/backups"
>> >>
>> >> errCode=${DEFCODE}
>> >> mailDir="/var/mail/zimbra"
>> >> zimbraDir="/opt/zimbra"
>> >>
>> >> Main()
>> >>   {
>> >>   service zimbra stop
>> >>
>> >>   RunMailRestore
>> >>
>> >>   service zimbra start
>> >>
>> >>   ExitScript ${errCode}
>> >>   }
>> >>
>> >> RunMailRestore()
>> >>   {
>> >>   EXTENSION=".tar"
>> >>
>> >>   dumpDir="${DUMPBASE}/mail"
>> >>   fifoDir="${dumpDir}/fifo"
>> >>
>> >>   RebuildFifoDir
>> >>
>> >>   ${MKFIFO} ${fifoDir}/mail${EXTENSION}
>> >>
>> >>   ${TAR} -xpf ${fifoDir}/mail${EXTENSION} 2>&1 </dev/null &
>> >>   }
>> >>
>> >> RebuildFifoDir()
>> >>   {
>> >>   if [ -d ${fifoDir} ]
>> >>   then
>> >>      ${RM} -rf ${fifoDir}
>> >>   fi
>> >>
>> >>   ${MKDIR} -p ${fifoDir}
>> >>   }
>> >>
>> >> ExitScript()
>> >>   {
>> >>   exit ${1}
>> >>   }
>> >>
>> >> Main
>> >>
>> >> The restore script simply does a tar xpf instead of a tar cpf.
>> >>
>> >>
>> >> Perhaps instead of doing that, just restore the data, and then do the
>> tar
>> >> xpf later.
>>
>
>
>
> ------------------------------------------------------------------------------
> Want excitement?
> Manually upgrade your production database.
> When you want reliability, choose Perforce
> Perforce version control. Predictably reliable.
>
> http://pubads.g.doubleclick.net/gampad/clk?id=157508191&iu=/4140/ostg.clktrk
> _______________________________________________
> Bacula-users mailing list
> Bacula-users@lists.sourceforge.net
> https://lists.sourceforge.net/lists/listinfo/bacula-users
>
>


-- 
============================================
Heitor Medrado de Faria | Need Bacula training? 10% discount coupon code at
Udemy: bacula-users
<https://www.udemy.com/bacula-backup-software/?couponCode=bacula-users>
+55 61 2021-8260
+55 61 8268-4220
Site: www.bacula.com.br
Facebook: heitor.faria <http://www.facebook.com/heitor.faria>
Gtalk: heitorfa...@gmail.com
============================================
------------------------------------------------------------------------------
Want excitement?
Manually upgrade your production database.
When you want reliability, choose Perforce
Perforce version control. Predictably reliable.
http://pubads.g.doubleclick.net/gampad/clk?id=157508191&iu=/4140/ostg.clktrk
_______________________________________________
Bacula-users mailing list
Bacula-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/bacula-users

Reply via email to