Video Screencast Help
Symantec to Separate Into Two Focused, Industry-Leading Technology Companies. Learn more.

Problème BMR client netbackup 7.1 avec Os AIX

Created: 21 Sep 2012 • Updated: 21 Sep 2012 | 13 comments
Bonjour
j'ai reçu le probleme BMR sur un cleint AIX 6.1
Plateforme: 
Master server : OS Solaris 10, Natbackup 7.1
Server à backup : agent client netbackup AIX 7.1, Os de AIX 6.1.
etat de client connecté est existe sur un reseau de backup dedié( reseau dédié pour  la palateforme de backup).
Merci de me aider à resoudre cette probleme vue la criticité du plateforme.
 
log de probleme:
 
09/21/2012 10:57:25 - Info nbjm (pid=17894) starting backup job (jobid=71653) for client KBPRORAC02D001, policy SOA_Prod_BMR, schedule Full-BMR
09/21/2012 10:57:25 - Info nbjm (pid=17894) requesting STANDARD_RESOURCE resources from RB for backup job (jobid=71653, request id:{5D25FC38-03CA-11E2-BD02-00144FA05E06})
09/21/2012 10:57:25 - requesting resource backup-hcart-robot-tld-0
09/21/2012 10:57:25 - requesting resource backup.NBU_CLIENT.MAXJOBS.KBPRORAC02D001
09/21/2012 10:57:25 - requesting resource backup.NBU_POLICY.MAXJOBS.SOA_Prod_BMR
09/21/2012 10:57:26 - granted resource  backup.NBU_CLIENT.MAXJOBS.KBPRORAC02D001
09/21/2012 10:57:26 - granted resource  backup.NBU_POLICY.MAXJOBS.SOA_Prod_BMR
09/21/2012 10:57:26 - granted resource  000227
09/21/2012 10:57:26 - granted resource  HP.ULTRIUM4-SCSI.001
09/21/2012 10:57:26 - granted resource  backup-hcart-robot-tld-0
09/21/2012 10:57:26 - estimated 15002642 kbytes needed
09/21/2012 10:57:26 - begin Parent Job
09/21/2012 10:57:26 - begin Bare Metal Restore: Start Notify Script
09/21/2012 10:57:26 - Info RUNCMD (pid=23519) started
09/21/2012 10:57:26 - Info RUNCMD (pid=23519) exiting with status: 0
Operation Status: 0
09/21/2012 10:57:26 - end Bare Metal Restore: Start Notify Script; elapsed time 0:00:00
09/21/2012 10:57:26 - begin Bare Metal Restore: Bare Metal Restore Save
09/21/2012 10:57:27 - Info bpbrm (pid=23526) KBPRORAC02D001 is the host to backup data from
09/21/2012 10:57:27 - started process bpbrm (pid=23526)
09/21/2012 10:57:29 - collecting BMR information
09/21/2012 10:57:29 - connecting
09/21/2012 10:57:29 - connected; connect time: 0:00:00
09/21/2012 10:57:29 - transfering BMR information to the master server
09/21/2012 10:57:29 - connecting
09/21/2012 10:57:29 - connected; connect time: 0:00:00
09/21/2012 10:57:59 - Error bpbrm (pid=23526) BMRERR: Received BMR error: Failed to import Config file. (1)
09/21/2012 10:57:59 - Error bpbrm (pid=23526) BMRERR: Received BMR error: Failed sending the discovery. (1)
09/21/2012 10:57:59 - Error bpbrm (pid=23526) BMRERR: Received BMR error: BMR information discovery failed. (1)
09/21/2012 10:57:59 - Info bmrsavecfg (pid=0) done. status: 1: the requested operation was partially successful
09/21/2012 10:57:59 - end writing
Operation Status: 1
09/21/2012 10:57:59 - end Bare Metal Restore: Bare Metal Restore Save; elapsed time 0:00:33
09/21/2012 10:57:59 - begin Bare Metal Restore: Policy Execution Manager Preprocessed
Operation Status: 0
09/21/2012 11:07:27 - end Bare Metal Restore: Policy Execution Manager Preprocessed; elapsed time 0:09:28
09/21/2012 11:07:27 - begin Bare Metal Restore: End Notify Script
09/21/2012 11:07:27 - Info RUNCMD (pid=23829) started
09/21/2012 11:07:27 - Info RUNCMD (pid=23829) exiting with status: 0
Operation Status: 0
09/21/2012 11:07:27 - end Bare Metal Restore: End Notify Script; elapsed time 0:00:00
Operation Status: 1
09/21/2012 11:07:27 - end Parent Job; elapsed time 0:10:01
the requested operation was partially successful  (1)
 
 
Merci d'avance.
 
 

Comments 13 CommentsJump to latest comment

mandar_khanolkar's picture

what is your netbackup master server and client version?

Please check debug level 6 logs of bmrsavecfg (generated on client) and bmrd (bmr master server log on your NB master server) to know exact failure cause.

thanks.

mandar

slim_Hannachi's picture
Plateforme Netbackup 7.1.
Master server : OS Solaris 10, Master Natbackup 7.1
Client Os : AIX 6.1,  Netbackup AIX 7.1, 
thank you for your reply
 Explain me please how to check debug level 6 logs of bmrsavecfg (generated on client OS AIX 6.1) and how the bmrd (bmr master server log NB master server on your Solaris 10 OS).
Thinks.
Slim
 
mandar_khanolkar's picture

Find /usr/openv/netbackup/nblog.conf file

Edit it to set debuglevel=6

Set it both on your master as well as on client. Re-run BMR backup to reproduce the problem. Open the logs. In case you are not getting the problem then attach them here.

thanks.

mandar

slim_Hannachi's picture

Hi,

thank you but how to  find the log file.

thanks in advance

mandar_khanolkar's picture

For BMRD (bmr master server log), on NB master server:

/usr/openv/logs/bmrd/*.log

and for bmrsavecfg logs, on your client:

/usr/openv/logs/bmrsavcfg/*.log

thanks.

mandar

slim_Hannachi's picture

Think you,

For BMRD (bmr master server log), on NB master server:

/usr/openv/logs/bmrd/51216-119-2887043588-120924-0000000000.log

and for bmrsavecfg logs, on your client:

/usr/openv/logs/bmrsavcfg/51216-121-2886994312-120924-0000000000.log

thanks.

Slim Hannachi

AttachmentSize
51216-121-2886994312-120924-0000000000.txt 17.4 KB
51216-119-2887043588-120924-0000000000.txt 533 bytes
mandar_khanolkar's picture

BTW the logs you provided are not generated with debug level 6.

The error message in BMRD logs indicates that client bmr configuration import failed while parsing /usr/sbin/lsjfs2 output on your client.

0,51216,128,119,2,1348503960612,15291,1,0:,50:Critical function failed. OutputType = LSJFS2_NUM.,51:ParseClientInformation.cpp:ParseClientInformation(),1
1,51216,128,119,3,1348503960627,15291,1,0:,0:,28:ImportCfg.cpp:ImportConfig(),1,(18|)

Please can you paste "/usr/sbin/lsjfs2" o/p on your client. Also need both bmrd and bmrsavecfg debug level 6 level log by updating both master and client nblog.conf as mentioned earlier. After increasing debugloglevel, re-fire bmr backup so sufficient logs will be generated.

Thanks.

Mandar

slim_Hannachi's picture

hi 

this is the result of /usr/sbin/lsjfs2

# /usr/sbin/lsjfs2
#MountPoint:Device:Vfs:Nodename:Type:Size:Options:AutoMount:Acct:OtherOptions:                            LvSize:FsSize:BlockSize:Sparse:InlineLog:InlineLogSz:EAformat:Quota:DMAPI:VIX:                            EFS:ISNAPSHOT:maxext
/:/dev/hd4:jfs2::bootfs:1048576:rw:oui:non::1048576:1048576:4096:oui:non:0:v1:                            non:non:oui:non:non:0:
/home:/dev/hd1:jfs2:::262144:rw:oui:non::262144:262144:4096:oui:non:0:v1:non:n                            on:oui:non:non:0:
/usr:/dev/hd2:jfs2::bootfs:25165824:rw:oui:non::25165824:25165824:4096:oui:non                            :0:v1:non:non:oui:non:non:0:
/var:/dev/hd9var:jfs2::bootfs:2097152:rw:oui:non::2097152:2097152:4096:oui:non                            :0:v1:non:non:oui:non:non:0:
/tmp:/dev/hd3:jfs2:::10485760:rw:oui:non::10485760:10485760:4096:oui:non:0:v1:                            non:non:oui:non:non:0:
/admin:/dev/hd11admin:jfs2:::262144:rw:oui:non::262144:262144:4096:oui:non:0:v                            1:non:non:oui:non:non:0:
/opt:/dev/hd10opt:jfs2:::786432:rw:oui:non::786432:786432:4096:oui:non:0:v1:no                            n:non:oui:non:non:0:
/var/adm/ras/livedump:/dev/livedump:jfs2:::524288:rw:oui:non::524288:524288:40                            96:oui:non:0:v1:non:non:oui:non:non:0:
/HA6:/dev/fslv00:jfs2:::1048576:rw:oui:non::1048576:1048576:4096:oui:non:0:v1:                            non:non:oui:non:non:0:
/orabin:/dev/fslv01:jfs2:::20971520:rw:oui:non::20971520:20971520:4096:oui:non                            :0:v1:non:non:oui:non:non:0:
 
just one remarque that this client is a cluster HACMP of IBM AIX and this is cluster of Oracle DB 10g with a partage ressources groupe  
# lspv
hdisk0          00f61d76c9dfa1ba                    rootvg          active
hdisk1          00f61d7611ffc3c7                    rootvg          active
hdisk2          00f62562ccf8833a                    vg_db100        concurrent
hdisk3          00f62562ccf8f5ce                    vg_db100        concurrent
hdisk4          00f62562ccf97beb                    vg_db100        concurrent
hdisk5          00f62562ccf9e080                    vg_db100        concurrent
hdisk6          00f61d76cb796b51                    vg_db100        concurrent
hdisk7          00f62562ccfa7ec3                    vg_db100        concurrent
hdisk8          00f61d76cb7a0566                    vg_bkpdb100     concurrent
hdisk9          00f62562ccfb20b1                    vg_db100        concurrent
hdisk10         00f62562ccfbc178                    vg_db100        concurrent
hdisk11         00f61d76cb7ac2ec                    vg_db100        concurrent
# lsdev -Cc disk
hdisk0  Disponible  Unité de disque SCSI virtuelle
hdisk1  Disponible  Unité de disque SCSI virtuelle
hdisk2  Disponible  Unité de disque SCSI virtuelle
hdisk3  Disponible  Unité de disque SCSI virtuelle
hdisk4  Disponible  Unité de disque SCSI virtuelle
hdisk5  Disponible  Unité de disque SCSI virtuelle
hdisk6  Disponible  Unité de disque SCSI virtuelle
hdisk7  Disponible  Unité de disque SCSI virtuelle
hdisk8  Disponible  Unité de disque SCSI virtuelle
hdisk9  Disponible  Unité de disque SCSI virtuelle
hdisk10 Disponible  Unité de disque SCSI virtuelle
hdisk11 Disponible  Unité de disque SCSI virtuelle
 
thinks
slim
 
mandar_khanolkar's picture

>> that this client is a cluster HACMP of IBM AIX

hmmm. this could be causing import issue failing to map file systems.

can you provide /usr/openv/netbackup/baremetal/client/data/bundle.dat file created on your client machine.

thanks.

mandar

slim_Hannachi's picture
Hello,
attached file request "bundle.dat".
thank you to unzip the file bundle.zip
thank you in advance.
 
Thinks
Hannachi Slim
AttachmentSize
bundle.zip 15.9 KB
mandar_khanolkar's picture

From your bundle.dat file, I found that below critical command "lsjfs2" is failed to run successfully on your client during BMR backup. This command tells which are jfs2 file systems on your client machine.

You can see ExitCode is 1.

Still it is listing (complete or partial) list of jfs2 file systems as well. Not sure if it is complete though until and unless you confirm on your client system.

======== failure:==========

[root@V-067819A data]# cat savcfg83
Name:./usr/openv/netbackup/baremetal/client/data/savcfg83
OS:aix
OSLevel:6.1.5.3
BMRVersion:7.1
Command:/usr/sbin/lsjfs2 2>/dev/null
OutputType:7
Key1:
Key2:
Key3:
Key4:
Key5:
ExitCode:1
+++ START BMR DATA +++
#MountPoint:Device:Vfs:Nodename:Type:Size:Options:AutoMount:Acct:OtherOptions:LvSize:FsSize:BlockSize:Sparse:InlineLog:InlineLogSz:EAformat:Quota:DMAPI:VIX:EFS:ISNAPSHOT:maxext
/:/dev/hd4:jfs2::bootfs:1048576:rw:yes:no::1048576:1048576:4096:yes:no:0:v1:no:no:yes:no:no:0:
/home:/dev/hd1:jfs2:::262144:rw:yes:no::262144:262144:4096:yes:no:0:v1:no:no:yes:no:no:0:
/usr:/dev/hd2:jfs2::bootfs:25165824:rw:yes:no::25165824:25165824:4096:yes:no:0:v1:no:no:yes:no:no:0:
/var:/dev/hd9var:jfs2::bootfs:2097152:rw:yes:no::2097152:2097152:4096:yes:no:0:v1:no:no:yes:no:no:0:
/tmp:/dev/hd3:jfs2:::10485760:rw:yes:no::10485760:10485760:4096:yes:no:0:v1:no:no:yes:no:no:0:
/admin:/dev/hd11admin:jfs2:::262144:rw:yes:no::262144:262144:4096:yes:no:0:v1:no:no:yes:no:no:0:
/opt:/dev/hd10opt:jfs2:::786432:rw:yes:no::786432:786432:4096:yes:no:0:v1:no:no:yes:no:no:0:
/var/adm/ras/livedump:/dev/livedump:jfs2:::524288:rw:yes:no::524288:524288:4096:yes:no:0:v1:no:no:yes:no:no:0:
/HA6:/dev/fslv00:jfs2:::1048576:rw:yes:no::1048576:1048576:4096:yes:no:0:v1:no:no:yes:no:no:0:
/orabin:/dev/fslv01:jfs2:::20971520:rw:yes:no::20971520:20971520:4096:yes:no:0:v1:no:no:yes:no:no:0:

======================

Try running below commands manually on your AIX client and confirm output.

a. /usr/sbin/lsjfs2

b. Immediately execute "echo $?" after executing command a.

Provide output here.

Thanks.

Mandar

mandar_khanolkar's picture

You mentioned about exit code 1 but not provided output displayed on screen which i asked earlier. Please provide command output displayed on your screen.

thanks.

mandar