CALLHOME status: Not able to upload callhome data
Hi All, I am getting the following error from one of our netbackup appliance. Can any one please let me know what is the actual problem (how to identify )and how to resolve the issue. I am new to Netbackup appliance :( CALLHOME status: Not able to upload callhome data Error: 500 SSL read timeout: Please contact Symantec Support to resolve this issue. Appliance Model 5200 V2.5 OS : Linux Redhat 7x1.6KViews1like4CommentsSpurious Alerts following Appliance Upgrade to 2.7.2
I have recently upgraded a number of 5230 Appliances to 2.7.2, and whilst all appear to be functioning as expected we are getting numerous alerts relating to disk failures which are 'false'. I am wondering whether there is a BIOS issue here (mismatch) ? Two questions: 1. How can I display the current BIOS level without rebooting the appliance ? 2. Is there a required BIOS update to be applied 'outside' of the regular software/firmware package used for 2.7.2 ? Thanks, AJ.1.9KViews1like4Commentsnetbackup unable to resolve hostname, no hostname alias added to pem name manager.
Hi, I see the below error on my problems report: netbackup unable to resolve hostname <hostname>, no hostname alias added to pem name manager. Now there's about 16 of these servers and they have all been decommissioned and they are not associated with any policy and they are not under client's host properties on the GUI. Upon searching under the /usr/openv/netbackup/db/client directory, all 16 of these client folders are still in there. Is it OK to just remove/delete these client folders on this directory? Or is there another way of getting rid of these error? Regards4.1KViews0likes4CommentsUnable to restore a oracle control file in RHEL
Hi Guys, Its me again :( I´m trying to restore a oracle database according to this technote HOWTO63971 but when I run this command: run { allocate channel c1 type 'sbt_tape'; send 'NB_ORA_SERV=nbu_master, NB_ORA_CLIENT=ora_clnt'; restore controlfile to '/tmp/cntrl.bak' from 'cntrl_624_1_770006541'; release channel c1; } Im trying to restore from 5230 appliance (dp_disk_ecbpprnb4)Solved1.4KViews0likes2CommentsNetBackup Appliance hangs then shuts down
Hi troops, IHAC who found one of the two 5220 2.6.1.1 appliance turned off twice and he had to power it off from the display since IMPI was not responfing either. They opened a case with Symantec but since in the logs there was nothing relevant, Support could not provide a RCA. I looked at the logs myself and apart from the gap in the messages logs I did not find any specific errors. Have you ever seen this ? Below is Support's responce and the messages log, the carsh must have occurre on the 29th August at about 21.42.01 var messages **************** Aug 29 21:42:01 nbami-5220 smbd[745]: [2015/08/29 21:42:01, 0] printing/print_cups.c:463(cups_async_callback) Aug 29 21:42:01 nbami-5220 smbd[745]: failed to retrieve printer list: NT_STATUS_UNSUCCESSFUL Aug 31 08:34:11 nbami-5220 syslog-ng[6216]: syslog-ng starting up; version='2.0.9' Aug 31 08:34:11 nbami-5220 SMmonitor: SMmonitor started. PID=6243 Aug 31 08:34:12 nbami-5220 logger: sisids.init: Using driver version 2.6.32.43/sisfim-x86_64-default.ko Aug 31 08:34:15 localhost StorageArray: ESM AutoSync Mode = Enabled, started on 8/31/15 8:34 AM Aug 31 08:34:15 nbami-5220 syslog-ng[6216]: Configuration reload request received, reloading configuration; Symantec Support reply: ***************************** • Appliance Details: Compute Node nbami-5220 NetBackup Model = 5220 Nbapp-release = 2.6.1.1 Serial Number = SYM1000789 Bios Version = S5500.86B.01.00.0057.031020111721 Firmware = 2.120.63-1242/ 1.41.372-2527 \\pun-evidence.pun.spt.symantec.com\evidence\pun\62\09190462\2015-07-29\ • Node does not have any errors. -> Time Monitoring Ran: Mon Jul 27 2015 13:01:26 UTC • There is nothing logged in the messages log since Jul 25 07:24:01 and Jul 26 19:42:50 Jul 25 07:24:01 nbami-5220 /usr/sbin/cron[7188]: (root) CMD (PATH=/opt/VRTSperl/bin:/sbin:/usr/sbin:$PATH; export PATH; . /opt/NBUAppliance/db/nbappdb_env.sh> /dev/null 2>&1; /opt/NBUAppliance/bin/perl -I/opt/NBUAppliance/scripts/ /opt/NBUAppliance/scripts//hwmon/callhome.pl> /dev/null 2>&1) Reboot occurred here Jul 26 19:42:50 nbami-5220 syslog-ng[6285]: syslog-ng starting up; version='2.0.9' Raid Card FW Term Log shows during that time CC check was running: \\pun-evidence.pun.spt.symantec.com\evidence\pun\62\09190462\2015-07-29\DataCollect\SYM1000789\Raid-Card-FW-Term-Log_07-27-2015_12.55.55.log 07/25/15 7:14:03: EVT#93269-07/25/15 7:14:03: 94=Patrol Read progress on PD 17(e0x18/s16) is 69.99%(14699s) 07/25/15 7:15:13: EVT#93270-07/25/15 7:15:13: 65=Consistency Check progress on VD 00/0 is 10.19%(15313s) 07/25/15 7:15:53: EVT#93271-07/25/15 7:15:53: 65=Consistency Check progress on VD 01/1 is 28.60%(15353s) 07/25/15 7:18:07: EVT#93272-07/25/15 7:18:07: 65=Consistency Check progress on VD 01/1 is 28.85%(15487s) • Alarm-Log_07-27-2015_12.55.55.log shows many repeated messages like below 137 06/30/2015 14:15:29 Voltage #0x16 Upper Non-critical going high 138 06/30/2015 15:34:19 Voltage #0x1c Upper Non-critical going high 155 06/30/2015 18:22:34 Voltage #0x1c Upper Critical going high 156 06/30/2015 18:22:51 Voltage #0x1c Upper Critical going high 197 06/30/2015 20:11:16 Voltage #0x1c Upper Critical going high 198 06/30/2015 20:13:13 Voltage #0x1c Upper Critical going high Getting Logged everyday till 13th July: e36 07/13/2015 03:54:37 Voltage #0x1c Upper Critical going high e37 07/13/2015 03:54:46 Voltage #0x1c Upper Critical going high • Checked on another customers machine. # ipmitool sdr list -verbose Sensor ID : BB +3.3V (0x16) Entity ID : 7.1 (System Board) Sensor Type (Analog) : Voltage Sensor ID : BB -12.0V (0x1c) Entity ID : 7.1 (System Board) Sensor Type (Analog) : Voltage 30 Requested for below outputs from customer # ipmiutil fru > /tmp/ipmi_events/ipmi_events fru.out # ipmiutil fru > /tmp/ipmi_events/ipmi_events fru.out # ipmiutil health > /tmp/ipmi_events/ipmi_events_health.out # ipmiutil sensor > /tmp/ipmi_events/sensor.out # ipmiutil sel > /tmp/ipmi_events/sel.out # ipmitool sdr list –verbose > /tmp/ipmi_events/sd_list_verbose.out # ipmitool sdr elist full > /tmp/ipmi_events/sdr_elist_full.out # ipmitool sdr list compact > /tmp/ipmi_events/srd_list_compact.out # ipmitool chassis status # ipmitool mc selftest 3/8 Customer confirmed that the backups are working fine and operation is normal but one LED on Power Supply is Blinking The backup operation are normal. One LED on Power Supply is blinking asked customer to confirm if the LED is blinking amber • LED is amber o Indication: The current power supply has failed. The other power supply in the unit is on. o Recommended actions: 1. Check if the AC power cord is unplugged. 2. Check if the AC is lost. Critical event causing an AC lost may include: over current protection (OCP), over voltage protection (OVP), over temperature protection (OVP). Customer mentioned that the LED is Blinking Green • LED is blinking green o Indication: AC present, only 12VSB on. Power supply is off or in cold redundant state. Customer provided DataCollect Log which is on fileshare ,, \\pun-evidence.pun.spt.symantec.com\evidence\pun\62\09190462\2015-08-03 06/08 checked datacollect and all hw components are showing ok checking further 14/08 emailed customer to know if the Appliance gone down again since last time.. these are old logs and nothing in particular seen during the time of hang - Also see message like below which may or may not have anything to do with the hung appliance that happened in July . messages/var/log/kernel_messages.out:May 14 11:53:38 nbami-5220 kernel: [ 1799.582514] :megasr[ahci]: port [1] encountered a hard error:[0x40000001] during NCQ operation, performing an explicit reset messages/var/log/kernel_messages.out:May 14 11:53:38 nbami-5220 kernel: [ 1799.582519] :megasr[ahci]: warning PortReset called for port[1] messages/var/log/kernel_messages.out:May 14 11:53:38 nbami-5220 kernel: [ 1799.594016] :megasr[ahci]: device on port:[1] online:[0x123] [0] milliseconds after reset messages/var/log/kernel_messages.out:May 14 11:53:38 nbami-5220 kernel: [ 1800.638618] :megasr[ahci]: converting NCQ hal packet [ffff880037876e48] to N-NCQ for error recovery owner:6 start block:0x681900 messages/var/log/kernel_messages.out:May 14 11:53:38 nbami-5220 kernel: [ 1800.638627] :megasr[ahci]: port [0x1] is paused for read log ext messages/var/log/kernel_messages.out:May 14 11:53:39 nbami-5220 kernel: [ 1800.649104] :megasr[ahci]: port [0] encountered a hard error:[0x40000001] during NCQ operation, performing an explicit reset messages/var/log/kernel_messages.out:May 14 11:53:39 nbami-5220 kernel: [ 1800.649107] :megasr[ahci]: warning PortReset called for port[0] messages/var/log/kernel_messages.out:May 14 11:53:39 nbami-5220 kernel: [ 1800.660604] :megasr[ahci]: device on port:[0] online:[0x123] [0] milliseconds after reset messages/var/log/kernel_messages.out:May 14 11:53:39 nbami-5220 kernel: [ 1801.705206] :megasr[ahci]: converting NCQ hal packet [ffff88003788c868] to N-NCQ for error recovery owner:4 start block:0x3621623a messages/var/log/kernel_messages.out:May 14 11:53:39 nbami-5220 kernel: [ 1801.705211] :megasr[ahci]: port [0x0] is paused for read log ext messages/var/log/kernel_messages.out:May 14 11:53:39 nbami-5220 kernel: [ 1801.705229] :megasr[ahci]: port [0x1] is restarted after read log ext messages/var/log/kernel_messages.out:May 14 11:53:39 nbami-5220 kernel: [ 1801.705637] :megasr[ahci]: port [0x0] is restarted after read log ext messages/var/log/kernel_messages.out:May 14 11:53:59 nbami-5220 kernel: [ 1819.537935] :megasr[ahci]: hal packet [ffff880037872bb8] owner:46 start block:0x1200 timeout detected. messages/var/log/kernel_messages.out:May 14 11:53:59 nbami-5220 kernel: [ 1819.537940] :megasr[ahci]: hal_pkt[ffff880037872bb8:ffff880037872bb8] in port[0] list [1] getting timed out messages/var/log/kernel_messages.out:May 14 11:53:59 nbami-5220 kernel: [ 1819.537945] :megasr[ahci]: hal_pkt[ffff880037872bb8:ffff880037851738] in port[0] list [2] getting timed out messages/var/log/kernel_messages.out:May 14 11:53:59 nbami-5220 kernel: [ 1819.537948] :megasr[ahci]: hal_pkt[ffff880037872bb8:ffff88003786e928] in port[0] list [2] getting timed out messages/var/log/kernel_messages.out:May 14 11:53:59 nbami-5220 kernel: [ 1819.537952] :megasr[ahci]: hal packet [ffff880037887f30] owner:46 start block:0x1200 timeout detected. messages/var/log/kernel_messages.out:May 14 11:54:32 nbami-5220 kernel: [ 1819.537956] :megasr[ahci]: hal_pkt[ffff880037887f30:ffff880037887f30] in port[1] list [1] getting timed out messages/var/log/kernel_messages.out:May 14 11:54:32 nbami-5220 kernel: [ 1819.537959] :megasr[ahci]: warning PortReset called for port[0] wikipedia Native Command Queuing (NCQ) is an extension of the Serial ATA protocol allowing hard disk drives to internally optimize the order in which received read and write commands are executed. This can reduce the amount of unnecessary drive head movement, resulting in increased performance (and slightly decreased wear of the drive) for workloads where multiple simultaneous read/write requests are outstanding, most often occurring in server-type applications. - messages log May 14 showed catalog volume full. messages/var/log/kernel_messages.out:May 14 17:38:25 nbami-5220 kernel: [22454.498786] vxfs: msgcnt 1 mesg 001: V-2-1: vx_nospace - /dev/vx/dsk/nbuapp/pdcatvol file system full (2 block extent) since there is nothing logged at the time of hang ,, unable to provide a root cause analysis. Its been 3 weeks since that the appliance is working fine hence customer gave consent to close the case . They will reference this case in future if happen to see the same scenario again. Thanks for your help, FrancescoSolved2KViews0likes5CommentsNB7.6.0.3 restore to different location issue
Hi. Evnironment: MasterServer: Linux SLES11 RemoteConsoleServer: Windows 2008R2 (virtual machine) - is use for manage environment, backup polices, etc MediaServers & Storage: 2x NetBacakup Appliance5230 NetBackup ver 7.6.0.3 isinstalld on all servers & clients. Problem: the restore is failed on the same host but for different location. 8/26/2015 8:29:49 AM - Info bpbrm(pid=11935) "ClientName" is the host to restore to 8/26/2015 8:29:49 AM - Info bpbrm(pid=11935) reading file list for client 8/26/2015 8:29:50 AM - Error bpbrm(pid=11935) Request from host "MediaServer_Name" to host "RemoteConsoleServer_Name" is not allowed access. Host "MediaSerever_Name" is not an authorized server for host "MasterServer_Name". 8/26/2015 8:29:50 AM - Error bpbrm(pid=11935) bpcd on "MasterServer_Name" exited with status 59: access to the client was not allowed 8/26/2015 8:29:50 AM - Error bpbrm(pid=11935) cannot put rename file on "ClientName" 8/26/2015 8:29:50 AM - Info tar(pid=0) done. status: 59: access to the client was not allowed 8/26/2015 8:29:50 AM - Error bpbrm(pid=11935) client restore EXIT STATUS 59: access to the client was not allowed 8/26/2015 8:41:35 AM - begin Restore 8/26/2015 8:41:35 AM - restoring image "ClientName"_1440178202 8/26/2015 8:41:36 AM - Info bprd(pid=1693) Restoring from copy 1 of image created Fri Aug 21 19:30:02 2015 from policy "PolicyName" 8/26/2015 8:41:36 AM - requesting resource @aaaab 8/26/2015 8:41:36 AM - granted resource MediaID=@aaaab;DiskVolume=PureDiskVolume;DiskPool="DiskPool_Name";Path=PureDiskVolume;StorageServer="StorageServer_Name";MediaServer="MediaServer_Name" 8/26/2015 8:41:39 AM - restored image "ClientName"_1440178202 - (access to the client was not allowed(59)); restore time 0:00:04 8/26/2015 8:41:39 AM - Warning bprd(pid=1693) Restore must be resumed prior to first image expiration on Mon Sep 21 19:30:02 2015 8/26/2015 8:41:39 AM - end Restore; elapsed time: 0:00:04 MS-Windows policy restore error(2808) What I've verified: Host Propeties: Master Server have all media & remote console on Additional Servers list Media Servershave all media & remote console on Additional Servers list Clienthave all media & remote console on Additional Servers list Restore to the source place (original) is succesfully completed(0), but the same error is existing in logs Problem existing for Windows & Linux clients. I don't check NDMP, SQL or VM yet.Solved1.6KViews0likes7Comments5220 as master/media, logs, ESX San method
Hi all, I'm testing this configuration to backup VM on SAN method. My 5220's is configured as Master/Media. Zoning is OK and appliance have access to datastore lun. When I run Vmware policy, i've see VMWare snapshot create but, Netbackup can't retrieve it. Status 6/23 appears in Admin Console. So I've setting loglevel in appliance for VxMS, to 7 but logs aren't very explicit. It seems to rest in log level 1... : Begin VxMS Session... Logmask set to 0x04000000 Current Environment Variables VFM_ROOT = /usr/openv VFM_PRIVATE_ROOT = /usr/openv VFM_MAP_API_LIB = NULL VFM_MAP_DIR = NULL VFM_UTIL_LIB = NULL 08/21/2015 14:33:41 : vdOpen:VixInterface.cpp:480 <ERROR> : VixDiskLib_Open() error: 13 08/21/2015 14:33:41 : vixMapObjCtl:VixCoordinator.cpp:976 <ERROR> : Returning: 23 08/21/2015 14:33:41 : vix_map_objctl:libvix.cpp:1206 <ERROR> : Returning: 23 Do you explain me if my debugg method is correct ? PS : NBD method is functionnal. ThanksSolved1.2KViews2likes8CommentsCatalog Backup Error Status 14
Hi everyone, I am having issues with our Catalog backup after weupgraded to 7.6.1.1. Thisenvironment is a simple 5220 Appliance which acts as our Master/Media server connected to an IBM library on the back end.When executing our Catalog backup everything seems to complete successfully except for the dr file save part of the process. I have verified that I can browse to the dr_file directory and have remapped with credentials successfully. Tried pointing dr file save to another directory which still failed, and also verified that there is plenty of space on the drive to complete. Not sure what else to try. I have included the job details below. Detailed Status 8/19/2015 9:24:11 AM - Info nbjm(pid=11786) starting backup job (jobid=186802) for client XXXXXXXXXXX, policy Catalog-Backup, schedule Catalog-Backup_Full 8/19/2015 9:24:11 AM - Info nbjm(pid=11786) requesting CATALOG_BACKUP_RESOURCE resources from RB for backup job (jobid=186802, request id:{941653FA-4675-11E5-8AFF-F1C3347D7AAD}) 8/19/2015 9:24:11 AM - requesting resource XXXXXXXXXXX.NBU_CATALOG.MAXJOBS 8/19/2015 9:24:11 AM - granted resource XXXXXXXXXXX.NBU_CATALOG.MAXJOBS 8/19/2015 9:24:11 AM - estimated 0 Kbytes needed 8/19/2015 9:24:11 AM - begin Parent Job 8/19/2015 9:24:11 AM - begin Catalog Backup, Start Notify Script 8/19/2015 9:24:11 AM - Info RUNCMD(pid=25799) started 8/19/2015 9:24:11 AM - Info RUNCMD(pid=25799) exiting with status: 0 Status 0 8/19/2015 9:24:11 AM - end Catalog Backup, Start Notify Script; elapsed time: 0:00:00 8/19/2015 9:24:11 AM - begin Catalog Backup, DBM Query 8/19/2015 9:29:26 AM - Error bpdbm(pid=25803) Unable to save disaster recovery information to /advdisk/dr_file: file write failed (14) Status 14 8/19/2015 9:29:27 AM - end Catalog Backup, DBM Query; elapsed time: 0:05:16 8/19/2015 9:29:27 AM - begin Catalog Backup, Stop On Error Status 0 8/19/2015 9:29:27 AM - end Catalog Backup, Stop On Error; elapsed time: 0:00:00 8/19/2015 9:29:27 AM - begin Catalog Backup, End Notify Script 8/19/2015 9:29:27 AM - Info RUNCMD(pid=26607) started 8/19/2015 9:29:27 AM - Info RUNCMD(pid=26607) exiting with status: 0 Status 0 8/19/2015 9:29:27 AM - end Catalog Backup, End Notify Script; elapsed time: 0:00:00 Status 14 8/19/2015 9:29:27 AM - end Parent Job; elapsed time: 0:05:16 file write failed(14)1KViews0likes2CommentsNetbackup and vCenter: Failed VM backup
Good day to all. I am new to Netbackup appliance, i would like to ask help how to fix the VM policy which failed to backup my VMs via vCenter 5.5. Configurationthat I did are; 1. Netbackup Appliance 5230 2.6.1.2 and NBU 7.6.1.2, configured as Master and Media Server. 2. Set the appliance as the VM backup host. 3. Set the vCenterCredential 4. Install NBU client 7.6.1.2 in vCenter server, define the NBU 5230 name as the master server. 5. Add vCenter IP and hostname in NBU 5230 host list 6. Created VM Policy, add the vCenter server name. - add client, schedule 7. No firewall between the appliance and vCenter. 8. NBD is the transport mode. 9. Installed and use Java admin console 7.6.1.2 to access the appliance from another VM machine. After I run the policy, I got the errro message "status: 6: the backup failed to back up the requested files" ===== 08/14/2015 17:00:46 - Info nbjm (pid=11458) starting backup job (jobid=127) for client MHQICT3100, policy test2, schedule incr 08/14/2015 17:00:46 - estimated 0 kbytes needed 08/14/2015 17:00:46 - Info nbjm (pid=11458) started backup (backupid=MHQICT3100_1439542846) job for client MHQICT3100, policy test2, schedule incr on storage unit stu_disk_prims-nbmaster-prod using backup host PRIMS-NBMASTER-PROD 08/14/2015 17:00:47 - Info bpbrm (pid=83138) MHQICT3100 is the host to backup data from 08/14/2015 17:00:47 - Info bpbrm (pid=83138) reading file list for client 08/14/2015 17:00:47 - Info bpbrm (pid=83138) accelerator enabled 08/14/2015 17:00:47 - Info bpbrm (pid=83138) Accelerator enabled backup with "Accelerator forced rescan", all data will be scanned and processed.Backup time will be longer than a normal Accelerator enabled backup. 08/14/2015 17:00:47 - Info bpbrm (pid=83138) There is no complete backup image match with track journal, a regular full backup will be performed. 08/14/2015 17:00:47 - Info bpbrm (pid=83138) starting bpbkar on client 08/14/2015 17:00:47 - started process bpbrm (pid=83138) 08/14/2015 17:00:47 - connecting 08/14/2015 17:00:47 - connected; connect time: 0:00:00 08/14/2015 17:00:47 - Info bpbkar (pid=83152) Backup started 08/14/2015 17:00:47 - Info bpbrm (pid=83138) bptm pid: 83161 08/14/2015 17:00:48 - Info bptm (pid=83161) start 08/14/2015 17:00:48 - Info bptm (pid=83161) using 262144 data buffer size 08/14/2015 17:00:48 - Info bptm (pid=83161) using 30 data buffers 08/14/2015 17:00:49 - Info bptm (pid=83161) start backup 08/14/2015 17:00:50 - begin writing 08/14/2015 17:00:55 - Error bpbrm (pid=83138) from client MHQICT3100: ERR - Error opening the snapshot disks using given transport mode: nbd Status 23 08/14/2015 17:00:56 - Critical bpbrm (pid=83138) from client MHQICT3100: FTL - cleanup() failed, status 6 08/14/2015 17:00:58 - Error bptm (pid=83161) media manager terminated by parent process 08/14/2015 17:01:03 - Info prims-nbmaster-prod (pid=83161) StorageServer=PureDisk:prims-nbmaster-prod; Report=PDDO Stats for (prims-nbmaster-prod): scanned: 2 KB, CR sent: 0 KB, CR sent over FC: 0 KB, dedup: 100.0%, cache disabled 08/14/2015 17:01:03 - Info bpbkar (pid=0) done. status: 6: the backup failed to back up the requested files 08/14/2015 17:01:03 - end writing; write time: 0:00:13 the backup failed to back up the requested files (6) ============ Appreciate much any advise. Thanks in advance.Solved1.5KViews0likes1CommentNetbackup appliance 5220 Fan failure error
Hi All, Can someone please guide on the below alert i have received today not sure If I need to escalate Symantec support to check it. +-------+ | Enclosure 1 Fan Information | |+-----+| ||ID | Status | Speed | LowWaterMark | State | Acknowledge || ||---+---+--------+---------++-------|| ||1 |Critical |0 RPM |2000 RPM |Failed |No || ||---+---+--------+---------++-------|| ||2 |Critical |0 RPM |2000 RPM |Failed |No || |+-----+| | Enclosure 1 Power Supply Information | |+-----+| || ID | Status | State | Acknowledge || ||-----+-+--------+--------|| ||1 |Critical |Warning |No || |+-----+| | Enclosure 1 Temperature Information | |+-----+| ||ID| Type | Temperature | HighWaterMark | State | Acknowledge || ||--+---------+---+------+-------+---|| ||1 |Backplane Temp 1 |- degrees C |- degrees C |Failed |No || ||--+---------+---+------+-------+---|| ||2 |Backplane Temp 2 |- degrees C |- degrees C |Failed |No || |+-----+| +-------+Solved1.1KViews0likes3Comments