cancel
Showing results for 
Search instead for 
Did you mean: 

kernel: st2: Error 18 (driver bt 0x0, host bt 0x0)

Marcelko
Level 4

Hello,

i didn't find any solution on forums

i would like to open top by sharing of tape drives between 2 media servers

i understand that reservation conflict is normal behaviour, few times per day or one daily we have error 18 or Control daemon connect or protocol error

on SAN switch is set failover for tape drives

media servers are RedHat with newest kernel and nbu is 7.5.0.7

the tape drives are never down and are working properly

have anybody suggestion?

 

Mar  6 05:00:05 gcs-amsdc1-nbmed02 kernel: st 1:0:1:0: reservation conflict
Mar  6 05:00:05 gcs-amsdc1-nbmed02 kernel: st 1:0:2:0: reservation conflict
Mar  6 05:00:05 gcs-amsdc1-nbmed02 kernel: st2: Error 18 (driver bt 0x0, host bt 0x0).
Mar  6 05:00:05 gcs-amsdc1-nbmed02 kernel: st 1:0:3:0: reservation conflict
Mar  6 05:00:05 gcs-amsdc1-nbmed02 kernel: st 1:0:3:0: reservation conflict

 

# more messages | grep "Error 18"
Mar  2 05:00:04 gcs-amsdc1-nbmed02 kernel: st3: Error 18 (driver bt 0x0, host bt 0x0).
Mar  4 05:00:05 gcs-amsdc1-nbmed02 kernel: st0: Error 18 (driver bt 0x0, host bt 0x0).
Mar  4 05:00:05 gcs-amsdc1-nbmed02 kernel: st2: Error 18 (driver bt 0x0, host bt 0x0).
Mar  5 05:00:05 gcs-amsdc1-nbmed02 kernel: st0: Error 18 (driver bt 0x0, host bt 0x0).
Mar  5 05:00:05 gcs-amsdc1-nbmed02 kernel: st1: Error 18 (driver bt 0x0, host bt 0x0).
Mar  6 05:00:05 gcs-amsdc1-nbmed02 kernel: st2: Error 18 (driver bt 0x0, host bt 0x0).

 

]# more messages | grep DOWN
Mar  2 13:56:33 gcs-amsdc1-nbmed02 tldd[14472]: TLD(1) going to DOWN state, status: Control daemon connect or protocol error
Mar  2 13:56:34 gcs-amsdc1-nbmed02 ltid[14446]: Request for media ID D177L5 is being rejected because mount requests are disabled (reason = robotic daemon going to DOWN state)
Mar  2 13:56:34 gcs-amsdc1-nbmed02 ltid[14446]: Request for media ID D146L5 is being rejected because mount requests are disabled (reason = robotic daemon going to DOWN state)
Mar  2 13:56:34 gcs-amsdc1-nbmed02 tldd[14472]: TLD(1) going to DOWN state, status: Control daemon connect or protocol error
Mar  2 14:00:11 gcs-amsdc1-nbmed02 tldd[14472]: TLD(1) going to DOWN state, status: Control daemon connect or protocol error
Mar  2 14:00:17 gcs-amsdc1-nbmed02 tldd[14472]: TLD(1) going to DOWN state, status: Control daemon connect or protocol error

 

more messages | grep error
Mar  2 13:56:33 gcs-amsdc1-nbmed02 tldd[14472]: TLD(1) going to DOWN state, status: Control daemon connect or protocol error
Mar  2 13:56:34 gcs-amsdc1-nbmed02 tldd[14472]: TLD(1) going to DOWN state, status: Control daemon connect or protocol error
Mar  2 14:00:11 gcs-amsdc1-nbmed02 tldd[14472]: TLD(1) going to DOWN state, status: Control daemon connect or protocol error
Mar  2 14:00:17 gcs-amsdc1-nbmed02 tldd[14472]: TLD(1) going to DOWN state, status: Control daemon connect or protocol error
Mar  2 14:02:11 gcs-amsdc1-nbmed02 tldd[14472]: TLD(1) unavailable: initialization failed: Control daemon connect or protocol error
Mar  2 14:07:42 gcs-amsdc1-nbmed02 tldd[14472]: TLD(1) unavailable: initialization failed: Control daemon connect or protocol error
Mar  2 14:13:13 gcs-amsdc1-nbmed02 tldd[14472]: TLD(1) unavailable: initialization failed: Control daemon connect or protocol error
Mar  2 14:18:44 gcs-amsdc1-nbmed02 tldd[14472]: TLD(1) unavailable: initialization failed: Control daemon connect or protocol error
Mar  2 14:24:16 gcs-amsdc1-nbmed02 tldd[14472]: TLD(1) unavailable: initialization failed: Control daemon connect or protocol error

 

 

 

23 REPLIES 23

Marcelko
Level 4

multipath to robotic path is ok

i re-added SSO license and rebooted media servers - because it was added and configured shared tape drives without reboot of media servers

now we dont see any problems

#cat messages

Mar 16 08:39:20 gcs-amsd2a-nbmed01_bck SEC_TEST: gcs-amsdc1-nbmed01 150316083920_67551
Mar 16 08:44:20 gcs-amsd2a-nbmed01_bck SEC_TEST: gcs-amsdc1-nbmed01 150316084420_71661
Mar 16 08:45:10 gcs-amsd2a-nbmed01_bck ntpd[7130]: synchronized to 160.118.49.21, stratum 2
Mar 16 08:49:20 gcs-amsd2a-nbmed01_bck SEC_TEST: gcs-amsdc1-nbmed01 150316084920_73583
Mar 16 08:54:20 gcs-amsd2a-nbmed01_bck SEC_TEST: gcs-amsdc1-nbmed01 150316085420_76897
Mar 16 08:59:20 gcs-amsd2a-nbmed01_bck SEC_TEST: gcs-amsdc1-nbmed01 150316085920_78722
Mar 16 09:04:20 gcs-amsd2a-nbmed01_bck SEC_TEST: gcs-amsdc1-nbmed01 150316090420_718
Mar 16 09:09:20 gcs-amsd2a-nbmed01_bck SEC_TEST: gcs-amsdc1-nbmed01 150316090920_2723
Mar 16 09:14:20 gcs-amsd2a-nbmed01_bck SEC_TEST: gcs-amsdc1-nbmed01 150316091420_6428
Mar 16 09:19:20 gcs-amsd2a-nbmed01_bck SEC_TEST: gcs-amsdc1-nbmed01 150316091920_8225
Mar 16 09:24:19 gcs-amsd2a-nbmed01_bck SEC_TEST: gcs-amsdc1-nbmed01 150316092419_11965
Mar 16 09:29:19 gcs-amsd2a-nbmed01_bck SEC_TEST: gcs-amsdc1-nbmed01 150316092919_13865
Mar 16 09:34:19 gcs-amsd2a-nbmed01_bck SEC_TEST: gcs-amsdc1-nbmed01 150316093419_17138
Mar 16 09:39:19 gcs-amsd2a-nbmed01_bck SEC_TEST: gcs-amsdc1-nbmed01 150316093919_19030
Mar 16 09:44:19 gcs-amsd2a-nbmed01_bck SEC_TEST: gcs-amsdc1-nbmed01 150316094419_22272
Mar 16 09:49:19 gcs-amsd2a-nbmed01_bck SEC_TEST: gcs-amsdc1-nbmed01 150316094919_24121
Mar 16 09:54:19 gcs-amsd2a-nbmed01_bck SEC_TEST: gcs-amsdc1-nbmed01 150316095419_27967
Mar 16 09:59:19 gcs-amsd2a-nbmed01_bck SEC_TEST: gcs-amsdc1-nbmed01 150316095919_29704
Mar 16 10:04:19 gcs-amsd2a-nbmed01_bck SEC_TEST: gcs-amsdc1-nbmed01 150316100419_33266
Mar 16 10:09:19 gcs-amsd2a-nbmed01_bck SEC_TEST: gcs-amsdc1-nbmed01 150316100919_35505
Mar 16 10:14:19 gcs-amsd2a-nbmed01_bck SEC_TEST: gcs-amsdc1-nbmed01 150316101419_38938
Mar 16 10:14:44 gcs-amsd2a-nbmed01_bck ntpd[7130]: synchronized to 160.118.49.23, stratum 2
Mar 16 10:19:19 gcs-amsd2a-nbmed01_bck SEC_TEST: gcs-amsdc1-nbmed01 150316101919_40655
Mar 16 10:24:19 gcs-amsd2a-nbmed01_bck SEC_TEST: gcs-amsdc1-nbmed01 150316102419_43852
Mar 16 10:29:19 gcs-amsd2a-nbmed01_bck SEC_TEST: gcs-amsdc1-nbmed01 150316102919_45569
Mar 16 10:34:19 gcs-amsd2a-nbmed01_bck SEC_TEST: gcs-amsdc1-nbmed01 150316103419_49139

 

Marcelko
Level 4

thanks sdo

it will be very helpfull by next troubleshooting

thank you

Marianne
Moderator
Moderator
Partner    VIP    Accredited Certified

So - my guess was right... As per the first line in my post of 7 March:

Something very wrong with your config. If SSO is setup properly you shouldn't see reservation conflicts.

as many people contacted me about the issue

i have  kernel: st 1:0:7:0: reservation conflict very offen

i found out that i have media server med01, where is robotic controller and med02, which has shared tape drives with med01

solution: when med01 is rebooted, then reservation conflict appear on med02, so med02 has to be rebooted as well after med01 and errors are gone