Scenario: see output below
Hardware:
V480 w/ two SCSI controllers (host)
A1000 (x2: disk arrays)
Configuration:
A1000:A and A1000:B are mirrored via disksuite
++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
One LUN (LUN0/disk) failed and the hotspare on Raid Manager 6.2 and Disksuite successfully picked-up
the failed LUN.
Can I safely assume that when I replace the failed HD (LUN 0) on the A1000, it should self-heal?
The hotspare on the A1000 successfully took over the failed LUN.
Here is what I'm going to do in disksuite to relieve the hotspare from the failed submirrors (d111 and d112).
a. metadb -d c3t5d0s4 (delete the replicas which has Write errors)
b. metadetach d110 d111 (detach the submirror where the hotspare is currently occupying)
c. metaclear -f d111 (completely remove the submirror)
d. boot to single user mode and replace the failed HD (LUN0) on A1000:A
e. prtvtoc -h /dev/dsk/c2t5d1s2 | fmthard -s - /dev/dsk/c3t5d0s2 (re-partion the newly replaced disk)
f. re-create the file system
newfs -r 10000 -i 2097152 -f 8192 /dev/rdsk/c3t5d0s6
newfs -r 10000 -i 2097152 -f 8192 /dev/rdsk/c3t5d0s4
g. metadb -a -f -c 3 c3t5d0s4 (re-create the database replicas)
h. metainit d111 1 1 /dev/dsk/c3t5d0s6 (re-create the submirror)
i. metattach d110 d111 (re-attach the newly re-created submirror)
j. wait for resync and do a boot -r from obp prompt
++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
Disksuite (Solaris Volume Manager) Output
=========================================
d110: Mirror
Submirror 0: d111
State: Okay
Submirror 1: d112
State: Okay
Pass: 1
Read option: roundrobin (default)
Write option: parallel (default)
Size: 285147392 blocks (135 GB)
d111: Submirror of d110
State: Okay
Hot spare pool: hsp001
Size: 285147392 blocks (135 GB)
Stripe 0:
Device Start Block Dbase State Reloc Hot Spare
c2t5d5s6 0 No Okay Yes
d112: Submirror of d110
State: Okay
Hot spare pool: hsp002
Size: 285147392 blocks (135 GB)
Stripe 0:
Device Start Block Dbase State Reloc Hot Spare
c3t5d5s6 0 No Okay Yes
==============================================================
hsp002: 2 hot spares
Device Status Length Reloc
c3t5d6s6 In use 285147392 blocks Yes
c2t5d6s6 Available 285147392 blocks Yes
hsp001: 2 hot spares
Device Status Length Reloc
c2t5d6s6 Available 285147392 blocks Yes
c3t5d6s6 In use 285147392 blocks Yes
==============================================================
# metadb
flags first blk block count
a m p luo 16 8192 /dev/dsk/c1t1d0s4
a p luo 8208 8192 /dev/dsk/c1t1d0s4
a p luo 16400 8192 /dev/dsk/c1t1d0s4
a p luo 16 8192 /dev/dsk/c1t0d0s4
a p luo 8208 8192 /dev/dsk/c1t0d0s4
a p luo 16400 8192 /dev/dsk/c1t0d0s4
a p luo 16 8192 /dev/dsk/c2t5d0s4
a p luo 8208 8192 /dev/dsk/c2t5d0s4
a p luo 16400 8192 /dev/dsk/c2t5d0s4
a p luo 16 8192 /dev/dsk/c2t5d1s4
a p luo 8208 8192 /dev/dsk/c2t5d1s4
a p luo 16400 8192 /dev/dsk/c2t5d1s4
a p luo 16 8192 /dev/dsk/c2t5d2s4
a p luo 8208 8192 /dev/dsk/c2t5d2s4
a p luo 16400 8192 /dev/dsk/c2t5d2s4
a p luo 16 8192 /dev/dsk/c2t5d3s4
a p luo 8208 8192 /dev/dsk/c2t5d3s4
a p luo 16400 8192 /dev/dsk/c2t5d3s4
a p luo 16 8192 /dev/dsk/c2t5d4s4
a p luo 8208 8192 /dev/dsk/c2t5d4s4
a p luo 16400 8192 /dev/dsk/c2t5d4s4
a p luo 16 8192 /dev/dsk/c2t5d5s4
a p luo 8208 8192 /dev/dsk/c2t5d5s4
a p luo 16400 8192 /dev/dsk/c2t5d5s4
a p luo 16 8192 /dev/dsk/c2t5d6s4
a p luo 8208 8192 /dev/dsk/c2t5d6s4
a p luo 16400 8192 /dev/dsk/c2t5d6s4
W p l 16 8192 /dev/dsk/c3t5d0s4
W p l 8208 8192 /dev/dsk/c3t5d0s4
W p l 16400 8192 /dev/dsk/c3t5d0s4
a p luo 16 8192 /dev/dsk/c3t5d1s4
a p luo 8208 8192 /dev/dsk/c3t5d1s4
a p luo 16400 8192 /dev/dsk/c3t5d1s4
a p luo 16 8192 /dev/dsk/c3t5d2s4
a p luo 8208 8192 /dev/dsk/c3t5d2s4
a p luo 16400 8192 /dev/dsk/c3t5d2s4
a p luo 16 8192 /dev/dsk/c3t5d3s4
a p luo 8208 8192 /dev/dsk/c3t5d3s4
a p luo 16400 8192 /dev/dsk/c3t5d3s4
a p luo 16 8192 /dev/dsk/c3t5d4s4
a p luo 8208 8192 /dev/dsk/c3t5d4s4
a p luo 16400 8192 /dev/dsk/c3t5d4s4
a p luo 16 8192 /dev/dsk/c3t5d5s4
a p luo 8208 8192 /dev/dsk/c3t5d5s4
a p luo 16400 8192 /dev/dsk/c3t5d5s4
a p luo 16 8192 /dev/dsk/c3t5d6s4
a p luo 8208 8192 /dev/dsk/c3t5d6s4
a p luo 16400 8192 /dev/dsk/c3t5d6s4
===============================================================================
Raid Manager Output
===================
# healthck -a
Health Check Summary Information
xxx_001: Optimal
xxx_002: Dead LUN at Drive [1,0]
healthck succeeded!
====================================================
# raidutil -c c3t5d0 -i
LUNs found on c3t5d0.
LUN 0 RAID 0 69967 MB
LUN 1 RAID 0 69967 MB
LUN 2 RAID 0 139934 MB
LUN 3 RAID 0 139934 MB
LUN 4 RAID 0 139934 MB
LUN 5 RAID 0 139934 MB
LUN 6 RAID 0 139934 MB
Vendor ID Symbios
ProductID StorEDGE A1000
Product Revision 0301
Boot Level 03.01.04.00
Boot Level Date 04/05/01
Firmware Level 03.01.04.68
Firmware Date 06/22/01
raidutil succeeded!
======================================================
# drivutil -i c3t5d0
Drive Information for xxx_002
Location Capacity Status Vendor Product Firmware Serial
(MB) ID Version Number
[1,0] 0 Failed
[2,0] 70007 Optimal SEAGATE ST373307LSUN72G 0507 3HZ7L7QZ00
[1,1] 70007 Optimal SEAGATE ST373307LSUN72G 0507 3HZ6XNL800
[2,1] 70007 Optimal SEAGATE ST373307LSUN72G 0507 3HZY5Y3T00
[1,2] 70007 Optimal SEAGATE ST373307LSUN72G 0507 3HZ14S0M00
[2,2] 70007 Optimal SEAGATE ST373307LSUN72G 0507 3HZY33C200
[1,3] 70007 Optimal SEAGATE ST373307LSUN72G 0507 3HZ7GBPG00
[2,3] 70007 Optimal SEAGATE ST373307LSUN72G 0507 3HZ6XM8700
[1,4] 70007 Optimal SEAGATE ST373307LSUN72G 0507 3HZ0JNM1ÿÿ
[2,4] 70007 Optimal SEAGATE ST373307LSUN72G 0507 3HZ1AFN5ÿÿ
[1,5] 70007 Optimal SEAGATE ST373307LSUN72G 0507 3HZY6DN000
[2,5] 70007 Optimal SEAGATE ST373307LSUN72G 0507 3HZX1S8M00
=====================================================
nsect = 64
Part Tag Flag Cylinders Size Blocks
0 unassigned wm 0 0 (0/0/0) 0
1 unassigned wm 0 0 (0/0/0) 0
2 backup wu 0 - 34980 68.32GB (34981/0/0) 143282176
3 unassigned wm 0 0 (0/0/0) 0
4 unassigned wm 0 - 12 26.00MB (13/0/0) 53248
5 unassigned wm 0 0 (0/0/0) 0
6 unassigned wm 13 - 34980 68.30GB (34968/0/0) 143228928
7 unassigned wm 0 0 (0/0/0) 0
==================================================