Tek-Tips is the largest IT community on the Internet today!

Members share and learn making Tek-Tips Forums the best source of peer-reviewed technical information on the Internet!

  • Congratulations derfloh on being selected by the Tek-Tips community for having the most helpful posts in the forums last week. Way to Go!

Diskuite/Raid Manager question (hostpare)

Status
Not open for further replies.

ponetguy2

MIS
Joined
Aug 28, 2002
Messages
442
Location
US


Scenario: see output below

Hardware:

V480 w/ two SCSI controllers (host)
A1000 (x2: disk arrays)

Configuration:

A1000:A and A1000:B are mirrored via disksuite

++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++

One LUN (LUN0/disk) failed and the hotspare on Raid Manager 6.2 and Disksuite successfully picked-up
the failed LUN.

Can I safely assume that when I replace the failed HD (LUN 0) on the A1000, it should self-heal?
The hotspare on the A1000 successfully took over the failed LUN.

Here is what I'm going to do in disksuite to relieve the hotspare from the failed submirrors (d111 and d112).

a. metadb -d c3t5d0s4 (delete the replicas which has Write errors)

b. metadetach d110 d111 (detach the submirror where the hotspare is currently occupying)

c. metaclear -f d111 (completely remove the submirror)

d. boot to single user mode and replace the failed HD (LUN0) on A1000:A

e. prtvtoc -h /dev/dsk/c2t5d1s2 | fmthard -s - /dev/dsk/c3t5d0s2 (re-partion the newly replaced disk)

f. re-create the file system

newfs -r 10000 -i 2097152 -f 8192 /dev/rdsk/c3t5d0s6
newfs -r 10000 -i 2097152 -f 8192 /dev/rdsk/c3t5d0s4

g. metadb -a -f -c 3 c3t5d0s4 (re-create the database replicas)

h. metainit d111 1 1 /dev/dsk/c3t5d0s6 (re-create the submirror)

i. metattach d110 d111 (re-attach the newly re-created submirror)

j. wait for resync and do a boot -r from obp prompt

++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++


Disksuite (Solaris Volume Manager) Output
=========================================

d110: Mirror
Submirror 0: d111
State: Okay
Submirror 1: d112
State: Okay
Pass: 1
Read option: roundrobin (default)
Write option: parallel (default)
Size: 285147392 blocks (135 GB)

d111: Submirror of d110
State: Okay
Hot spare pool: hsp001
Size: 285147392 blocks (135 GB)
Stripe 0:
Device Start Block Dbase State Reloc Hot Spare
c2t5d5s6 0 No Okay Yes


d112: Submirror of d110
State: Okay
Hot spare pool: hsp002
Size: 285147392 blocks (135 GB)
Stripe 0:
Device Start Block Dbase State Reloc Hot Spare
c3t5d5s6 0 No Okay Yes

==============================================================

hsp002: 2 hot spares
Device Status Length Reloc
c3t5d6s6 In use 285147392 blocks Yes
c2t5d6s6 Available 285147392 blocks Yes

hsp001: 2 hot spares
Device Status Length Reloc
c2t5d6s6 Available 285147392 blocks Yes
c3t5d6s6 In use 285147392 blocks Yes

==============================================================

# metadb
flags first blk block count
a m p luo 16 8192 /dev/dsk/c1t1d0s4
a p luo 8208 8192 /dev/dsk/c1t1d0s4
a p luo 16400 8192 /dev/dsk/c1t1d0s4
a p luo 16 8192 /dev/dsk/c1t0d0s4
a p luo 8208 8192 /dev/dsk/c1t0d0s4
a p luo 16400 8192 /dev/dsk/c1t0d0s4
a p luo 16 8192 /dev/dsk/c2t5d0s4
a p luo 8208 8192 /dev/dsk/c2t5d0s4
a p luo 16400 8192 /dev/dsk/c2t5d0s4
a p luo 16 8192 /dev/dsk/c2t5d1s4
a p luo 8208 8192 /dev/dsk/c2t5d1s4
a p luo 16400 8192 /dev/dsk/c2t5d1s4
a p luo 16 8192 /dev/dsk/c2t5d2s4
a p luo 8208 8192 /dev/dsk/c2t5d2s4
a p luo 16400 8192 /dev/dsk/c2t5d2s4
a p luo 16 8192 /dev/dsk/c2t5d3s4
a p luo 8208 8192 /dev/dsk/c2t5d3s4
a p luo 16400 8192 /dev/dsk/c2t5d3s4
a p luo 16 8192 /dev/dsk/c2t5d4s4
a p luo 8208 8192 /dev/dsk/c2t5d4s4
a p luo 16400 8192 /dev/dsk/c2t5d4s4
a p luo 16 8192 /dev/dsk/c2t5d5s4
a p luo 8208 8192 /dev/dsk/c2t5d5s4
a p luo 16400 8192 /dev/dsk/c2t5d5s4
a p luo 16 8192 /dev/dsk/c2t5d6s4
a p luo 8208 8192 /dev/dsk/c2t5d6s4
a p luo 16400 8192 /dev/dsk/c2t5d6s4
W p l 16 8192 /dev/dsk/c3t5d0s4
W p l 8208 8192 /dev/dsk/c3t5d0s4
W p l 16400 8192 /dev/dsk/c3t5d0s4
a p luo 16 8192 /dev/dsk/c3t5d1s4
a p luo 8208 8192 /dev/dsk/c3t5d1s4
a p luo 16400 8192 /dev/dsk/c3t5d1s4
a p luo 16 8192 /dev/dsk/c3t5d2s4
a p luo 8208 8192 /dev/dsk/c3t5d2s4
a p luo 16400 8192 /dev/dsk/c3t5d2s4
a p luo 16 8192 /dev/dsk/c3t5d3s4
a p luo 8208 8192 /dev/dsk/c3t5d3s4
a p luo 16400 8192 /dev/dsk/c3t5d3s4
a p luo 16 8192 /dev/dsk/c3t5d4s4
a p luo 8208 8192 /dev/dsk/c3t5d4s4
a p luo 16400 8192 /dev/dsk/c3t5d4s4
a p luo 16 8192 /dev/dsk/c3t5d5s4
a p luo 8208 8192 /dev/dsk/c3t5d5s4
a p luo 16400 8192 /dev/dsk/c3t5d5s4
a p luo 16 8192 /dev/dsk/c3t5d6s4
a p luo 8208 8192 /dev/dsk/c3t5d6s4
a p luo 16400 8192 /dev/dsk/c3t5d6s4

===============================================================================

Raid Manager Output
===================

# healthck -a

Health Check Summary Information

xxx_001: Optimal
xxx_002: Dead LUN at Drive [1,0]

healthck succeeded!

====================================================

# raidutil -c c3t5d0 -i
LUNs found on c3t5d0.
LUN 0 RAID 0 69967 MB
LUN 1 RAID 0 69967 MB
LUN 2 RAID 0 139934 MB
LUN 3 RAID 0 139934 MB
LUN 4 RAID 0 139934 MB
LUN 5 RAID 0 139934 MB
LUN 6 RAID 0 139934 MB

Vendor ID Symbios
ProductID StorEDGE A1000
Product Revision 0301
Boot Level 03.01.04.00
Boot Level Date 04/05/01
Firmware Level 03.01.04.68
Firmware Date 06/22/01
raidutil succeeded!

======================================================

# drivutil -i c3t5d0

Drive Information for xxx_002


Location Capacity Status Vendor Product Firmware Serial
(MB) ID Version Number
[1,0] 0 Failed
[2,0] 70007 Optimal SEAGATE ST373307LSUN72G 0507 3HZ7L7QZ00
[1,1] 70007 Optimal SEAGATE ST373307LSUN72G 0507 3HZ6XNL800
[2,1] 70007 Optimal SEAGATE ST373307LSUN72G 0507 3HZY5Y3T00
[1,2] 70007 Optimal SEAGATE ST373307LSUN72G 0507 3HZ14S0M00
[2,2] 70007 Optimal SEAGATE ST373307LSUN72G 0507 3HZY33C200
[1,3] 70007 Optimal SEAGATE ST373307LSUN72G 0507 3HZ7GBPG00
[2,3] 70007 Optimal SEAGATE ST373307LSUN72G 0507 3HZ6XM8700
[1,4] 70007 Optimal SEAGATE ST373307LSUN72G 0507 3HZ0JNM1ÿÿ
[2,4] 70007 Optimal SEAGATE ST373307LSUN72G 0507 3HZ1AFN5ÿÿ
[1,5] 70007 Optimal SEAGATE ST373307LSUN72G 0507 3HZY6DN000
[2,5] 70007 Optimal SEAGATE ST373307LSUN72G 0507 3HZX1S8M00

=====================================================

nsect = 64
Part Tag Flag Cylinders Size Blocks
0 unassigned wm 0 0 (0/0/0) 0
1 unassigned wm 0 0 (0/0/0) 0
2 backup wu 0 - 34980 68.32GB (34981/0/0) 143282176
3 unassigned wm 0 0 (0/0/0) 0
4 unassigned wm 0 - 12 26.00MB (13/0/0) 53248
5 unassigned wm 0 0 (0/0/0) 0
6 unassigned wm 13 - 34980 68.30GB (34968/0/0) 143228928
7 unassigned wm 0 0 (0/0/0) 0

==================================================
 
correction. no hot spare setup on a1000/raid manager.
 
# drivutil -l c3t5d0
Logical Unit Information for xxx_002

LUN Group Device RAID Capacity Status
Name Level (MB)

0 1 c3t5d0 0 69967 Dead
1 2 c3t5d1 0 69967 Optimal
2 3 c3t5d2 0 139934 Optimal
3 4 c3t5d3 0 139934 Optimal
4 5 c3t5d4 0 139934 Optimal
5 6 c3t5d5 0 139934 Optimal
6 7 c3t5d6 0 139934 Optimal
 
Why not just use 'metareplace -e d111 c3t5d0s6' if you just need to replaced a failed drive?
 
Please forgive my ignorance. I'm still new to SDS and RM 6.2.

One LUN (LUN0/disk 1,0) failed on one of the A1000 and the there is no hotspare setup on Raid Manager 6.2.

raidutil -c c3t5d0 -D 0 (delete LUN 0)

raidutil -c c3t5d0 -n 0 -l 0 -s 69967 -g 10 (recreate LUN 0)
 
any advise on how i should deal with the dead lun?

thank you in advance :)
 
Status
Not open for further replies.

Part and Inventory Search

Sponsor

Back
Top