Subject: Q: T2000: raidctl vs. zpool status



Hi all,

IHACWHAC (I have a colleague who has a customer - hello, if you're listening :-) who's trying to build and test a scenario where he can salvage the data off the (internal ?) disks of a T2000 in case the sysboard and with it the on-board raid controller dies.

If I understood correctly, he replaces the motherboard, does some magic to get the raid config back, but even when raidctl says "I'm fine", zpool complains that it cannot open one of the replicas:

# raidctl
RAID Volume RAID RAID Disk
Volume Type Status Disk Status
------------------------------------------------------
c0t0d0 IM OK c0t0d0 OK
c0t1d0 OK
c0t2d0 IM OK c0t2d0 OK
c0t3d0 OK

# zpool status -x
pool: dpool
state: FAULTED
status: One or more devices could not be opened. There are insufficient
replicas for the pool to continue functioning.
action: Attach the missing device and online it using 'zpool online'.
see: http://www.sun.com/msg/ZFS-8000-D3
scrub: none requested
config:

NAME STATE READ WRITE CKSUM
dpool UNAVAIL 0 0 0 insufficient replicas
c0t2d0 UNAVAIL 0 0 0 cannot open
#

what the customer does to achieve this is documented in the attachment (sorry about the German comments, but I thought translating them would have been a bit much to ask).

TIA for any comments, etc.
Michael
--
Michael Schuster (+49 89) 46008-2974 / x62974
visit the online support center: http://www.sun.com/osc/

Recursion, n.: see 'Recursion'
-------- Original Message --------

ich habe das system booten können habe! Aber alle ZFS-Daten inkl. darauf
liegender Zonen sind dabei verloren gegangen:

{0} ok select [email protected][email protected][email protected][email protected] scsi-pci1000,56 LSI,1064E
scsi-pci1000,56 ?
{0} ok show-volumes
Volume 0 Target 83 Type IM (Integrated Mirroring)
Optimal Enabled Inactive
2 Members 143112591 Blocks, 73 GB
Disk 1
Primary Online
Target 2 FUJITSU MAY2073RCSUN72G 0401
Disk 0
Secondary Online
Target 3 FUJITSU MAY2073RCSUN72G 0401
Volume 1 Target 82 Type IM (Integrated Mirroring)
Optimal Enabled Inactive
2 Members 143112591 Blocks, 73 GB
Disk 2
Primary Online
Target 0 FUJITSU MAY2073RCSUN72G 0401
Disk 3
Secondary Online
Target 1 FUJITSU MAY2073RCSUN72G 0401
{0} ok
{0} ok 0 activate-volume
Volume 0 is now activated
{0} ok show-volumes
Volume 0 Target 82 Type IM (Integrated Mirroring)
Optimal Enabled Inactive
2 Members 143112591 Blocks, 73 GB
Disk 2
Primary Online
Target 0 FUJITSU MAY2073RCSUN72G 0401
Disk 3
Secondary Online
Target 1 FUJITSU MAY2073RCSUN72G 0401
Volume 1 Target 2 Type IM (Integrated Mirroring)
Degraded Enabled Resync In Progress <-----
Anzeigefehler? (Volume 1 statt 0)
2 Members 143112591 Blocks, 73 GB
Disk 1
Primary Online
Target 4 FUJITSU MAY2073RCSUN72G 0401
Disk 0
Secondary Online Out Of Sync
Target 3 FUJITSU MAY2073RCSUN72G 0401
{0} ok 1 activate-volume
RAID volume request failed, ioc-status = 0007 <----- deswegen
diese Fehlermeldung!
Volume 1 could not be activated
{0} ok 0 activate-volume <----- deswegen noch
mal Volume 0 !!!
Volume 0 is now activated
{0} ok show-volumes
Volume 0 Target 2 Type IM (Integrated Mirroring)
Degraded Enabled Resync In Progress <----- jetzt stimmt
die Anzeige wieder!
2 Members 143112591 Blocks, 73 GB
Disk 1
Primary Online
Target 4 FUJITSU MAY2073RCSUN72G 0401
Disk 0
Secondary Online Out Of Sync
Target 3 FUJITSU MAY2073RCSUN72G 0401
Volume 1 Target 0 Type IM (Integrated Mirroring)
Degraded Enabled <----- hier auch.
Die Spiegelung von Volume 1 beginnt erst wenn die von Volume 0 abgeschlossen
wird.
2 Members 143112591 Blocks, 73 GB
Disk 2
Primary Online
Target 5 FUJITSU MAY2073RCSUN72G 0401
Disk 3
Secondary Online Out Of Sync
Target 1 FUJITSU MAY2073RCSUN72G 0401
{0} ok
.
.
.
{0} ok show-volumes
Volume 0 Target 2 Type IM (Integrated Mirroring)
Optimal Enabled <----- Die
Spiegelung von Volume 0 ist abgeschlossen.
2 Members 143112591 Blocks, 73 GB
Disk 1
Primary Online
Target 4 FUJITSU MAY2073RCSUN72G 0401
Disk 0
Secondary Online
Target 3 FUJITSU MAY2073RCSUN72G 0401
Volume 1 Target 0 Type IM (Integrated Mirroring)
Degraded Enabled Resync In Progress <----- Die
Spiegelung von Volume 1 ist nun angelaufen.
2 Members 143112591 Blocks, 73 GB
Disk 2
Primary Online
Target 5 FUJITSU MAY2073RCSUN72G 0401
Disk 3
Secondary Online Out Of Sync
Target 1 FUJITSU MAY2073RCSUN72G 0401
{0} ok boot
Boot device: disk File and args:
ERROR: [email protected]: Last Trap: Fast Data Access MMU Miss <----- Bootversuche
scheitern obwohl die Boorplatten (Volume 0) optimal sind!
[Exception handlers interrupted, please file a bug]
[type 'resume' to attempt a normal recovery]
{0} ok
.
.
.
{0} ok show-volumes
Volume 0 Target 2 Type IM (Integrated Mirroring)
Optimal Enabled
2 Members 143112591 Blocks, 73 GB
Disk 1
Primary Online
Target 4 FUJITSU MAY2073RCSUN72G 0401
Disk 0
Secondary Online
Target 3 FUJITSU MAY2073RCSUN72G 0401
Volume 1 Target 0 Type IM (Integrated Mirroring)
Optimal Enabled
2 Members 143112591 Blocks, 73 GB
Disk 2
Primary Online
Target 5 FUJITSU MAY2073RCSUN72G 0401
Disk 3
Secondary Online
Target 1 FUJITSU MAY2073RCSUN72G 0401
{0} ok boot
Boot device: disk File and args:
ERROR: [email protected]: Last Trap: Fast Data Access MMU Miss <----- Bootversuche
scheitern auch alle Volumes optimal sind!
[Exception handlers interrupted, please file a bug]
[type 'resume' to attempt a normal recovery]
{0} ok
{0} ok reset-all <----- mit
reset-all...

SC Alert: Host System has Reset

SC Alert: Host system has shut down.
cpu cpu cpu cpu cpu cpu cpu cpu cpu cpu cpu cpu cpu cpu cpu cpu cpu cpu cpu cpu
cpu cpu cpu cpu Device: pci
[email protected]: Device 0 pci
.
.
.
Sun Fire T200, No Keyboard
Copyright 2006 Sun Microsystems, Inc. All rights reserved.
OpenBoot 4.20.4, 8184 MB memory available, Serial #64571856.
Ethernet address 0:3:ba:d9:49:d0, Host ID: 83d949d0.



{0} ok boot
Boot device: disk File and args:
Loading ufs-file-system package 1.4 04 Aug 1995 13:02:54.
FCode UFS Reader 1.12 00/07/17 15:48:16.
Loading: /platform/SUNW,Sun-Fire-T200/ufsboot
Loading: /platform/sun4v/ufsboot
SunOS Release 5.10 Version Generic_118833-17 64-bit <-----
Server bootet nun...
Copyright 1983-2005 Sun Microsystems, Inc. All rights reserved.
Use is subject to license terms.
Hostname: fmuw31
checking ufs filesystems
/dev/rdsk/c0t0d0s5: is logging.

fmuw31 console login: Jul 17 18:28:48 fmuw31 sendmail[315]: My unqualified host
name (fmuw31) unknown; sleeping for retry
Jul 17 18:28:48 fmuw31 sendmail[316]: My unqualified host name (fmuw31)
unknown; sleeping for retry

SUNW-MSG-ID: ZFS-8000-CS, TYPE: Fault, VER: 1, SEVERITY: Major <-----
Aber es gibt ein ZFS-Problem!
EVENT-TIME: Mon Jul 17 18:28:50 CEST 2006
PLATFORM: SUNW,Sun-Fire-T200, CSN: -, HOSTNAME: fmuw31
SOURCE: zfs-diagnosis, REV: 1.0
EVENT-ID: 37c8cdbf-addb-4d34-c180-d1b2379e27b2
DESC: A ZFS pool failed to open. Refer to http://sun.com/msg/ZFS-8000-CS for
more information.
AUTO-RESPONSE: No automated response will occur.
IMPACT: The pool data is unavailable
REC-ACTION: Run 'zpool status -x' and either attach the missing device or
restore from backup.

SC Alert: Host detected fault, MSGID: ZFS-8000-CS

fmuw31 console login: root
Password:
Jul 17 18:29:48 fmuw31 login: ROOT LOGIN /dev/console
Last login: Fri Jul 14 17:02:59 from 53.1.224.107
Sun Microsystems Inc. SunOS 5.10 Generic January 2005
#
#
# zpool list
NAME SIZE USED AVAIL CAP HEALTH ALTROOT
dpool - - - - FAULTED - <-----
Datenverlust!!!
# zfs list
no datasets available <-----
Datenverlust!!!
#


-----> Bemerkung zur ZFS Konfiguration:
Auf dem zweiten Spiegelpaar das über den onboard RAID-Controller auf dem
urspünglichen Server erstellt wurde,
wurde ein ZFS-Pool der die ganze Platte / Volume erfaste (68GB) erzeugt und
ZFS-Filesystemeangelegt:

# zpool create dpool c0t2d0
# zpool status
pool: dpool
state: ONLINE
scrub: none requested
config:

NAME STATE READ WRITE CKSUM
dpool ONLINE 0 0 0
c0t2d0 ONLINE 0 0 0

errors: No known data errors
# zpool list
NAME SIZE USED AVAIL CAP HEALTH ALTROOT
dpool 68G 79K 68.0G 0% ONLINE -
# zfs create dpool/zones
# zfs list
NAME USED AVAIL REFER MOUNTPOINT
dpool 106K 66.9G 26.5K /dpool
dpool/zones 24.5K 66.9G 24.5K /dpool/zones
# zfs set mountpoint=/zones dpool/zones
# zfs create dpool/zones/fmuw31-vs01
# zfs create dpool/zones/fmuw31-vs02
# zfs create dpool/zones/fmuw31-vs03
# zfs list
NAME USED AVAIL REFER MOUNTPOINT
dpool 204K 66.9G 24.5K /dpool
dpool/zones 102K 66.9G 28.5K /zones
dpool/zones/fmuw31-vs01 24.5K 66.9G 24.5K /zones/fmuw31-vs01
dpool/zones/fmuw31-vs02 24.5K 66.9G 24.5K /zones/fmuw31-vs02
dpool/zones/fmuw31-vs03 24.5K 66.9G 24.5K /zones/fmuw31-vs03
#
...
es wurden Virtuelle-server / Zonen auf diesen ZFSs erstellt... die
Plattenbelegung sah wie folgt aus:

# df -k
Filesystem kbytes used avail capacity Mounted on
/dev/dsk/c0t0d0s0 20654577 3680834 16767198 19% /
/devices 0 0 0 0% /devices
ctfs 0 0 0 0% /system/contract
proc 0 0 0 0% /proc
mnttab 0 0 0 0% /etc/mnttab
swap 22868632 1032 22867600 1% /etc/svc/volatile
objfs 0 0 0 0% /system/object
/platform/SUNW,Sun-Fire-T200/lib/libc_psr/libc_psr_hwcap1.so.1
20654577 3680834 16767198 19%
/platform/sun4v/lib/libc_psr.so.1
/platform/SUNW,Sun-Fire-T200/lib/sparcv9/libc_psr/libc_psr_hwcap1.so.1
20654577 3680834 16767198 19%
/platform/sun4v/lib/sparcv9/libc_psr.so.1
fd 0 0 0 0% /dev/fd
/dev/dsk/c0t0d0s4 8262325 70108 8109594 1% /var
swap 22868040 440 22867600 1% /tmp
swap 22867656 56 22867600 1% /var/run
/dev/dsk/c0t0d0s5 24677176 24489 24405916 1% /vol1
dpool 70189056 24 70026796 1% /dpool
dpool/zones 70189056 28 70026796 1% /zones
dpool/zones/fmuw31-vs01
20971520 80148 20891371 1% /zones/fmuw31-vs01
dpool/zones/fmuw31-vs02
70189056 81881 70026796 1% /zones/fmuw31-vs02
dpool/zones/fmuw31-vs03
70189056 24 70026796 1% /zones/fmuw31-vs03
#

... und jetzt ist das alles weg!



_______________________________________________
zfs-discuss mailing list
[email protected]
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss



Programming list archiving by: Enterprise Git Hosting