Current-Users archive
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index][Old Index]
Re: kern/54289 hosed my RAID. Recovery possible?
On 2019-08-15 00:03, jdbaker%consolidated.net@localhost wrote:
The SiI3214 SATALink card suffers from the identify problem in netbsd-9
and -current (PR kern/54289).
Booting a netbsd-9 kernel, the drives failed to identify which caused
RAIDframe to mark the 4 drives on that card (of 8) in my RAID as
FAILED.
Rebooting netbsd-8, the drives identify properly, but are still marked
as
FAILED.
Is there any way to unmark them so the raid will configure and recover?
Normally 'raidctl -C' is used during first time configuration. Could it
be used to force configuration, ignoring the FAILED status? Would the
RAID
be recoverable with parity rebuild afterwards?
This seems to have worked. The disks not being correctly
identified/attached
under netbsd-9 apparently had them recorded as failed on the components
that
did attach (on the machine's on-board intel ahcisata ports). Rebooting
netbsd-8, although the drives identified and attached properly, they
were
still considered failed components.
Being a multiple-disk failure is usually fatal to a RAID, but the
components
weren't actually failed. Un-configuring with 'raidctl -u' then forcing
a
config with 'raidctl -C /path/to/config' did not show any fatal errors
and
subsequent 'raidctl -s' showed all component labels (w/serial number)
intact.
Parity rewrite took a long time.
Afterwards, 'gpt show raid0d' and 'dkctl raid0d listwedges' showed
things to
be intact that far. Rebooting the machine, the RAID properly
autoconfigured.
'fsck' reported the filesystem as clean (since it never got mounted
after the
failed reboot into netbsd-9). An 'fsck -f' run is in progress.
John D. Baker
Home |
Main Index |
Thread Index |
Old Index