Re: ZFS: Rescue FAULTED Pool

From: A FreeBSD User <freebsd_at_walstatt-de.de>
Date: Mon, 03 Feb 2025 18:24:10 UTC
Am Sat, 1 Feb 2025 09:10:25 -0500
Dennis Clarke <dclarke@blastwave.org> schrieb:

> >>
> >> The most useful thing to share right now would be the output of `zpool
> >> import` (with no pool name) on the rebooted system.
> >>
> >> That will show where the issues are, and suggest how they might be solved.
> >>  
> > 
> > Hello, this exactly happens when trying to import the pool. Prior to the loss, device da1p1
> > has been faulted with numbers in the colum/columns "corrupted data"/further not seen now.
> > 
> > 
> >   ~# zpool import
> >     pool: BUNKER00
> >       id: XXXXXXXXXXXXXXXXXXXX
> >    state: FAULTED
> > status: The pool metadata is corrupted.
> >   action: The pool cannot be imported due to damaged devices or data.
> >          The pool may be active on another system, but can be imported using
> >          the '-f' flag.
> >     see: https://openzfs.github.io/openzfs-docs/msg/ZFS-8000-72
> >   config:
> > 
> >          BUNKER00    FAULTED  corrupted data
> >            raidz1-0  ONLINE
> >              da2p1   ONLINE
> >              da3p1   ONLINE
> >              da4p1   ONLINE
> >              da7p1   ONLINE
> >              da6p1   ONLINE
> >              da1p1   ONLINE
> >              da5p1   ONLINE
> > 
> > 
> >   ~# zpool import -f BUNKER00
> > cannot import 'BUNKER00': I/O error
> >          Destroy and re-create the pool from
> >          a backup source.
> > 
> > 
> > ~# zpool import -F BUNKER00
> > cannot import 'BUNKER00': one or more devices is currently unavailable
> >   
> 
>      This is indeed a sad situation. You have a raidz1 pool with one or
> MORE devices that seem to have left the stage. I suspect more than one.
> 
>      I can only guess what you see from "camcontrol devlist" as well as
> data from "gpart show -l" where we would see the partition data along
> with and GPT labels. If in fact you used GPT scheme. You have a list of
> devices that all say "p1" there and so I guess you made some sort of a
> partition table. ZFS does not need that but it can be nice to have. In
> any case, it really does look like you have _more_ than one failure in
> there somewhere and only dmesg and some separate tests on each device
> would reveal the truth.
> 
> 
> --
> Dennis Clarke
> RISC-V/SPARC/PPC/ARM/CISC
> UNIX and Linux spoken
> 
> 


Hello all!

Thank you for your tips!

Luckily, "zpool import -FX" as suggested herein did after a while (60-80 minutes) the trick!
There might be some data losses - but compared to the alternative bareable.

Thank you very much!

Kind regards,

Oliver


-- 

A FreeBSD user