Given that the replace was crawling, I did the following:
Ensured that the degraded filesystem was
noautoin/etc/fstabRebooted the machine (which took about 20 minutes due to I/O hangs)
Disabled the LVM VG containing the btrfs fs on the failed drive:
sudo vgchange -an <failed-vg>Disabled the failed device:
echo 1 | sudo tee /sys/block/sdb/device/deleteMounted the filesystem
-o ro,degraded(degradedcan only be used once)Checked
replace statusand saw it was suspended:Started on 26.Jan 00:36:12, suspended on 26.Jan 10:13:30 at 4.1%, 0 write errs, 0Mounted
-o remount,rwand saw thereplacecontinue:kernel: BTRFS info (device dm-5): continuing dev_replace from <missing disk> (devid 2) to target /dev/mapper/vg6TBd1-ark @4%
As I'm writing this:
replace statusshows a healthy 0.1% progress every 30 seconds or soiostat -d 1 -m <target-dev>shows about 145MB/s (Seagate advertises 160MB/s)
Update:
After completion, I noticed that btrfs device usage /mountpoint was showing some Data,DUP and Metadata,single, rather than only RAID1, so I rebalanced:
btrfs balance start -dconvert=raid1,soft -mconvert=raid1,soft /mountpoint Also, consider resizeing if both devices now contain slack:
btrfs filesystem resize max /mountpoint I would also recommend that you scrub as I had 262016 correctable csum errors as detailed here262016 correctable csum errors seemingly related to the interrupted replace, which seems to be related to the interrupted replace.