Hallo,

I originally had RAID with six 4TB drives, which was more than 80
percent full. So now I bought
a 10TB drive, added it to the Array and gave the command to remove the
oldest drive in the array.

 btrfs device delete /dev/sda /mnt/btrfs-raid

I kept a terminal with "watch btrfs fi show" open and It showed that
the size of /dev/sda had been set to zero and that data was being
redistributed to the other drives. All seemed well, but now the
process stalls at 8GB being left on /dev/sda/. It also seems that the
size of the drive has been reset the original value of 3,64TiB.

Label: none  uuid: 1609e4e1-4037-4d31-bf12-f84a691db5d8
        Total devices 7 FS bytes used 8.07TiB
        devid    1 size 3.64TiB used 8.00GiB path /dev/sda
        devid    2 size 3.64TiB used 2.73TiB path /dev/sdc
        devid    3 size 3.64TiB used 2.73TiB path /dev/sdd
        devid    4 size 3.64TiB used 2.73TiB path /dev/sde
        devid    5 size 3.64TiB used 2.73TiB path /dev/sdf
        devid    6 size 3.64TiB used 2.73TiB path /dev/sdg
        devid    7 size 9.10TiB used 2.50TiB path /dev/sdb

I see no more btrfs worker processes and no more activity in iotop.
How do I proceed? I am using a current debian stretch which uses
Kernel 4.9.0-8 and btrfs-progs 4.7.3-1.

How should I proceed? I have a Backup but would prefer an easier and
less time-comsuming way out of this mess.

Yours
Stefan

Reply via email to