It's about 800Mb, I think I could upload that. I ran it with the -s parameter, is that enough to remove all personal info from the image? Also, I had to run it with -w because otherwise it died on the same corrupt node.
On Fri, Apr 1, 2016 at 2:25 AM, Qu Wenruo <quwen...@cn.fujitsu.com> wrote: > > > Ivan P wrote on 2016/03/31 18:04 +0200: >> >> Ok, it will take a while until I can attempt repairing it, since I >> will have to order a spare HDD to copy the data to. >> Should I take some sort of debug snapshot of the fs so you can take a >> look at it? I think I read something about a snapshot that only >> contains the fs but not the data that somewhere. > > That's btrfs-image. > > It would be good, but if your metadata is over 3G, I think it's would take a > lot of time uploading. > > Thanks, > Qu > >> >> Regards, >> Ivan. >> >> On Tue, Mar 29, 2016 at 3:57 AM, Qu Wenruo <quwen...@cn.fujitsu.com> >> wrote: >>> >>> >>> >>> Ivan P wrote on 2016/03/28 23:21 +0200: >>>> >>>> >>>> Well, the file in this inode is fine, I was able to copy it off the >>>> disk. However, rm-ing the file causes a segmentation fault. Shortly >>>> after that, I get a kernel oops. Same thing happens if I attempt to >>>> re-run scrub. >>>> >>>> How can I delete that inode? Could deleting it destroy the filesystem >>>> beyond repair? >>> >>> >>> >>> The kernel oops should protect you from completely destroying the fs. >>> >>> However it seems that the problem is beyond kernel's handle (kernel >>> oops). >>> >>> So no safe recovery method now. >>> >>> From now on, any repair advice from me *MAY* *destroy* your fs. >>> So please do backup when you still can. >>> >>> >>> The best possible try would be "btrfsck --init-extent-tree --repair". >>> >>> If it works, then mount it and run "btrfs balance start <mnt>". >>> Lastly, umount and use btrfsck to re-check if it fixes the problem. >>> >>> Thanks, >>> Qu >>> >>> >>>> >>>> Regards, >>>> Ivan >>>> >>>> On Mon, Mar 28, 2016 at 3:10 AM, Qu Wenruo <quwenruo.bt...@gmx.com> >>>> wrote: >>>>> >>>>> >>>>> >>>>> >>>>> Ivan P wrote on 2016/03/27 16:31 +0200: >>>>>> >>>>>> >>>>>> >>>>>> Thanks for the reply, >>>>>> >>>>>> the raid1 array was created from scratch, so not converted from ext*. >>>>>> I used btrfs-progs version 4.2.3 on kernel 4.2.5 to create the array, >>>>>> btw. >>>>> >>>>> >>>>> >>>>> >>>>> I don't remember any strange behavior after 4.0, so no clue here. >>>>> >>>>> Go to the subvolume 5 (the top-level subvolume), find inode 71723 and >>>>> try >>>>> to >>>>> remove it. >>>>> Then, use 'btrfs filesystem sync <mount point>' to sync the inode >>>>> removal. >>>>> >>>>> Finally use latest btrfs-progs to check if the problem disappears. >>>>> >>>>> This problem seems to be quite strange, so I can't locate the root >>>>> cause, >>>>> but try to remove the file and hopes kernel can handle it. >>>>> >>>>> Thanks, >>>>> Qu >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> Is there a way to fix the current situation without taking the whole >>>>>> data off the disk? >>>>>> I'm not familiar with file systems terms, so what exactly could I have >>>>>> lost, if anything? >>>>>> >>>>>> Regards, >>>>>> Ivan >>>>>> >>>>>> On Sun, Mar 27, 2016 at 4:23 PM, Qu Wenruo <quwenruo.bt...@gmx.com >>>>>> <mailto:quwenruo.bt...@gmx.com>> wrote: >>>>>> >>>>>> >>>>>> >>>>>> On 03/27/2016 05:54 PM, Ivan P wrote: >>>>>> >>>>>> Read the info on the wiki, here's the rest of the requested >>>>>> information: >>>>>> >>>>>> # uname -r >>>>>> 4.4.5-1-ARCH >>>>>> >>>>>> # btrfs fi show >>>>>> Label: 'ArchVault' uuid: >>>>>> cd8a92b6-c5b5-4b19-b5e6-a839828d12d8 >>>>>> Total devices 1 FS bytes used 2.10GiB >>>>>> devid 1 size 14.92GiB used 4.02GiB path >>>>>> /dev/sdc1 >>>>>> >>>>>> Label: 'Vault' uuid: 013cda95-8aab-4cb2-acdd-2f0f78036e02 >>>>>> Total devices 2 FS bytes used 800.72GiB >>>>>> devid 1 size 931.51GiB used 808.01GiB path >>>>>> /dev/sda >>>>>> devid 2 size 931.51GiB used 808.01GiB path >>>>>> /dev/sdb >>>>>> >>>>>> # btrfs fi df /mnt/vault/ >>>>>> Data, RAID1: total=806.00GiB, used=799.81GiB >>>>>> System, RAID1: total=8.00MiB, used=128.00KiB >>>>>> Metadata, RAID1: total=2.00GiB, used=936.20MiB >>>>>> GlobalReserve, single: total=320.00MiB, used=0.00B >>>>>> >>>>>> On Fri, Mar 25, 2016 at 3:16 PM, Ivan P >>>>>> <chrnosphe...@gmail.com >>>>>> <mailto:chrnosphe...@gmail.com>> wrote: >>>>>> >>>>>> Hello, >>>>>> >>>>>> using kernel 4.4.5 and btrfs-progs 4.4.1, I today ran a >>>>>> scrub on my >>>>>> 2x1Tb btrfs raid1 array and it finished with 36 >>>>>> unrecoverable errors >>>>>> [1], all blaming the treeblock 741942071296. Running >>>>>> "btrfs >>>>>> check >>>>>> --readonly" on one of the devices lists that extent as >>>>>> corrupted [2]. >>>>>> >>>>>> How can I recover, how much did I really lose, and how >>>>>> can >>>>>> I >>>>>> prevent >>>>>> it from happening again? >>>>>> If you need me to provide more info, do tell. >>>>>> >>>>>> [1] http://cwillu.com:8080/188.110.141.36/1 >>>>>> >>>>>> >>>>>> This message itself is normal, it just means a tree block is >>>>>> crossing 64K stripe boundary. >>>>>> And due to scrub limit, it can check if it's good or bad. >>>>>> But.... >>>>>> >>>>>> [2] http://pastebin.com/xA5zezqw >>>>>> >>>>>> This one is much more meaningful, showing several strange bugs. >>>>>> >>>>>> 1. corrupt extent record: key 741942071296 168 1114112 >>>>>> This means, this is a EXTENT_ITEM(168), and according to the >>>>>> offset, >>>>>> it means the length of the extent is, 1088K, definitely not a >>>>>> valid >>>>>> tree block size. >>>>>> >>>>>> But according to [1], kernel think it's a tree block, which is >>>>>> quite >>>>>> strange. >>>>>> Normally, such mismatch only happens in fs converted from ext*. >>>>>> >>>>>> 2. Backref 741942071296 root 5 owner 71723 offset 2589392896 >>>>>> num_refs 0 not found in extent tree >>>>>> >>>>>> num_refs 0, this is also strange, normal backref won't have a >>>>>> zero >>>>>> refrence number. >>>>>> >>>>>> 3. bad metadata [741942071296, 741943185408) crossing stripe >>>>>> boundary >>>>>> It could be a false warning fixed in latest btrfsck. >>>>>> But you're using 4.4.1, so I think that's the problem. >>>>>> >>>>>> 4. bad extent [741942071296, 741943185408), type mismatch with >>>>>> chunk >>>>>> This seems to explain the problem, a data extent appears in a >>>>>> metadata chunk. >>>>>> It seems that you're really using converted btrfs. >>>>>> >>>>>> If so, just roll it back to ext*. Current btrfs-convert has >>>>>> known >>>>>> bug but fix is still under review. >>>>>> >>>>>> If want to use btrfs, use a newly created one instead of >>>>>> btrfs-convert. >>>>>> >>>>>> Thanks, >>>>>> Qu >>>>>> >>>>>> >>>>>> Regards, >>>>>> Soukyuu >>>>>> >>>>>> P.S.: please add me to CC when replying as I did not >>>>>> subscribe to the >>>>>> mailing list. Majordomo won't let me use my hotmail >>>>>> address >>>>>> and I >>>>>> don't want that much traffic on this address. >>>>>> >>>>>> -- >>>>>> To unsubscribe from this list: send the line "unsubscribe >>>>>> linux-btrfs" in >>>>>> the body of a message to majord...@vger.kernel.org >>>>>> <mailto:majord...@vger.kernel.org> >>>>>> More majordomo info at >>>>>> http://vger.kernel.org/majordomo-info.html >>>>>> >>>>>> >>>>> >>>> -- >>>> To unsubscribe from this list: send the line "unsubscribe linux-btrfs" >>>> in >>>> the body of a message to majord...@vger.kernel.org >>>> More majordomo info at http://vger.kernel.org/majordomo-info.html >>>> >>>> >>> >>> >> >> > > -- To unsubscribe from this list: send the line "unsubscribe linux-btrfs" in the body of a message to majord...@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html