Good Morning, We have a large cluster with 50 gfs2 SAN mounts. The mounts range in size from 1TB to 15TB each. We have some with 6-8TB of data but most average around 3TB used right now. We were doing network testing a while back to check our redundancy incase of a switch failure, and the tests failed.. multiple times. We ended having the SAN mounts yanked out from under the cluster. Long story short, we seem to have corruption. I can still bring the volumes up with the cluster but when i take everything down and do a fsck I get the following:
(ran with fsck -n /dev/$device) Found a copy of the root directory in a journal at block: 0x501ca. Damaged root dinode not fixed. The root dinode should be at block 0x2f3b98b7 but it seems to be destroyed. Found a copy of the root directory in a journal at block: 0x501d2. Damaged root dinode not fixed. The root dinode should be at block 0x28a3ac7f but it seems to be destroyed. Found a copy of the root directory in a journal at block: 0x501da. Damaged root dinode not fixed. Unable to locate the root directory. Can't find any dinodes that might be the root; using master - 1. Found a possible root at: 0x16 The root dinode block is destroyed. At this point I recommend reinitializing it. Hopefully everything will later be put into lost+found. The root dinode was not reinitialized; aborting. This particular device had 4698 "seems to be destroyed.. found a copy" messages before the final, "Can't find any dinodes" message. I fear that we have a number of mounts in this state. Is there any way to recover? Thanks in advance. -- Linux-cluster mailing list Linux-cluster@redhat.com https://www.redhat.com/mailman/listinfo/linux-cluster