Hi Andrei,
Thanks for the info, sorry about the improper terminology. In
better news, I discovered that one disk wasn't getting recognized by the
OS at certain outputs of one of my mini-sas to sata cables, so I got a
new one, and the disk works fine on that. So I'm that's at least one
bad cable, I still have to check the other 3, though.
------ Corey
On 07/08/2016 10:40 PM, Andrei Borzenkov wrote:
07.07.2016 09:40, Corey Coughlin пишет:
Hi Tomasz,
Thanks for the response! I should clear some things up, though.
On 07/06/2016 03:59 PM, Tomasz Kusmierz wrote:
On 6 Jul 2016, at 23:14, Corey Coughlin
<corey.coughlin....@gmail.com> wrote:
Hi all,
Hoping you all can help, have a strange problem, think I know
what's going on, but could use some verification. I set up a raid1
type btrfs filesystem on an Ubuntu 16.04 system, here's what it looks
like:
btrfs fi show
Label: none uuid: 597ee185-36ac-4b68-8961-d4adc13f95d4
Total devices 10 FS bytes used 3.42TiB
devid 1 size 1.82TiB used 1.18TiB path /dev/sdd
devid 2 size 698.64GiB used 47.00GiB path /dev/sdk
devid 3 size 931.51GiB used 280.03GiB path /dev/sdm
devid 4 size 931.51GiB used 280.00GiB path /dev/sdl
devid 5 size 1.82TiB used 1.17TiB path /dev/sdi
devid 6 size 1.82TiB used 823.03GiB path /dev/sdj
devid 7 size 698.64GiB used 47.00GiB path /dev/sdg
devid 8 size 1.82TiB used 1.18TiB path /dev/sda
devid 9 size 1.82TiB used 1.18TiB path /dev/sdb
devid 10 size 1.36TiB used 745.03GiB path /dev/sdh
Now when I say that the drives mount points change, I'm not saying they
change when I reboot. They change while the system is running. For
instance, here's the fi show after I ran a "check --repair" run this
afternoon:
btrfs fi show
Label: none uuid: 597ee185-36ac-4b68-8961-d4adc13f95d4
Total devices 10 FS bytes used 3.42TiB
devid 1 size 1.82TiB used 1.18TiB path /dev/sdd
devid 2 size 698.64GiB used 47.00GiB path /dev/sdk
devid 3 size 931.51GiB used 280.03GiB path /dev/sdm
devid 4 size 931.51GiB used 280.00GiB path /dev/sdl
devid 5 size 1.82TiB used 1.17TiB path /dev/sdi
devid 6 size 1.82TiB used 823.03GiB path /dev/sds
devid 7 size 698.64GiB used 47.00GiB path /dev/sdg
devid 8 size 1.82TiB used 1.18TiB path /dev/sda
devid 9 size 1.82TiB used 1.18TiB path /dev/sdb
devid 10 size 1.36TiB used 745.03GiB path /dev/sdh
Notice that /dev/sdj in the previous run changed to /dev/sds. There was
no reboot, the mount just changed. I don't know why that is happening,
but it seems like the majority of the errors are on that drive. But
given that I've fixed the start/stop issue on that disk, it probably
isn't a WD Green issue.
It's not "mount point", it is just device names. Do not make it sound
more confusing than it already is :)
This implies that disks drop off and reappear. Do you have "dmesg" or
log (/var/log/syslog or /var/log/messages or journalctl) for the same
period of time?
--
To unsubscribe from this list: send the line "unsubscribe linux-btrfs" in
the body of a message to majord...@vger.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html