Hi Darren,
On 08/30/12 11:07, Anonymous wrote:
Hi. I have a spare off the shelf consumer PC and was thinking about loading
Solaris on it for a development box since I use Studio @work and like it
better than gcc. I was thinking maybe it isn't so smart to use ZFS since it
has only one
Thank you.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
It depends on the model. Consumer models are less likely to
immediately flush. My understanding that this is done in part to do
some write coalescing and reduce the number of P/E cycles. Enterprise
models should either flush, or contain a super capacitor that provides
enough power for the
You wrote:
2012-07-23 18:37, Anonymous wrote:
Really, it would be so helpful to know which drives we can buy with
confidence and which should be avoided...is there any way to know from the
manufacturers web sites or do you have to actually buy one and see what it
does? Thanks to
I've been watching the heat control issue carefully since I had to take a
job offshore (cough reverse H1B cough) in a place without adequate AC and I
was able to get them to ship my servers and some other gear. Then I read
Intel is guaranteeing their servers will work up to 100 degrees F ambient
Hello all,
Trying to reply to everyone so far in one post.
casper@oracle.com said
Did you try:
iostat -En
I issued that command and I see (soft) errors from all 4 drives. There is a
serial no. field in the message headers but it is has no contents.
messages in
Richard Elling said
If the errors bubble up to ZFS, then they will be shown in the output of
zpool status
On the console I was seeing retryable read errors that eventually
failed. The block number and drive path were included but not any info I
could relate to the actual disk.
zpool status
Thank you all for your answers and links :-)
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
On Solaris 10 If I install using ZFS root on only one drive is there a way
to add another drive as a mirror later? Sorry if this was discussed
already. I searched the archives and couldn't find the answer. Thank you.
___
zfs-discuss mailing list
You wrote:
Hi Roy, things got alot worse since my first email. I don't know what
happened but I can't import the old pool at all. It shows no errors but when
I import it I get a kernel panic from assertion failed: zvol_get_stats(os,
nv) which looks like is fixed by patch 6801926 which
I am having a problem after a new install of Solaris 10. The installed rpool
works fine when I have only those disks connected. When I connect disks from
an rpool I created during a previous installation, my newly installed rpool
is ignored even though the BIOS (x86) is set to boot only from the
Hi Roy, things got alot worse since my first email. I don't know what
happened but I can't import the old pool at all. It shows no errors but when
I import it I get a kernel panic from assertion failed: zvol_get_stats(os,
nv) which looks like is fixed by patch 6801926 which is applied in Solaris
Hi Dave,
Hi Cindy.
Consider the easiest configuration first and it will probably save
you time and money in the long run, like this:
73g x 73g mirror (one large s0 on each disk) - rpool
73g x 73g mirror (use whole disks) - data pool
Then, get yourself two replacement disks, a good
If you plan to generate a lot of data, why use the root pool? You can put
the /home and /proj filesystems (/export/...) on a separate pool, thus
off-loading the root pool.
I don't, it's a development box with not alot happening.
My two cents,
thanks
14 matches
Mail list logo