Re: NetBSD-10.0RC

2023-11-27 Thread Mr Roooster
On Mon, 20 Nov 2023 at 03:20, Todd Gruhn  wrote:
>
> On the bottom of nv(4) it says "GeForce GTX" .
>
> I have a GTX-1660 -- will NetBSD now work with this?

It doesn't look like it.

You have a Turing based card, with a TU116 core.

The latest card that seems to be supported on NetBSD is the 10xx
series, via the nouveau driver[1], these use the Pascal
microarchitecture. The Turing stuff is the successor to this.

Cheers,

Ian

1 - https://man.netbsd.org/nouveau.4


Re: zfs pool behavior - is it ever freed?

2023-07-28 Thread Mr Roooster
On Thu, 27 Jul 2023 at 19:28, Greg Troxel  wrote:
>
> Mike Pumford  writes:
>
[snip]
>
> > If I've read it right there needs to be a mechanism for memory
> > pressure to force ZFS to release memory. Doing it after all the
> > processes have been swapped to disk is way too late as the chances are
> > the system will become non-responsive by then. From memory this was a
> > problem FreeBSD had to solve as well.
>
> It would be interesting to read a description of what they did.  That
> seems easier than figuring it out from scratch.
>
I'm not sure they did a lot more than expose the ARC limit as a sysctl.

I moved to FreeBSD from Net a few years ago (mainly to get ZFS), and
have had similar issues under heavy load with a large ARC. It wouldn't
crash or hang, but it would always favour killing something over
flushing the ARC under pressure. I did a little bit of digging and got
the impression this was the way it was intended to work. (Although
reading this thread it may be a little more complex than that. :) )

Once I limited my ARC my problems went away. I limited mine to 16 gig
on a 96 gig system, but I was running some processes with high memory
usage. I've not had cause to increase it though, and the system runs
reliably. It has a few zpools, and I'm running a VM of an iSCSI
exposed ZVOL, so it get a decent amount of use.

(This is my home system, not a production system, however it does have
something like 10 HDDs in, so is often quite I/O loaded).

Cheers,

Ian


Re: Running NetBSD installer on VM to install on real disk [Was: NetBSD 9.3 UEFI...]

2022-12-09 Thread Mr Roooster
On Mon, 5 Dec 2022 at 12:48, Mayuresh  wrote:
>
> On Sat, Dec 03, 2022 at 02:50:00PM +0100, Martin Husemann wrote:
> > I would not go via the VM detour, but instead:
>
> Just curious, what can go wrong this way.
>
I have installed OSs (Not NetBSD) using this method in the past,
giving the VM full access to the physical disk. however the disk
booted in the VM wasn't in use by the host. (Infact, as I was in
Windows there's no way to give the VM this low level access to the OSs
boot disk).

It does seem to be possible to boot QEMU off the host disk under Linux
(by pointing it at the block device rather than a partition), but you
could really set yourself up for problems. (Qemu writing to the wrong
bit of your disk may end badly. :D )

If you try I suspect you could get away with creating the parition in
linux first, and setting the partition type, so the NetBSD VM doesn't
have to update the partition table at all.

It's probably worth unmounting the /boot EFI partitoin in linux too,
just to be safe.

Oh, and run a 'sync' in linux too, just to be sure, it really does
like to cache writes. (I mean it's shouldn't cache writes to critical
file system things, but it's best to be sure).

You could then probably newfs the partition in the VM and manually
expand the base sets and bung on a kernel. (It's been a couple of
years since I installed NetBSD, but I assume nothing's changed that
would prevent this from working since?)

As others have said I'd then manually copy the necessary files to the
EFI partiton and setup whatever current bootloader you're using to
boot NetBSD.

Ian


Re: Feed facility/priority to logger(1) via stdin - desirable extension or bad idea?

2022-10-20 Thread Mr Roooster
On Fri, 7 Oct 2022 at 14:19, Matthias Petermann  wrote:
>
>
> - Can what I have in mind already be solved (differently or more
> elegantly) with existing tools from the base system?
>
It's not elegent, but depending on your shell you can abuse tee to do
something like:

$ echo -e "alert|alert thing\ninfo|less important\n" |
 tee >(grep "^info|" | cut -c6- | logger -puser.notice) |
 grep "^alert|" | cut -c7- | logger -plocal0.notice

cheers,

Ian


Re: NetBSD on a NUC

2021-10-04 Thread Mr Roooster
On Wed, 29 Sept 2021 at 00:26, Carl Brewer  wrote:
>
> On 28/09/2021 5:59 am, Bob Bernstein wrote:
>
> > My question: other than replacing the NUC's drive, are there useful
> > directions in which I might point my efforts?
>
> NUC's are famous for overheating, SSDs for getting damaged by heat.
> Replace the SSD drive, they're cheap, it takes 10 minutes and all you
> need is a screwdriver.  If it doesn't fix it, you have a spare SSD,
> that's a win anyway.
>nn
Depending on the model of NUC it might also have an onboard SATA
header. If you've a spare HD you could try that before buying new
hardware.

I had a NUC with a cheap samsung SSD that would very occasionally
refuse to boot, never fully tracked that one down.

Ian


Re: ZFS RAIDZ2 and wd uncorrectable data error - why does ZFS not notice the hardware error?

2021-07-19 Thread Mr Roooster
On Sun, 18 Jul 2021 at 00:30, Greg Troxel  wrote:
>
>
[snip]
>
> Ah, interesting point.  I find this confusing, because I thought an
> uncorrectable read error would, for disks I've dealt with, cause the
> sector to be marked as permanently failed and pending reallocation.
>
It depends where the failure occurs I expect. A drive could read just
fine, but then a damaged cable may cause enough noise that the data
doesn't always make it to the controller correctly.

> I also didn't realize that wd(4) would issue aother read when there is a
> failure, but maybe that's in zfs glue code.
>
wd has retried for years I think, it certanly used to do that with the
soft RAID code.

Looks to be set at 5 in the source[1], if I'm looking in the right place. :D

I expect if you just use wd devices for ZFS there may be some merit in
setting the retries to 1 and letting ZFS deal with it, it'd stop the
slow I/O, with the effect of ZFS failing the drive.

[snip]
> >>5 200  140 yes online  positiveReallocated sector count0
> >
> > I was expecting to see this value greater than 0 if the drive was
> > failing, is the drive bad or the cabling?
>
> Sectors get marked as failed, and then they actually get reallocated
> when you write.
> I bet after a dd of /dev/zero that will go up.

This is useful to know!. :)

Ian

1. 
https://github.com/NetBSD/src/blob/05082e19134c05f2f4b6eca73223cdc6b5ab09bf/sys/dev/ata/wd.c#L94


Re: ZFS RAIDZ2 and wd uncorrectable data error - why does ZFS not notice the hardware error?

2021-07-17 Thread Mr Roooster
On Wed, 14 Jul 2021 at 12:07, Matthias Petermann  wrote:
>
> Hello all,
>
>
> ```
> [ 87240.313853] wd2: (uncorrectable data error)
> [ 87240.313853] wd2d: error reading fsbn 5707914328 of
> 5707914328-5707914455 (wd2 bn 5707914328; cn 5662613 tn 6 sn 46)
> [ 87465.637977] wd2d: error reading fsbn 5710464152 of
> 5710464152-5710464215 (wd2 bn 5710464152; cn 5665143 tn 0 sn 8), xfer
> 338, retry 0
> [ 87465.637977] wd2: (uncorrectable data error)
> [ 87475.561683] wd2: soft error (corrected) xfer 338
> [ 87506.393194] wd2d: error reading fsbn 5710555128 of
> 5710555128-5710555255 (wd2 bn 5710555128; cn 5665233 tn 4 sn 12), xfer
> 40, retry 0
> [ 87506.393194] wd2: (uncorrectable data error)
> [ 87515.156465] wd2d: error reading fsbn 5710555128 of
> 5710555128-5710555255 (wd2 bn 5710555128; cn 5665233 tn 4 sn 12), xfer
> 40, retry 1
> ```
>
> The whole syslog is full of these messages. What surprises me is that
> there are "uncorrectable" data errors in the syslog. Nevertheless, the
> data can still be read - albeit very slowly. My assumption was that the
> redundancies of RAID2 are being used to compensate for the defects. To
> my surprise, ZFS does not seem to have noticed any of these defects:
>
The wd driver is retrying, (IIRC it retries 3 times) and suceeding on
the second or 3rd attempt. (See xfer 338, retry 0, followed by a 'soft
error corrected' with the same xfer number 10 seconds later. This is
the retry suceeding).

This sits below ZFS and therefore ZFS never sees the error. If the
read failed 3 times you'd probably get a data error in ZFS.

>
> For the sake of completeness, here is the issue of S.M.A.R.T. - even if
> I find it difficult to interpret:
>
> ```
> saturn$ doas atactl wd2 smart status
> SMART supported, SMART enabled
> id value thresh crit collect reliability description raw
>1 197   51 yes online  positiveRaw read error rate 38669
>3 176   21 yes online  positiveSpin-up time6158
>4 1000 no  online  positiveStart/stop count510
>5 200  140 yes online  positiveReallocated sector count0

I was expecting to see this value greater than 0 if the drive was
failing, is the drive bad or the cabling?

Cheers,

Ian


Re: GeForce GTX 680

2021-06-18 Thread Mr Roooster
On Fri, 18 Jun 2021 at 14:04, Todd Gruhn  wrote:
>
> I just recieved a GTX 680. I saw *680 in the nouveau(4) page.
> Aparrently that does not also apply to GTX 680. Is anything
> known about support for the nVidia GTX 680?

If the Wikipedia[1] page is correct that's a GK104, which is listed as
supported in 9 and -current.[2]

1 - https://en.wikipedia.org/wiki/GeForce_600_series
2 - https://man.netbsd.org/nouveau.4