Re: Removing IPv6 address causes all IPv6 traffic to stop working

2018-07-12 Thread Frank de Bot (lists)
The routes are not affected. the strangest thing I find that the host
and jails are accessible from outside, and it can reach outside hosts
via v6, but everything that's  staying on the server fails.

Alan Somers wrote:
> How did you assign the jails' IPv6 addresses in the first place?  The usual
> way is to assign them as /128 aliases, in which case the command to remove
> them would include a "/128", not "/48".  I think when you're deleting a
> "/48" you're also removing some routes that the jail host is using.
> -Alan
> 
> On Thu, Jul 12, 2018 at 4:30 PM, Frank de Bot (lists) 
> wrote:
> 
>> On a older server running FreeBSD 10.2 i have a number of jails. For
>> migration to FreeBSD I'm planning to shutdown the jail, move the data to
>> the new server and spin up the jail there. IP addresses are alse moved.
>>
>> When I remove an IPv6 address with the following command 'ifconfig em0
>> inet6 v6address/48 delete', all IPv6 traffic on the server stops
>> working. I can reach IPv6 from outside to the server and from the server
>> to outside, but everything on the server (jail to jail for example)
>> immediatly stops working. IPv4 works normally
>> Unfortunately it's not feasible to bing-bang migrate everything,
>> ignoring ipv6 is not possible because a lot of the jails interact with
>> each other with ipv6 (some even exclusivly)
>>
>> What can cause this?
>>
>>
>> Regards,
>>
>> Frank de Bot
>> ___
>> freebsd-stable@freebsd.org mailing list
>> https://lists.freebsd.org/mailman/listinfo/freebsd-stable
>> To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"
>>
> ___
> freebsd-stable@freebsd.org mailing list
> https://lists.freebsd.org/mailman/listinfo/freebsd-stable
> To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"
> 

___
freebsd-stable@freebsd.org mailing list
https://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"


Removing IPv6 address causes all IPv6 traffic to stop working

2018-07-12 Thread Frank de Bot (lists)
On a older server running FreeBSD 10.2 i have a number of jails. For
migration to FreeBSD I'm planning to shutdown the jail, move the data to
the new server and spin up the jail there. IP addresses are alse moved.

When I remove an IPv6 address with the following command 'ifconfig em0
inet6 v6address/48 delete', all IPv6 traffic on the server stops
working. I can reach IPv6 from outside to the server and from the server
to outside, but everything on the server (jail to jail for example)
immediatly stops working. IPv4 works normally
Unfortunately it's not feasible to bing-bang migrate everything,
ignoring ipv6 is not possible because a lot of the jails interact with
each other with ipv6 (some even exclusivly)

What can cause this?


Regards,

Frank de Bot
___
freebsd-stable@freebsd.org mailing list
https://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"


Only 1 initiator per target

2017-12-06 Thread Frank de Bot (lists)
Hi,

Is it possible to limit the number of initiators connected to a target
to 1 besides the 'initiator-name' config on the target? All servers
should be able to connect, but for safety I want to prevent a second
initiator from connecting to that target. A single iniator should be
able to connect several times if they use multi-path.


Regards,

Frank de Bot
___
freebsd-stable@freebsd.org mailing list
https://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"


Re: Reloading iscsi target, iscsi initiator panics

2015-10-18 Thread Frank de Bot (lists)
Frank de Bot wrote:
> Edward Tomasz Napierała wrote:
>> On 1015T2005, Frank de Bot (lists) wrote:
>>> Edward Tomasz Napierała wrote:
>>>> On 1014T2316, Frank de Bot (lists) wrote:
>>>>> Hello,
>>>>>
>>>>> I have a FreeBSD 10.2 server running as iSCSI target. Another FreeBSD
>>>>> 10.2 is an initiator and has several targets used.
>>>>>
>>>>> When I add a target and reload its config with 'service ctld reload',
>>>>> the FreeBSD initiator server panics. It's reproducable (I have a
>>>>> coredump, but I think it's clear what the problem is)
>>>>
>>>> How easy it is to reproduce, eg. does it happen about every time?
>>>> Could you paste the backtrace?
>>>
>>> The first time I didn't understand what happened, the second time I
>>> could relate it directly to the reloading of ctld on the iSCSI, within
>>> moments the server paniced.
>>>
>>> I got 2 backtraces:
>>>
>>> First one:
>>>
>>> Fatal trap 12: page fault while in kernel mode
>>> cpuid = 1; softdep_deallocate_dependencies: got error 6 while accessing
>>
>> This is the FFS on the initiator side panicing because the device it's
>> on went away.  The softupdates code can't handle that very well.
>>
>> I have no idea why the devices went away and then reappeared, as visible
>> in the logs.  What has the changed in the ctl.conf?  Do you have any
>> iscsi-related sysctls set on the initiator side?
>>
> 
> I've added a new target in the ctl.conf . On the linux server I also see
> a brief disconnect,  but a reconnect is handled well.
> 
> I haven't set any sysctl's related to iscsi
> 
> My ctl.conf is (again anonymized):
> 
> auth-group my-auth {
> chap "myiscsi" "verysecret"
> }
> portal-group pg0 {
> discovery-auth-group my-auth
> listen 10.13.37.2
> }
> 
> target iqn.2015-03.lan.my.nas:vmstorage-29 {
> auth-group my-auth
> portal-group pg0
> lun 0 {
> path /tank/images/iscsi/vmstorage-29/vmstorage-29.img
> size 20484M
> blocksize 4096
> option unmap on
> }
> }
> 
> target iqn.2015-03.lan.my.nas:vmstorage-44 {
> auth-group my-auth
> portal-group pg0
> lun 0 {
> path /tank/images/iscsi/vmstorage-44/vmstorage-44.img
> size 102404M
> blocksize 4096
> option unmap on
> }
> }
> 
> target iqn.2015-03.lan.my.nas:keyserver.my.nl {
> auth-group my-auth
> portal-group pg0
> lun 0 {
> path /dev/zvol/tank/hosting_images/keyserver.my.nl
> blocksize 4096
> option unmap on
> }
> }
> 
> the vmstorage-44 is last added to the config
> 

I've started up my test environment, but I could not reproduce is there.
When reloading the target, a
SCSI sense: UNIT ATTENTION asc:3f,e (Reported LUNs data has changed) is
reported, but no device is destroyed. All servers are running the same
kernel and userland 10.2-RELEASE FreeBSD 10.2-RELEASE #0 r28

What can those device disconnects cause?

> 
> 
>> ___
>> freebsd-stable@freebsd.org mailing list
>> https://lists.freebsd.org/mailman/listinfo/freebsd-stable
>> To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"
>>
> 
> ___
> freebsd-stable@freebsd.org mailing list
> https://lists.freebsd.org/mailman/listinfo/freebsd-stable
> To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"
> 

___
freebsd-stable@freebsd.org mailing list
https://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"

Re: Reloading iscsi target, iscsi initiator panics

2015-10-15 Thread Frank de Bot (lists)
Edward Tomasz Napierała wrote:
> On 1014T2316, Frank de Bot (lists) wrote:
>> Hello,
>>
>> I have a FreeBSD 10.2 server running as iSCSI target. Another FreeBSD
>> 10.2 is an initiator and has several targets used.
>>
>> When I add a target and reload its config with 'service ctld reload',
>> the FreeBSD initiator server panics. It's reproducable (I have a
>> coredump, but I think it's clear what the problem is)
> 
> How easy it is to reproduce, eg. does it happen about every time?
> Could you paste the backtrace?

The first time I didn't understand what happened, the second time I
could relate it directly to the reloading of ctld on the iSCSI, within
moments the server paniced.

I got 2 backtraces:

First one:

Fatal trap 12: page fault while in kernel mode
cpuid = 1; softdep_deallocate_dependencies: got error 6 while accessing
filesystem
apic id = 01
fault virtual address   = 0x0
fault code  = supervisor read data, page not present
instruction pointer = 0x20:0x808a86b5
stack pointer   = 0x28:0xfe023b5f1200
frame pointer   = 0x28:0xfe023b5f1260
code segment= base 0x0, limit 0xf, type 0x1b
= DPL 0, pres 1, long 1, def32 0, gran 1
processor eflags= interrupt enabled, resume, IOPL = 0
current process = 42663 (php)
trap number = 12
panic: page fault
cpuid = 1
KDB: stack backtrace:
#0 0x80984e30 at kdb_backtrace+0x60
#1 0x809489e6 at vpanic+0x126
#2 0x809488b3 at panic+0x43
#3 0x80d4aadb at trap_fatal+0x36b
#4 0x80d4addd at trap_pfault+0x2ed
#5 0x80d4a47a at trap+0x47a
#6 0x80d307f2 at calltrap+0x8
#7 0x80b73208 at ffs_blkfree+0x168
#8 0x80b8c747 at handle_workitem_freefrag+0x127
#9 0x80b73e28 at ffs_reallocblks+0xc08
#10 0x80e74117 at VOP_REALLOCBLKS_APV+0xa7
#11 0x809da65b at cluster_write+0x3eb
#12 0x80ba4dcb at ffs_write+0x45b
#13 0x80e72495 at VOP_WRITE_APV+0x145
#14 0x809fc3f9 at vn_write+0x259
#15 0x809fc632 at vn_io_fault_doio+0x22
#16 0x809fa17c at vn_io_fault1+0x7c
#17 0x809f864b at vn_io_fault+0x18b
Uptime: 16d1h28m10s
(da3:iscsi10:0:0:0): Synchronize cache failed
Dumping 1154 out of 8163
MB:..2%..12%..21%..31%..41%..52%..61%..71%..81%..91%


Second:
Device resolver went missing before all of the data could be written to
it; expect data loss.
/dev: got error 6 while accessing filesystem
panic: softdep_deallocate_dependencies: dangling deps
cpuid = 0
KDB: stack backtrace:
#0 0x80984e30 at kdb_backtrace+0x60
#1 0x809489e6 at vpanic+0x126
#2 0x809488b3 at panic+0x43
#3 0x80b866da at softdep_deallocate_dependencies+0x6a
#4 0x809d1d95 at brelse+0x145
#5 0x809e9ee7 at flushbuflist+0x137
#6 0x809e9b5b at bufobj_invalbuf+0x1cb
#7 0x809ec4be at vgonel+0x17e
#8 0x809ec8ec at vgone+0x4c
#9 0x8082b4e8 at devfs_delete+0x218
#10 0x8082b95c at devfs_populate_loop+0x1fc
#11 0x8082b74a at devfs_populate+0x2a
#12 0x8082fa6b at devfs_populate_vp+0x9b
#13 0x808303bc at devfs_lookup+0x2c
#14 0x80e71621 at VOP_LOOKUP_APV+0xa1
#15 0x809e0951 at lookup+0x5a1
#16 0x809e00b4 at namei+0x4d4
#17 0x809f9135 at vn_open_cred+0xd5
Uptime: 38m45s
(da3:iscsi4:0:0:0): Synchronize cache failed
(da4:iscsi5:0:0:0): Synchronize cache failed
Dumping 815 out of 8163 MB:..2%..12%..22%..32%..42%..52%..61%..71%..81%..91%

I've put the whole core.txt.%d files here (with my ip addresses and
hostnames anonymized):
 - http://frankdebot.nl/tmp/core.txt.0
 - http://frankdebot.nl/tmp/core.txt.1


> 
> ___
> freebsd-stable@freebsd.org mailing list
> https://lists.freebsd.org/mailman/listinfo/freebsd-stable
> To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"
> 

___
freebsd-stable@freebsd.org mailing list
https://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"

Reloading iscsi target, iscsi initiator panics

2015-10-14 Thread Frank de Bot (lists)
Hello,

I have a FreeBSD 10.2 server running as iSCSI target. Another FreeBSD
10.2 is an initiator and has several targets used.

When I add a target and reload its config with 'service ctld reload',
the FreeBSD initiator server panics. It's reproducable (I have a
coredump, but I think it's clear what the problem is)

How can I have the target reloads it config, without all initiators
crashing? Right before the crash the iscsi session drops and reinits


Regards,

Frank de Bot


Related dmesg from the initiator:
Oct 14 20:59:49 host004 kernel: (da0:iscsi1:0:0:0): WRITE(10). CDB: 2a
00 00 29 93 c0 00 00 08 00
Oct 14 20:59:49 host004 kernel: (da0:iscsi1:0:0:0): CAM status: SCSI
Status Error
Oct 14 20:59:49 host004 kernel: (da0:iscsi1:0:0:0): SCSI status: Check
Condition
Oct 14 20:59:49 host004 kernel: (da0:iscsi1:0:0:0): SCSI sense: UNIT
ATTENTION asc:3f,e (Reported LUNs data has changed)
Oct 14 20:59:49 host004 kernel: (da0:iscsi1:0:0:0): Retrying command
(per sense data)
Oct 14 20:59:49 host004 kernel: da0 at iscsi1 bus 0 scbus7 target 0 lun 0
Oct 14 20:59:49 host004 kernel: da0:  s/n MYSERIAL
  2 detached
Oct 14 20:59:49 host004 kernel: (da0:iscsi1:0:0:0): Periph destroyed
Oct 14 20:59:49 host004 kernel: da0 at iscsi1 bus 0 scbus7 target 0 lun 0
Oct 14 20:59:49 host004 kernel: da0:  Fixed Direct
Access SPC-4 SCSI device
Oct 14 20:59:49 host004 kernel: da0: Serial Number MYSERIAL   2
Oct 14 20:59:49 host004 kernel: da0: 150.000MB/s transfers
Oct 14 20:59:49 host004 kernel: da0: Command Queueing enabled
Oct 14 20:59:49 host004 kernel: da0: 20480MB (5242880 4096 byte sectors:
255H 63S/T 326C)
___
freebsd-stable@freebsd.org mailing list
https://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"


Re: em0 and em1 watchdog timed out

2015-10-05 Thread Frank de Bot (lists)
The connection is fine when working without lagg/lacp. As soon I
configure it, on load it the network devices will have watchdog
timeouts. The switch I use is a HP 1810-24G, PL.1.9

One of the other threads mentioned he used an other router (/switch) and
problems didn't occur anymore.

Unfortunately  I don't  have an other switch lying around to test with.


Frank


Frank de Bot (lists) wrote:
> FreeBSD 10.2 r28 . The drivers is Intel(R) PRO/1000 Network
> Connection 7.4.2 I will start some tests without lagg configured.
> 
> 
> Regards,
> 
> Frank de Bot
> 
> 
> Jack Vogel wrote:
>> You missed the all-important details: OS version, driver version.
>> And another question I can think of, do these interfaces watchdog if they
>> are not configured with lagg?
>>
>> Cheers,
>>
>> Jack
>>
>>
>> On Fri, Oct 2, 2015 at 10:37 AM, Frank de Bot (lists) <li...@searchy.net>
>> wrote:
>>
>>> On a server I have 2 interfaces configured in a lagg. It seemed to be
>>> running stable until now (about a week), both interfaces repeatable go
>>> down after a watchdog timeout.
>>>
>>> Output from /var/log/messages
>>>
>>> Oct  2 10:15:11 nas kernel: em0: Watchdog timeout Queue[0]-- resetting
>>> Oct  2 10:15:11 nas kernel: Interface is RUNNING and ACTIVE
>>> Oct  2 10:15:11 nas kernel: em0: TX Queue 0 --
>>> Oct  2 10:15:11 nas kernel: em0: hw tdh = 804, hw tdt = 905
>>> Oct  2 10:15:11 nas kernel: em0: Tx Queue Status = -2147483648
>>> Oct  2 10:15:11 nas kernel: em0: TX descriptors avail = 915
>>> Oct  2 10:15:11 nas kernel: em0: Tx Descriptors avail failure = 0
>>> Oct  2 10:15:11 nas kernel: em0: RX Queue 0 --
>>> Oct  2 10:15:11 nas kernel: em0: hw rdh = 912, hw rdt = 911
>>> Oct  2 10:15:11 nas kernel: em0: RX discarded packets = 0
>>> Oct  2 10:15:11 nas kernel: em0: RX Next to Check = 912
>>> Oct  2 10:15:11 nas kernel: em0: RX Next to Refresh = 911
>>> Oct  2 10:15:11 nas kernel: em0: link state changed to DOWN
>>> Oct  2 10:15:13 nas kernel: em0: link state changed to UP
>>> Oct  2 10:15:13 nas devd: Executing '/etc/rc.d/dhclient quietstart em0'
>>> Oct  2 10:21:56 nas kernel: em1: Interface stopped DISTRIBUTING,
>>> possible flapping
>>> Oct  2 10:21:59 nas kernel: em1: Watchdog timeout Queue[0]-- resetting
>>> Oct  2 10:21:59 nas kernel: Interface is RUNNING and ACTIVE
>>> Oct  2 10:21:59 nas kernel: em1: TX Queue 0 --
>>> Oct  2 10:21:59 nas kernel: em1: hw tdh = 686, hw tdt = 806
>>> Oct  2 10:21:59 nas kernel: em1: Tx Queue Status = -2147483648
>>> Oct  2 10:21:59 nas kernel: em1: TX descriptors avail = 896
>>> Oct  2 10:21:59 nas kernel: em1: Tx Descriptors avail failure = 0
>>> Oct  2 10:21:59 nas kernel: em1: RX Queue 0 --
>>> Oct  2 10:21:59 nas kernel: em1: hw rdh = 167, hw rdt = 166
>>> Oct  2 10:21:59 nas kernel: em1: RX discarded packets = 0
>>> Oct  2 10:21:59 nas kernel: em1: RX Next to Check = 167
>>> Oct  2 10:21:59 nas kernel: em1: RX Next to Refresh = 166
>>> Oct  2 10:21:59 nas kernel: em1: link state changed to DOWN
>>> Oct  2 10:22:02 nas kernel: em1: link state changed to UP
>>> Oct  2 10:22:02 nas devd: Executing '/etc/rc.d/dhclient quietstart em1'
>>>
>>> it seems to occur when the interface are loaded. On the same server I
>>> also have a PCI card, the same issue occurs here. em0 and em1 are both
>>> onboard 80003ES2LAN. the PCI card is 82571EB.
>>>
>>> I've found different threads, but I couldn't manage to find  a fix for
>>> this.
>>>
>>> What else can be wrong?
>>>
>>> Regards,
>>>
>>> Frank de Bot
>>> ___
>>> freebsd-stable@freebsd.org mailing list
>>> https://lists.freebsd.org/mailman/listinfo/freebsd-stable
>>> To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"
>>>
>> ___
>> freebsd-stable@freebsd.org mailing list
>> https://lists.freebsd.org/mailman/listinfo/freebsd-stable
>> To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"
>>
> 
> ___
> freebsd-stable@freebsd.org mailing list
> https://lists.freebsd.org/mailman/listinfo/freebsd-stable
> To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"
> 

___
freebsd-stable@freebsd.org mailing list
https://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"


Re: em0 and em1 watchdog timed out

2015-10-04 Thread Frank de Bot (lists)
FreeBSD 10.2 r28 . The drivers is Intel(R) PRO/1000 Network
Connection 7.4.2 I will start some tests without lagg configured.


Regards,

Frank de Bot


Jack Vogel wrote:
> You missed the all-important details: OS version, driver version.
> And another question I can think of, do these interfaces watchdog if they
> are not configured with lagg?
> 
> Cheers,
> 
> Jack
> 
> 
> On Fri, Oct 2, 2015 at 10:37 AM, Frank de Bot (lists) <li...@searchy.net>
> wrote:
> 
>> On a server I have 2 interfaces configured in a lagg. It seemed to be
>> running stable until now (about a week), both interfaces repeatable go
>> down after a watchdog timeout.
>>
>> Output from /var/log/messages
>>
>> Oct  2 10:15:11 nas kernel: em0: Watchdog timeout Queue[0]-- resetting
>> Oct  2 10:15:11 nas kernel: Interface is RUNNING and ACTIVE
>> Oct  2 10:15:11 nas kernel: em0: TX Queue 0 --
>> Oct  2 10:15:11 nas kernel: em0: hw tdh = 804, hw tdt = 905
>> Oct  2 10:15:11 nas kernel: em0: Tx Queue Status = -2147483648
>> Oct  2 10:15:11 nas kernel: em0: TX descriptors avail = 915
>> Oct  2 10:15:11 nas kernel: em0: Tx Descriptors avail failure = 0
>> Oct  2 10:15:11 nas kernel: em0: RX Queue 0 --
>> Oct  2 10:15:11 nas kernel: em0: hw rdh = 912, hw rdt = 911
>> Oct  2 10:15:11 nas kernel: em0: RX discarded packets = 0
>> Oct  2 10:15:11 nas kernel: em0: RX Next to Check = 912
>> Oct  2 10:15:11 nas kernel: em0: RX Next to Refresh = 911
>> Oct  2 10:15:11 nas kernel: em0: link state changed to DOWN
>> Oct  2 10:15:13 nas kernel: em0: link state changed to UP
>> Oct  2 10:15:13 nas devd: Executing '/etc/rc.d/dhclient quietstart em0'
>> Oct  2 10:21:56 nas kernel: em1: Interface stopped DISTRIBUTING,
>> possible flapping
>> Oct  2 10:21:59 nas kernel: em1: Watchdog timeout Queue[0]-- resetting
>> Oct  2 10:21:59 nas kernel: Interface is RUNNING and ACTIVE
>> Oct  2 10:21:59 nas kernel: em1: TX Queue 0 --
>> Oct  2 10:21:59 nas kernel: em1: hw tdh = 686, hw tdt = 806
>> Oct  2 10:21:59 nas kernel: em1: Tx Queue Status = -2147483648
>> Oct  2 10:21:59 nas kernel: em1: TX descriptors avail = 896
>> Oct  2 10:21:59 nas kernel: em1: Tx Descriptors avail failure = 0
>> Oct  2 10:21:59 nas kernel: em1: RX Queue 0 --
>> Oct  2 10:21:59 nas kernel: em1: hw rdh = 167, hw rdt = 166
>> Oct  2 10:21:59 nas kernel: em1: RX discarded packets = 0
>> Oct  2 10:21:59 nas kernel: em1: RX Next to Check = 167
>> Oct  2 10:21:59 nas kernel: em1: RX Next to Refresh = 166
>> Oct  2 10:21:59 nas kernel: em1: link state changed to DOWN
>> Oct  2 10:22:02 nas kernel: em1: link state changed to UP
>> Oct  2 10:22:02 nas devd: Executing '/etc/rc.d/dhclient quietstart em1'
>>
>> it seems to occur when the interface are loaded. On the same server I
>> also have a PCI card, the same issue occurs here. em0 and em1 are both
>> onboard 80003ES2LAN. the PCI card is 82571EB.
>>
>> I've found different threads, but I couldn't manage to find  a fix for
>> this.
>>
>> What else can be wrong?
>>
>> Regards,
>>
>> Frank de Bot
>> ___
>> freebsd-stable@freebsd.org mailing list
>> https://lists.freebsd.org/mailman/listinfo/freebsd-stable
>> To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"
>>
> ___
> freebsd-stable@freebsd.org mailing list
> https://lists.freebsd.org/mailman/listinfo/freebsd-stable
> To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"
> 

___
freebsd-stable@freebsd.org mailing list
https://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"


em0 and em1 watchdog timed out

2015-10-02 Thread Frank de Bot (lists)
On a server I have 2 interfaces configured in a lagg. It seemed to be
running stable until now (about a week), both interfaces repeatable go
down after a watchdog timeout.

Output from /var/log/messages

Oct  2 10:15:11 nas kernel: em0: Watchdog timeout Queue[0]-- resetting
Oct  2 10:15:11 nas kernel: Interface is RUNNING and ACTIVE
Oct  2 10:15:11 nas kernel: em0: TX Queue 0 --
Oct  2 10:15:11 nas kernel: em0: hw tdh = 804, hw tdt = 905
Oct  2 10:15:11 nas kernel: em0: Tx Queue Status = -2147483648
Oct  2 10:15:11 nas kernel: em0: TX descriptors avail = 915
Oct  2 10:15:11 nas kernel: em0: Tx Descriptors avail failure = 0
Oct  2 10:15:11 nas kernel: em0: RX Queue 0 --
Oct  2 10:15:11 nas kernel: em0: hw rdh = 912, hw rdt = 911
Oct  2 10:15:11 nas kernel: em0: RX discarded packets = 0
Oct  2 10:15:11 nas kernel: em0: RX Next to Check = 912
Oct  2 10:15:11 nas kernel: em0: RX Next to Refresh = 911
Oct  2 10:15:11 nas kernel: em0: link state changed to DOWN
Oct  2 10:15:13 nas kernel: em0: link state changed to UP
Oct  2 10:15:13 nas devd: Executing '/etc/rc.d/dhclient quietstart em0'
Oct  2 10:21:56 nas kernel: em1: Interface stopped DISTRIBUTING,
possible flapping
Oct  2 10:21:59 nas kernel: em1: Watchdog timeout Queue[0]-- resetting
Oct  2 10:21:59 nas kernel: Interface is RUNNING and ACTIVE
Oct  2 10:21:59 nas kernel: em1: TX Queue 0 --
Oct  2 10:21:59 nas kernel: em1: hw tdh = 686, hw tdt = 806
Oct  2 10:21:59 nas kernel: em1: Tx Queue Status = -2147483648
Oct  2 10:21:59 nas kernel: em1: TX descriptors avail = 896
Oct  2 10:21:59 nas kernel: em1: Tx Descriptors avail failure = 0
Oct  2 10:21:59 nas kernel: em1: RX Queue 0 --
Oct  2 10:21:59 nas kernel: em1: hw rdh = 167, hw rdt = 166
Oct  2 10:21:59 nas kernel: em1: RX discarded packets = 0
Oct  2 10:21:59 nas kernel: em1: RX Next to Check = 167
Oct  2 10:21:59 nas kernel: em1: RX Next to Refresh = 166
Oct  2 10:21:59 nas kernel: em1: link state changed to DOWN
Oct  2 10:22:02 nas kernel: em1: link state changed to UP
Oct  2 10:22:02 nas devd: Executing '/etc/rc.d/dhclient quietstart em1'

it seems to occur when the interface are loaded. On the same server I
also have a PCI card, the same issue occurs here. em0 and em1 are both
onboard 80003ES2LAN. the PCI card is 82571EB.

I've found different threads, but I couldn't manage to find  a fix for this.

What else can be wrong?

Regards,

Frank de Bot
___
freebsd-stable@freebsd.org mailing list
https://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"


NFS Client changing it's source address? (FreeBSD 10.2)

2015-08-29 Thread Frank de Bot (lists)
Hello,

I have a server running FreeBSD 10.2. It has several NFS mounts.
Frequently my NFS mount hang (v3). After a little investigation it looks
like FreeBSD has chosen a wrong source address for it's connections and
all packets are departing from the wrong interface.

Sockstat output:

[root@host004 ~]# sockstat -4 | grep 2049
root ssh14689 3  tcp4   10.4.2.4:2049910.4.2.5:22
??  ? ?  tcp4   10.13.37.4:67210.13.37.2:2049
??  ? ?  tcp4   79.x.x.210:90510.13.37.2:2049
??  ? ?  tcp4   79.x.x.210:99210.13.37.2:2049

tcpdump confirms nfs connection are trying to get out via the 79.x.x.x
interface

My fstab for the nfs mounts look like:

10.13.37.2:/tank/hostingbase  /opt/jails/hostingbase   nfs
nfsv3,ro,noatime,async,noauto0   0

/opt/jails/hostingbase  /opt/jails/test01   nullfs
ro,noatime,noauto  0   0
10.13.37.2:/tank/hosting/test   /opt/jails/test01/opt   nfs
nfsv3,noatime,async,rw,noauto   0   0
tmpfs   /opt/jails/test01/shm   tmpfs
rw,size=51200,noauto  0   0

/opt/jails/hostingbase  /opt/jails/test2   nullfs
ro,noatime,noauto  0   0
10.13.37.2:/tank/hosting/test2   /opt/jails/test2/opt   nfs
nfsv3,noatime,async,rw,noauto   0   0
tmpfs   /opt/jails/test2/shm   tmpfs
rw,size=51200,noauto  0   0


The change of source address looks to be happening after a nfs
connection  is re-established. At first everything works, I leave the
server idling (it's a test  server) and after that the mounts are hanging

10.2-RELEASE #0 r28 is the current running version.

Regards,

Frank de Bot
___
freebsd-stable@freebsd.org mailing list
https://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to freebsd-stable-unsubscr...@freebsd.org


Multiple IP/subnet in jail, source address for connections

2015-08-24 Thread Frank de Bot (lists)
Hello,

I'm trying to have jail with a public and a private IP address.  Both
are on the same interface. The public is called 79.x.x.213 and private
10.4.3.6
Out from ifconfig within the jail is:

inet 79.x.x.213 netmask 0x broadcast 79.x.x.213
inet 10.4.3.6 netmask 0x broadcast 10.4.3.6

When I try to reach a host on the 10.4.3.0/24 network, it will use the
source address 79.x.x.123 (seen with tcpdump)
When done outside of the jail on the server, it does have the right
source address.
How can I get my jail to have the right source address? Some tools
provide a way to define a source address, like telnet -s,  but it's not
workable.


Frank de Bot
___
freebsd-stable@freebsd.org mailing list
https://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to freebsd-stable-unsubscr...@freebsd.org


kernel process [nfscl] high cpu

2015-05-02 Thread Frank de Bot (lists)
Hi,

On a 10.1-RELEASE-p9 server I have several NFS mounts used for a jail.
Because it's a server only to test, there is a low load. But the [nfscl]
process is hogging a CPU after a while. This happens pretty fast, within
1 or 2 days. I'm noticing the high CPU of the process when I want to do
some test after a little while (those 1 or 2 days).

My jail.conf look like:

exec.start = /bin/sh /etc/rc;
exec.stop = /bin/sh /etc/rc.shutdown;
exec.clean;
mount.devfs;
exec.consolelog = /var/log/jail.$name.log;
#mount.fstab = /usr/local/etc/jail.fstab.$name;

test01 {
host.hostname = test01_hosting;
ip4.addr = somepublicaddress;
ip4.addr += someprivateaddress;

mount = 10.13.37.2:/tank/hostingbase  /opt/jails/test01
   nfs nfsv4,minorversion=1,pnfs,ro,noatime0   0;
mount +=  10.13.37.2:/tank/hosting/test
/opt/jails/test01/opt   nfs nfsv4,minorversion=1,pnfs,noatime
 0   0;

path = /opt/jails/test01;
}

Last test was with NFS 4.1, I also worked with NFS 4.(0) with the same
result. In the readonly nfs share there are symbolic links point to the
read-write share for logging, storing .run files, etc. When I monitor my
network interface with tcpdump, there is little nfs traffic, only when I
do try to access the shares there is activity.

What is causing nfscl to run around in circles, hogging the CPU (it
makes the system slow to respond too) or how can I found out what's the
cause?


Regards,

Frank de Bot
___
freebsd-stable@freebsd.org mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to freebsd-stable-unsubscr...@freebsd.org


Re: ZFS L2arc 16.0E size

2015-03-05 Thread Frank de Bot (lists)
Frank de Bot (lists) wrote:
 Hello,
 
 I have a FreeBSD 10.1 system with a raidz2 zfs configuration with 2ssd's
 for l2arc . It is running '10.1-STABLE FreeBSD 10.1-STABLE #0 r278805'
 Currently I'm running tests before it can go to production, but I have
 the following issue. After a while the l2arc devices indicate 16.0E free
 space and it starts 'consuming' more than it can hold
 

I've tried to 'debug' with dtrace, I found out different things:

- l2arc_write_buffers sometimes caused the vdev-vdev_stat.vs_alloc to
grow larger than vdev-vdev_asize.
- l2arc_dev-l2ad_end is larger than vdev-vdev_asize
- At some point l2arc_eviction isn't doing anything, but
l2ard_dev-l2ad_evict is higher than l2ard_dev-l2ad_hand . taddr is
matching l2ard_dev-l2ad_evict . I would assume it should evict that
space. l2arc_write_buffers will continue because there seems te be room
enough, I guess this would be caused by vdev_asize - vs_alloc is
negative and indicating a 16.0E freespace.

It could be that I'm assuming wrong things or interpret things wrong.
Please let me know.


Frank de Bot



___
freebsd-stable@freebsd.org mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to freebsd-stable-unsubscr...@freebsd.org