Re: [OmniOS-discuss] RSF-1/ZFS panics node when offlining one iSCSI storage mirror

2016-03-14 Thread Guenther Alka

about iscsiadm modify initiator-node -T conn-login-max=60

You can set lower values but tuning settings depend on rsp-timeout and 
login-delay

so you must set seqentially or together.

http://docs.oracle.com/cd/E36784_01/html/E36836/iscsi-19.html

ps
If you use napp-it pro (16.03 dev), I have added initiator tuning as a 
menu under Comstar > Initiator > Settings



Gea


Am 07.03.2016 um 18:51 schrieb Stephan Budach:

Hi Dan,

Am 07.03.16 um 15:41 schrieb Dan McDonald:
On Mar 6, 2016, at 9:44 AM, Stephan Budach  
wrote:


when I noted that one node would panic,
AS A RULE -- if you have an OmniOS box panic, you should save off the 
corefile (vmdump.N) and be able to share it with the list.  I 
understand this may be an RSF-1 panic, BUT if it's not, it'd be nice 
to know.


You can upload it to uploads.omniti.com if you wish, just request an 
upload token.


Dan

thanks - I will keep that in mind and I actually had a core dump 
available, but since I was testing around, I didn't mean to occupy 
anyone's time more than absolutely necessary and so I dumoed them.
Speaking of that incident, I have lowered the iSCSI connection timeout 
to 60s, which seems to be the lowest value supported by issueing a


iscsiadm modify initiator-node -T conn-login-max=60

and afterwards I used stmfadm offline target on the storage node to 
cut the target off. This time, the initiator timed out after 60s and 
that particular zpool changed it's status to degraded without anything 
happening. I still have to test that under load, but I will probably 
push that to next weekend.


Thanks,
Stephan
___
OmniOS-discuss mailing list
OmniOS-discuss@lists.omniti.com
http://lists.omniti.com/mailman/listinfo/omnios-discuss


--
H  f   G
Hochschule für Gestaltung
university of design

Schwäbisch Gmünd
Rektor-Klaus Str. 100
73525 Schwäbisch Gmünd

Guenther Alka, Dipl.-Ing. (FH)
Leiter des Rechenzentrums
head of computer center

Tel 07171 602 627
Fax 07171 69259
guenther.a...@hfg-gmuend.de
http://rz.hfg-gmuend.de

___
OmniOS-discuss mailing list
OmniOS-discuss@lists.omniti.com
http://lists.omniti.com/mailman/listinfo/omnios-discuss


Re: [OmniOS-discuss] RSF-1/ZFS panics node when offlining one iSCSI storage mirror

2016-03-07 Thread Stephan Budach

Hi Dan,

Am 07.03.16 um 15:41 schrieb Dan McDonald:

On Mar 6, 2016, at 9:44 AM, Stephan Budach  wrote:

when I noted that one node would panic,

AS A RULE -- if you have an OmniOS box panic, you should save off the corefile 
(vmdump.N) and be able to share it with the list.  I understand this may be an 
RSF-1 panic, BUT if it's not, it'd be nice to know.

You can upload it to uploads.omniti.com if you wish, just request an upload 
token.

Dan

thanks - I will keep that in mind and I actually had a core dump 
available, but since I was testing around, I didn't mean to occupy 
anyone's time more than absolutely necessary and so I dumoed them.
Speaking of that incident, I have lowered the iSCSI connection timeout 
to 60s, which seems to be the lowest value supported by issueing a


iscsiadm modify initiator-node -T conn-login-max=60

and afterwards I used stmfadm offline target on the storage node to cut 
the target off. This time, the initiator timed out after 60s and that 
particular zpool changed it's status to degraded without anything 
happening. I still have to test that under load, but I will probably 
push that to next weekend.


Thanks,
Stephan
___
OmniOS-discuss mailing list
OmniOS-discuss@lists.omniti.com
http://lists.omniti.com/mailman/listinfo/omnios-discuss


Re: [OmniOS-discuss] RSF-1/ZFS panics node when offlining one iSCSI storage mirror

2016-03-07 Thread Dan McDonald

> On Mar 6, 2016, at 9:44 AM, Stephan Budach  wrote:
> 
> when I noted that one node would panic,

AS A RULE -- if you have an OmniOS box panic, you should save off the corefile 
(vmdump.N) and be able to share it with the list.  I understand this may be an 
RSF-1 panic, BUT if it's not, it'd be nice to know.

You can upload it to uploads.omniti.com if you wish, just request an upload 
token.

Dan

___
OmniOS-discuss mailing list
OmniOS-discuss@lists.omniti.com
http://lists.omniti.com/mailman/listinfo/omnios-discuss


[OmniOS-discuss] RSF-1/ZFS panics node when offlining one iSCSI storage mirror

2016-03-06 Thread Stephan Budach

Hi,

I have set up a rather simple RSF-1 project, where two RSF-1 nodes 
connect to two storage heads via iSCSI. I have deployed one network and 
two disc heatbeats and I was trying all sorts of possible failures, when 
I noted that one node would panic, if I offlined an iSCSI target on one 
storage node and thus shutting down one side of a zpool mirror 
completely. Issueing a zpool status would't return and after a while the 
host got nuked.


I then onlined the target again and waited until the node returned and 
than removed the local iSCSI initiator on the RSF-1 node instead, which 
resulted in a degraded, but functional zpool and this time, the node 
didn't get nuked.


What is the difference between these two approaches and can I setup my 
systems such as that offlining a target doesn't lead to this behaviour? 
I'd imagine, that a target failure might as well occur as any other 
sofware fault.


Thanks,
Stephan
___
OmniOS-discuss mailing list
OmniOS-discuss@lists.omniti.com
http://lists.omniti.com/mailman/listinfo/omnios-discuss