Re: [OmniOS-discuss] RSF-1/ZFS panics node when offlining one iSCSI storage mirror
about iscsiadm modify initiator-node -T conn-login-max=60 You can set lower values but tuning settings depend on rsp-timeout and login-delay so you must set seqentially or together. http://docs.oracle.com/cd/E36784_01/html/E36836/iscsi-19.html ps If you use napp-it pro (16.03 dev), I have added initiator tuning as a menu under Comstar > Initiator > Settings Gea Am 07.03.2016 um 18:51 schrieb Stephan Budach: Hi Dan, Am 07.03.16 um 15:41 schrieb Dan McDonald: On Mar 6, 2016, at 9:44 AM, Stephan Budachwrote: when I noted that one node would panic, AS A RULE -- if you have an OmniOS box panic, you should save off the corefile (vmdump.N) and be able to share it with the list. I understand this may be an RSF-1 panic, BUT if it's not, it'd be nice to know. You can upload it to uploads.omniti.com if you wish, just request an upload token. Dan thanks - I will keep that in mind and I actually had a core dump available, but since I was testing around, I didn't mean to occupy anyone's time more than absolutely necessary and so I dumoed them. Speaking of that incident, I have lowered the iSCSI connection timeout to 60s, which seems to be the lowest value supported by issueing a iscsiadm modify initiator-node -T conn-login-max=60 and afterwards I used stmfadm offline target on the storage node to cut the target off. This time, the initiator timed out after 60s and that particular zpool changed it's status to degraded without anything happening. I still have to test that under load, but I will probably push that to next weekend. Thanks, Stephan ___ OmniOS-discuss mailing list OmniOS-discuss@lists.omniti.com http://lists.omniti.com/mailman/listinfo/omnios-discuss -- H f G Hochschule für Gestaltung university of design Schwäbisch Gmünd Rektor-Klaus Str. 100 73525 Schwäbisch Gmünd Guenther Alka, Dipl.-Ing. (FH) Leiter des Rechenzentrums head of computer center Tel 07171 602 627 Fax 07171 69259 guenther.a...@hfg-gmuend.de http://rz.hfg-gmuend.de ___ OmniOS-discuss mailing list OmniOS-discuss@lists.omniti.com http://lists.omniti.com/mailman/listinfo/omnios-discuss
Re: [OmniOS-discuss] RSF-1/ZFS panics node when offlining one iSCSI storage mirror
Hi Dan, Am 07.03.16 um 15:41 schrieb Dan McDonald: On Mar 6, 2016, at 9:44 AM, Stephan Budachwrote: when I noted that one node would panic, AS A RULE -- if you have an OmniOS box panic, you should save off the corefile (vmdump.N) and be able to share it with the list. I understand this may be an RSF-1 panic, BUT if it's not, it'd be nice to know. You can upload it to uploads.omniti.com if you wish, just request an upload token. Dan thanks - I will keep that in mind and I actually had a core dump available, but since I was testing around, I didn't mean to occupy anyone's time more than absolutely necessary and so I dumoed them. Speaking of that incident, I have lowered the iSCSI connection timeout to 60s, which seems to be the lowest value supported by issueing a iscsiadm modify initiator-node -T conn-login-max=60 and afterwards I used stmfadm offline target on the storage node to cut the target off. This time, the initiator timed out after 60s and that particular zpool changed it's status to degraded without anything happening. I still have to test that under load, but I will probably push that to next weekend. Thanks, Stephan ___ OmniOS-discuss mailing list OmniOS-discuss@lists.omniti.com http://lists.omniti.com/mailman/listinfo/omnios-discuss
Re: [OmniOS-discuss] RSF-1/ZFS panics node when offlining one iSCSI storage mirror
> On Mar 6, 2016, at 9:44 AM, Stephan Budachwrote: > > when I noted that one node would panic, AS A RULE -- if you have an OmniOS box panic, you should save off the corefile (vmdump.N) and be able to share it with the list. I understand this may be an RSF-1 panic, BUT if it's not, it'd be nice to know. You can upload it to uploads.omniti.com if you wish, just request an upload token. Dan ___ OmniOS-discuss mailing list OmniOS-discuss@lists.omniti.com http://lists.omniti.com/mailman/listinfo/omnios-discuss
[OmniOS-discuss] RSF-1/ZFS panics node when offlining one iSCSI storage mirror
Hi, I have set up a rather simple RSF-1 project, where two RSF-1 nodes connect to two storage heads via iSCSI. I have deployed one network and two disc heatbeats and I was trying all sorts of possible failures, when I noted that one node would panic, if I offlined an iSCSI target on one storage node and thus shutting down one side of a zpool mirror completely. Issueing a zpool status would't return and after a while the host got nuked. I then onlined the target again and waited until the node returned and than removed the local iSCSI initiator on the RSF-1 node instead, which resulted in a degraded, but functional zpool and this time, the node didn't get nuked. What is the difference between these two approaches and can I setup my systems such as that offlining a target doesn't lead to this behaviour? I'd imagine, that a target failure might as well occur as any other sofware fault. Thanks, Stephan ___ OmniOS-discuss mailing list OmniOS-discuss@lists.omniti.com http://lists.omniti.com/mailman/listinfo/omnios-discuss