Btw, I can ping node2 via interconnects from node1.
I can't login to node2 neither from console nor via sshd.



Piotr Jasiukajtis pisze:
> Hi,
> 
> I installed Solaris Cluster Express 12/08 on 2 pvm virtual nodes (on 2
> physical SXCE104 machines).
> 
> Each virtual node has 3 NICs (3 VNICs from xVM dom0),
> but each physical node has only one physical NIC connected to the one
> physical switch. I know about security issues and such...
> 
> There is a problem with interconnects so I can't create the cluster.
> 
> Any idea? :)
> 
> 
> 
> [root at node1 ~]# /usr/cluster/bin/clnode status -v
> 
> === Cluster Nodes ===
> 
> --- Node Status ---
> 
> Node Name                                                       Status
> ---------                                                       ------
> node1                                                           Online
> node2                                                           Offline
> 
> 
> --- Node IPMP Group Status ---
> 
> Node Name        Group Name        Status        Adapter        Status
> ---------        ----------        ------        -------        ------
> node1            sc_ipmp0          Online        xnf0           Online
> 
> 
> [root at node1 ~]# /usr/cluster/bin/clintr show
> 
> === Transport Cables ===
> 
> Transport Cable:                                node1:xnf1,switch1 at 1
>   Endpoint1:                                       node1:xnf1
>   Endpoint2:                                       switch1 at 1
>   State:                                           Enabled
> 
> Transport Cable:                                node1:xnf2,switch2 at 1
>   Endpoint1:                                       node1:xnf2
>   Endpoint2:                                       switch2 at 1
>   State:                                           Enabled
> 
> Transport Cable:                                node2:xnf1,switch1 at 2
>   Endpoint1:                                       node2:xnf1
>   Endpoint2:                                       switch1 at 2
>   State:                                           Enabled
> 
> Transport Cable:                                node2:xnf2,switch2 at 2
>   Endpoint1:                                       node2:xnf2
>   Endpoint2:                                       switch2 at 2
>   State:                                           Enabled
> 
> 
> === Transport Switches ===
> 
> Transport Switch:                               switch1
>   State:                                           Enabled
>   Type:                                            switch
>   Port Names:                                      1 2
>   Port State(1):                                   Enabled
>   Port State(2):                                   Enabled
> 
> Transport Switch:                               switch2
>   State:                                           Enabled
>   Type:                                            switch
>   Port Names:                                      1 2
>   Port State(1):                                   Enabled
>   Port State(2):                                   Enabled
> 
> 
> --- Transport Adapters for node1 ---
> 
> Transport Adapter:                              xnf1
>   State:                                           Enabled
>   Transport Type:                                  dlpi
>   device_name:                                     xnf
>   device_instance:                                 1
>   lazy_free:                                       1
>   dlpi_heartbeat_timeout:                          10000
>   dlpi_heartbeat_quantum:                          1000
>   nw_bandwidth:                                    80
>   bandwidth:                                       70
>   ip_address:                                      172.16.0.129
>   netmask:                                         255.255.255.128
>   Port Names:                                      0
>   Port State(0):                                   Enabled
> 
> Transport Adapter:                              xnf2
>   State:                                           Enabled
>   Transport Type:                                  dlpi
>   device_name:                                     xnf
>   device_instance:                                 2
>   lazy_free:                                       1
>   dlpi_heartbeat_timeout:                          10000
>   dlpi_heartbeat_quantum:                          1000
>   nw_bandwidth:                                    80
>   bandwidth:                                       70
>   ip_address:                                      172.16.1.1
>   netmask:                                         255.255.255.128
>   Port Names:                                      0
>   Port State(0):                                   Enabled
> 
> 
> --- Transport Adapters for node2 ---
> 
> Transport Adapter:                              xnf1
>   State:                                           Enabled
>   Transport Type:                                  dlpi
>   device_name:                                     xnf
>   device_instance:                                 1
>   lazy_free:                                       1
>   dlpi_heartbeat_timeout:                          10000
>   dlpi_heartbeat_quantum:                          1000
>   nw_bandwidth:                                    80
>   bandwidth:                                       70
>   ip_address:                                      172.16.0.130
>   netmask:                                         255.255.255.128
>   Port Names:                                      0
>   Port State(0):                                   Enabled
> 
> Transport Adapter:                              xnf2
>   State:                                           Enabled
>   Transport Type:                                  dlpi
>   device_name:                                     xnf
>   device_instance:                                 2
>   lazy_free:                                       1
>   dlpi_heartbeat_timeout:                          10000
>   dlpi_heartbeat_quantum:                          1000
>   nw_bandwidth:                                    80
>   bandwidth:                                       70
>   ip_address:                                      172.16.1.2
>   netmask:                                         255.255.255.128
>   Port Names:                                      0
>   Port State(0):                                   Enabled
> 
> 
> [root at node1 ~]# /usr/cluster/bin/clintr status -v
> 
> === Cluster Transport Paths ===
> 
> Endpoint1                       Endpoint2                       Status
> ---------                       ---------                       ------
> node1:xnf2                      node2:xnf2                      faulted
> node1:xnf1                      node2:xnf1                      faulted
> 
> 
> 
> Jan 10 11:22:14 node1 genunix: [ID 965873 kern.notice] NOTICE: CMM: Node
> node1 (nodeid = 1) with votecount = 1 added.
> Jan 10 11:22:14 node1 genunix: [ID 843983 kern.notice] NOTICE: CMM: Node
> node1: attempting to join cluster.
> Jan 10 11:22:14 node1 genunix: [ID 525628 kern.notice] NOTICE: CMM:
> Cluster has reached quorum.
> Jan 10 11:22:14 node1 genunix: [ID 377347 kern.notice] NOTICE: CMM: Node
> node1 (nodeid = 1) is up; new incarnation number = 1231582933.
> Jan 10 11:22:14 node1 genunix: [ID 108990 kern.notice] NOTICE: CMM:
> Cluster members: node1.
> Jan 10 11:22:14 node1 genunix: [ID 279084 kern.notice] NOTICE: CMM: node
> reconfiguration #1 completed.
> Jan 10 11:22:17 node1 genunix: [ID 499756 kern.notice] NOTICE: CMM: Node
> node1: joined cluster.
> Jan 10 11:22:17 node1 ip: [ID 856290 kern.notice] ip: joining multicasts
> failed (18) on clprivnet0 - will use link layer broadcasts for multicast
> Jan 10 11:22:28 node1 Cluster.CCR: [ID 914260 daemon.warning] Failed to
> retrieve global fencing status from the global name server
> Jan 10 11:22:28 node1 last message repeated 1 time
> Jan 10 11:22:48 node1 Cluster.CCR: [ID 409585 daemon.error]
> /usr/cluster/bin/scgdevs: Cannot register devices as HA.
> Jan 10 11:22:53 node1 xntpd[909]: [ID 702911 daemon.notice] xntpd
> 3-5.93e+sun 03/08/29 16:23:05 (1.4)
> Jan 10 11:22:53 node1 xntpd[909]: [ID 301315 daemon.notice] tickadj = 5,
> tick = 10000, tvu_maxslew = 495, est. hz = 100
> Jan 10 11:22:53 node1 xntpd[909]: [ID 266339 daemon.notice] using kernel
> phase-lock loop 0041, drift correction 0.00000
> Jan 10 11:22:53 node1 last message repeated 1 time
> Jan 10 11:23:00 node1 : [ID 386282 daemon.error] ccr_initialize failure
> Jan 10 11:23:04 node1 last message repeated 8 times
> Jan 10 11:23:04 node1 svc.startd[8]: [ID 748625 daemon.error]
> system/cluster/scdpm:default failed repeatedly: transitioned to
> maintenance (see 'svcs -xv' for details)
> Jan 10 11:24:01 node1 xpvd: [ID 395608 kern.info] xenbus at 0, xenbus0
> Jan 10 11:24:01 node1 genunix: [ID 936769 kern.info] xenbus0 is
> /xpvd/xenbus at 0
> Jan 10 11:28:15 node1 genunix: [ID 965873 kern.notice] NOTICE: CMM: Node
> node2 (nodeid = 2) with votecount = 0 added.
> Jan 10 11:28:15 node1 genunix: [ID 108990 kern.notice] NOTICE: CMM:
> Cluster members: node1.
> Jan 10 11:28:15 node1 genunix: [ID 279084 kern.notice] NOTICE: CMM: node
> reconfiguration #2 completed.
> Jan 10 11:28:16 node1 genunix: [ID 884114 kern.notice] NOTICE: clcomm:
> Adapter xnf1 constructed
> Jan 10 11:28:16 node1 ip: [ID 856290 kern.notice] ip: joining multicasts
> failed (18) on clprivnet0 - will use link layer broadcasts for multicast
> Jan 10 11:28:16 node1 genunix: [ID 884114 kern.notice] NOTICE: clcomm:
> Adapter xnf2 constructed
> Jan 10 11:28:25 node1 rpc_scadmd[1196]: [ID 801593 daemon.notice] stdout:
> Jan 10 11:28:25 node1 rpc_scadmd[1196]: [ID 801593 daemon.notice] stderr:
> Jan 10 11:28:26 node1 rpc_scadmd[1196]: [ID 801593 daemon.notice] stdout:
> Jan 10 11:28:26 node1 rpc_scadmd[1196]: [ID 801593 daemon.notice] stderr:
> Jan 10 11:29:16 node1 genunix: [ID 604153 kern.notice] NOTICE: clcomm:
> Path node1:xnf1 - node2:xnf1 errors during initiation
> Jan 10 11:29:16 node1 genunix: [ID 618107 kern.warning] WARNING: Path
> node1:xnf1 - node2:xnf1 initiation encountered errors, errno = 62.
> Remote node may be down or unreachable through this path.
> Jan 10 11:29:16 node1 genunix: [ID 604153 kern.notice] NOTICE: clcomm:
> Path node1:xnf2 - node2:xnf2 errors during initiation
> Jan 10 11:29:16 node1 genunix: [ID 618107 kern.warning] WARNING: Path
> node1:xnf2 - node2:xnf2 initiation encountered errors, errno = 62.
> Remote node may be down or unreachable through this path.
> Jan 10 11:30:24 node1 genunix: [ID 537175 kern.notice] NOTICE: CMM: Node
> node2 (nodeid: 2, incarnation #: 1231583261) has become reachable.
> Jan 10 11:30:24 node1 xnf: [ID 601036 kern.warning] WARNING: xnf2:
> oversized packet (1518 bytes) dropped
> Jan 10 11:30:24 node1 last message repeated 1 time
> Jan 10 11:30:24 node1 genunix: [ID 387288 kern.notice] NOTICE: clcomm:
> Path node1:xnf2 - node2:xnf2 online
> Jan 10 11:30:24 node1 genunix: [ID 387288 kern.notice] NOTICE: clcomm:
> Path node1:xnf1 - node2:xnf1 online
> Jan 10 11:30:28 node1 xnf: [ID 601036 kern.warning] WARNING: xnf2:
> oversized packet (1518 bytes) dropped
> Jan 10 11:30:54 node1 last message repeated 2 times
> Jan 10 11:30:54 node1 xnf: [ID 601036 kern.warning] WARNING: xnf1:
> oversized packet (1518 bytes) dropped
> Jan 10 11:31:08 node1 last message repeated 2 times
> Jan 10 11:31:28 node1 xnf: [ID 601036 kern.warning] WARNING: xnf2:
> oversized packet (1518 bytes) dropped
> Jan 10 11:32:28 node1 last message repeated 1 time
> Jan 10 11:33:29 node1 genunix: [ID 489438 kern.notice] NOTICE: clcomm:
> Path node1:xnf2 - node2:xnf2 being drained
> Jan 10 11:33:29 node1 genunix: [ID 387288 kern.notice] NOTICE: clcomm:
> Path node1:xnf2 - node2:xnf2 online
> 
> 
> 
> 
> [root at node1 ~]# uname -a
> SunOS node1 5.11 snv_101a i86pc i386 i86xpv
> 
> 
> 
> # xm start -c node2
> v3.1.4-xvm chgset 'Mon Nov 24 22:48:21 2008 -0800 15909:8ac8abf844b5'
> SunOS Release 5.11 Version snv_101a 64-bit
> Copyright 1983-2008 Sun Microsystems, Inc.  All rights reserved.
> Use is subject to license terms.
> Hostname: node2
> Configuring devices.
> /usr/cluster/bin/scdidadm:  Could not load DID instance list.
> /usr/cluster/bin/scdidadm:  Cannot open
> /etc/cluster/ccr/global/did_instances.
> Booting as part of a cluster
> name is non-existent for this module
> for a list of valid names, use name '?'
> NOTICE: CMM: Node node1 (nodeid = 1) with votecount = 1 added.
> NOTICE: CMM: Node node2 (nodeid = 2) with votecount = 0 added.
> NOTICE: clcomm: Adapter xnf2 constructed
> NOTICE: clcomm: Adapter xnf1 constructed
> NOTICE: CMM: Node node2: attempting to join cluster.
> NOTICE: CMM: Node node1 (nodeid: 1, incarnation #: 1231582933) has
> become reachable.
> WARNING: xnf1: oversized packet (1518 bytes) dropped
> NOTICE: clcomm: Path node2:xnf1 - node1:xnf1 online
> NOTICE: clcomm: Path node2:xnf2 - node1:xnf2 online
> WARNING: xnf1: oversized packet (1518 bytes) dropped
> WARNING: xnf1: oversized packet (1518 bytes) dropped
> WARNING: xnf2: oversized packet (1518 bytes) dropped
> WARNING: xnf2: oversized packet (1518 bytes) dropped
> WARNING: xnf2: oversized packet (1518 bytes) dropped
> 
> 
> # uname -srvi
> SunOS 5.11 snv_104 i86xpv
> 
> # dladm show-link
> LINK        CLASS    MTU    STATE    OVER
> bge0        phys     1500   unknown  --
> e1000g0     phys     1500   up       --
> vnic1       vnic     1500   unknown  e1000g0
> vnic2       vnic     1500   unknown  e1000g0
> vnic18      vnic     1500   unknown  e1000g0
> vnic19      vnic     1500   unknown  e1000g0
> vnic20      vnic     1500   unknown  e1000g0
> 
> 
> 


-- 
Regards,
Piotr Jasiukajtis | estibi | SCA OS0072
http://estseg.blogspot.com

Reply via email to