I think it's strange:

[root at node1 ~]# snoop -v -r -d xnf1
Using device xnf1 (promiscuous mode)
ETHER:  ----- Ether Header -----
ETHER:
ETHER:  Packet 1 arrived at 23:01:0.56727
ETHER:  Packet size = 60 bytes
ETHER:  Destination = 0:16:3e:46:23:9b,
ETHER:  Source      = 0:16:3e:46:31:9b,
ETHER:  Ethertype = 0833 (Unknown)
ETHER:

ETHER:  ----- Ether Header -----
ETHER:
ETHER:  Packet 2 arrived at 23:01:1.05863
ETHER:  Packet size = 42 bytes
ETHER:  Destination = 0:16:3e:46:31:9b,
ETHER:  Source      = 0:16:3e:46:23:9b,
ETHER:  VLAN ID     = 0
ETHER:  VLAN Priority = 7
ETHER:  Ethertype = 0833 (Unknown)
ETHER:

ETHER:  ----- Ether Header -----
ETHER:
ETHER:  Packet 3 arrived at 23:01:1.26870
ETHER:  Packet size = 60 bytes
ETHER:  Destination = 1:80:c2:0:0:e, (multicast)
ETHER:  Source      = 0:1a:70:20:c3:f9,
ETHER:  Ethertype = 88CC (Unknown)
ETHER:


[root at node1 ~]# arp -an | grep 172.16
xnf1   172.16.0.129         255.255.255.255 SPLA     00:16:3e:46:23:9b
xnf1   172.16.0.130         255.255.255.255          00:16:3e:46:31:9b
clprivnet0 172.16.4.1           255.255.255.255 SPLA     00:00:00:00:00:01
xnf2   172.16.1.1           255.255.255.255 SPLA     00:16:82:74:29:ba
xnf2   172.16.1.2           255.255.255.255          00:16:3e:74:2a:ba





Piotr Jasiukajtis pisze:
> Btw, I can ping node2 via interconnects from node1.
> I can't login to node2 neither from console nor via sshd.
> 
> 
> 
> Piotr Jasiukajtis pisze:
>> Hi,
>>
>> I installed Solaris Cluster Express 12/08 on 2 pvm virtual nodes (on 2
>> physical SXCE104 machines).
>>
>> Each virtual node has 3 NICs (3 VNICs from xVM dom0),
>> but each physical node has only one physical NIC connected to the one
>> physical switch. I know about security issues and such...
>>
>> There is a problem with interconnects so I can't create the cluster.
>>
>> Any idea? :)
>>
>>
>>
>> [root at node1 ~]# /usr/cluster/bin/clnode status -v
>>
>> === Cluster Nodes ===
>>
>> --- Node Status ---
>>
>> Node Name                                                       Status
>> ---------                                                       ------
>> node1                                                           Online
>> node2                                                           Offline
>>
>>
>> --- Node IPMP Group Status ---
>>
>> Node Name        Group Name        Status        Adapter        Status
>> ---------        ----------        ------        -------        ------
>> node1            sc_ipmp0          Online        xnf0           Online
>>
>>
>> [root at node1 ~]# /usr/cluster/bin/clintr show
>>
>> === Transport Cables ===
>>
>> Transport Cable:                                node1:xnf1,switch1 at 1
>>   Endpoint1:                                       node1:xnf1
>>   Endpoint2:                                       switch1 at 1
>>   State:                                           Enabled
>>
>> Transport Cable:                                node1:xnf2,switch2 at 1
>>   Endpoint1:                                       node1:xnf2
>>   Endpoint2:                                       switch2 at 1
>>   State:                                           Enabled
>>
>> Transport Cable:                                node2:xnf1,switch1 at 2
>>   Endpoint1:                                       node2:xnf1
>>   Endpoint2:                                       switch1 at 2
>>   State:                                           Enabled
>>
>> Transport Cable:                                node2:xnf2,switch2 at 2
>>   Endpoint1:                                       node2:xnf2
>>   Endpoint2:                                       switch2 at 2
>>   State:                                           Enabled
>>
>>
>> === Transport Switches ===
>>
>> Transport Switch:                               switch1
>>   State:                                           Enabled
>>   Type:                                            switch
>>   Port Names:                                      1 2
>>   Port State(1):                                   Enabled
>>   Port State(2):                                   Enabled
>>
>> Transport Switch:                               switch2
>>   State:                                           Enabled
>>   Type:                                            switch
>>   Port Names:                                      1 2
>>   Port State(1):                                   Enabled
>>   Port State(2):                                   Enabled
>>
>>
>> --- Transport Adapters for node1 ---
>>
>> Transport Adapter:                              xnf1
>>   State:                                           Enabled
>>   Transport Type:                                  dlpi
>>   device_name:                                     xnf
>>   device_instance:                                 1
>>   lazy_free:                                       1
>>   dlpi_heartbeat_timeout:                          10000
>>   dlpi_heartbeat_quantum:                          1000
>>   nw_bandwidth:                                    80
>>   bandwidth:                                       70
>>   ip_address:                                      172.16.0.129
>>   netmask:                                         255.255.255.128
>>   Port Names:                                      0
>>   Port State(0):                                   Enabled
>>
>> Transport Adapter:                              xnf2
>>   State:                                           Enabled
>>   Transport Type:                                  dlpi
>>   device_name:                                     xnf
>>   device_instance:                                 2
>>   lazy_free:                                       1
>>   dlpi_heartbeat_timeout:                          10000
>>   dlpi_heartbeat_quantum:                          1000
>>   nw_bandwidth:                                    80
>>   bandwidth:                                       70
>>   ip_address:                                      172.16.1.1
>>   netmask:                                         255.255.255.128
>>   Port Names:                                      0
>>   Port State(0):                                   Enabled
>>
>>
>> --- Transport Adapters for node2 ---
>>
>> Transport Adapter:                              xnf1
>>   State:                                           Enabled
>>   Transport Type:                                  dlpi
>>   device_name:                                     xnf
>>   device_instance:                                 1
>>   lazy_free:                                       1
>>   dlpi_heartbeat_timeout:                          10000
>>   dlpi_heartbeat_quantum:                          1000
>>   nw_bandwidth:                                    80
>>   bandwidth:                                       70
>>   ip_address:                                      172.16.0.130
>>   netmask:                                         255.255.255.128
>>   Port Names:                                      0
>>   Port State(0):                                   Enabled
>>
>> Transport Adapter:                              xnf2
>>   State:                                           Enabled
>>   Transport Type:                                  dlpi
>>   device_name:                                     xnf
>>   device_instance:                                 2
>>   lazy_free:                                       1
>>   dlpi_heartbeat_timeout:                          10000
>>   dlpi_heartbeat_quantum:                          1000
>>   nw_bandwidth:                                    80
>>   bandwidth:                                       70
>>   ip_address:                                      172.16.1.2
>>   netmask:                                         255.255.255.128
>>   Port Names:                                      0
>>   Port State(0):                                   Enabled
>>
>>
>> [root at node1 ~]# /usr/cluster/bin/clintr status -v
>>
>> === Cluster Transport Paths ===
>>
>> Endpoint1                       Endpoint2                       Status
>> ---------                       ---------                       ------
>> node1:xnf2                      node2:xnf2                      faulted
>> node1:xnf1                      node2:xnf1                      faulted
>>
>>
>>
>> Jan 10 11:22:14 node1 genunix: [ID 965873 kern.notice] NOTICE: CMM: Node
>> node1 (nodeid = 1) with votecount = 1 added.
>> Jan 10 11:22:14 node1 genunix: [ID 843983 kern.notice] NOTICE: CMM: Node
>> node1: attempting to join cluster.
>> Jan 10 11:22:14 node1 genunix: [ID 525628 kern.notice] NOTICE: CMM:
>> Cluster has reached quorum.
>> Jan 10 11:22:14 node1 genunix: [ID 377347 kern.notice] NOTICE: CMM: Node
>> node1 (nodeid = 1) is up; new incarnation number = 1231582933.
>> Jan 10 11:22:14 node1 genunix: [ID 108990 kern.notice] NOTICE: CMM:
>> Cluster members: node1.
>> Jan 10 11:22:14 node1 genunix: [ID 279084 kern.notice] NOTICE: CMM: node
>> reconfiguration #1 completed.
>> Jan 10 11:22:17 node1 genunix: [ID 499756 kern.notice] NOTICE: CMM: Node
>> node1: joined cluster.
>> Jan 10 11:22:17 node1 ip: [ID 856290 kern.notice] ip: joining multicasts
>> failed (18) on clprivnet0 - will use link layer broadcasts for multicast
>> Jan 10 11:22:28 node1 Cluster.CCR: [ID 914260 daemon.warning] Failed to
>> retrieve global fencing status from the global name server
>> Jan 10 11:22:28 node1 last message repeated 1 time
>> Jan 10 11:22:48 node1 Cluster.CCR: [ID 409585 daemon.error]
>> /usr/cluster/bin/scgdevs: Cannot register devices as HA.
>> Jan 10 11:22:53 node1 xntpd[909]: [ID 702911 daemon.notice] xntpd
>> 3-5.93e+sun 03/08/29 16:23:05 (1.4)
>> Jan 10 11:22:53 node1 xntpd[909]: [ID 301315 daemon.notice] tickadj = 5,
>> tick = 10000, tvu_maxslew = 495, est. hz = 100
>> Jan 10 11:22:53 node1 xntpd[909]: [ID 266339 daemon.notice] using kernel
>> phase-lock loop 0041, drift correction 0.00000
>> Jan 10 11:22:53 node1 last message repeated 1 time
>> Jan 10 11:23:00 node1 : [ID 386282 daemon.error] ccr_initialize failure
>> Jan 10 11:23:04 node1 last message repeated 8 times
>> Jan 10 11:23:04 node1 svc.startd[8]: [ID 748625 daemon.error]
>> system/cluster/scdpm:default failed repeatedly: transitioned to
>> maintenance (see 'svcs -xv' for details)
>> Jan 10 11:24:01 node1 xpvd: [ID 395608 kern.info] xenbus at 0, xenbus0
>> Jan 10 11:24:01 node1 genunix: [ID 936769 kern.info] xenbus0 is
>> /xpvd/xenbus at 0
>> Jan 10 11:28:15 node1 genunix: [ID 965873 kern.notice] NOTICE: CMM: Node
>> node2 (nodeid = 2) with votecount = 0 added.
>> Jan 10 11:28:15 node1 genunix: [ID 108990 kern.notice] NOTICE: CMM:
>> Cluster members: node1.
>> Jan 10 11:28:15 node1 genunix: [ID 279084 kern.notice] NOTICE: CMM: node
>> reconfiguration #2 completed.
>> Jan 10 11:28:16 node1 genunix: [ID 884114 kern.notice] NOTICE: clcomm:
>> Adapter xnf1 constructed
>> Jan 10 11:28:16 node1 ip: [ID 856290 kern.notice] ip: joining multicasts
>> failed (18) on clprivnet0 - will use link layer broadcasts for multicast
>> Jan 10 11:28:16 node1 genunix: [ID 884114 kern.notice] NOTICE: clcomm:
>> Adapter xnf2 constructed
>> Jan 10 11:28:25 node1 rpc_scadmd[1196]: [ID 801593 daemon.notice] stdout:
>> Jan 10 11:28:25 node1 rpc_scadmd[1196]: [ID 801593 daemon.notice] stderr:
>> Jan 10 11:28:26 node1 rpc_scadmd[1196]: [ID 801593 daemon.notice] stdout:
>> Jan 10 11:28:26 node1 rpc_scadmd[1196]: [ID 801593 daemon.notice] stderr:
>> Jan 10 11:29:16 node1 genunix: [ID 604153 kern.notice] NOTICE: clcomm:
>> Path node1:xnf1 - node2:xnf1 errors during initiation
>> Jan 10 11:29:16 node1 genunix: [ID 618107 kern.warning] WARNING: Path
>> node1:xnf1 - node2:xnf1 initiation encountered errors, errno = 62.
>> Remote node may be down or unreachable through this path.
>> Jan 10 11:29:16 node1 genunix: [ID 604153 kern.notice] NOTICE: clcomm:
>> Path node1:xnf2 - node2:xnf2 errors during initiation
>> Jan 10 11:29:16 node1 genunix: [ID 618107 kern.warning] WARNING: Path
>> node1:xnf2 - node2:xnf2 initiation encountered errors, errno = 62.
>> Remote node may be down or unreachable through this path.
>> Jan 10 11:30:24 node1 genunix: [ID 537175 kern.notice] NOTICE: CMM: Node
>> node2 (nodeid: 2, incarnation #: 1231583261) has become reachable.
>> Jan 10 11:30:24 node1 xnf: [ID 601036 kern.warning] WARNING: xnf2:
>> oversized packet (1518 bytes) dropped
>> Jan 10 11:30:24 node1 last message repeated 1 time
>> Jan 10 11:30:24 node1 genunix: [ID 387288 kern.notice] NOTICE: clcomm:
>> Path node1:xnf2 - node2:xnf2 online
>> Jan 10 11:30:24 node1 genunix: [ID 387288 kern.notice] NOTICE: clcomm:
>> Path node1:xnf1 - node2:xnf1 online
>> Jan 10 11:30:28 node1 xnf: [ID 601036 kern.warning] WARNING: xnf2:
>> oversized packet (1518 bytes) dropped
>> Jan 10 11:30:54 node1 last message repeated 2 times
>> Jan 10 11:30:54 node1 xnf: [ID 601036 kern.warning] WARNING: xnf1:
>> oversized packet (1518 bytes) dropped
>> Jan 10 11:31:08 node1 last message repeated 2 times
>> Jan 10 11:31:28 node1 xnf: [ID 601036 kern.warning] WARNING: xnf2:
>> oversized packet (1518 bytes) dropped
>> Jan 10 11:32:28 node1 last message repeated 1 time
>> Jan 10 11:33:29 node1 genunix: [ID 489438 kern.notice] NOTICE: clcomm:
>> Path node1:xnf2 - node2:xnf2 being drained
>> Jan 10 11:33:29 node1 genunix: [ID 387288 kern.notice] NOTICE: clcomm:
>> Path node1:xnf2 - node2:xnf2 online
>>
>>
>>
>>
>> [root at node1 ~]# uname -a
>> SunOS node1 5.11 snv_101a i86pc i386 i86xpv
>>
>>
>>
>> # xm start -c node2
>> v3.1.4-xvm chgset 'Mon Nov 24 22:48:21 2008 -0800 15909:8ac8abf844b5'
>> SunOS Release 5.11 Version snv_101a 64-bit
>> Copyright 1983-2008 Sun Microsystems, Inc.  All rights reserved.
>> Use is subject to license terms.
>> Hostname: node2
>> Configuring devices.
>> /usr/cluster/bin/scdidadm:  Could not load DID instance list.
>> /usr/cluster/bin/scdidadm:  Cannot open
>> /etc/cluster/ccr/global/did_instances.
>> Booting as part of a cluster
>> name is non-existent for this module
>> for a list of valid names, use name '?'
>> NOTICE: CMM: Node node1 (nodeid = 1) with votecount = 1 added.
>> NOTICE: CMM: Node node2 (nodeid = 2) with votecount = 0 added.
>> NOTICE: clcomm: Adapter xnf2 constructed
>> NOTICE: clcomm: Adapter xnf1 constructed
>> NOTICE: CMM: Node node2: attempting to join cluster.
>> NOTICE: CMM: Node node1 (nodeid: 1, incarnation #: 1231582933) has
>> become reachable.
>> WARNING: xnf1: oversized packet (1518 bytes) dropped
>> NOTICE: clcomm: Path node2:xnf1 - node1:xnf1 online
>> NOTICE: clcomm: Path node2:xnf2 - node1:xnf2 online
>> WARNING: xnf1: oversized packet (1518 bytes) dropped
>> WARNING: xnf1: oversized packet (1518 bytes) dropped
>> WARNING: xnf2: oversized packet (1518 bytes) dropped
>> WARNING: xnf2: oversized packet (1518 bytes) dropped
>> WARNING: xnf2: oversized packet (1518 bytes) dropped
>>
>>
>> # uname -srvi
>> SunOS 5.11 snv_104 i86xpv
>>
>> # dladm show-link
>> LINK        CLASS    MTU    STATE    OVER
>> bge0        phys     1500   unknown  --
>> e1000g0     phys     1500   up       --
>> vnic1       vnic     1500   unknown  e1000g0
>> vnic2       vnic     1500   unknown  e1000g0
>> vnic18      vnic     1500   unknown  e1000g0
>> vnic19      vnic     1500   unknown  e1000g0
>> vnic20      vnic     1500   unknown  e1000g0
>>
>>
>>
> 
> 


-- 
Regards,
Piotr Jasiukajtis | estibi | SCA OS0072
http://estseg.blogspot.com

Reply via email to