And it's configuration of node2 (node1 is almost the same):

# virsh dumpxml node2
<domain type='xen' id='13'>
  <name>node2</name>
  <uuid>ae1223b6-af23-dcdb-60b6-f11dc86b12d8</uuid>
  <bootloader>/usr/lib/xen/bin/pygrub</bootloader>
  <os>
    <type>linux</type>
  </os>
  <memory>786432</memory>
  <vcpu>1</vcpu>
  <on_poweroff>destroy</on_poweroff>
  <on_reboot>restart</on_reboot>
  <on_crash>restart</on_crash>
  <distro name='solaris'/>
  <clock offset='localtime'/>
  <devices>
    <interface type='bridge'>
      <source bridge='e1000g0 '/>
      <target dev='vif13.0'/>
      <mac address='00:16:3e:43:73:23'/>
      <script path='vif-vnic'/>
    </interface>
    <interface type='bridge'>
      <source bridge='e1000g0'/>
      <target dev='vif13.1'/>
      <mac address='00:16:3e:46:31:9b'/>
      <script path='vif-vnic'/>
    </interface>
    <interface type='bridge'>
      <source bridge='e1000g0'/>
      <target dev='vif13.2'/>
      <mac address='00:16:3e:74:2a:ba'/>
      <script path='vif-vnic'/>
    </interface>
    <disk type='block' device='disk'>
      <driver name='phy'/>
      <source dev='/dev/zvol/dsk/data/export/xvm/node2'/>
      <target dev='xvda'/>
    </disk>
    <console tty='/dev/pts/1'/>
  </devices>
</domain>

Btw, I CCed the xen-discuss list.

Piotr Jasiukajtis pisze:
> I think it's strange:
> 
> [root at node1 ~]# snoop -v -r -d xnf1
> Using device xnf1 (promiscuous mode)
> ETHER:  ----- Ether Header -----
> ETHER:
> ETHER:  Packet 1 arrived at 23:01:0.56727
> ETHER:  Packet size = 60 bytes
> ETHER:  Destination = 0:16:3e:46:23:9b,
> ETHER:  Source      = 0:16:3e:46:31:9b,
> ETHER:  Ethertype = 0833 (Unknown)
> ETHER:
> 
> ETHER:  ----- Ether Header -----
> ETHER:
> ETHER:  Packet 2 arrived at 23:01:1.05863
> ETHER:  Packet size = 42 bytes
> ETHER:  Destination = 0:16:3e:46:31:9b,
> ETHER:  Source      = 0:16:3e:46:23:9b,
> ETHER:  VLAN ID     = 0
> ETHER:  VLAN Priority = 7
> ETHER:  Ethertype = 0833 (Unknown)
> ETHER:
> 
> ETHER:  ----- Ether Header -----
> ETHER:
> ETHER:  Packet 3 arrived at 23:01:1.26870
> ETHER:  Packet size = 60 bytes
> ETHER:  Destination = 1:80:c2:0:0:e, (multicast)
> ETHER:  Source      = 0:1a:70:20:c3:f9,
> ETHER:  Ethertype = 88CC (Unknown)
> ETHER:
> 
> 
> [root at node1 ~]# arp -an | grep 172.16
> xnf1   172.16.0.129         255.255.255.255 SPLA     00:16:3e:46:23:9b
> xnf1   172.16.0.130         255.255.255.255          00:16:3e:46:31:9b
> clprivnet0 172.16.4.1           255.255.255.255 SPLA     00:00:00:00:00:01
> xnf2   172.16.1.1           255.255.255.255 SPLA     00:16:82:74:29:ba
> xnf2   172.16.1.2           255.255.255.255          00:16:3e:74:2a:ba
> 
> 
> 
> 
> 
> Piotr Jasiukajtis pisze:
>> Btw, I can ping node2 via interconnects from node1.
>> I can't login to node2 neither from console nor via sshd.
>>
>>
>>
>> Piotr Jasiukajtis pisze:
>>> Hi,
>>>
>>> I installed Solaris Cluster Express 12/08 on 2 pvm virtual nodes (on 2
>>> physical SXCE104 machines).
>>>
>>> Each virtual node has 3 NICs (3 VNICs from xVM dom0),
>>> but each physical node has only one physical NIC connected to the one
>>> physical switch. I know about security issues and such...
>>>
>>> There is a problem with interconnects so I can't create the cluster.
>>>
>>> Any idea? :)
>>>
>>>
>>>
>>> [root at node1 ~]# /usr/cluster/bin/clnode status -v
>>>
>>> === Cluster Nodes ===
>>>
>>> --- Node Status ---
>>>
>>> Node Name                                                       Status
>>> ---------                                                       ------
>>> node1                                                           Online
>>> node2                                                           Offline
>>>
>>>
>>> --- Node IPMP Group Status ---
>>>
>>> Node Name        Group Name        Status        Adapter        Status
>>> ---------        ----------        ------        -------        ------
>>> node1            sc_ipmp0          Online        xnf0           Online
>>>
>>>
>>> [root at node1 ~]# /usr/cluster/bin/clintr show
>>>
>>> === Transport Cables ===
>>>
>>> Transport Cable:                                node1:xnf1,switch1 at 1
>>>   Endpoint1:                                       node1:xnf1
>>>   Endpoint2:                                       switch1 at 1
>>>   State:                                           Enabled
>>>
>>> Transport Cable:                                node1:xnf2,switch2 at 1
>>>   Endpoint1:                                       node1:xnf2
>>>   Endpoint2:                                       switch2 at 1
>>>   State:                                           Enabled
>>>
>>> Transport Cable:                                node2:xnf1,switch1 at 2
>>>   Endpoint1:                                       node2:xnf1
>>>   Endpoint2:                                       switch1 at 2
>>>   State:                                           Enabled
>>>
>>> Transport Cable:                                node2:xnf2,switch2 at 2
>>>   Endpoint1:                                       node2:xnf2
>>>   Endpoint2:                                       switch2 at 2
>>>   State:                                           Enabled
>>>
>>>
>>> === Transport Switches ===
>>>
>>> Transport Switch:                               switch1
>>>   State:                                           Enabled
>>>   Type:                                            switch
>>>   Port Names:                                      1 2
>>>   Port State(1):                                   Enabled
>>>   Port State(2):                                   Enabled
>>>
>>> Transport Switch:                               switch2
>>>   State:                                           Enabled
>>>   Type:                                            switch
>>>   Port Names:                                      1 2
>>>   Port State(1):                                   Enabled
>>>   Port State(2):                                   Enabled
>>>
>>>
>>> --- Transport Adapters for node1 ---
>>>
>>> Transport Adapter:                              xnf1
>>>   State:                                           Enabled
>>>   Transport Type:                                  dlpi
>>>   device_name:                                     xnf
>>>   device_instance:                                 1
>>>   lazy_free:                                       1
>>>   dlpi_heartbeat_timeout:                          10000
>>>   dlpi_heartbeat_quantum:                          1000
>>>   nw_bandwidth:                                    80
>>>   bandwidth:                                       70
>>>   ip_address:                                      172.16.0.129
>>>   netmask:                                         255.255.255.128
>>>   Port Names:                                      0
>>>   Port State(0):                                   Enabled
>>>
>>> Transport Adapter:                              xnf2
>>>   State:                                           Enabled
>>>   Transport Type:                                  dlpi
>>>   device_name:                                     xnf
>>>   device_instance:                                 2
>>>   lazy_free:                                       1
>>>   dlpi_heartbeat_timeout:                          10000
>>>   dlpi_heartbeat_quantum:                          1000
>>>   nw_bandwidth:                                    80
>>>   bandwidth:                                       70
>>>   ip_address:                                      172.16.1.1
>>>   netmask:                                         255.255.255.128
>>>   Port Names:                                      0
>>>   Port State(0):                                   Enabled
>>>
>>>
>>> --- Transport Adapters for node2 ---
>>>
>>> Transport Adapter:                              xnf1
>>>   State:                                           Enabled
>>>   Transport Type:                                  dlpi
>>>   device_name:                                     xnf
>>>   device_instance:                                 1
>>>   lazy_free:                                       1
>>>   dlpi_heartbeat_timeout:                          10000
>>>   dlpi_heartbeat_quantum:                          1000
>>>   nw_bandwidth:                                    80
>>>   bandwidth:                                       70
>>>   ip_address:                                      172.16.0.130
>>>   netmask:                                         255.255.255.128
>>>   Port Names:                                      0
>>>   Port State(0):                                   Enabled
>>>
>>> Transport Adapter:                              xnf2
>>>   State:                                           Enabled
>>>   Transport Type:                                  dlpi
>>>   device_name:                                     xnf
>>>   device_instance:                                 2
>>>   lazy_free:                                       1
>>>   dlpi_heartbeat_timeout:                          10000
>>>   dlpi_heartbeat_quantum:                          1000
>>>   nw_bandwidth:                                    80
>>>   bandwidth:                                       70
>>>   ip_address:                                      172.16.1.2
>>>   netmask:                                         255.255.255.128
>>>   Port Names:                                      0
>>>   Port State(0):                                   Enabled
>>>
>>>
>>> [root at node1 ~]# /usr/cluster/bin/clintr status -v
>>>
>>> === Cluster Transport Paths ===
>>>
>>> Endpoint1                       Endpoint2                       Status
>>> ---------                       ---------                       ------
>>> node1:xnf2                      node2:xnf2                      faulted
>>> node1:xnf1                      node2:xnf1                      faulted
>>>
>>>
>>>
>>> Jan 10 11:22:14 node1 genunix: [ID 965873 kern.notice] NOTICE: CMM: Node
>>> node1 (nodeid = 1) with votecount = 1 added.
>>> Jan 10 11:22:14 node1 genunix: [ID 843983 kern.notice] NOTICE: CMM: Node
>>> node1: attempting to join cluster.
>>> Jan 10 11:22:14 node1 genunix: [ID 525628 kern.notice] NOTICE: CMM:
>>> Cluster has reached quorum.
>>> Jan 10 11:22:14 node1 genunix: [ID 377347 kern.notice] NOTICE: CMM: Node
>>> node1 (nodeid = 1) is up; new incarnation number = 1231582933.
>>> Jan 10 11:22:14 node1 genunix: [ID 108990 kern.notice] NOTICE: CMM:
>>> Cluster members: node1.
>>> Jan 10 11:22:14 node1 genunix: [ID 279084 kern.notice] NOTICE: CMM: node
>>> reconfiguration #1 completed.
>>> Jan 10 11:22:17 node1 genunix: [ID 499756 kern.notice] NOTICE: CMM: Node
>>> node1: joined cluster.
>>> Jan 10 11:22:17 node1 ip: [ID 856290 kern.notice] ip: joining multicasts
>>> failed (18) on clprivnet0 - will use link layer broadcasts for multicast
>>> Jan 10 11:22:28 node1 Cluster.CCR: [ID 914260 daemon.warning] Failed to
>>> retrieve global fencing status from the global name server
>>> Jan 10 11:22:28 node1 last message repeated 1 time
>>> Jan 10 11:22:48 node1 Cluster.CCR: [ID 409585 daemon.error]
>>> /usr/cluster/bin/scgdevs: Cannot register devices as HA.
>>> Jan 10 11:22:53 node1 xntpd[909]: [ID 702911 daemon.notice] xntpd
>>> 3-5.93e+sun 03/08/29 16:23:05 (1.4)
>>> Jan 10 11:22:53 node1 xntpd[909]: [ID 301315 daemon.notice] tickadj = 5,
>>> tick = 10000, tvu_maxslew = 495, est. hz = 100
>>> Jan 10 11:22:53 node1 xntpd[909]: [ID 266339 daemon.notice] using kernel
>>> phase-lock loop 0041, drift correction 0.00000
>>> Jan 10 11:22:53 node1 last message repeated 1 time
>>> Jan 10 11:23:00 node1 : [ID 386282 daemon.error] ccr_initialize failure
>>> Jan 10 11:23:04 node1 last message repeated 8 times
>>> Jan 10 11:23:04 node1 svc.startd[8]: [ID 748625 daemon.error]
>>> system/cluster/scdpm:default failed repeatedly: transitioned to
>>> maintenance (see 'svcs -xv' for details)
>>> Jan 10 11:24:01 node1 xpvd: [ID 395608 kern.info] xenbus at 0, xenbus0
>>> Jan 10 11:24:01 node1 genunix: [ID 936769 kern.info] xenbus0 is
>>> /xpvd/xenbus at 0
>>> Jan 10 11:28:15 node1 genunix: [ID 965873 kern.notice] NOTICE: CMM: Node
>>> node2 (nodeid = 2) with votecount = 0 added.
>>> Jan 10 11:28:15 node1 genunix: [ID 108990 kern.notice] NOTICE: CMM:
>>> Cluster members: node1.
>>> Jan 10 11:28:15 node1 genunix: [ID 279084 kern.notice] NOTICE: CMM: node
>>> reconfiguration #2 completed.
>>> Jan 10 11:28:16 node1 genunix: [ID 884114 kern.notice] NOTICE: clcomm:
>>> Adapter xnf1 constructed
>>> Jan 10 11:28:16 node1 ip: [ID 856290 kern.notice] ip: joining multicasts
>>> failed (18) on clprivnet0 - will use link layer broadcasts for multicast
>>> Jan 10 11:28:16 node1 genunix: [ID 884114 kern.notice] NOTICE: clcomm:
>>> Adapter xnf2 constructed
>>> Jan 10 11:28:25 node1 rpc_scadmd[1196]: [ID 801593 daemon.notice] stdout:
>>> Jan 10 11:28:25 node1 rpc_scadmd[1196]: [ID 801593 daemon.notice] stderr:
>>> Jan 10 11:28:26 node1 rpc_scadmd[1196]: [ID 801593 daemon.notice] stdout:
>>> Jan 10 11:28:26 node1 rpc_scadmd[1196]: [ID 801593 daemon.notice] stderr:
>>> Jan 10 11:29:16 node1 genunix: [ID 604153 kern.notice] NOTICE: clcomm:
>>> Path node1:xnf1 - node2:xnf1 errors during initiation
>>> Jan 10 11:29:16 node1 genunix: [ID 618107 kern.warning] WARNING: Path
>>> node1:xnf1 - node2:xnf1 initiation encountered errors, errno = 62.
>>> Remote node may be down or unreachable through this path.
>>> Jan 10 11:29:16 node1 genunix: [ID 604153 kern.notice] NOTICE: clcomm:
>>> Path node1:xnf2 - node2:xnf2 errors during initiation
>>> Jan 10 11:29:16 node1 genunix: [ID 618107 kern.warning] WARNING: Path
>>> node1:xnf2 - node2:xnf2 initiation encountered errors, errno = 62.
>>> Remote node may be down or unreachable through this path.
>>> Jan 10 11:30:24 node1 genunix: [ID 537175 kern.notice] NOTICE: CMM: Node
>>> node2 (nodeid: 2, incarnation #: 1231583261) has become reachable.
>>> Jan 10 11:30:24 node1 xnf: [ID 601036 kern.warning] WARNING: xnf2:
>>> oversized packet (1518 bytes) dropped
>>> Jan 10 11:30:24 node1 last message repeated 1 time
>>> Jan 10 11:30:24 node1 genunix: [ID 387288 kern.notice] NOTICE: clcomm:
>>> Path node1:xnf2 - node2:xnf2 online
>>> Jan 10 11:30:24 node1 genunix: [ID 387288 kern.notice] NOTICE: clcomm:
>>> Path node1:xnf1 - node2:xnf1 online
>>> Jan 10 11:30:28 node1 xnf: [ID 601036 kern.warning] WARNING: xnf2:
>>> oversized packet (1518 bytes) dropped
>>> Jan 10 11:30:54 node1 last message repeated 2 times
>>> Jan 10 11:30:54 node1 xnf: [ID 601036 kern.warning] WARNING: xnf1:
>>> oversized packet (1518 bytes) dropped
>>> Jan 10 11:31:08 node1 last message repeated 2 times
>>> Jan 10 11:31:28 node1 xnf: [ID 601036 kern.warning] WARNING: xnf2:
>>> oversized packet (1518 bytes) dropped
>>> Jan 10 11:32:28 node1 last message repeated 1 time
>>> Jan 10 11:33:29 node1 genunix: [ID 489438 kern.notice] NOTICE: clcomm:
>>> Path node1:xnf2 - node2:xnf2 being drained
>>> Jan 10 11:33:29 node1 genunix: [ID 387288 kern.notice] NOTICE: clcomm:
>>> Path node1:xnf2 - node2:xnf2 online
>>>
>>>
>>>
>>>
>>> [root at node1 ~]# uname -a
>>> SunOS node1 5.11 snv_101a i86pc i386 i86xpv
>>>
>>>
>>>
>>> # xm start -c node2
>>> v3.1.4-xvm chgset 'Mon Nov 24 22:48:21 2008 -0800 15909:8ac8abf844b5'
>>> SunOS Release 5.11 Version snv_101a 64-bit
>>> Copyright 1983-2008 Sun Microsystems, Inc.  All rights reserved.
>>> Use is subject to license terms.
>>> Hostname: node2
>>> Configuring devices.
>>> /usr/cluster/bin/scdidadm:  Could not load DID instance list.
>>> /usr/cluster/bin/scdidadm:  Cannot open
>>> /etc/cluster/ccr/global/did_instances.
>>> Booting as part of a cluster
>>> name is non-existent for this module
>>> for a list of valid names, use name '?'
>>> NOTICE: CMM: Node node1 (nodeid = 1) with votecount = 1 added.
>>> NOTICE: CMM: Node node2 (nodeid = 2) with votecount = 0 added.
>>> NOTICE: clcomm: Adapter xnf2 constructed
>>> NOTICE: clcomm: Adapter xnf1 constructed
>>> NOTICE: CMM: Node node2: attempting to join cluster.
>>> NOTICE: CMM: Node node1 (nodeid: 1, incarnation #: 1231582933) has
>>> become reachable.
>>> WARNING: xnf1: oversized packet (1518 bytes) dropped
>>> NOTICE: clcomm: Path node2:xnf1 - node1:xnf1 online
>>> NOTICE: clcomm: Path node2:xnf2 - node1:xnf2 online
>>> WARNING: xnf1: oversized packet (1518 bytes) dropped
>>> WARNING: xnf1: oversized packet (1518 bytes) dropped
>>> WARNING: xnf2: oversized packet (1518 bytes) dropped
>>> WARNING: xnf2: oversized packet (1518 bytes) dropped
>>> WARNING: xnf2: oversized packet (1518 bytes) dropped
>>>
>>>
>>> # uname -srvi
>>> SunOS 5.11 snv_104 i86xpv
>>>
>>> # dladm show-link
>>> LINK        CLASS    MTU    STATE    OVER
>>> bge0        phys     1500   unknown  --
>>> e1000g0     phys     1500   up       --
>>> vnic1       vnic     1500   unknown  e1000g0
>>> vnic2       vnic     1500   unknown  e1000g0
>>> vnic18      vnic     1500   unknown  e1000g0
>>> vnic19      vnic     1500   unknown  e1000g0
>>> vnic20      vnic     1500   unknown  e1000g0
>>>
>>>
>>>
>>
> 
> 


-- 
Regards,
Piotr Jasiukajtis | estibi | SCA OS0072
http://estseg.blogspot.com

Reply via email to