Hi Kaushal,

Thanks for the detailed reply....let me explain my setup first :-

1. Ovirt Engine
2. 4* host as well as storage machine (Host and gluster combined)
3. Every host has 24 bricks...

Now whenever the host machine reboot...it can come up but can not join the
cluster again and through the following error "Gluster command [<UNKNOWN>]
failed on server.."

Please check my comment in line :-

1. Use the same string for doing the peer probe and for the brick address
during volume create/add-brick. Ideally, we suggest you use properly
resolvable FQDNs everywhere. If that is not possible, then use only IP
addresses. Try to avoid short names.
---------------
[root@cpu05 ~]# gluster peer status
Number of Peers: 3

Hostname: cpu03.stack.com
Uuid: 5729b8c4-e80d-4353-b456-6f467bddbdfb
State: Peer in Cluster (Connected)

Hostname: cpu04.stack.com
Uuid: d272b790-c4b2-4bed-ba68-793656e6d7b0
State: Peer in Cluster (Connected)
Other names:
10.10.0.8

Hostname: cpu02.stack.com
Uuid: 8d8a7041-950e-40d0-85f9-58d14340ca25
State: Peer in Cluster (Connected)
[root@cpu05 ~]#
----------------
2. During boot up, make sure to launch glusterd only after the network is
up. This will allow the new peer identification mechanism to do its
job correctly.
>> I think the service itself doing the same job....

[root@cpu05 ~]# cat /usr/lib/systemd/system/glusterd.service
[Unit]
Description=GlusterFS, a clustered file-system server
After=network.target rpcbind.service
Before=network-online.target

[Service]
Type=forking
PIDFile=/var/run/glusterd.pid
LimitNOFILE=65536
ExecStart=/usr/sbin/glusterd -p /var/run/glusterd.pid
KillMode=process

[Install]
WantedBy=multi-user.target
[root@cpu05 ~]#
--------------------

gluster logs :-

[2014-11-24 09:22:22.147471] I [MSGID: 100030] [glusterfsd.c:2018:main]
0-/usr/sbin/glusterd: Started running /usr/sbin/glusterd version 3.6.1
(args: /usr/sbin/glusterd -p /var/run/glusterd.pid)
[2014-11-24 09:22:22.151565] I [glusterd.c:1214:init] 0-management: Maximum
allowed open file descriptors set to 65536
[2014-11-24 09:22:22.151599] I [glusterd.c:1259:init] 0-management: Using
/var/lib/glusterd as working directory
[2014-11-24 09:22:22.155216] W [rdma.c:4195:__gf_rdma_ctx_create]
0-rpc-transport/rdma: rdma_cm event channel creation failed (No such device)
[2014-11-24 09:22:22.155264] E [rdma.c:4483:init] 0-rdma.management: Failed
to initialize IB Device
[2014-11-24 09:22:22.155285] E [rpc-transport.c:333:rpc_transport_load]
0-rpc-transport: 'rdma' initialization failed
[2014-11-24 09:22:22.155354] W [rpcsvc.c:1524:rpcsvc_transport_create]
0-rpc-service: cannot create listener, initing the transport failed
[2014-11-24 09:22:22.156290] I
[glusterd.c:413:glusterd_check_gsync_present] 0-glusterd: geo-replication
module not installed in the system
[2014-11-24 09:22:22.161318] I
[glusterd-store.c:2043:glusterd_restore_op_version] 0-glusterd: retrieved
op-version: 30600
[2014-11-24 09:22:22.821800] I
[glusterd-handler.c:3146:glusterd_friend_add_from_peerinfo] 0-management:
connect returned 0
[2014-11-24 09:22:22.825810] I
[glusterd-handler.c:3146:glusterd_friend_add_from_peerinfo] 0-management:
connect returned 0
[2014-11-24 09:22:22.828705] I
[glusterd-handler.c:3146:glusterd_friend_add_from_peerinfo] 0-management:
connect returned 0
[2014-11-24 09:22:22.828771] I [rpc-clnt.c:969:rpc_clnt_connection_init]
0-management: setting frame-timeout to 600
[2014-11-24 09:22:22.832670] I [rpc-clnt.c:969:rpc_clnt_connection_init]
0-management: setting frame-timeout to 600
[2014-11-24 09:22:22.835919] I [rpc-clnt.c:969:rpc_clnt_connection_init]
0-management: setting frame-timeout to 600
[2014-11-24 09:22:22.840209] E
[glusterd-store.c:4248:glusterd_resolve_all_bricks] 0-glusterd: resolve
brick failed in restore
[2014-11-24 09:22:22.840233] E [xlator.c:425:xlator_init] 0-management:
Initialization of volume 'management' failed, review your volfile again
[2014-11-24 09:22:22.840245] E [graph.c:322:glusterfs_graph_init]
0-management: initializing translator failed
[2014-11-24 09:22:22.840264] E [graph.c:525:glusterfs_graph_activate]
0-graph: init failed
[2014-11-24 09:22:22.840754] W [glusterfsd.c:1194:cleanup_and_exit] (-->
0-: received signum (0), shutting down

Thanks,
Punit




On Wed, Nov 26, 2014 at 7:14 PM, Kaushal M <kshlms...@gmail.com> wrote:

> Based on the logs I can guess that glusterd is being started before
> the network has come up and that the addresses given to bricks do not
> directly match the addresses used in during peer probe.
>
> The gluster_after_reboot log has the line "[2014-11-25
> 06:46:09.972113] E [glusterd-store.c:2632:glusterd_resolve_all_bricks]
> 0-glusterd: resolve brick failed in restore".
>
> Brick resolution fails when glusterd cannot match the address for the
> brick, with one of the peers. Brick resolution happens in two phases,
> 1. We first try to identify the peer by performing string comparisions
> with the brick address and the peer addresses (The peer names will be
> the names/addresses that were given when the peer was probed).
> 2. If we don't find a match from step 1, we will then resolve all the
> brick address and the peer addresses into addrinfo structs, and then
> compare these structs to find a match. This process should generally
> find a match if available. This will fail only if the network is not
> up yet as we cannot resolve addresses.
>
> The above steps are applicable only to glusterfs versions >=3.6. They
> were introduced to reduce problems with peer identification, like the
> one you encountered
>
> Since both of the steps failed to find a match in one run, but
> succeeded later, we can come to the conclusion that,
> a) the bricks don't have the exact same string used in peer probe for
> their addresses as step 1 failed, and
> b) the network was not up in the initial run, as step 2 failed during
> the initial run, but passed in the second run.
>
> Please let me know if my conclusion is correct.
>
> If it is, you can solve your problem in two ways.
> 1. Use the same string for doing the peer probe and for the brick
> address during volume create/add-brick. Ideally, we suggest you use
> properly resolvable FQDNs everywhere. If that is not possible, then
> use only IP addresses. Try to avoid short names.
> 2. During boot up, make sure to launch glusterd only after the network
> is up. This will allow the new peer identification mechanism to do its
> job correctly.
>
>
> If you have already followed these steps and yet still hit the
> problem, then please provide more information (setup, logs, etc.). It
> could be much different problem that you are facing.
>
> ~kaushal
>
> On Wed, Nov 26, 2014 at 4:01 PM, Punit Dambiwal <hypu...@gmail.com> wrote:
> > Is there any one can help on this ??
> >
> > Thanks,
> > punit
> >
> > On Wed, Nov 26, 2014 at 9:42 AM, Punit Dambiwal <hypu...@gmail.com>
> wrote:
> >>
> >> Hi,
> >>
> >> My Glusterfs version is :- glusterfs-3.6.1-1.el7
> >>
> >> On Wed, Nov 26, 2014 at 1:59 AM, Kanagaraj Mayilsamy <
> kmayi...@redhat.com>
> >> wrote:
> >>>
> >>> [+gluster-us...@gluster.org]
> >>>
> >>> "Initialization of volume 'management' failed, review your volfile
> >>> again", glusterd throws this error when the service is started
> automatically
> >>> after the reboot. But the service is successfully started later
> manually by
> >>> the user.
> >>>
> >>> can somebody from gluster-users please help on this?
> >>>
> >>> glusterfs version: 3.5.1
> >>>
> >>> Thanks,
> >>> Kanagaraj
> >>>
> >>> ----- Original Message -----
> >>> > From: "Punit Dambiwal" <hypu...@gmail.com>
> >>> > To: "Kanagaraj" <kmayi...@redhat.com>
> >>> > Cc: users@ovirt.org
> >>> > Sent: Tuesday, November 25, 2014 7:24:45 PM
> >>> > Subject: Re: [ovirt-users] Gluster command [<UNKNOWN>] failed on
> >>> > server...
> >>> >
> >>> > Hi Kanagraj,
> >>> >
> >>> > Please check the attached log files....i didn't find any thing
> >>> > special....
> >>> >
> >>> > On Tue, Nov 25, 2014 at 12:12 PM, Kanagaraj <kmayi...@redhat.com>
> >>> > wrote:
> >>> >
> >>> > >  Do you see any errors in
> >>> > > /var/log/glusterfs/etc-glusterfs-glusterd.vol.log or vdsm.log when
> >>> > > the
> >>> > > service is trying to start automatically after the reboot?
> >>> > >
> >>> > > Thanks,
> >>> > > Kanagaraj
> >>> > >
> >>> > >
> >>> > > On 11/24/2014 08:13 PM, Punit Dambiwal wrote:
> >>> > >
> >>> > > Hi Kanagaraj,
> >>> > >
> >>> > >  Yes...once i will start the gluster service and then vdsmd ...the
> >>> > > host
> >>> > > can connect to cluster...but the question is why it's not started
> >>> > > even it
> >>> > > has chkconfig enabled...
> >>> > >
> >>> > >  I have tested it in two host cluster environment...(Centos 6.6 and
> >>> > > centos 7.0) on both hypervisior cluster..it's failed to reconnect
> in
> >>> > > to
> >>> > > cluster after reboot....
> >>> > >
> >>> > >  In both the environment glusterd enabled for next boot....but it's
> >>> > > failed with the same error....seems it's bug in either gluster or
> >>> > > Ovirt ??
> >>> > >
> >>> > >  Please help me to find the workaround here if can not resolve
> >>> > > it...as
> >>> > > without this the Host machine can not connect after reboot....that
> >>> > > means
> >>> > > engine will consider it as down and every time need to manually
> start
> >>> > > the
> >>> > > gluster service and vdsmd... ??
> >>> > >
> >>> > >  Thanks,
> >>> > > Punit
> >>> > >
> >>> > > On Mon, Nov 24, 2014 at 10:20 PM, Kanagaraj <kmayi...@redhat.com>
> >>> > > wrote:
> >>> > >
> >>> > >>  From vdsm.log "error: Connection failed. Please check if gluster
> >>> > >> daemon
> >>> > >> is operational."
> >>> > >>
> >>> > >> Starting glusterd service should fix this issue. 'service glusterd
> >>> > >> start'
> >>> > >> But i am wondering why the glusterd was not started automatically
> >>> > >> after
> >>> > >> the reboot.
> >>> > >>
> >>> > >> Thanks,
> >>> > >> Kanagaraj
> >>> > >>
> >>> > >>
> >>> > >>
> >>> > >> On 11/24/2014 07:18 PM, Punit Dambiwal wrote:
> >>> > >>
> >>> > >> Hi Kanagaraj,
> >>> > >>
> >>> > >>  Please find the attached VDSM logs :-
> >>> > >>
> >>> > >>  ----------------
> >>> > >>  Thread-13::DEBUG::2014-11-24
> >>> > >>
> >>> > >>
> 21:41:17,182::resourceManager::977::Storage.ResourceManager.Owner::(cancelAll)
> >>> > >> Owner.cancelAll requests {}
> >>> > >> Thread-13::DEBUG::2014-11-24
> >>> > >> 21:41:17,182::task::993::Storage.TaskManager.Task::(_decref)
> >>> > >> Task=`1691d409-9b27-4585-8281-5ec26154367a`::ref 0 aborting False
> >>> > >> Thread-13::DEBUG::2014-11-24
> >>> > >> 21:41:32,393::task::595::Storage.TaskManager.Task::(_updateState)
> >>> > >> Task=`994c7bc3-a236-4d03-a732-e068c7ed9ed4`::moving from state
> init
> >>> > >> ->
> >>> > >> state preparing
> >>> > >> Thread-13::INFO::2014-11-24
> >>> > >> 21:41:32,393::logUtils::44::dispatcher::(wrapper) Run and protect:
> >>> > >> repoStats(options=None)
> >>> > >> Thread-13::INFO::2014-11-24
> >>> > >> 21:41:32,393::logUtils::47::dispatcher::(wrapper) Run and protect:
> >>> > >> repoStats, Return response: {}
> >>> > >> Thread-13::DEBUG::2014-11-24
> >>> > >> 21:41:32,393::task::1191::Storage.TaskManager.Task::(prepare)
> >>> > >> Task=`994c7bc3-a236-4d03-a732-e068c7ed9ed4`::finished: {}
> >>> > >> Thread-13::DEBUG::2014-11-24
> >>> > >> 21:41:32,394::task::595::Storage.TaskManager.Task::(_updateState)
> >>> > >> Task=`994c7bc3-a236-4d03-a732-e068c7ed9ed4`::moving from state
> >>> > >> preparing
> >>> > >> ->
> >>> > >> state finished
> >>> > >> Thread-13::DEBUG::2014-11-24
> >>> > >>
> >>> > >>
> 21:41:32,394::resourceManager::940::Storage.ResourceManager.Owner::(releaseAll)
> >>> > >> Owner.releaseAll requests {} resources {}
> >>> > >> Thread-13::DEBUG::2014-11-24
> >>> > >>
> >>> > >>
> 21:41:32,394::resourceManager::977::Storage.ResourceManager.Owner::(cancelAll)
> >>> > >> Owner.cancelAll requests {}
> >>> > >> Thread-13::DEBUG::2014-11-24
> >>> > >> 21:41:32,394::task::993::Storage.TaskManager.Task::(_decref)
> >>> > >> Task=`994c7bc3-a236-4d03-a732-e068c7ed9ed4`::ref 0 aborting False
> >>> > >> Thread-13::DEBUG::2014-11-24
> >>> > >> 21:41:41,550::BindingXMLRPC::1132::vds::(wrapper) client
> >>> > >> [10.10.10.2]::call
> >>> > >> getCapabilities with () {}
> >>> > >> Thread-13::DEBUG::2014-11-24
> >>> > >> 21:41:41,553::utils::738::root::(execCmd)
> >>> > >> /sbin/ip route show to 0.0.0.0/0 table all (cwd None)
> >>> > >> Thread-13::DEBUG::2014-11-24
> >>> > >> 21:41:41,560::utils::758::root::(execCmd)
> >>> > >> SUCCESS: <err> = ''; <rc> = 0
> >>> > >> Thread-13::DEBUG::2014-11-24
> >>> > >> 21:41:41,588::caps::728::root::(_getKeyPackages) rpm package
> >>> > >> ('gluster-swift',) not found
> >>> > >> Thread-13::DEBUG::2014-11-24
> >>> > >> 21:41:41,592::caps::728::root::(_getKeyPackages) rpm package
> >>> > >> ('gluster-swift-object',) not found
> >>> > >> Thread-13::DEBUG::2014-11-24
> >>> > >> 21:41:41,593::caps::728::root::(_getKeyPackages) rpm package
> >>> > >> ('gluster-swift-plugin',) not found
> >>> > >> Thread-13::DEBUG::2014-11-24
> >>> > >> 21:41:41,598::caps::728::root::(_getKeyPackages) rpm package
> >>> > >> ('gluster-swift-account',) not found
> >>> > >> Thread-13::DEBUG::2014-11-24
> >>> > >> 21:41:41,598::caps::728::root::(_getKeyPackages) rpm package
> >>> > >> ('gluster-swift-proxy',) not found
> >>> > >> Thread-13::DEBUG::2014-11-24
> >>> > >> 21:41:41,598::caps::728::root::(_getKeyPackages) rpm package
> >>> > >> ('gluster-swift-doc',) not found
> >>> > >> Thread-13::DEBUG::2014-11-24
> >>> > >> 21:41:41,599::caps::728::root::(_getKeyPackages) rpm package
> >>> > >> ('gluster-swift-container',) not found
> >>> > >> Thread-13::DEBUG::2014-11-24
> >>> > >> 21:41:41,599::caps::728::root::(_getKeyPackages) rpm package
> >>> > >> ('glusterfs-geo-replication',) not found
> >>> > >> Thread-13::DEBUG::2014-11-24 21:41:41,600::caps::646::root::(get)
> >>> > >> VirtioRNG DISABLED: libvirt version 0.10.2-29.el6_5.9 required >=
> >>> > >> 0.10.2-31
> >>> > >> Thread-13::DEBUG::2014-11-24
> >>> > >> 21:41:41,603::BindingXMLRPC::1139::vds::(wrapper) return
> >>> > >> getCapabilities
> >>> > >> with {'status': {'message': 'Done', 'code': 0}, 'info':
> >>> > >> {'HBAInventory':
> >>> > >> {'iSCSI': [{'InitiatorName':
> >>> > >> 'iqn.1994-05.com.redhat:32151ce183c8'}],
> >>> > >> 'FC':
> >>> > >> []}, 'packages2': {'kernel': {'release': '431.el6.x86_64',
> >>> > >> 'buildtime':
> >>> > >> 1385061309.0, 'version': '2.6.32'}, 'glusterfs-rdma': {'release':
> >>> > >> '1.el6',
> >>> > >> 'buildtime': 1403622628L, 'version': '3.5.1'}, 'glusterfs-fuse':
> >>> > >> {'release': '1.el6', 'buildtime': 1403622628L, 'version':
> '3.5.1'},
> >>> > >> 'spice-server': {'release': '6.el6_5.2', 'buildtime': 1402324637L,
> >>> > >> 'version': '0.12.4'}, 'vdsm': {'release': '1.gitdb83943.el6',
> >>> > >> 'buildtime':
> >>> > >> 1412784567L, 'version': '4.16.7'}, 'qemu-kvm': {'release':
> >>> > >> '2.415.el6_5.10', 'buildtime': 1402435700L, 'version':
> '0.12.1.2'},
> >>> > >> 'qemu-img': {'release': '2.415.el6_5.10', 'buildtime':
> 1402435700L,
> >>> > >> 'version': '0.12.1.2'}, 'libvirt': {'release': '29.el6_5.9',
> >>> > >> 'buildtime':
> >>> > >> 1402404612L, 'version': '0.10.2'}, 'glusterfs': {'release':
> '1.el6',
> >>> > >> 'buildtime': 1403622628L, 'version': '3.5.1'}, 'mom': {'release':
> >>> > >> '2.el6',
> >>> > >> 'buildtime': 1403794344L, 'version': '0.4.1'}, 'glusterfs-server':
> >>> > >> {'release': '1.el6', 'buildtime': 1403622628L, 'version':
> '3.5.1'}},
> >>> > >> 'numaNodeDistance': {'1': [20, 10], '0': [10, 20]}, 'cpuModel':
> >>> > >> 'Intel(R)
> >>> > >> Xeon(R) CPU           X5650  @ 2.67GHz', 'liveMerge': 'false',
> >>> > >> 'hooks':
> >>> > >> {},
> >>> > >> 'cpuSockets': '2', 'vmTypes': ['kvm'], 'selinux': {'mode': '1'},
> >>> > >> 'kdumpStatus': 0, 'supportedProtocols': ['2.2', '2.3'],
> 'networks':
> >>> > >> {'ovirtmgmt': {'iface': u'bond0.10', 'addr': '43.252.176.16',
> >>> > >> 'bridged':
> >>> > >> False, 'ipv6addrs': ['fe80::62eb:69ff:fe20:b46c/64'], 'mtu':
> '1500',
> >>> > >> 'bootproto4': 'none', 'netmask': '255.255.255.0', 'ipv4addrs': ['
> >>> > >> 43.252.176.16/24' <http://43.252.176.16/24%27>], 'interface':
> >>> > >> u'bond0.10', 'ipv6gateway': '::', 'gateway': '43.25.17.1'},
> >>> > >> 'Internal':
> >>> > >> {'iface': 'Internal', 'addr': '', 'cfg': {'DEFROUTE': 'no',
> >>> > >> 'HOTPLUG':
> >>> > >> 'no', 'MTU': '9000', 'DELAY': '0', 'NM_CONTROLLED': 'no',
> >>> > >> 'BOOTPROTO':
> >>> > >> 'none', 'STP': 'off', 'DEVICE': 'Internal', 'TYPE': 'Bridge',
> >>> > >> 'ONBOOT':
> >>> > >> 'no'}, 'bridged': True, 'ipv6addrs':
> >>> > >> ['fe80::210:18ff:fecd:daac/64'],
> >>> > >> 'gateway': '', 'bootproto4': 'none', 'netmask': '', 'stp': 'off',
> >>> > >> 'ipv4addrs': [], 'mtu': '9000', 'ipv6gateway': '::', 'ports':
> >>> > >> ['bond1.100']}, 'storage': {'iface': u'bond1', 'addr':
> '10.10.10.6',
> >>> > >> 'bridged': False, 'ipv6addrs': ['fe80::210:18ff:fecd:daac/64'],
> >>> > >> 'mtu':
> >>> > >> '9000', 'bootproto4': 'none', 'netmask': '255.255.255.0',
> >>> > >> 'ipv4addrs': ['
> >>> > >> 10.10.10.6/24' <http://10.10.10.6/24%27>], 'interface': u'bond1',
> >>> > >> 'ipv6gateway': '::', 'gateway': ''}, 'VMNetwork': {'iface':
> >>> > >> 'VMNetwork',
> >>> > >> 'addr': '', 'cfg': {'DEFROUTE': 'no', 'HOTPLUG': 'no', 'MTU':
> >>> > >> '1500',
> >>> > >> 'DELAY': '0', 'NM_CONTROLLED': 'no', 'BOOTPROTO': 'none', 'STP':
> >>> > >> 'off',
> >>> > >> 'DEVICE': 'VMNetwork', 'TYPE': 'Bridge', 'ONBOOT': 'no'},
> 'bridged':
> >>> > >> True,
> >>> > >> 'ipv6addrs': ['fe80::62eb:69ff:fe20:b46c/64'], 'gateway': '',
> >>> > >> 'bootproto4':
> >>> > >> 'none', 'netmask': '', 'stp': 'off', 'ipv4addrs': [], 'mtu':
> '1500',
> >>> > >> 'ipv6gateway': '::', 'ports': ['bond0.36']}}, 'bridges':
> >>> > >> {'Internal':
> >>> > >> {'addr': '', 'cfg': {'DEFROUTE': 'no', 'HOTPLUG': 'no', 'MTU':
> >>> > >> '9000',
> >>> > >> 'DELAY': '0', 'NM_CONTROLLED': 'no', 'BOOTPROTO': 'none', 'STP':
> >>> > >> 'off',
> >>> > >> 'DEVICE': 'Internal', 'TYPE': 'Bridge', 'ONBOOT': 'no'},
> >>> > >> 'ipv6addrs':
> >>> > >> ['fe80::210:18ff:fecd:daac/64'], 'mtu': '9000', 'netmask': '',
> >>> > >> 'stp':
> >>> > >> 'off', 'ipv4addrs': [], 'ipv6gateway': '::', 'gateway': '',
> 'opts':
> >>> > >> {'topology_change_detected': '0', 'multicast_last_member_count':
> >>> > >> '2',
> >>> > >> 'hash_elasticity': '4', 'multicast_query_response_interval':
> '999',
> >>> > >> 'multicast_snooping': '1', 'multicast_startup_query_interval':
> >>> > >> '3124',
> >>> > >> 'hello_timer': '31', 'multicast_querier_interval': '25496',
> >>> > >> 'max_age':
> >>> > >> '1999', 'hash_max': '512', 'stp_state': '0', 'root_id':
> >>> > >> '8000.001018cddaac', 'priority': '32768',
> >>> > >> 'multicast_membership_interval':
> >>> > >> '25996', 'root_path_cost': '0', 'root_port': '0',
> >>> > >> 'multicast_querier':
> >>> > >> '0',
> >>> > >> 'multicast_startup_query_count': '2', 'hello_time': '199',
> >>> > >> 'topology_change': '0', 'bridge_id': '8000.001018cddaac',
> >>> > >> 'topology_change_timer': '0', 'ageing_time': '29995', 'gc_timer':
> >>> > >> '31',
> >>> > >> 'group_addr': '1:80:c2:0:0:0', 'tcn_timer': '0',
> >>> > >> 'multicast_query_interval': '12498',
> >>> > >> 'multicast_last_member_interval':
> >>> > >> '99', 'multicast_router': '1', 'forward_delay': '0'}, 'ports':
> >>> > >> ['bond1.100']}, 'VMNetwork': {'addr': '', 'cfg': {'DEFROUTE':
> 'no',
> >>> > >> 'HOTPLUG': 'no', 'MTU': '1500', 'DELAY': '0', 'NM_CONTROLLED':
> 'no',
> >>> > >> 'BOOTPROTO': 'none', 'STP': 'off', 'DEVICE': 'VMNetwork', 'TYPE':
> >>> > >> 'Bridge',
> >>> > >> 'ONBOOT': 'no'}, 'ipv6addrs': ['fe80::62eb:69ff:fe20:b46c/64'],
> >>> > >> 'mtu':
> >>> > >> '1500', 'netmask': '', 'stp': 'off', 'ipv4addrs': [],
> 'ipv6gateway':
> >>> > >> '::',
> >>> > >> 'gateway': '', 'opts': {'topology_change_detected': '0',
> >>> > >> 'multicast_last_member_count': '2', 'hash_elasticity': '4',
> >>> > >> 'multicast_query_response_interval': '999', 'multicast_snooping':
> >>> > >> '1',
> >>> > >> 'multicast_startup_query_interval': '3124', 'hello_timer': '131',
> >>> > >> 'multicast_querier_interval': '25496', 'max_age': '1999',
> >>> > >> 'hash_max':
> >>> > >> '512', 'stp_state': '0', 'root_id': '8000.60eb6920b46c',
> 'priority':
> >>> > >> '32768', 'multicast_membership_interval': '25996',
> 'root_path_cost':
> >>> > >> '0',
> >>> > >> 'root_port': '0', 'multicast_querier': '0',
> >>> > >> 'multicast_startup_query_count': '2', 'hello_time': '199',
> >>> > >> 'topology_change': '0', 'bridge_id': '8000.60eb6920b46c',
> >>> > >> 'topology_change_timer': '0', 'ageing_time': '29995', 'gc_timer':
> >>> > >> '31',
> >>> > >> 'group_addr': '1:80:c2:0:0:0', 'tcn_timer': '0',
> >>> > >> 'multicast_query_interval': '12498',
> >>> > >> 'multicast_last_member_interval':
> >>> > >> '99', 'multicast_router': '1', 'forward_delay': '0'}, 'ports':
> >>> > >> ['bond0.36']}}, 'uuid': '44454C4C-4C00-1057-8053-B7C04F504E31',
> >>> > >> 'lastClientIface': 'bond1', 'nics': {'eth3': {'permhwaddr':
> >>> > >> '00:10:18:cd:da:ae', 'addr': '', 'cfg': {'SLAVE': 'yes',
> >>> > >> 'NM_CONTROLLED':
> >>> > >> 'no', 'MTU': '9000', 'HWADDR': '00:10:18:cd:da:ae', 'MASTER':
> >>> > >> 'bond1',
> >>> > >> 'DEVICE': 'eth3', 'ONBOOT': 'no'}, 'ipv6addrs': [], 'mtu': '9000',
> >>> > >> 'netmask': '', 'ipv4addrs': [], 'hwaddr': '00:10:18:cd:da:ac',
> >>> > >> 'speed':
> >>> > >> 1000}, 'eth2': {'permhwaddr': '00:10:18:cd:da:ac', 'addr': '',
> >>> > >> 'cfg':
> >>> > >> {'SLAVE': 'yes', 'NM_CONTROLLED': 'no', 'MTU': '9000', 'HWADDR':
> >>> > >> '00:10:18:cd:da:ac', 'MASTER': 'bond1', 'DEVICE': 'eth2',
> 'ONBOOT':
> >>> > >> 'no'},
> >>> > >> 'ipv6addrs': [], 'mtu': '9000', 'netmask': '', 'ipv4addrs': [],
> >>> > >> 'hwaddr':
> >>> > >> '00:10:18:cd:da:ac', 'speed': 1000}, 'eth1': {'permhwaddr':
> >>> > >> '60:eb:69:20:b4:6d', 'addr': '', 'cfg': {'SLAVE': 'yes',
> >>> > >> 'NM_CONTROLLED':
> >>> > >> 'no', 'MTU': '1500', 'HWADDR': '60:eb:69:20:b4:6d', 'MASTER':
> >>> > >> 'bond0',
> >>> > >> 'DEVICE': 'eth1', 'ONBOOT': 'yes'}, 'ipv6addrs': [], 'mtu':
> '1500',
> >>> > >> 'netmask': '', 'ipv4addrs': [], 'hwaddr': '60:eb:69:20:b4:6c',
> >>> > >> 'speed':
> >>> > >> 1000}, 'eth0': {'permhwaddr': '60:eb:69:20:b4:6c', 'addr': '',
> >>> > >> 'cfg':
> >>> > >> {'SLAVE': 'yes', 'NM_CONTROLLED': 'no', 'MTU': '1500', 'HWADDR':
> >>> > >> '60:eb:69:20:b4:6c', 'MASTER': 'bond0', 'DEVICE': 'eth0',
> 'ONBOOT':
> >>> > >> 'yes'},
> >>> > >> 'ipv6addrs': [], 'mtu': '1500', 'netmask': '', 'ipv4addrs': [],
> >>> > >> 'hwaddr':
> >>> > >> '60:eb:69:20:b4:6c', 'speed': 1000}}, 'software_revision': '1',
> >>> > >> 'clusterLevels': ['3.0', '3.1', '3.2', '3.3', '3.4', '3.5'],
> >>> > >> 'cpuFlags':
> >>> > >>
> >>> > >>
> u'fpu,vme,de,pse,tsc,msr,pae,mce,cx8,apic,sep,mtrr,pge,mca,cmov,pat,pse36,clflush,dts,acpi,mmx,fxsr,sse,sse2,ss,ht,tm,pbe,syscall,nx,pdpe1gb,rdtscp,lm,constant_tsc,arch_perfmon,pebs,bts,rep_good,xtopology,nonstop_tsc,pni,pclmulqdq,dtes64,monitor,ds_cpl,vmx,smx,est,tm2,ssse3,cx16,xtpr,pdcm,pcid,dca,sse4_1,sse4_2,popcnt,aes,lahf_lm,tpr_shadow,vnmi,flexpriority,ept,vpid,model_Nehalem,model_Conroe,model_coreduo,model_core2duo,model_Penryn,model_Westmere,model_n270',
> >>> > >> 'ISCSIInitiatorName': 'iqn.1994-05.com.redhat:32151ce183c8',
> >>> > >> 'netConfigDirty': 'False', 'supportedENGINEs': ['3.0', '3.1',
> '3.2',
> >>> > >> '3.3',
> >>> > >> '3.4', '3.5'], 'autoNumaBalancing': 2, 'reservedMem': '321',
> >>> > >> 'bondings':
> >>> > >> {'bond4': {'addr': '', 'cfg': {}, 'mtu': '1500', 'netmask': '',
> >>> > >> 'slaves':
> >>> > >> [], 'hwaddr': '00:00:00:00:00:00'}, 'bond0': {'addr': '', 'cfg':
> >>> > >> {'HOTPLUG': 'no', 'MTU': '1500', 'NM_CONTROLLED': 'no',
> >>> > >> 'BONDING_OPTS':
> >>> > >> 'mode=4 miimon=100', 'DEVICE': 'bond0', 'ONBOOT': 'yes'},
> >>> > >> 'ipv6addrs':
> >>> > >> ['fe80::62eb:69ff:fe20:b46c/64'], 'mtu': '1500', 'netmask': '',
> >>> > >> 'ipv4addrs': [], 'hwaddr': '60:eb:69:20:b4:6c', 'slaves': ['eth0',
> >>> > >> 'eth1'],
> >>> > >> 'opts': {'miimon': '100', 'mode': '4'}}, 'bond1': {'addr':
> >>> > >> '10.10.10.6',
> >>> > >> 'cfg': {'DEFROUTE': 'no', 'IPADDR': '10.10.10.6', 'HOTPLUG': 'no',
> >>> > >> 'MTU':
> >>> > >> '9000', 'NM_CONTROLLED': 'no', 'NETMASK': '255.255.255.0',
> >>> > >> 'BOOTPROTO':
> >>> > >> 'none', 'BONDING_OPTS': 'mode=4 miimon=100', 'DEVICE': 'bond1',
> >>> > >> 'ONBOOT':
> >>> > >> 'no'}, 'ipv6addrs': ['fe80::210:18ff:fecd:daac/64'], 'mtu':
> '9000',
> >>> > >> 'netmask': '255.255.255.0', 'ipv4addrs': ['10.10.10.6/24'
> >>> > >> <http://10.10.10.6/24%27>], 'hwaddr': '00:10:18:cd:da:ac',
> 'slaves':
> >>> > >> ['eth2', 'eth3'], 'opts': {'miimon': '100', 'mode': '4'}},
> 'bond2':
> >>> > >> {'addr': '', 'cfg': {}, 'mtu': '1500', 'netmask': '', 'slaves':
> [],
> >>> > >> 'hwaddr': '00:00:00:00:00:00'}, 'bond3': {'addr': '', 'cfg': {},
> >>> > >> 'mtu':
> >>> > >> '1500', 'netmask': '', 'slaves': [], 'hwaddr':
> >>> > >> '00:00:00:00:00:00'}},
> >>> > >> 'software_version': '4.16', 'memSize': '24019', 'cpuSpeed':
> >>> > >> '2667.000',
> >>> > >> 'numaNodes': {u'1': {'totalMemory': '12288', 'cpus': [6, 7, 8, 9,
> >>> > >> 10, 11,
> >>> > >> 18, 19, 20, 21, 22, 23]}, u'0': {'totalMemory': '12278', 'cpus':
> [0,
> >>> > >> 1, 2,
> >>> > >> 3, 4, 5, 12, 13, 14, 15, 16, 17]}}, 'version_name': 'Snow Man',
> >>> > >> 'vlans':
> >>> > >> {'bond0.10': {'iface': 'bond0', 'addr': '43.25.17.16', 'cfg':
> >>> > >> {'DEFROUTE':
> >>> > >> 'yes', 'VLAN': 'yes', 'IPADDR': '43.25.17.16', 'HOTPLUG': 'no',
> >>> > >> 'GATEWAY':
> >>> > >> '43.25.17.1', 'NM_CONTROLLED': 'no', 'NETMASK': '255.255.255.0',
> >>> > >> 'BOOTPROTO': 'none', 'DEVICE': 'bond0.10', 'MTU': '1500',
> 'ONBOOT':
> >>> > >> 'yes'},
> >>> > >> 'ipv6addrs': ['fe80::62eb:69ff:fe20:b46c/64'], 'vlanid': 10,
> 'mtu':
> >>> > >> '1500',
> >>> > >> 'netmask': '255.255.255.0', 'ipv4addrs': ['43.25.17.16/24']
> >>> > >> <http://43.25.17.16/24%27%5D>}, 'bond0.36': {'iface': 'bond0',
> >>> > >> 'addr':
> >>> > >> '', 'cfg': {'BRIDGE': 'VMNetwork', 'VLAN': 'yes', 'HOTPLUG': 'no',
> >>> > >> 'MTU':
> >>> > >> '1500', 'NM_CONTROLLED': 'no', 'DEVICE': 'bond0.36', 'ONBOOT':
> >>> > >> 'no'},
> >>> > >> 'ipv6addrs': ['fe80::62eb:69ff:fe20:b46c/64'], 'vlanid': 36,
> 'mtu':
> >>> > >> '1500',
> >>> > >> 'netmask': '', 'ipv4addrs': []}, 'bond1.100': {'iface': 'bond1',
> >>> > >> 'addr':
> >>> > >> '', 'cfg': {'BRIDGE': 'Internal', 'VLAN': 'yes', 'HOTPLUG': 'no',
> >>> > >> 'MTU':
> >>> > >> '9000', 'NM_CONTROLLED': 'no', 'DEVICE': 'bond1.100', 'ONBOOT':
> >>> > >> 'no'},
> >>> > >> 'ipv6addrs': ['fe80::210:18ff:fecd:daac/64'], 'vlanid': 100,
> 'mtu':
> >>> > >> '9000',
> >>> > >> 'netmask': '', 'ipv4addrs': []}}, 'cpuCores': '12', 'kvmEnabled':
> >>> > >> 'true',
> >>> > >> 'guestOverhead': '65', 'cpuThreads': '24', 'emulatedMachines':
> >>> > >> [u'rhel6.5.0', u'pc', u'rhel6.4.0', u'rhel6.3.0', u'rhel6.2.0',
> >>> > >> u'rhel6.1.0', u'rhel6.0.0', u'rhel5.5.0', u'rhel5.4.4',
> >>> > >> u'rhel5.4.0'],
> >>> > >> 'operatingSystem': {'release': '5.el6.centos.11.1', 'version':
> '6',
> >>> > >> 'name':
> >>> > >> 'RHEL'}, 'lastClient': '10.10.10.2'}}
> >>> > >> Thread-13::DEBUG::2014-11-24
> >>> > >> 21:41:41,620::BindingXMLRPC::1132::vds::(wrapper) client
> >>> > >> [10.10.10.2]::call
> >>> > >> getHardwareInfo with () {}
> >>> > >> Thread-13::DEBUG::2014-11-24
> >>> > >> 21:41:41,621::BindingXMLRPC::1139::vds::(wrapper) return
> >>> > >> getHardwareInfo
> >>> > >> with {'status': {'message': 'Done', 'code': 0}, 'info':
> >>> > >> {'systemProductName': 'CS24-TY', 'systemSerialNumber': '7LWSPN1',
> >>> > >> 'systemFamily': 'Server', 'systemVersion': 'A00', 'systemUUID':
> >>> > >> '44454c4c-4c00-1057-8053-b7c04f504e31', 'systemManufacturer':
> >>> > >> 'Dell'}}
> >>> > >> Thread-13::DEBUG::2014-11-24
> >>> > >> 21:41:41,733::BindingXMLRPC::1132::vds::(wrapper) client
> >>> > >> [10.10.10.2]::call
> >>> > >> hostsList with () {} flowID [222e8036]
> >>> > >> Thread-13::ERROR::2014-11-24
> >>> > >> 21:41:44,753::BindingXMLRPC::1148::vds::(wrapper) vdsm exception
> >>> > >> occured
> >>> > >> Traceback (most recent call last):
> >>> > >>   File "/usr/share/vdsm/rpc/BindingXMLRPC.py", line 1135, in
> wrapper
> >>> > >>     res = f(*args, **kwargs)
> >>> > >>   File "/usr/share/vdsm/gluster/api.py", line 54, in wrapper
> >>> > >>     rv = func(*args, **kwargs)
> >>> > >>   File "/usr/share/vdsm/gluster/api.py", line 251, in hostsList
> >>> > >>     return {'hosts': self.svdsmProxy.glusterPeerStatus()}
> >>> > >>   File "/usr/share/vdsm/supervdsm.py", line 50, in __call__
> >>> > >>     return callMethod()
> >>> > >>   File "/usr/share/vdsm/supervdsm.py", line 48, in <lambda>
> >>> > >>     **kwargs)
> >>> > >>   File "<string>", line 2, in glusterPeerStatus
> >>> > >>   File "/usr/lib64/python2.6/multiprocessing/managers.py", line
> 740,
> >>> > >> in
> >>> > >> _callmethod
> >>> > >>     raise convert_to_error(kind, result)
> >>> > >> GlusterCmdExecFailedException: Command execution failed
> >>> > >> error: Connection failed. Please check if gluster daemon is
> >>> > >> operational.
> >>> > >> return code: 1
> >>> > >> Thread-13::DEBUG::2014-11-24
> >>> > >> 21:41:50,949::task::595::Storage.TaskManager.Task::(_updateState)
> >>> > >> Task=`c9042986-c978-4b08-adb2-616f5299e115`::moving from state
> init
> >>> > >> ->
> >>> > >> state preparing
> >>> > >> Thread-13::INFO::2014-11-24
> >>> > >> 21:41:50,950::logUtils::44::dispatcher::(wrapper) Run and protect:
> >>> > >> repoStats(options=None)
> >>> > >> Thread-13::INFO::2014-11-24
> >>> > >> 21:41:50,950::logUtils::47::dispatcher::(wrapper) Run and protect:
> >>> > >> repoStats, Return response: {}
> >>> > >> Thread-13::DEBUG::2014-11-24
> >>> > >> 21:41:50,950::task::1191::Storage.TaskManager.Task::(prepare)
> >>> > >> Task=`c9042986-c978-4b08-adb2-616f5299e115`::finished: {}
> >>> > >> Thread-13::DEBUG::2014-11-24
> >>> > >> 21:41:50,950::task::595::Storage.TaskManager.Task::(_updateState)
> >>> > >> Task=`c9042986-c978-4b08-adb2-616f5299e115`::moving from state
> >>> > >> preparing
> >>> > >> ->
> >>> > >> state finished
> >>> > >> Thread-13::DEBUG::2014-11-24
> >>> > >>
> >>> > >>
> 21:41:50,951::resourceManager::940::Storage.ResourceManager.Owner::(releaseAll)
> >>> > >> Owner.releaseAll requests {} resources {}
> >>> > >> Thread-13::DEBUG::2014-11-24
> >>> > >>
> >>> > >>
> 21:41:50,951::resourceManager::977::Storage.ResourceManager.Owner::(cancelAll)
> >>> > >> Owner.cancelAll requests {}
> >>> > >> Thread-13::DEBUG::2014-11-24
> >>> > >> 21:41:50,951::task::993::Storage.TaskManager.Task::(_decref)
> >>> > >> Task=`c9042986-c978-4b08-adb2-616f5299e115`::ref 0 aborting False
> >>> > >>  -------------------------------
> >>> > >>
> >>> > >>  [root@compute4 ~]# service glusterd status
> >>> > >> glusterd is stopped
> >>> > >> [root@compute4 ~]# chkconfig --list | grep glusterd
> >>> > >> glusterd        0:off   1:off   2:on    3:on    4:on    5:on
> >>> > >> 6:off
> >>> > >> [root@compute4 ~]#
> >>> > >>
> >>> > >>  Thanks,
> >>> > >> Punit
> >>> > >>
> >>> > >> On Mon, Nov 24, 2014 at 6:36 PM, Kanagaraj <kmayi...@redhat.com>
> >>> > >> wrote:
> >>> > >>
> >>> > >>>  Can you send the corresponding error in vdsm.log from the host?
> >>> > >>>
> >>> > >>> Also check if glusterd service is running.
> >>> > >>>
> >>> > >>> Thanks,
> >>> > >>> Kanagaraj
> >>> > >>>
> >>> > >>>
> >>> > >>> On 11/24/2014 03:39 PM, Punit Dambiwal wrote:
> >>> > >>>
> >>> > >>>   Hi,
> >>> > >>>
> >>> > >>>  After reboot my Hypervisior host can not activate again in the
> >>> > >>> cluster
> >>> > >>> and failed with the following error :-
> >>> > >>>
> >>> > >>>  Gluster command [<UNKNOWN>] failed on server...
> >>> > >>>
> >>> > >>>  Engine logs :-
> >>> > >>>
> >>> > >>>  2014-11-24 18:05:28,397 INFO
> >>> > >>>
> >>> > >>>
> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
> >>> > >>> (DefaultQuartzScheduler_Worker-64) START,
> >>> > >>> GlusterVolumesListVDSCommand(HostName = Compute4, HostId =
> >>> > >>> 33648a90-200c-45ca-89d5-1ce305d79a6a), log id: 5f251c90
> >>> > >>> 2014-11-24 18:05:30,609 INFO
> >>> > >>>
> >>> > >>>
> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
> >>> > >>> (DefaultQuartzScheduler_Worker-64) FINISH,
> >>> > >>> GlusterVolumesListVDSCommand,
> >>> > >>> return:
> >>> > >>>
> >>> > >>>
> {26ae1672-ee09-4a38-8fd2-72dd9974cc2b=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@d95203e0
> },
> >>> > >>> log id: 5f251c90
> >>> > >>> 2014-11-24 18:05:33,768 INFO
> >>> > >>>  [org.ovirt.engine.core.bll.ActivateVdsCommand]
> >>> > >>> (ajp--127.0.0.1-8702-8)
> >>> > >>> [287d570d] Lock Acquired to object EngineLock [exclusiveLocks=
> key:
> >>> > >>> 0bf6b00f-7947-4411-b55a-cc5eea2b381a value: VDS
> >>> > >>> , sharedLocks= ]
> >>> > >>> 2014-11-24 18:05:33,795 INFO
> >>> > >>>  [org.ovirt.engine.core.bll.ActivateVdsCommand]
> >>> > >>> (org.ovirt.thread.pool-8-thread-45) [287d570d] Running command:
> >>> > >>> ActivateVdsCommand internal: false. Entities affected :  ID:
> >>> > >>> 0bf6b00f-7947-4411-b55a-cc5eea2b381a Type: VDSAction group
> >>> > >>> MANIPULATE_HOST
> >>> > >>> with role type ADMIN
> >>> > >>> 2014-11-24 18:05:33,796 INFO
> >>> > >>>  [org.ovirt.engine.core.bll.ActivateVdsCommand]
> >>> > >>> (org.ovirt.thread.pool-8-thread-45) [287d570d] Before acquiring
> >>> > >>> lock in
> >>> > >>> order to prevent monitoring for host Compute5 from data-center
> >>> > >>> SV_WTC
> >>> > >>> 2014-11-24 18:05:33,797 INFO
> >>> > >>>  [org.ovirt.engine.core.bll.ActivateVdsCommand]
> >>> > >>> (org.ovirt.thread.pool-8-thread-45) [287d570d] Lock acquired,
> from
> >>> > >>> now a
> >>> > >>> monitoring of host will be skipped for host Compute5 from
> >>> > >>> data-center
> >>> > >>> SV_WTC
> >>> > >>> 2014-11-24 18:05:33,817 INFO
> >>> > >>>  [org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand]
> >>> > >>> (org.ovirt.thread.pool-8-thread-45) [287d570d] START,
> >>> > >>> SetVdsStatusVDSCommand(HostName = Compute5, HostId =
> >>> > >>> 0bf6b00f-7947-4411-b55a-cc5eea2b381a, status=Unassigned,
> >>> > >>> nonOperationalReason=NONE, stopSpmFailureLogged=false), log id:
> >>> > >>> 1cbc7311
> >>> > >>> 2014-11-24 18:05:33,820 INFO
> >>> > >>>  [org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand]
> >>> > >>> (org.ovirt.thread.pool-8-thread-45) [287d570d] FINISH,
> >>> > >>> SetVdsStatusVDSCommand, log id: 1cbc7311
> >>> > >>> 2014-11-24 18:05:34,086 INFO
> >>> > >>>  [org.ovirt.engine.core.bll.ActivateVdsCommand]
> >>> > >>> (org.ovirt.thread.pool-8-thread-45) Activate finished. Lock
> >>> > >>> released.
> >>> > >>> Monitoring can run now for host Compute5 from data-center SV_WTC
> >>> > >>> 2014-11-24 18:05:34,088 INFO
> >>> > >>>
> >>> > >>>
> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
> >>> > >>> (org.ovirt.thread.pool-8-thread-45) Correlation ID: 287d570d, Job
> >>> > >>> ID:
> >>> > >>> 5ef8e4d6-b2bc-469e-8e81-7ef74b2a001a, Call Stack: null, Custom
> >>> > >>> Event ID:
> >>> > >>> -1, Message: Host Compute5 was activated by admin.
> >>> > >>> 2014-11-24 18:05:34,090 INFO
> >>> > >>>  [org.ovirt.engine.core.bll.ActivateVdsCommand]
> >>> > >>> (org.ovirt.thread.pool-8-thread-45) Lock freed to object
> EngineLock
> >>> > >>> [exclusiveLocks= key: 0bf6b00f-7947-4411-b55a-cc5eea2b381a value:
> >>> > >>> VDS
> >>> > >>> , sharedLocks= ]
> >>> > >>> 2014-11-24 18:05:35,792 INFO
> >>> > >>>
> >>> > >>>
> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
> >>> > >>> (DefaultQuartzScheduler_Worker-55) [3706e836] START,
> >>> > >>> GlusterVolumesListVDSCommand(HostName = Compute4, HostId =
> >>> > >>> 33648a90-200c-45ca-89d5-1ce305d79a6a), log id: 48a0c832
> >>> > >>> 2014-11-24 18:05:37,064 INFO
> >>> > >>>
> >>> > >>>
> [org.ovirt.engine.core.vdsbroker.vdsbroker.GetHardwareInfoVDSCommand]
> >>> > >>> (DefaultQuartzScheduler_Worker-69) START,
> >>> > >>> GetHardwareInfoVDSCommand(HostName = Compute5, HostId =
> >>> > >>> 0bf6b00f-7947-4411-b55a-cc5eea2b381a,
> >>> > >>> vds=Host[Compute5,0bf6b00f-7947-4411-b55a-cc5eea2b381a]), log id:
> >>> > >>> 6d560cc2
> >>> > >>> 2014-11-24 18:05:37,074 INFO
> >>> > >>>
> >>> > >>>
> [org.ovirt.engine.core.vdsbroker.vdsbroker.GetHardwareInfoVDSCommand]
> >>> > >>> (DefaultQuartzScheduler_Worker-69) FINISH,
> >>> > >>> GetHardwareInfoVDSCommand, log
> >>> > >>> id: 6d560cc2
> >>> > >>> 2014-11-24 18:05:37,093 WARN
> >>> > >>>  [org.ovirt.engine.core.vdsbroker.VdsManager]
> >>> > >>> (DefaultQuartzScheduler_Worker-69) Host Compute5 is running with
> >>> > >>> disabled
> >>> > >>> SELinux.
> >>> > >>> 2014-11-24 18:05:37,127 INFO
> >>> > >>>
> >>> > >>>
> [org.ovirt.engine.core.bll.HandleVdsCpuFlagsOrClusterChangedCommand]
> >>> > >>> (DefaultQuartzScheduler_Worker-69) [2b4a51cf] Running command:
> >>> > >>> HandleVdsCpuFlagsOrClusterChangedCommand internal: true. Entities
> >>> > >>> affected
> >>> > >>> :  ID: 0bf6b00f-7947-4411-b55a-cc5eea2b381a Type: VDS
> >>> > >>> 2014-11-24 18:05:37,147 INFO
> >>> > >>>
> >>> > >>>
> [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
> >>> > >>> (DefaultQuartzScheduler_Worker-69) [2b4a51cf] START,
> >>> > >>> GlusterServersListVDSCommand(HostName = Compute5, HostId =
> >>> > >>> 0bf6b00f-7947-4411-b55a-cc5eea2b381a), log id: 4faed87
> >>> > >>> 2014-11-24 18:05:37,164 INFO
> >>> > >>>
> >>> > >>>
> [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
> >>> > >>> (DefaultQuartzScheduler_Worker-69) [2b4a51cf] FINISH,
> >>> > >>> GlusterServersListVDSCommand, log id: 4faed87
> >>> > >>> 2014-11-24 18:05:37,189 INFO
> >>> > >>>  [org.ovirt.engine.core.bll.SetNonOperationalVdsCommand]
> >>> > >>> (DefaultQuartzScheduler_Worker-69) [4a84c4e5] Running command:
> >>> > >>> SetNonOperationalVdsCommand internal: true. Entities affected :
> >>> > >>> ID:
> >>> > >>> 0bf6b00f-7947-4411-b55a-cc5eea2b381a Type: VDS
> >>> > >>> 2014-11-24 18:05:37,206 INFO
> >>> > >>>  [org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand]
> >>> > >>> (DefaultQuartzScheduler_Worker-69) [4a84c4e5] START,
> >>> > >>> SetVdsStatusVDSCommand(HostName = Compute5, HostId =
> >>> > >>> 0bf6b00f-7947-4411-b55a-cc5eea2b381a, status=NonOperational,
> >>> > >>> nonOperationalReason=GLUSTER_COMMAND_FAILED,
> >>> > >>> stopSpmFailureLogged=false),
> >>> > >>> log id: fed5617
> >>> > >>> 2014-11-24 18:05:37,209 INFO
> >>> > >>>  [org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand]
> >>> > >>> (DefaultQuartzScheduler_Worker-69) [4a84c4e5] FINISH,
> >>> > >>> SetVdsStatusVDSCommand, log id: fed5617
> >>> > >>> 2014-11-24 18:05:37,223 ERROR
> >>> > >>>
> >>> > >>>
> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
> >>> > >>> (DefaultQuartzScheduler_Worker-69) [4a84c4e5] Correlation ID:
> >>> > >>> 4a84c4e5,
> >>> > >>> Job
> >>> > >>> ID: 4bfd4a6d-c3ef-468f-a40e-a3a6ca13011b, Call Stack: null,
> Custom
> >>> > >>> Event
> >>> > >>> ID: -1, Message: Gluster command [<UNKNOWN>] failed on server
> >>> > >>> Compute5.
> >>> > >>> 2014-11-24 18:05:37,243 INFO
> >>> > >>>
> >>> > >>>
> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
> >>> > >>> (DefaultQuartzScheduler_Worker-69) [4a84c4e5] Correlation ID:
> null,
> >>> > >>> Call
> >>> > >>> Stack: null, Custom Event ID: -1, Message: Status of host
> Compute5
> >>> > >>> was
> >>> > >>> set
> >>> > >>> to NonOperational.
> >>> > >>> 2014-11-24 18:05:37,272 INFO
> >>> > >>>  [org.ovirt.engine.core.bll.HandleVdsVersionCommand]
> >>> > >>> (DefaultQuartzScheduler_Worker-69) [a0c8a7f] Running command:
> >>> > >>> HandleVdsVersionCommand internal: true. Entities affected :  ID:
> >>> > >>> 0bf6b00f-7947-4411-b55a-cc5eea2b381a Type: VDS
> >>> > >>> 2014-11-24 18:05:37,274 INFO
> >>> > >>>  [org.ovirt.engine.core.vdsbroker.VdsUpdateRunTimeInfo]
> >>> > >>> (DefaultQuartzScheduler_Worker-69) [a0c8a7f] Host
> >>> > >>> 0bf6b00f-7947-4411-b55a-cc5eea2b381a : Compute5 is already in
> >>> > >>> NonOperational status for reason GLUSTER_COMMAND_FAILED.
> >>> > >>> SetNonOperationalVds command is skipped.
> >>> > >>> 2014-11-24 18:05:38,065 INFO
> >>> > >>>
> >>> > >>>
> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
> >>> > >>> (DefaultQuartzScheduler_Worker-55) [3706e836] FINISH,
> >>> > >>> GlusterVolumesListVDSCommand, return:
> >>> > >>>
> >>> > >>>
> {26ae1672-ee09-4a38-8fd2-72dd9974cc2b=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@4e72a1b1
> },
> >>> > >>> log id: 48a0c832
> >>> > >>> 2014-11-24 18:05:43,243 INFO
> >>> > >>>
> >>> > >>>
> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
> >>> > >>> (DefaultQuartzScheduler_Worker-35) START,
> >>> > >>> GlusterVolumesListVDSCommand(HostName = Compute4, HostId =
> >>> > >>> 33648a90-200c-45ca-89d5-1ce305d79a6a), log id: 3ce13ebc
> >>> > >>> ^C
> >>> > >>> [root@ccr01 ~]#
> >>> > >>>
> >>> > >>>  Thanks,
> >>> > >>> Punit
> >>> > >>>
> >>> > >>>
> >>> > >>>  _______________________________________________
> >>> > >>> Users mailing
> >>> > >>> listUsers@ovirt.orghttp://lists.ovirt.org/mailman/listinfo/users
> >>> > >>>
> >>> > >>>
> >>> > >>>
> >>> > >>
> >>> > >>
> >>> > >
> >>> > >
> >>> >
> >>
> >>
> >
> >
> > _______________________________________________
> > Gluster-users mailing list
> > gluster-us...@gluster.org
> > http://supercolony.gluster.org/mailman/listinfo/gluster-users
>
_______________________________________________
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users

Reply via email to