On 11/18/2015 10:22 PM, p...@email.cz wrote:
Hello,
yes, I'm talking about gluster volumes.
"storages" not defined yet.
The main problem is about how to remove all definitions from gluster
configs on nodes and in ovirt too ( maybe oVirt will update
automaticaly , as U wrote before ).
1) nodes are in maintenance mode, glustred is running with errors
*# systemctl status glusterd*
glusterd.service - GlusterFS, a clustered file-system server
Loaded: loaded (/usr/lib/systemd/system/glusterd.service; enabled)
Active: *active (running)* since St 2015-11-18 14:12:26 CET; 3h
16min ago
Process: 4465 ExecStart=/usr/sbin/glusterd -p /var/run/glusterd.pid
--log-level $LOG_LEVEL $GLUSTERD_OPTIONS (code=exited, status=0/SUCCESS)
Main PID: 4466 (glusterd)
CGroup: /system.slice/glusterd.service
├─4466 /usr/sbin/glusterd -p /var/run/glusterd.pid
--log-level INFO
└─4612 /usr/sbin/glusterfs -s localhost --volfile-id
gluster/glustershd -p /var/lib/glusterd/glustershd/run/glustershd.pid
-l /var/log/glusterfs/glustershd.log -S /var/run/glus...
lis 18 17:25:44 1hp2.algocloud.net etc-glusterfs-glusterd.vol[4466]:
[2015-11-18 16:25:44.288734] C
[rpc-clnt-ping.c:165:rpc_clnt_ping_timer_expired] 0-management: server
16.0.0...onnecting.
lis 18 17:26:23 1hp2.algocloud.net etc-glusterfs-glusterd.vol[4466]:
[2015-11-18 16:26:23.297273] C
[rpc-clnt-ping.c:165:rpc_clnt_ping_timer_expired] 0-management: server
16.0.0...onnecting.
lis 18 17:26:41 1hp2.algocloud.net etc-glusterfs-glusterd.vol[4466]:
[2015-11-18 16:26:41.302793] C
[rpc-clnt-ping.c:165:rpc_clnt_ping_timer_expired] 0-management: server
16.0.0...onnecting.
lis 18 17:26:54 1hp2.algocloud.net etc-glusterfs-glusterd.vol[4466]:
[2015-11-18 16:26:54.307579] C
[rpc-clnt-ping.c:165:rpc_clnt_ping_timer_expired] 0-management: server
16.0.0...onnecting.
lis 18 17:27:33 1hp2.algocloud.net etc-glusterfs-glusterd.vol[4466]:
[2015-11-18 16:27:33.316049] C
[rpc-clnt-ping.c:165:rpc_clnt_ping_timer_expired] 0-management: server
16.0.0...onnecting.
lis 18 17:27:51 1hp2.algocloud.net etc-glusterfs-glusterd.vol[4466]:
[2015-11-18 16:27:51.321659] C
[rpc-clnt-ping.c:165:rpc_clnt_ping_timer_expired] 0-management: server
16.0.0...onnecting.
lis 18 17:28:04 1hp2.algocloud.net etc-glusterfs-glusterd.vol[4466]:
[2015-11-18 16:28:04.326615] C
[rpc-clnt-ping.c:165:rpc_clnt_ping_timer_expired] 0-management: server
16.0.0...onnecting.
lis 18 17:28:43 1hp2.algocloud.net etc-glusterfs-glusterd.vol[4466]:
[2015-11-18 16:28:43.335278] C
[rpc-clnt-ping.c:165:rpc_clnt_ping_timer_expired] 0-management: server
16.0.0...onnecting.
lis 18 17:29:01 1hp2.algocloud.net etc-glusterfs-glusterd.vol[4466]:
[2015-11-18 16:29:01.340909] C
[rpc-clnt-ping.c:165:rpc_clnt_ping_timer_expired] 0-management: server
16.0.0...onnecting.
lis 18 17:29:14 1hp2.algocloud.net etc-glusterfs-glusterd.vol[4466]:
[2015-11-18 16:29:14.345827] C
[rpc-clnt-ping.c:165:rpc_clnt_ping_timer_expired] 0-management: server
16.0.0...onnecting.
Hint: Some lines were ellipsized, use -l to show in full.
The log at /var/log/glusterfs/etc-glusterfs-glusterd.vol.log will give
you more information on the errors
2) all gluster data was cleared from filesystem ( meaning "
.glusterfs" , VM's data , etc. ( rm -rf ./.* ; rm -rf ./* = really
cleaned ) )
3) from command line :
#*gluster volume info 1HP-R2P1**
*
Volume Name: 1HP-R2P1
Type: Replicate
Volume ID: 8b667651-7104-4db9-a006-4effa40524e6
Status: Started
Number of Bricks: 1 x 2 = 2
Transport-type: tcp
Bricks:
Brick1: 1hp1-san:/STORAGE/p1/G
Brick2: 1hp2-san:/STORAGE/p1/G
Options Reconfigured:
performance.readdir-ahead: on
*# cat /etc/hosts*
172.16.5.151 1hp1
172.16.5.152 1hp2
172.16.5.153 2hp1
172.16.5.154 2hp2
16.0.0.151 1hp1-SAN
16.0.0.152 1hp2-SAN
16.0.0.153 2hp1-SAN
16.0.0.154 2hp2-SAN
# *gluster peer status* ( in Ovirt nodes defined in 172.16.5.0 (mgmt -
1Gb ), but bricks in 16.0.0.0 network ( VM's - 10Gb ( repl/move) )
Number of Peers: 4
Hostname: *172.16.5.152*
Uuid: 47b030ab-75d8-49ec-b67d-650e22dc2271
State: Peer in Cluster (Connected)
Other names:
*1hp2* *
which of them are correct - both ??* - would I mix it ?? ( peers in
the same net, of course ( 16.0.0.0 ) )
If you want to add bricks using the 16.0.0.0 network, from ovirt you
will need to set this up like below
1. Define a network in the cluster with "gluster" network role.
2. After you add the hosts to ovirt using the 172.16.. network, assign
the "gluster" network to the 16.0.. interface using the "Setup Networks"
dialog.
Now when you create the volume from oVirt, the 16.0 network will be used
to add the bricks.
But in your case it looks like the same host is known as 2 peers - 1hp2
and 1hp2-SAN? Did you set this up from gluster CLI?
You could try peer detaching 1hp2-SAN and peer probing it again from
another host. (1hp2-SAN should be shown as other name for 1hp2)
Hostname: 1hp2-SAN
Uuid: 47b030ab-75d8-49ec-b67d-650e22dc2271
State: Peer in Cluster (Connected)
Hostname: 2hp2-SAN
Uuid: f98ff1e1-c866-4af8-a6fa-3e8141a207cd
State: Peer in Cluster (Connected)
Hostname: 2hp1-SAN
Uuid: 7dcd603f-052f-4188-94fa-9dbca6cd19b3
State: Peer in Cluster (Connected)
#*gluster volume delete 1HP-R2P1*
Deleting volume will erase all information about the volume. Do you
want to continue? (y/n) y
*Error : Request timed out*
Please attach gluster log for identifying the issue.
*node info :* all in current version
OS Version:RHEL - 7 - 1.1503.el7.centos.2.8
Kernel Version 3.10.0 - 229.20.1.el7.x86_64
KVM Version:2.3.0 - 29.1.el7
LIBVIRT Version:libvirt-1.2.8-16.el7_1.5
VDSM Version:vdsm-4.17.999-152.git84c0adc.el7
SPICE Version:0.12.4 - 9.el7_1.3
GlusterFS Version:glusterfs-3.7.6-1.el7
oVirt : 3.6
regs.
pavel
On 18.11.2015 17:17, Sahina Bose wrote:
Are you talking of gluster volumes shown in Volumes tab?
If you have removed only the gluster volumes and not the gluster
nodes - the oVirt engine will update the configuration with backend
gluster.
However, if the gluster nodes are also removed from backend - the
nodes should be in Non-responsive state in the UI?
You could put all nodes in gluster cluster in maintenance mode, and
forceremove(checkbox provided) the nodes.
On 11/18/2015 07:26 PM, p...@email.cz wrote:
Hello,
howto remove volume definition from oVirt DB ( & from nodes gluster
config ) if volume totaly cleaned in background in running mode ??
regs.
Paf1
_______________________________________________
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users
_______________________________________________
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users