Hi,
I would like to ask if anybody knows how to handle the gwcli status below.
- Disks state in gwcli shows as "Unknowm"
- Clients still mounting the "Unknown" disks and seems working normally.
Two of the rbd disks show "Unknown" instead of "Online" in gwcli.
==============================================================================================================
# gwcli ls /disks
o- disks
.......................................................................................................
[77312G, Disks: 10]
o- ssd-rf2
......................................................................................................
[ssd-rf2 (6.0T)]
| o- iscsi_01
.................................................................................
[ssd-rf2/iscsi_01 (Unknown, 3.0T)]
| o- iscsi_02
.................................................................................
[ssd-rf2/iscsi_02 (Unknown, 3.0T)]
o- ssd-rf3
......................................................................................................
[ssd-rf3 (8.0T)]
o- iscsi_pool_01
........................................................................
[ssd-rf3/iscsi_pool_01
(Online, 4.0T)]
o- iscsi_pool_02
........................................................................
[ssd-rf3/iscsi_pool_02
(Online, 4.0T)]
==============================================================================================================
Both "Lock Owner" and "State" are "Unknown" inside info session.
==============================================================================================================
# gwcli /disks/ssd-rf2/iscsi_01 info
Image .. iscsi_01
Ceph Cluster .. ceph
Pool .. ssd-rf2
Wwn .. 7b441630-2868-47d2-94f1-35efea4cf258
Size H .. 3.0T
Feature List .. RBD_FEATURE_LAYERING
RBD_FEATURE_EXCLUSIVE_LOCK
RBD_FEATURE_OBJECT_MAP
RBD_FEATURE_FAST_DIFF
RBD_FEATURE_DEEP_FLATTEN
Snapshots ..
Owner .. sds-ctt-gw1
Lock Owner .. Unknown
State .. Unknown
Backstore .. user:rbd
Backstore Object Name .. ssd-rf2.iscsi_01
Control Values
- hw_max_sectors .. 1024
- max_data_area_mb .. 8
- osd_op_timeout .. 30
- qfull_timeout .. 5
==============================================================================================================
Below is reference output from a noral rbd disk.
==============================================================================================================
# gwcli /disks/ssd-rf3/iscsi_pool_01 info
Image .. iscsi_pool_01
Ceph Cluster .. ceph
Pool .. ssd-rf3
Wwn .. 20396fed-2aba-422d-99c2-8353b8910fa4
Size H .. 4.0T
Feature List .. RBD_FEATURE_LAYERING
RBD_FEATURE_EXCLUSIVE_LOCK
RBD_FEATURE_OBJECT_MAP
RBD_FEATURE_FAST_DIFF
RBD_FEATURE_DEEP_FLATTEN
Snapshots ..
Owner .. sds-ctt-gw2
Lock Owner .. sds-ctt-gw2
State .. Online
Backstore .. user:rbd
Backstore Object Name .. ssd-rf3.iscsi_pool_01
Control Values
- hw_max_sectors .. 1024
- max_data_area_mb .. 8
- osd_op_timeout .. 30
- qfull_timeout .. 5
==============================================================================================================
Nothing special found in the rbd setting.
==============================================================================================================
root@sds-ctt-mon1:/# rbd ls -p ssd-rf2
iscsi_01
iscsi_02
root@sds-ctt-mon1:/# rbd -p ssd-rf2 info iscsi_01
rbd image 'iscsi_01':
size 3 TiB in 3145728 objects
order 20 (1 MiB objects)
snapshot_count: 0
id: 272654e71f95e9
block_name_prefix: rbd_data.272654e71f95e9
format: 2
features: layering, exclusive-lock, object-map, fast-diff,
deep-flatten
op_features:
flags:
create_timestamp: Mon Mar 7 05:28:55 2022
access_timestamp: Tue May 17 02:17:16 2022
modify_timestamp: Tue May 17 02:17:16 2022
root@sds-ctt-mon1:/# rbd -p ssd-rf3 info iscsi_pool_01
rbd image 'iscsi_pool_01':
size 4 TiB in 4194304 objects
order 20 (1 MiB objects)
snapshot_count: 0
id: 29bebcd9d3b6aa
block_name_prefix: rbd_data.29bebcd9d3b6aa
format: 2
features: layering, exclusive-lock, object-map, fast-diff,
deep-flatten
op_features:
flags:
create_timestamp: Tue Aug 11 02:32:37 2020
access_timestamp: Tue May 17 02:17:31 2022
modify_timestamp: Tue May 17 02:17:39 2022
root@sds-ctt-mon1:/#
==============================================================================================================
Cluster working healthly.
==============================================================================================================
# ceph health detail
HEALTH_OK
==============================================================================================================
Looking forward to any suggestions.
Thanks.
Regs,
Icy
_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]