Unfortunately nothing....
Is there any way to make it more verbose?
On 14.02.22, 11:48, "Eugen Block" <[email protected]> wrote:
What does 'dmesg' reveal?
Zitat von Lo Re Giuseppe <[email protected]>:
> root@fulen-w006:~# ll client.fulen.keyring
> -rw-r--r-- 1 root root 69 Feb 11 15:30 client.fulen.keyring
> root@fulen-w006:~# ll ceph.conf
> -rw-r--r-- 1 root root 118 Feb 11 19:15 ceph.conf
> root@fulen-w006:~# rbd -c ceph.conf --id fulen --keyring
> client.fulen.keyring map fulen-nvme-meta/test-loreg-3
> rbd: sysfs write failed
> In some cases useful info is found in syslog - try "dmesg | tail".
> rbd: map failed: (1) Operation not permitted
>
> where test-loreg-3 is an image in a EC pool.
>
> root@fulen-w006:~# rbd -c ceph.conf --id fulen --keyring
> client.fulen.keyring info fulen-nvme-meta/test-loreg-3
> rbd image 'test-loreg-3':
> size 1 GiB in 256 objects
> order 22 (4 MiB objects)
> snapshot_count: 0
> id: e2375bbddf414a
> data_pool: fulen-hdd-data
> block_name_prefix: rbd_data.36.e2375bbddf414a
> format: 2
> features: layering, exclusive-lock, data-pool
> op_features:
> flags:
> create_timestamp: Thu Feb 10 18:17:42 2022
> access_timestamp: Thu Feb 10 18:17:42 2022
> modify_timestamp: Thu Feb 10 18:17:42 2022
>
> Giuseppe
>
> On 11.02.22, 14:52, "Eugen Block" <[email protected]> wrote:
>
> How are the permissions of the client keyring on both systems?
>
> Zitat von Lo Re Giuseppe <[email protected]>:
>
> > Hi,
> >
> > It's a single ceph cluster, I'm testing from 2 different client
nodes.
> > The caps are below.
> > I think is unlikely that caps are the cause as they work from one
> > client node, same ceph user, and not from the other one...
> >
> > Cheers,
> >
> > Giuseppe
> >
> >
> > [root@naret-monitor01 ~]# ceph auth get client.fulen
> > exported keyring for client.fulen
> > [client.fulen]
> > key = xxxxxxxx
> > caps mgr = "profile rbd pool=fulen-hdd, profile rbd
> > pool=fulen-nvme, profile rbd pool=fulen-dcache, profile rbd
> > pool=fulen-dcache-data, profile rbd pool=fulen-dcache-meta, profile
> > rbd pool=fulen-hdd-data, profile rbd pool=fulen-nvme-meta"
> > caps mon = "profile rbd"
> > caps osd = "profile rbd pool=fulen-hdd, profile rbd
> > pool=fulen-nvme, profile rbd pool=fulen-dcache, profile rbd
> > pool=fulen-dcache-data, profile rbd pool=fulen-dcache-meta, profile
> > rbd pool=fulen-hdd-data, profile rbd pool=fulen-nvme-meta"
> >
> >
> >
> > On 11.02.22, 13:22, "Eugen Block" <[email protected]> wrote:
> >
> > Hi,
> >
> > the first thing coming to mind are the user's caps. Which
> permissions
> > do they have? Have you compared 'ceph auth get
> client.fulen' on both
> > clusters? Please paste the output from both clusters and redact
> > sensitive information.
> >
> >
> > Zitat von Lo Re Giuseppe <[email protected]>:
> >
> > > Hi all,
> > >
> > > This is my first post to this user group, I’m not a ceph
expert,
> > > sorry if I say/ask anything trivial.
> > >
> > > On a Kubernetes cluster I have an issue in creating
> volumes from a
> > > (csi) ceph EC pool.
> > >
> > > I can reproduce the problem from rbd cli like this from
> one of the
> > > k8s worker nodes:
> > >
> > > “””
> > > root@fulen-w006:~# ceph -v
> > > ceph version 15.2.14
(cd3bb7e87a2f62c1b862ff3fd8b1eec13391a5be)
> > > octopus (stable)
> > >
> > > root@fulen-w006:~# rbd -m 148.187.20.141:6789 --id fulen
> --keyfile
> > > key create test-loreg-3 --data-pool fulen-hdd-data --pool
> > > fulen-nvme-meta --size 1G
> > >
> > > root@fulen-w006:~# rbd -m 148.187.20.141:6789 --id fulen
> --keyfile
> > > key map fulen-nvme-meta/test-loreg-3
> > > ...
> > > rbd: sysfs write failed
> > > ...
> > > In some cases useful info is found in syslog - try
> "dmesg | tail".
> > > rbd: map failed: (1) Operation not permitted
> > > “””
> > >
> > > The same sequence of operations works on a different
> node (not part
> > > of the k8s cluster, completely different setup):
> > >
> > > “””
> > > root@storagesmw: # ceph -v
> > > ceph version 15.2.13
(c44bc49e7a57a87d84dfff2a077a2058aa2172e2)
> > > octopus (stable)
> > >
> > > root@storagesmw: # rbd -m 148.187.20.141:6789 --id fulen
> --keyfile
> > > client.fulen.key create test-loreg-4 --data-pool
fulen-hdd-data
> > > --pool fulen-nvme-meta --size 1G
> > >
> > > root@storagesmw: # rbd -m 148.187.20.141:6789 --id fulen
> --keyfile
> > > client.fulen.key info fulen-nvme-meta/test-loreg-4
> > > rbd image 'test-loreg-4':
> > > size 1 GiB in 256 objects
> > > order 22 (4 MiB objects)
> > > snapshot_count: 0
> > > id: cafc436ff3573
> > > data_pool: fulen-hdd-data
> > > block_name_prefix: rbd_data.36.cafc436ff3573
> > > format: 2
> > > features: layering, exclusive-lock, object-map,
> fast-diff,
> > > deep-flatten, data-pool
> > > op_features:
> > > flags:
> > > create_timestamp: Thu Feb 10 18:23:26 2022
> > > access_timestamp: Thu Feb 10 18:23:26 2022
> > > modify_timestamp: Thu Feb 10 18:23:26 2022
> > >
> > > root@storagesmw: # rbd -m 148.187.20.141:6789 --id fulen
> --keyfile
> > > client.fulen.key map fulen-nvme-meta/test-loreg-4
> > > RBD image feature set mismatch. You can disable features
> unsupported
> > > by the kernel with "rbd feature disable
> fulen-nvme-meta/test-loreg-4
> > > object-map fast-diff deep-flatten".
> > > In some cases useful info is found in syslog - try
> "dmesg | tail".
> > > rbd: map failed: (6) No such device or address
> > >
> > > root@storagesmw: # rbd -m 148.187.20.141:6789 --id fulen
> --keyfile
> > > client.fulen.key feature disable fulen-nvme-meta/test-loreg-4
> > > object-map fast-diff deep-flatten
> > >
> > > root@storagesmw: # rbd -m 148.187.20.141:6789 --id fulen
> --keyfile
> > > client.fulen.key map fulen-nvme-meta/test-loreg-4
> > > /dev/rbd0
> > > “””
> > >
> > > The 2 nodes OS release and kernel are below.
> > >
> > > Does anyone have any advice on how to debug this?
> > >
> > > Thanks in advance,
> > >
> > > Giuseppe
> > >
> > > ============================================== Fulen-w006:
> > >
> > > root@fulen-w006:~# cat /etc/os-release
> > > NAME="Ubuntu"
> > > VERSION="20.04.3 LTS (Focal Fossa)"
> > > ID=ubuntu
> > > ID_LIKE=debian
> > > PRETTY_NAME="Ubuntu 20.04.3 LTS"
> > > VERSION_ID="20.04"
> > > HOME_URL="https://www.ubuntu.com/"
> > > SUPPORT_URL="https://help.ubuntu.com/"
> > > BUG_REPORT_URL="https://bugs.launchpad.net/ubuntu/"
> > >
> >
>
PRIVACY_POLICY_URL="https://www.ubuntu.com/legal/terms-and-policies/privacy-policy"
> > > VERSION_CODENAME=focal
> > > UBUNTU_CODENAME=focal
> > > root@fulen-w006:~# uname -a
> > > Linux fulen-w006.cscs.ch 5.4.0-96-generic #109-Ubuntu
> SMP Wed Jan 12
> > > 16:49:16 UTC 2022 x86_64 x86_64 x86_64 GNU/Linux
> > >
> > > Storagesmw:
> > >
> > > root@storagesmw:~/loreg/ceph_conf # cat /etc/os-release
> > > NAME="Red Hat Enterprise Linux Server"
> > > VERSION="7.8 (Maipo)"
> > > ID="rhel"
> > > ID_LIKE="fedora"
> > > VARIANT="Server"
> > > VARIANT_ID="server"
> > > VERSION_ID="7.8"
> > > PRETTY_NAME="Red Hat Enterprise Linux"
> > > ANSI_COLOR="0;31"
> > > CPE_NAME="cpe:/o:redhat:enterprise_linux:7.8:GA:server"
> > > HOME_URL="https://www.redhat.com/"
> > > BUG_REPORT_URL="https://bugzilla.redhat.com/"
> > >
> > > REDHAT_BUGZILLA_PRODUCT="Red Hat Enterprise Linux 7"
> > > REDHAT_BUGZILLA_PRODUCT_VERSION=7.8
> > > REDHAT_SUPPORT_PRODUCT="Red Hat Enterprise Linux"
> > > REDHAT_SUPPORT_PRODUCT_VERSION="7.8"
> > > root@storagesmw:~/loreg/ceph_conf # uname -a
> > > Linux storagesmw.cscs.ch 3.10.0-1127.19.1.el7.x86_64 #1
> SMP Tue Aug
> > > 11 19:12:04 EDT 2020 x86_64 x86_64 x86_64 GNU/Linux
> > >
> > >
> > >
> > >
> > >
> > > _______________________________________________
> > > ceph-users mailing list -- [email protected]
> > > To unsubscribe send an email to [email protected]
> >
> >
> >
> > _______________________________________________
> > ceph-users mailing list -- [email protected]
> > To unsubscribe send an email to [email protected]
_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]