Hi Devender,

I think the problem is that you are using the old syntax, try with the new
one [1]

mount -t ceph admin@.van2-dev=/ /cephfs/ -o
noatime,ms_mode=prefer-crc,_netdev

Regards.

[1] https://docs.ceph.com/en/squid/man/8/mount.ceph/

On Wed, Jul 16, 2025 at 12:11 AM Devender Singh <deven...@netskrt.io> wrote:

> Tried below too but nothing works, any suggestion?
>
> # mount -t ceph 10.101.1.246,10.101.1.247,10.101.1.248:/ /mnt/cephfs   -o
> name=admin,fs=van2.dev,noatime,_netdev,ms_mode=prefer-crc
>
> mount error: no mds (Metadata Server) is up. The cluster might be laggy,
> or you may not be authorized
>
> # mount -t ceph :/ /cephfs/ -o
> noatime,ms_mode=prefer-crc,_netdev,fs=van2-dev
> mount error: no mds (Metadata Server) is up. The cluster might be laggy,
> or you may not be authorized
>
> Regards
> Dev
>
> > On Jul 15, 2025, at 2:17 PM, Devender Singh <deven...@netskrt.io> wrote:
> >
> > Hello all
> >
> > I have cephfs working fine with running ads but still not able to mount,
> getting error mount error: no mds (Metadata Server) is up
> >
> > Below are details
> >
> >
> > # ceph osd dump | grep min_compat_client
> > require_min_compat_client reef
> > min_compat_client reef
> >
> > # ceph balancer status
> > {
> >     "active": true,
> >     "last_optimize_duration": "0:00:00.002882",
> >     "last_optimize_started": "Tue Jul 15 21:16:41 2025",
> >     "mode": "upmap-read",
> >     "no_optimization_needed": true,
> >     "optimize_result": "Unable to find further optimization, or pool(s)
> pg_num is decreasing, or distribution is already perfect",
> >     "plans": []
> > }
> >
> > root@van2-host06n:~# dpkg -l ceph-common
> > Desired=Unknown/Install/Remove/Purge/Hold
> > |
> Status=Not/Inst/Conf-files/Unpacked/halF-conf/Half-inst/trig-aWait/Trig-pend
> > |/ Err?=(none)/Reinst-required (Status,Err: uppercase=bad)
> > ||/ Name           Version       Architecture Description
> >
> +++-==============-=============-============-==================================================================
> > ii  ceph-common    19.2.2-1jammy amd64        common utilities to mount
> and interact with a ceph storage cluster
> >
> > root@van2-host06n:~# ceph fs status
> > van2.dev - 0 clients
> > ========
> > RANK  STATE                  MDS                    ACTIVITY     DNS
> INOS   DIRS   CAPS
> >  0    active  van2.dev.van2-host06n.wqttsi  Reqs:    0 /s    10     13
>    12      0
> >  1    active  van2.dev.van2-host07n.fngvlt  Reqs:    0 /s    10     13
>    11      0
> >         POOL            TYPE     USED  AVAIL
> > cephfs.van2.dev.meta  metadata   230k  53.0T
> > cephfs.van2.dev.data    data       0   53.0T
> >            STANDBY MDS
> > van2-dev.van2-host07n.jnacmh
> > van2-dev.van2-host06n.wpzshe
> > van2-dev.van2-host08n.slrfej
> > MDS version: ceph version 19.2.2
> (0eceb0defba60152a8182f7bd87d164b639885b8) squid (stable)
> >
> >
> > SYstemd file.
> > ===========
> >
> > # cat cephfs.mount
> > #
> > # this mounts on /cephfs
> >
> > [Unit]
> > Description=mount /cephfs
> > After=network-online.target
> > Requires=network-online.target
> >
> > [Mount]
> > Where=/cephfs
> > What=10.101.1.246:6789,10.101.1.247:6789,10.101.1.248:6789:/
> > Type=ceph
> > Options=fs=van2.dev,name=admin,noatime,_netdev,x-systemd.automount
> >
> > [Install]
> > WantedBy=multi-user.target
> >
> >
> > root@van2-host06n:~# journalctl -u cephfs.mount
> > Jul 15 19:31:10 van2-host06n.van2.example.com systemd[1]: Mounting
> mount /cephfs...
> > Jul 15 19:32:12 van2-host06n.van2.example.com mount[1809096]: mount
> error: no mds (Metadata Server) is up. The clu>
> > Jul 15 19:32:12 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
> Mount process exited, code=exited, status=>
> > Jul 15 19:32:12 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
> Failed with result 'exit-code'.
> > Jul 15 19:32:12 van2-host06n.van2.example.com systemd[1]: Failed to
> mount mount /cephfs.
> > Jul 15 19:34:59 van2-host06n.van2.example.com systemd[1]: Mounting
> mount /cephfs...
> > Jul 15 19:35:59 van2-host06n.van2.example.com mount[1810584]: mount
> error: no mds (Metadata Server) is up. The clu>
> > Jul 15 19:35:59 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
> Mount process exited, code=exited, status=>
> > Jul 15 19:35:59 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
> Failed with result 'exit-code'.
> > Jul 15 19:35:59 van2-host06n.van2.example.com systemd[1]: Failed to
> mount mount /cephfs.
> > Jul 15 19:36:16 van2-host06n.van2.example.com systemd[1]: Mounting
> mount /cephfs...
> > Jul 15 19:37:17 van2-host06n.van2.example.com mount[1811498]: mount
> error: no mds (Metadata Server) is up. The clu>
> > Jul 15 19:37:17 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
> Mount process exited, code=exited, status=>
> > Jul 15 19:37:17 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
> Failed with result 'exit-code'.
> > Jul 15 19:37:17 van2-host06n.van2.example.com systemd[1]: Failed to
> mount mount /cephfs.
> > Jul 15 19:49:38 van2-host06n.van2.example.com systemd[1]: Mounting
> mount /cephfs...
> > Jul 15 19:50:40 van2-host06n.van2.example.com mount[1817486]: mount
> error: no mds (Metadata Server) is up. The clu>
> > Jul 15 19:50:40 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
> Mount process exited, code=exited, status=>
> > Jul 15 19:50:40 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
> Failed with result 'exit-code'.
> > Jul 15 19:50:40 van2-host06n.van2.example.com systemd[1]: Failed to
> mount mount /cephfs.
> > Jul 15 19:51:58 van2-host06n.van2.example.com systemd[1]: Mounting
> mount /cephfs...
> > Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
> Mount process exited, code=killed, status=>
> > Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
> Failed with result 'signal'.
> > Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
> Unit process 1817555 (mount.ceph) remains >
> > Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: Unmounted
> mount /cephfs.
> > Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
> Found left-over process 1817555 (mount.cep>
> > Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: This usually
> indicates unclean termination of a previous>
> > Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: Mounting
> mount /cephfs...
> > Jul 15 19:52:59 van2-host06n.van2.example.com mount[1817555]: mount
> error: no mds (Metadata Server) is up. The clu>
> > lines 1-29...skipping...
> > Jul 15 19:31:10 van2-host06n.van2.example.com systemd[1]: Mounting
> mount /cephfs...
> > Jul 15 19:32:12 van2-host06n.van2.example.com mount[1809096]: mount
> error: no mds (Metadata Server) is up. The cluster might be laggy, or you
> may not be>
> > Jul 15 19:32:12 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
> Mount process exited, code=exited, status=32/n/a
> > Jul 15 19:32:12 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
> Failed with result 'exit-code'.
> > Jul 15 19:32:12 van2-host06n.van2.example.com systemd[1]: Failed to
> mount mount /cephfs.
> > Jul 15 19:34:59 van2-host06n.van2.example.com systemd[1]: Mounting
> mount /cephfs...
> > Jul 15 19:35:59 van2-host06n.van2.example.com mount[1810584]: mount
> error: no mds (Metadata Server) is up. The cluster might be laggy, or you
> may not be>
> > Jul 15 19:35:59 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
> Mount process exited, code=exited, status=32/n/a
> > Jul 15 19:35:59 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
> Failed with result 'exit-code'.
> > Jul 15 19:35:59 van2-host06n.van2.example.com systemd[1]: Failed to
> mount mount /cephfs.
> > Jul 15 19:36:16 van2-host06n.van2.example.com systemd[1]: Mounting
> mount /cephfs...
> > Jul 15 19:37:17 van2-host06n.van2.example.com mount[1811498]: mount
> error: no mds (Metadata Server) is up. The cluster might be laggy, or you
> may not be>
> > Jul 15 19:37:17 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
> Mount process exited, code=exited, status=32/n/a
> > Jul 15 19:37:17 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
> Failed with result 'exit-code'.
> > Jul 15 19:37:17 van2-host06n.van2.example.com systemd[1]: Failed to
> mount mount /cephfs.
> > Jul 15 19:49:38 van2-host06n.van2.example.com systemd[1]: Mounting
> mount /cephfs...
> > Jul 15 19:50:40 van2-host06n.van2.example.com mount[1817486]: mount
> error: no mds (Metadata Server) is up. The cluster might be laggy, or you
> may not be>
> > Jul 15 19:50:40 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
> Mount process exited, code=exited, status=32/n/a
> > Jul 15 19:50:40 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
> Failed with result 'exit-code'.
> > Jul 15 19:50:40 van2-host06n.van2.example.com systemd[1]: Failed to
> mount mount /cephfs.
> > Jul 15 19:51:58 van2-host06n.van2.example.com systemd[1]: Mounting
> mount /cephfs...
> > Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
> Mount process exited, code=killed, status=15/TERM
> > Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
> Failed with result 'signal'.
> > Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
> Unit process 1817555 (mount.ceph) remains running after unit stopped.
> > Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: Unmounted
> mount /cephfs.
> > Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
> Found left-over process 1817555 (mount.ceph) in control group while
> starting uni>
> > Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: This usually
> indicates unclean termination of a previous run, or service implementation
> defici>
> > Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: Mounting
> mount /cephfs...
> > Jul 15 19:52:59 van2-host06n.van2.example.com mount[1817555]: mount
> error: no mds (Metadata Server) is up. The cluster might be laggy, or you
> may not be>
> > Jul 15 19:53:26 van2-host06n.van2.example.com mount[1817572]: mount
> error: no mds (Metadata Server) is up. The cluster might be laggy, or you
> may not be>
> > Jul 15 19:53:26 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
> Mount process exited, code=exited, status=32/n/a
> > Jul 15 19:53:26 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
> Failed with result 'exit-code'.
> > Jul 15 19:53:26 van2-host06n.van2.example.com systemd[1]: Failed to
> mount mount /cephfs.
> > Jul 15 19:56:29 van2-host06n.van2.example.com systemd[1]: Mounting
> mount /cephfs...
> > Jul 15 19:57:29 van2-host06n.van2.example.com mount[1817655]: mount
> error: no mds (Metadata Server) is up. The cluster might be laggy, or you
> may not be>
> > Jul 15 19:57:29 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
> Mount process exited, code=exited, status=32/n/a
> > Jul 15 19:57:29 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
> Failed with result 'exit-code'.
> > Jul 15 19:57:29 van2-host06n.van2.example.com systemd[1]: Failed to
> mount mount /cephfs.
> > lines 1-38...skipping...
> > Jul 15 19:31:10 van2-host06n.van2.example.com systemd[1]: Mounting
> mount /cephfs...
> > Jul 15 19:32:12 van2-host06n.van2.example.com mount[1809096]: mount
> error: no mds (Metadata Server) is up. The cluster might be laggy, or you
> may not be authorized
> > Jul 15 19:32:12 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
> Mount process exited, code=exited, status=32/n/a
> > Jul 15 19:32:12 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
> Failed with result 'exit-code'.
> > Jul 15 19:32:12 van2-host06n.van2.example.com systemd[1]: Failed to
> mount mount /cephfs.
> > Jul 15 19:34:59 van2-host06n.van2.example.com systemd[1]: Mounting
> mount /cephfs...
> > Jul 15 19:35:59 van2-host06n.van2.example.com mount[1810584]: mount
> error: no mds (Metadata Server) is up. The cluster might be laggy, or you
> may not be authorized
> > Jul 15 19:35:59 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
> Mount process exited, code=exited, status=32/n/a
> > Jul 15 19:35:59 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
> Failed with result 'exit-code'.
> > Jul 15 19:35:59 van2-host06n.van2.example.com systemd[1]: Failed to
> mount mount /cephfs.
> > Jul 15 19:36:16 van2-host06n.van2.example.com systemd[1]: Mounting
> mount /cephfs...
> > Jul 15 19:37:17 van2-host06n.van2.example.com mount[1811498]: mount
> error: no mds (Metadata Server) is up. The cluster might be laggy, or you
> may not be authorized
> > Jul 15 19:37:17 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
> Mount process exited, code=exited, status=32/n/a
> > Jul 15 19:37:17 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
> Failed with result 'exit-code'.
> > Jul 15 19:37:17 van2-host06n.van2.example.com systemd[1]: Failed to
> mount mount /cephfs.
> > Jul 15 19:49:38 van2-host06n.van2.example.com systemd[1]: Mounting
> mount /cephfs...
> > Jul 15 19:50:40 van2-host06n.van2.example.com mount[1817486]: mount
> error: no mds (Metadata Server) is up. The cluster might be laggy, or you
> may not be authorized
> > Jul 15 19:50:40 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
> Mount process exited, code=exited, status=32/n/a
> > Jul 15 19:50:40 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
> Failed with result 'exit-code'.
> > Jul 15 19:50:40 van2-host06n.van2.example.com systemd[1]: Failed to
> mount mount /cephfs.
> > Jul 15 19:51:58 van2-host06n.van2.example.com systemd[1]: Mounting
> mount /cephfs...
> > Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
> Mount process exited, code=killed, status=15/TERM
> > Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
> Failed with result 'signal'.
> > Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
> Unit process 1817555 (mount.ceph) remains running after unit stopped.
> > Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: Unmounted
> mount /cephfs.
> > Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
> Found left-over process 1817555 (mount.ceph) in control group while
> starting unit. Ignoring.
> > Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: This usually
> indicates unclean termination of a previous run, or service implementation
> deficiencies.
> > Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: Mounting
> mount /cephfs...
> > Jul 15 19:52:59 van2-host06n.van2.example.com mount[1817555]: mount
> error: no mds (Metadata Server) is up. The cluster might be laggy, or you
> may not be authorized
> > Jul 15 19:53:26 van2-host06n.van2.example.com mount[1817572]: mount
> error: no mds (Metadata Server) is up. The cluster might be laggy, or you
> may not be authorized
> > Jul 15 19:53:26 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
> Mount process exited, code=exited, status=32/n/a
> > Jul 15 19:53:26 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
> Failed with result 'exit-code'.
> > Jul 15 19:53:26 van2-host06n.van2.example.com systemd[1]: Failed to
> mount mount /cephfs.
> > Jul 15 19:56:29 van2-host06n.van2.example.com systemd[1]: Mounting
> mount /cephfs...
> > Jul 15 19:57:29 van2-host06n.van2.example.com mount[1817655]: mount
> error: no mds (Metadata Server) is up. The cluster might be laggy, or you
> may not be authorized
> > Jul 15 19:57:29 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
> Mount process exited, code=exited, status=32/n/a
> > Jul 15 19:57:29 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
> Failed with result 'exit-code'.
> > Jul 15 19:57:29 van2-host06n.van2.example.com systemd[1]: Failed to
> mount mount /cephfs.
> > Jul 15 19:58:50 van2-host06n.van2.example.com systemd[1]: Mounting
> mount /cephfs...
> > Jul 15 19:59:51 van2-host06n.van2.example.com mount[1818287]: mount
> error: no mds (Metadata Server) is up. The cluster might be laggy, or you
> may not be authorized
> > Jul 15 19:59:51 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
> Mount process exited, code=exited, status=32/n/a
> > Jul 15 19:59:51 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
> Failed with result 'exit-code'.
> > Jul 15 19:59:51 van2-host06n.van2.example.com systemd[1]: Failed to
> mount mount /cephfs.
> > Jul 15 20:05:43 van2-host06n.van2.example.com systemd[1]: Mounting
> mount /cephfs...
> > Jul 15 20:06:44 van2-host06n.van2.example.com mount[1818912]: mount
> error: no mds (Metadata Server) is up. The cluster might be laggy, or you
> may not be authorized
> > lines 1-45
> >
>
> _______________________________________________
> ceph-users mailing list -- ceph-users@ceph.io
> To unsubscribe send an email to ceph-users-le...@ceph.io
>
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to