Hello Dario

Thanks for your reply. 
No luck using those too.. 

Also, I want to mount cephfs to proxmox, there also getting issue mounting.. 

# mount -t ceph admin@.van2-dev=/ /cephfs/ -o noatime,ms_mode=prefer-crc,_netdev

mount error: no mds (Metadata Server) is up. The cluster might be laggy, or you 
may not be authorized


I am using ceph version 19.2.2 and client is also 19.2.2.

Regards
Dev

> On Jul 18, 2025, at 8:11 AM, Dario Graña <dgr...@pic.es> wrote:
> 
> Hi Devender,
> 
> I think the problem is that you are using the old syntax, try with the new
> one [1]
> 
> mount -t ceph admin@.van2-dev=/ /cephfs/ -o
> noatime,ms_mode=prefer-crc,_netdev
> 
> Regards.
> 
> [1] 
> https://www.google.com/url?q=https://docs.ceph.com/en/squid/man/8/mount.ceph/&source=gmail-imap&ust=1753457087000000&usg=AOvVaw3RPvx2e7p7nZypeiYOOf9c
> 
> On Wed, Jul 16, 2025 at 12:11 AM Devender Singh <deven...@netskrt.io> wrote:
> 
>> Tried below too but nothing works, any suggestion?
>> 
>> # mount -t ceph 10.101.1.246,10.101.1.247,10.101.1.248:/ /mnt/cephfs   -o
>> name=admin,fs=van2.dev,noatime,_netdev,ms_mode=prefer-crc
>> 
>> mount error: no mds (Metadata Server) is up. The cluster might be laggy,
>> or you may not be authorized
>> 
>> # mount -t ceph :/ /cephfs/ -o
>> noatime,ms_mode=prefer-crc,_netdev,fs=van2-dev
>> mount error: no mds (Metadata Server) is up. The cluster might be laggy,
>> or you may not be authorized
>> 
>> Regards
>> Dev
>> 
>>> On Jul 15, 2025, at 2:17 PM, Devender Singh <deven...@netskrt.io> wrote:
>>> 
>>> Hello all
>>> 
>>> I have cephfs working fine with running ads but still not able to mount,
>> getting error mount error: no mds (Metadata Server) is up
>>> 
>>> Below are details
>>> 
>>> 
>>> # ceph osd dump | grep min_compat_client
>>> require_min_compat_client reef
>>> min_compat_client reef
>>> 
>>> # ceph balancer status
>>> {
>>>    "active": true,
>>>    "last_optimize_duration": "0:00:00.002882",
>>>    "last_optimize_started": "Tue Jul 15 21:16:41 2025",
>>>    "mode": "upmap-read",
>>>    "no_optimization_needed": true,
>>>    "optimize_result": "Unable to find further optimization, or pool(s)
>> pg_num is decreasing, or distribution is already perfect",
>>>    "plans": []
>>> }
>>> 
>>> root@van2-host06n:~# dpkg -l ceph-common
>>> Desired=Unknown/Install/Remove/Purge/Hold
>>> |
>> Status=Not/Inst/Conf-files/Unpacked/halF-conf/Half-inst/trig-aWait/Trig-pend
>>> |/ Err?=(none)/Reinst-required (Status,Err: uppercase=bad)
>>> ||/ Name           Version       Architecture Description
>>> 
>> +++-==============-=============-============-==================================================================
>>> ii  ceph-common    19.2.2-1jammy amd64        common utilities to mount
>> and interact with a ceph storage cluster
>>> 
>>> root@van2-host06n:~# ceph fs status
>>> van2.dev - 0 clients
>>> ========
>>> RANK  STATE                  MDS                    ACTIVITY     DNS
>> INOS   DIRS   CAPS
>>> 0    active  van2.dev.van2-host06n.wqttsi  Reqs:    0 /s    10     13
>>   12      0
>>> 1    active  van2.dev.van2-host07n.fngvlt  Reqs:    0 /s    10     13
>>   11      0
>>>        POOL            TYPE     USED  AVAIL
>>> cephfs.van2.dev.meta  metadata   230k  53.0T
>>> cephfs.van2.dev.data    data       0   53.0T
>>>           STANDBY MDS
>>> van2-dev.van2-host07n.jnacmh
>>> van2-dev.van2-host06n.wpzshe
>>> van2-dev.van2-host08n.slrfej
>>> MDS version: ceph version 19.2.2
>> (0eceb0defba60152a8182f7bd87d164b639885b8) squid (stable)
>>> 
>>> 
>>> SYstemd file.
>>> ===========
>>> 
>>> # cat cephfs.mount
>>> #
>>> # this mounts on /cephfs
>>> 
>>> [Unit]
>>> Description=mount /cephfs
>>> After=network-online.target
>>> Requires=network-online.target
>>> 
>>> [Mount]
>>> Where=/cephfs
>>> What=10.101.1.246:6789,10.101.1.247:6789,10.101.1.248:6789:/
>>> Type=ceph
>>> Options=fs=van2.dev,name=admin,noatime,_netdev,x-systemd.automount
>>> 
>>> [Install]
>>> WantedBy=multi-user.target
>>> 
>>> 
>>> root@van2-host06n:~# journalctl -u cephfs.mount
>>> Jul 15 19:31:10 van2-host06n.van2.example.com systemd[1]: Mounting
>> mount /cephfs...
>>> Jul 15 19:32:12 van2-host06n.van2.example.com mount[1809096]: mount
>> error: no mds (Metadata Server) is up. The clu>
>>> Jul 15 19:32:12 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
>> Mount process exited, code=exited, status=>
>>> Jul 15 19:32:12 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
>> Failed with result 'exit-code'.
>>> Jul 15 19:32:12 van2-host06n.van2.example.com systemd[1]: Failed to
>> mount mount /cephfs.
>>> Jul 15 19:34:59 van2-host06n.van2.example.com systemd[1]: Mounting
>> mount /cephfs...
>>> Jul 15 19:35:59 van2-host06n.van2.example.com mount[1810584]: mount
>> error: no mds (Metadata Server) is up. The clu>
>>> Jul 15 19:35:59 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
>> Mount process exited, code=exited, status=>
>>> Jul 15 19:35:59 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
>> Failed with result 'exit-code'.
>>> Jul 15 19:35:59 van2-host06n.van2.example.com systemd[1]: Failed to
>> mount mount /cephfs.
>>> Jul 15 19:36:16 van2-host06n.van2.example.com systemd[1]: Mounting
>> mount /cephfs...
>>> Jul 15 19:37:17 van2-host06n.van2.example.com mount[1811498]: mount
>> error: no mds (Metadata Server) is up. The clu>
>>> Jul 15 19:37:17 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
>> Mount process exited, code=exited, status=>
>>> Jul 15 19:37:17 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
>> Failed with result 'exit-code'.
>>> Jul 15 19:37:17 van2-host06n.van2.example.com systemd[1]: Failed to
>> mount mount /cephfs.
>>> Jul 15 19:49:38 van2-host06n.van2.example.com systemd[1]: Mounting
>> mount /cephfs...
>>> Jul 15 19:50:40 van2-host06n.van2.example.com mount[1817486]: mount
>> error: no mds (Metadata Server) is up. The clu>
>>> Jul 15 19:50:40 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
>> Mount process exited, code=exited, status=>
>>> Jul 15 19:50:40 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
>> Failed with result 'exit-code'.
>>> Jul 15 19:50:40 van2-host06n.van2.example.com systemd[1]: Failed to
>> mount mount /cephfs.
>>> Jul 15 19:51:58 van2-host06n.van2.example.com systemd[1]: Mounting
>> mount /cephfs...
>>> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
>> Mount process exited, code=killed, status=>
>>> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
>> Failed with result 'signal'.
>>> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
>> Unit process 1817555 (mount.ceph) remains >
>>> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: Unmounted
>> mount /cephfs.
>>> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
>> Found left-over process 1817555 (mount.cep>
>>> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: This usually
>> indicates unclean termination of a previous>
>>> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: Mounting
>> mount /cephfs...
>>> Jul 15 19:52:59 van2-host06n.van2.example.com mount[1817555]: mount
>> error: no mds (Metadata Server) is up. The clu>
>>> lines 1-29...skipping...
>>> Jul 15 19:31:10 van2-host06n.van2.example.com systemd[1]: Mounting
>> mount /cephfs...
>>> Jul 15 19:32:12 van2-host06n.van2.example.com mount[1809096]: mount
>> error: no mds (Metadata Server) is up. The cluster might be laggy, or you
>> may not be>
>>> Jul 15 19:32:12 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
>> Mount process exited, code=exited, status=32/n/a
>>> Jul 15 19:32:12 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
>> Failed with result 'exit-code'.
>>> Jul 15 19:32:12 van2-host06n.van2.example.com systemd[1]: Failed to
>> mount mount /cephfs.
>>> Jul 15 19:34:59 van2-host06n.van2.example.com systemd[1]: Mounting
>> mount /cephfs...
>>> Jul 15 19:35:59 van2-host06n.van2.example.com mount[1810584]: mount
>> error: no mds (Metadata Server) is up. The cluster might be laggy, or you
>> may not be>
>>> Jul 15 19:35:59 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
>> Mount process exited, code=exited, status=32/n/a
>>> Jul 15 19:35:59 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
>> Failed with result 'exit-code'.
>>> Jul 15 19:35:59 van2-host06n.van2.example.com systemd[1]: Failed to
>> mount mount /cephfs.
>>> Jul 15 19:36:16 van2-host06n.van2.example.com systemd[1]: Mounting
>> mount /cephfs...
>>> Jul 15 19:37:17 van2-host06n.van2.example.com mount[1811498]: mount
>> error: no mds (Metadata Server) is up. The cluster might be laggy, or you
>> may not be>
>>> Jul 15 19:37:17 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
>> Mount process exited, code=exited, status=32/n/a
>>> Jul 15 19:37:17 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
>> Failed with result 'exit-code'.
>>> Jul 15 19:37:17 van2-host06n.van2.example.com systemd[1]: Failed to
>> mount mount /cephfs.
>>> Jul 15 19:49:38 van2-host06n.van2.example.com systemd[1]: Mounting
>> mount /cephfs...
>>> Jul 15 19:50:40 van2-host06n.van2.example.com mount[1817486]: mount
>> error: no mds (Metadata Server) is up. The cluster might be laggy, or you
>> may not be>
>>> Jul 15 19:50:40 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
>> Mount process exited, code=exited, status=32/n/a
>>> Jul 15 19:50:40 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
>> Failed with result 'exit-code'.
>>> Jul 15 19:50:40 van2-host06n.van2.example.com systemd[1]: Failed to
>> mount mount /cephfs.
>>> Jul 15 19:51:58 van2-host06n.van2.example.com systemd[1]: Mounting
>> mount /cephfs...
>>> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
>> Mount process exited, code=killed, status=15/TERM
>>> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
>> Failed with result 'signal'.
>>> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
>> Unit process 1817555 (mount.ceph) remains running after unit stopped.
>>> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: Unmounted
>> mount /cephfs.
>>> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
>> Found left-over process 1817555 (mount.ceph) in control group while
>> starting uni>
>>> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: This usually
>> indicates unclean termination of a previous run, or service implementation
>> defici>
>>> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: Mounting
>> mount /cephfs...
>>> Jul 15 19:52:59 van2-host06n.van2.example.com mount[1817555]: mount
>> error: no mds (Metadata Server) is up. The cluster might be laggy, or you
>> may not be>
>>> Jul 15 19:53:26 van2-host06n.van2.example.com mount[1817572]: mount
>> error: no mds (Metadata Server) is up. The cluster might be laggy, or you
>> may not be>
>>> Jul 15 19:53:26 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
>> Mount process exited, code=exited, status=32/n/a
>>> Jul 15 19:53:26 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
>> Failed with result 'exit-code'.
>>> Jul 15 19:53:26 van2-host06n.van2.example.com systemd[1]: Failed to
>> mount mount /cephfs.
>>> Jul 15 19:56:29 van2-host06n.van2.example.com systemd[1]: Mounting
>> mount /cephfs...
>>> Jul 15 19:57:29 van2-host06n.van2.example.com mount[1817655]: mount
>> error: no mds (Metadata Server) is up. The cluster might be laggy, or you
>> may not be>
>>> Jul 15 19:57:29 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
>> Mount process exited, code=exited, status=32/n/a
>>> Jul 15 19:57:29 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
>> Failed with result 'exit-code'.
>>> Jul 15 19:57:29 van2-host06n.van2.example.com systemd[1]: Failed to
>> mount mount /cephfs.
>>> lines 1-38...skipping...
>>> Jul 15 19:31:10 van2-host06n.van2.example.com systemd[1]: Mounting
>> mount /cephfs...
>>> Jul 15 19:32:12 van2-host06n.van2.example.com mount[1809096]: mount
>> error: no mds (Metadata Server) is up. The cluster might be laggy, or you
>> may not be authorized
>>> Jul 15 19:32:12 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
>> Mount process exited, code=exited, status=32/n/a
>>> Jul 15 19:32:12 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
>> Failed with result 'exit-code'.
>>> Jul 15 19:32:12 van2-host06n.van2.example.com systemd[1]: Failed to
>> mount mount /cephfs.
>>> Jul 15 19:34:59 van2-host06n.van2.example.com systemd[1]: Mounting
>> mount /cephfs...
>>> Jul 15 19:35:59 van2-host06n.van2.example.com mount[1810584]: mount
>> error: no mds (Metadata Server) is up. The cluster might be laggy, or you
>> may not be authorized
>>> Jul 15 19:35:59 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
>> Mount process exited, code=exited, status=32/n/a
>>> Jul 15 19:35:59 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
>> Failed with result 'exit-code'.
>>> Jul 15 19:35:59 van2-host06n.van2.example.com systemd[1]: Failed to
>> mount mount /cephfs.
>>> Jul 15 19:36:16 van2-host06n.van2.example.com systemd[1]: Mounting
>> mount /cephfs...
>>> Jul 15 19:37:17 van2-host06n.van2.example.com mount[1811498]: mount
>> error: no mds (Metadata Server) is up. The cluster might be laggy, or you
>> may not be authorized
>>> Jul 15 19:37:17 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
>> Mount process exited, code=exited, status=32/n/a
>>> Jul 15 19:37:17 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
>> Failed with result 'exit-code'.
>>> Jul 15 19:37:17 van2-host06n.van2.example.com systemd[1]: Failed to
>> mount mount /cephfs.
>>> Jul 15 19:49:38 van2-host06n.van2.example.com systemd[1]: Mounting
>> mount /cephfs...
>>> Jul 15 19:50:40 van2-host06n.van2.example.com mount[1817486]: mount
>> error: no mds (Metadata Server) is up. The cluster might be laggy, or you
>> may not be authorized
>>> Jul 15 19:50:40 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
>> Mount process exited, code=exited, status=32/n/a
>>> Jul 15 19:50:40 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
>> Failed with result 'exit-code'.
>>> Jul 15 19:50:40 van2-host06n.van2.example.com systemd[1]: Failed to
>> mount mount /cephfs.
>>> Jul 15 19:51:58 van2-host06n.van2.example.com systemd[1]: Mounting
>> mount /cephfs...
>>> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
>> Mount process exited, code=killed, status=15/TERM
>>> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
>> Failed with result 'signal'.
>>> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
>> Unit process 1817555 (mount.ceph) remains running after unit stopped.
>>> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: Unmounted
>> mount /cephfs.
>>> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
>> Found left-over process 1817555 (mount.ceph) in control group while
>> starting unit. Ignoring.
>>> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: This usually
>> indicates unclean termination of a previous run, or service implementation
>> deficiencies.
>>> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: Mounting
>> mount /cephfs...
>>> Jul 15 19:52:59 van2-host06n.van2.example.com mount[1817555]: mount
>> error: no mds (Metadata Server) is up. The cluster might be laggy, or you
>> may not be authorized
>>> Jul 15 19:53:26 van2-host06n.van2.example.com mount[1817572]: mount
>> error: no mds (Metadata Server) is up. The cluster might be laggy, or you
>> may not be authorized
>>> Jul 15 19:53:26 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
>> Mount process exited, code=exited, status=32/n/a
>>> Jul 15 19:53:26 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
>> Failed with result 'exit-code'.
>>> Jul 15 19:53:26 van2-host06n.van2.example.com systemd[1]: Failed to
>> mount mount /cephfs.
>>> Jul 15 19:56:29 van2-host06n.van2.example.com systemd[1]: Mounting
>> mount /cephfs...
>>> Jul 15 19:57:29 van2-host06n.van2.example.com mount[1817655]: mount
>> error: no mds (Metadata Server) is up. The cluster might be laggy, or you
>> may not be authorized
>>> Jul 15 19:57:29 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
>> Mount process exited, code=exited, status=32/n/a
>>> Jul 15 19:57:29 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
>> Failed with result 'exit-code'.
>>> Jul 15 19:57:29 van2-host06n.van2.example.com systemd[1]: Failed to
>> mount mount /cephfs.
>>> Jul 15 19:58:50 van2-host06n.van2.example.com systemd[1]: Mounting
>> mount /cephfs...
>>> Jul 15 19:59:51 van2-host06n.van2.example.com mount[1818287]: mount
>> error: no mds (Metadata Server) is up. The cluster might be laggy, or you
>> may not be authorized
>>> Jul 15 19:59:51 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
>> Mount process exited, code=exited, status=32/n/a
>>> Jul 15 19:59:51 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
>> Failed with result 'exit-code'.
>>> Jul 15 19:59:51 van2-host06n.van2.example.com systemd[1]: Failed to
>> mount mount /cephfs.
>>> Jul 15 20:05:43 van2-host06n.van2.example.com systemd[1]: Mounting
>> mount /cephfs...
>>> Jul 15 20:06:44 van2-host06n.van2.example.com mount[1818912]: mount
>> error: no mds (Metadata Server) is up. The cluster might be laggy, or you
>> may not be authorized
>>> lines 1-45
>>> 
>> 
>> _______________________________________________
>> ceph-users mailing list -- ceph-users@ceph.io
>> To unsubscribe send an email to ceph-users-le...@ceph.io
>> 
> _______________________________________________
> ceph-users mailing list -- ceph-users@ceph.io
> To unsubscribe send an email to ceph-users-le...@ceph.io

_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to