Sorry I couldn’t answer back sooner, I made it with
/etc/systemd/system/rbd-{ceph_pool}-{ceph_image}.service, I had some mistakes
in the name of the config file:
[Unit]
Description=RADOS block device mapping for "{ceph_pool}"/"{ceph_image}"
Conflicts=shutdown.target
Wants=network-online.target
After=NetworkManager-wait-online.service
[Service]
Type=oneshot
ExecStart=/sbin/modprobe rbd
ExecStart=/bin/sh -c "/bin/echo {ceph_mon_ip}
name={ceph_admin},secret={ceph_key} {ceph_pool} {ceph_image} >/sys/bus/rbd/add"
TimeoutSec=0
RemainAfterExit=yes
[Install]
WantedBy=multi-user.target
Yes I did use systemctl and normal mount points in ftstab now is working =)
Thank you so mucho for all your help, Id like to take this message to ask you
the last thing that is concerned to me, I did a failure test shutting down 1
server, but while the server is down I cannot write in the filesystem, is that
a normal behaviour? I checked the crush algorithm policy and it has “host”
taken for the replica that means that each replica is going to be in different
hosts…
Thank you everyone!
[cid:[email protected]]
Jesus Chavez
SYSTEMS ENGINEER-C.SALES
[email protected]<mailto:[email protected]>
Phone: +52 55 5267 3146
Mobile: +51 1 5538883255
CCIE - 44433
Cisco.com<http://www.cisco.com/>
[cid:[email protected]]
Think before you print.
This email may contain confidential and privileged material for the sole use of
the intended recipient. Any review, use, distribution or disclosure by others
is strictly prohibited. If you are not the intended recipient (or authorized to
receive for the recipient), please contact the sender by reply email and delete
all copies of this message.
Please click
here<http://www.cisco.com/web/about/doing_business/legal/cri/index.html> for
Company Registration Information.
On Mar 7, 2015, at 3:15 AM, Gian
<[email protected]<mailto:[email protected]>> wrote:
Hi,
are you using /etc/ceph/rbdmount as a 'mapping fstab' plus your mountpoints in
normal fstab plus the systemctl service ?
Gian
On 07 Mar 2015, at 05:26, Jesus Chavez (jeschave)
<[email protected]<mailto:[email protected]>> wrote:
Still not working does anybody know show to automap and Mount rbd image on
redhat?
Regards
Jesus Chavez
SYSTEMS ENGINEER-C.SALES
[email protected]<mailto:[email protected]>
Phone: +52 55 5267 3146<tel:+52%2055%205267%203146>
Mobile: +51 1 5538883255<tel:+51%201%205538883255>
CCIE - 44433
On Mar 2, 2015, at 4:52 AM, Jesus Chavez (jeschave)
<[email protected]<mailto:[email protected]>> wrote:
Thank you so much Alexandre! :)
Jesus Chavez
SYSTEMS ENGINEER-C.SALES
[email protected]<mailto:[email protected]>
Phone: +52 55 5267 3146<tel:+52%2055%205267%203146>
Mobile: +51 1 5538883255<tel:+51%201%205538883255>
CCIE - 44433
On Mar 2, 2015, at 4:26 AM, Alexandre DERUMIER
<[email protected]<mailto:[email protected]>> wrote:
Hi,
maybe this can help you:
http://www.sebastien-han.fr/blog/2013/11/22/map-slash-unmap-rbd-device-on-boot-slash-shutdown/
Regards,
Alexandre
----- Mail original -----
De: "Jesus Chavez (jeschave)" <[email protected]<mailto:[email protected]>>
À: "ceph-users" <[email protected]<mailto:[email protected]>>
Envoyé: Lundi 2 Mars 2015 11:14:49
Objet: [ceph-users] Permanente Mount RBD blocs device RHEL7
Hi all! I have been trying to get permanent my fs maked by the rbd device
mapping on rhel7 modifying /etc/fstab but everytime I reboot the server I lose
the mapping to the pool so the server gets stuck since It didnt find the
/dev/rbd0 device, does anybody know if there any procedure to not lose the
mapping or make the filesystem permanent?
Thanks!
Jesus Chavez
SYSTEMS ENGINEER-C.SALES
[email protected]<mailto:[email protected]>
Phone : +52 55 5267 3146
Mobile: +51 1 5538883255
CCIE - 44433
_______________________________________________
ceph-users mailing list
[email protected]<mailto:[email protected]>
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
[email protected]<mailto:[email protected]>
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
[email protected]<mailto:[email protected]>
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com