Re: [ceph-users] iSCSI Multipath (Load Balancing) vs RBD Exclusive Lock
-- From: "Jason Dillaman"Sent: Sunday, March 11, 2018 1:46 AM To: "shadow_lin" Cc: "Lazuardi Nasution" ; "Ceph Users" Subject: Re: [ceph-users] iSCSI Multipath (Load Balancing) vs RBD Exclusive Lock On Sat, Mar 10, 2018 at 10:11 AM, shadow_lin wrote: Hi Jason, As discussed in this thread, for active/passive, upon initiator failover, we used the RBD exclusive-lock feature to blacklist the old "active" iSCSI target gateway so that it cannot talk w/ the Ceph cluster before new writes are accepted on the new target gateway. I can get during the new active target gateway was talking to rbd the old active target gateway cannot write because of the RBD exclusive-lock But after the new target gateway done the writes,if the old target gateway had some blocked io during the failover,cant it then get the lock and overwrite the new writes? Negative -- it's blacklisted so it cannot talk to the cluster. PS: Petasan say they can do active/active iscsi with patched suse kernel. I'll let them comment on these corner cases. We are not currently handling these corner cases. We have not hit this in practice but will work on it. We need to account for in-flight time early in the target stack before reaching krbd/tcmu. /Maged 2018-03-10 shadowlin 发件人:Jason Dillaman 发送时间:2018-03-10 21:40 主题:Re: [ceph-users] iSCSI Multipath (Load Balancing) vs RBD Exclusive Lock 收件人:"shadow_lin" 抄送:"Mike Christie" ,"Lazuardi Nasution" ,"Ceph Users" On Sat, Mar 10, 2018 at 7:42 AM, shadow_lin wrote: Hi Mike, So for now only suse kernel with target_rbd_core and tcmu-runner can run active/passive multipath safely? Negative, the LIO / tcmu-runner implementation documented here [1] is safe for active/passive. I am a newbie to iscsi. I think the stuck io get excuted cause overwrite problem can happen with both active/active and active/passive. What makes the active/passive safer than active/active? As discussed in this thread, for active/passive, upon initiator failover, we used the RBD exclusive-lock feature to blacklist the old "active" iSCSI target gateway so that it cannot talk w/ the Ceph cluster before new writes are accepted on the new target gateway. What mechanism should be implement to avoid the problem with active/passive and active/active multipath? Active/passive it solved as discussed above. For active/active, we don't have a solution that is known safe under all failure conditions. If LIO supported MCS (multiple connections per session) instead of just MPIO (multipath IO), the initiator would provide enough context to the target to detect IOs from a failover situation. 2018-03-10 shadowlin 发件人:Mike Christie 发送时间:2018-03-09 00:54 主题:Re: [ceph-users] iSCSI Multipath (Load Balancing) vs RBD Exclusive Lock 收件人:"shadow_lin" ,"Lazuardi Nasution" ,"Ceph Users" 抄送: On 03/07/2018 09:24 AM, shadow_lin wrote: Hi Christie, Is it safe to use active/passive multipath with krbd with exclusive lock for lio/tgt/scst/tcmu? No. We tried to use lio and krbd initially, but there is a issue where IO might get stuck in the target/block layer and get executed after new IO. So for lio, tgt and tcmu it is not safe as is right now. We could add some code tcmu's file_example handler which can be used with krbd so it works like the rbd one. I do know enough about SCST right now. Is it safe to use active/active multipath If use suse kernel with target_core_rbd? Thanks. 2018-03-07 shadowlin *发件人:*Mike Christie *发送时间:*2018-03-07 03:51 *主题:*Re: [ceph-users] iSCSI Multipath (Load Balancing) vs RBD Exclusive Lock *收件人:*"Lazuardi Nasution" ,"Ceph Users" *抄送:* On 03/06/2018 01:17 PM, Lazuardi Nasution wrote: > Hi, > > I want to do load balanced multipathing (multiple iSCSI gateway/exporter > nodes) of iSCSI backed with RBD images. Should I disable exclusive lock > feature? What if I don't disable that feature? I'm using TGT (manual > way) since I get so many CPU stuck error messages when I was using LIO. > You are using LIO/TGT with krbd right? You cannot or shouldn't do active/active multipathing. If you have the lock enabled then it bounces between paths for each IO and will be slow. If you do not
[ceph-users] rbd-nbd not resizing even after kernel tweaks
I am running into the problem described in https://lkml.org/lkml/2018/2/19/565 and https://tracker.ceph.com/issues/23137 I went ahead and built a custom kernel reverting the change https://github.com/torvalds/linux/commit/639812a1ed9bf49ae2c026086fbf975339cd1eef After that a resize shows in lsblk and /sys/block/nbdX/size, but not in parted for a mounted filesystem. Unmapping and remapping the NBD device shows the size in parted. Thank you for any help -- Alex Gorbachev Storcium ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
Re: [ceph-users] iSCSI Multipath (Load Balancing) vs RBD Exclusive Lock
On Sat, Mar 10, 2018 at 10:11 AM, shadow_linwrote: > Hi Jason, > >>As discussed in this thread, for active/passive, upon initiator >>failover, we used the RBD exclusive-lock feature to blacklist the old >>"active" iSCSI target gateway so that it cannot talk w/ the Ceph >>cluster before new writes are accepted on the new target gateway. > > I can get during the new active target gateway was talking to rbd the old > active target gateway cannot write because of the RBD exclusive-lock > But after the new target gateway done the writes,if the old target gateway > had some blocked io during the failover,cant it then get the lock and > overwrite the new writes? Negative -- it's blacklisted so it cannot talk to the cluster. > PS: > Petasan say they can do active/active iscsi with patched suse kernel. I'll let them comment on these corner cases. > 2018-03-10 > > shadowlin > > > > 发件人:Jason Dillaman > 发送时间:2018-03-10 21:40 > 主题:Re: [ceph-users] iSCSI Multipath (Load Balancing) vs RBD Exclusive Lock > 收件人:"shadow_lin" > 抄送:"Mike Christie" ,"Lazuardi > Nasution" ,"Ceph Users" > > On Sat, Mar 10, 2018 at 7:42 AM, shadow_lin wrote: >> Hi Mike, >> So for now only suse kernel with target_rbd_core and tcmu-runner can run >> active/passive multipath safely? > > Negative, the LIO / tcmu-runner implementation documented here [1] is > safe for active/passive. > >> I am a newbie to iscsi. I think the stuck io get excuted cause overwrite >> problem can happen with both active/active and active/passive. >> What makes the active/passive safer than active/active? > > As discussed in this thread, for active/passive, upon initiator > failover, we used the RBD exclusive-lock feature to blacklist the old > "active" iSCSI target gateway so that it cannot talk w/ the Ceph > cluster before new writes are accepted on the new target gateway. > >> What mechanism should be implement to avoid the problem with >> active/passive >> and active/active multipath? > > Active/passive it solved as discussed above. For active/active, we > don't have a solution that is known safe under all failure conditions. > If LIO supported MCS (multiple connections per session) instead of > just MPIO (multipath IO), the initiator would provide enough context > to the target to detect IOs from a failover situation. > >> 2018-03-10 >> >> shadowlin >> >> >> >> 发件人:Mike Christie >> 发送时间:2018-03-09 00:54 >> 主题:Re: [ceph-users] iSCSI Multipath (Load Balancing) vs RBD Exclusive Lock >> 收件人:"shadow_lin" ,"Lazuardi >> Nasution" ,"Ceph Users" >> 抄送: >> >> On 03/07/2018 09:24 AM, shadow_lin wrote: >>> Hi Christie, >>> Is it safe to use active/passive multipath with krbd with exclusive lock >>> for lio/tgt/scst/tcmu? >> >> No. We tried to use lio and krbd initially, but there is a issue where >> IO might get stuck in the target/block layer and get executed after new >> IO. So for lio, tgt and tcmu it is not safe as is right now. We could >> add some code tcmu's file_example handler which can be used with krbd so >> it works like the rbd one. >> >> I do know enough about SCST right now. >> >> >>> Is it safe to use active/active multipath If use suse kernel with >>> target_core_rbd? >>> Thanks. >>> >>> 2018-03-07 >>> >>> shadowlin >>> >>> >>> >>> *发件人:*Mike Christie >>> *发送时间:*2018-03-07 03:51 >>> *主题:*Re: [ceph-users] iSCSI Multipath (Load Balancing) vs RBD >>> Exclusive Lock >>> *收件人:*"Lazuardi Nasution" ,"Ceph >>> Users" >>> *抄送:* >>> >>> On 03/06/2018 01:17 PM, Lazuardi Nasution wrote: >>> > Hi, >>> > >>> > I want to do load balanced multipathing (multiple iSCSI >>> gateway/exporter >>> > nodes) of iSCSI backed with RBD images. Should I disable exclusive >>> lock >>> > feature? What if I don't disable that feature? I'm using TGT >>> (manual >>> > way) since I get so many CPU stuck error messages when I was using >>> LIO. >>> > >>> >>> You are using LIO/TGT with krbd right? >>> >>> You cannot or shouldn't do active/active multipathing. If you have >>> the >>> lock enabled then it bounces between paths for each IO and will be >>> slow. >>> If you do not have it enabled then you can end up with stale IO >>> overwriting current data. >>> >>> >>> >>> >> >> >> ___ >> ceph-users mailing list >> ceph-users@lists.ceph.com >>
Re: [ceph-users] (no subject)
Hi Nathan, this indeed appears to be a Gentoo-specific issue. They install the file at: /usr/libexec/ceph/ceph-osd-prestart.sh instead of /usr/lib/ceph/ceph-osd-prestart.sh It depends on how you strongly you follow FHS ( http://refspecs.linuxfoundation.org/FHS_3.0/fhs/ch04s07.html ) which is the actual correct place to be used. It seems Gentoo packagers took FHS 3.0 seriously to decide where to install things, but did not patch the code accordingly, so this surely warrants a Gentoo bug. Distro packagers should then decide whether they want to change install location (and rather follow Ceph upstream than FHS 3.0), or patch the code to enforce FHS 3.0. https://bugs.gentoo.org/632028 is also related to that. Cheers, Oliver Am 11.03.2018 um 00:34 schrieb Nathan Dehnel: > Trying to create an OSD: > > gentooserver ~ # ceph-volume lvm create --data /dev/sdb > Running command: ceph-authtool --gen-print-key > Running command: ceph --cluster ceph --name client.bootstrap-osd --keyring > /var/lib/ceph/bootstrap-osd/ceph.keyring -i - osd new > e70500fe-0d51-48c3-a607-414957886726 > Running command: vgcreate --force --yes > ceph-a736559a-92d1-483e-9289-d2c7feed510f /dev/sdb > stdout: Volume group "ceph-a736559a-92d1-483e-9289-d2c7feed510f" > successfully created > Running command: lvcreate --yes -l 100%FREE -n > osd-block-e70500fe-0d51-48c3-a607-414957886726 > ceph-a736559a-92d1-483e-9289-d2c7feed510f > stdout: Logical volume "osd-block-e70500fe-0d51-48c3-a607-414957886726" > created. > Running command: ceph-authtool --gen-print-key > Running command: mount -t tmpfs tmpfs /var/lib/ceph/osd/ceph-0 > Running command: chown -R ceph:ceph /dev/dm-0 > Running command: ln -s > /dev/ceph-a736559a-92d1-483e-9289-d2c7feed510f/osd-block-e70500fe-0d51-48c3-a607-414957886726 > /var/lib/ceph/osd/ceph-0/block > Running command: ceph --cluster ceph --name client.bootstrap-osd --keyring > /var/lib/ceph/bootstrap-osd/ceph.keyring mon getmap -o > /var/lib/ceph/osd/ceph-0/activate.monmap > stderr: got monmap epoch 1 > Running command: ceph-authtool /var/lib/ceph/osd/ceph-0/keyring > --create-keyring --name osd.0 --add-key > AQBEZqRalRoRCBAA03R6VshykLcZjMgQnFKDtg== > stdout: creating /var/lib/ceph/osd/ceph-0/keyring > stdout: added entity osd.0 auth auth(auid = 18446744073709551615 > key=AQBEZqRalRoRCBAA03R6VshykLcZjMgQnFKDtg== with 0 caps) > Running command: chown -R ceph:ceph /var/lib/ceph/osd/ceph-0/keyring > Running command: chown -R ceph:ceph /var/lib/ceph/osd/ceph-0/ > Running command: ceph-osd --cluster ceph --osd-objectstore bluestore --mkfs > -i 0 --monmap /var/lib/ceph/osd/ceph-0/activate.monmap --keyfile - --osd-data > /var/lib/ceph/osd/ceph-0/ --osd-uuid e70500fe-0d51-48c3-a607-414957886726 > --setuser ceph --setgroup ceph > --> ceph-volume lvm prepare successful for: /dev/sdb > Running command: ceph-bluestore-tool --cluster=ceph prime-osd-dir --dev > /dev/ceph-a736559a-92d1-483e-9289-d2c7feed510f/osd-block-e70500fe-0d51-48c3-a607-414957886726 > --path /var/lib/ceph/osd/ceph-0 > Running command: ln -snf > /dev/ceph-a736559a-92d1-483e-9289-d2c7feed510f/osd-block-e70500fe-0d51-48c3-a607-414957886726 > /var/lib/ceph/osd/ceph-0/block > Running command: chown -R ceph:ceph /dev/dm-0 > Running command: chown -R ceph:ceph /var/lib/ceph/osd/ceph-0 > Running command: systemctl enable > ceph-volume@lvm-0-e70500fe-0d51-48c3-a607-414957886726 > stderr: Created symlink > /etc/systemd/system/multi-user.target.wants/ceph-volume@lvm-0-e70500fe-0d51-48c3-a607-414957886726.service > → /lib/systemd/system/ceph-volume@.service. > Running command: systemctl start ceph-osd@0 > stderr: Job for ceph-osd@0.service failed because the control process exited > with error code. > See "systemctl status ceph-osd@0.service" and "journalctl -xe" for details. > --> Was unable to complete a new OSD, will rollback changes > --> OSD will be fully purged from the cluster, because the ID was generated > Running command: ceph osd purge osd.0 --yes-i-really-mean-it > stderr: purged osd.0 > --> RuntimeError: command returned non-zero exit status: 1 > > journalctl -xe > -- Unit ceph-osd@0.service has begun starting up. > Mar 10 17:14:34 gentooserver systemd[3977]: ceph-osd@0.service: Failed to > execute command: No such file or directory > Mar 10 17:14:34 gentooserver systemd[3977]: ceph-osd@0.service: Failed at > step EXEC spawning /usr/lib/ceph/ceph-osd-prestart.sh: No such file or > directory > -- Subject: Process /usr/lib/ceph/ceph-osd-prestart.sh could not be executed > -- Defined-By: systemd > -- Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel > -- > -- The process /usr/lib/ceph/ceph-osd-prestart.sh could not be executed and > failed. > -- > -- The error number returned by this process is 2. > Mar 10 17:14:34 gentooserver systemd[1]: ceph-osd@0.service: Control process > exited, code=exited status=203 > Mar 10 17:14:34 gentooserver systemd[1]: ceph-osd@0.service: Failed with >
[ceph-users] (no subject)
Trying to create an OSD: gentooserver ~ # ceph-volume lvm create --data /dev/sdb Running command: ceph-authtool --gen-print-key Running command: ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring -i - osd new e70500fe-0d51-48c3-a607-414957886726 Running command: vgcreate --force --yes ceph-a736559a-92d1-483e-9289-d2c7feed510f /dev/sdb stdout: Volume group "ceph-a736559a-92d1-483e-9289-d2c7feed510f" successfully created Running command: lvcreate --yes -l 100%FREE -n osd-block-e70500fe-0d51-48c3-a607-414957886726 ceph-a736559a-92d1-483e-9289-d2c7feed510f stdout: Logical volume "osd-block-e70500fe-0d51-48c3-a607-414957886726" created. Running command: ceph-authtool --gen-print-key Running command: mount -t tmpfs tmpfs /var/lib/ceph/osd/ceph-0 Running command: chown -R ceph:ceph /dev/dm-0 Running command: ln -s /dev/ceph-a736559a-92d1-483e-9289-d2c7feed510f/osd-block-e70500fe-0d51-48c3-a607-414957886726 /var/lib/ceph/osd/ceph-0/block Running command: ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring mon getmap -o /var/lib/ceph/osd/ceph-0/activate.monmap stderr: got monmap epoch 1 Running command: ceph-authtool /var/lib/ceph/osd/ceph-0/keyring --create-keyring --name osd.0 --add-key AQBEZqRalRoRCBAA03R6VshykLcZjMgQnFKDtg== stdout: creating /var/lib/ceph/osd/ceph-0/keyring stdout: added entity osd.0 auth auth(auid = 18446744073709551615 key=AQBEZqRalRoRCBAA03R6VshykLcZjMgQnFKDtg== with 0 caps) Running command: chown -R ceph:ceph /var/lib/ceph/osd/ceph-0/keyring Running command: chown -R ceph:ceph /var/lib/ceph/osd/ceph-0/ Running command: ceph-osd --cluster ceph --osd-objectstore bluestore --mkfs -i 0 --monmap /var/lib/ceph/osd/ceph-0/activate.monmap --keyfile - --osd-data /var/lib/ceph/osd/ceph-0/ --osd-uuid e70500fe-0d51-48c3-a607-414957886726 --setuser ceph --setgroup ceph --> ceph-volume lvm prepare successful for: /dev/sdb Running command: ceph-bluestore-tool --cluster=ceph prime-osd-dir --dev /dev/ceph-a736559a-92d1-483e-9289-d2c7feed510f/osd-block-e70500fe-0d51-48c3-a607-414957886726 --path /var/lib/ceph/osd/ceph-0 Running command: ln -snf /dev/ceph-a736559a-92d1-483e-9289-d2c7feed510f/osd-block-e70500fe-0d51-48c3-a607-414957886726 /var/lib/ceph/osd/ceph-0/block Running command: chown -R ceph:ceph /dev/dm-0 Running command: chown -R ceph:ceph /var/lib/ceph/osd/ceph-0 Running command: systemctl enable ceph-volume@lvm-0-e70500fe-0d51-48c3-a607-414957886726 stderr: Created symlink /etc/systemd/system/multi-user.target.wants/ceph-volume@lvm-0-e70500fe-0d51-48c3-a607-414957886726.service → /lib/systemd/system/ceph-volume@.service. Running command: systemctl start ceph-osd@0 stderr: Job for ceph-osd@0.service failed because the control process exited with error code. See "systemctl status ceph-osd@0.service" and "journalctl -xe" for details. --> Was unable to complete a new OSD, will rollback changes --> OSD will be fully purged from the cluster, because the ID was generated Running command: ceph osd purge osd.0 --yes-i-really-mean-it stderr: purged osd.0 --> RuntimeError: command returned non-zero exit status: 1 journalctl -xe -- Unit ceph-osd@0.service has begun starting up. Mar 10 17:14:34 gentooserver systemd[3977]: ceph-osd@0.service: Failed to execute command: No such file or directory Mar 10 17:14:34 gentooserver systemd[3977]: ceph-osd@0.service: Failed at step EXEC spawning /usr/lib/ceph/ceph-osd-prestart.sh: No such file or directory -- Subject: Process /usr/lib/ceph/ceph-osd-prestart.sh could not be executed -- Defined-By: systemd -- Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel -- -- The process /usr/lib/ceph/ceph-osd-prestart.sh could not be executed and failed. -- -- The error number returned by this process is 2. Mar 10 17:14:34 gentooserver systemd[1]: ceph-osd@0.service: Control process exited, code=exited status=203 Mar 10 17:14:34 gentooserver systemd[1]: ceph-osd@0.service: Failed with result 'exit-code'. Mar 10 17:14:34 gentooserver systemd[1]: Failed to start Ceph object storage daemon osd.0. -- Subject: Unit ceph-osd@0.service has failed -- Defined-By: systemd -- Support: https://lists.freedesktop.org/mailman/listinfo/systemd-devel -- -- Unit ceph-osd@0.service has failed. -- -- The result is RESULT. Why is this file missing? Should I file a bug with my distro's packager? ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
Re: [ceph-users] New Ceph cluster design
Hi, As i understand it, you'll have one RAID1 of two SSDs for 12 HDDs. A WAL is used for all writes on your host. If you have good SSDs, they can handle 450-550 MBpsc. Your 12 HDDs SATA can handle 12 x 100 MBps that is to say 1200 GBps. So your RAID 1 will be the bootleneck with this design. A good design would be to have one SSD for 4 or 5 HDD. In your case, the best option would be to start with 3 SSDs for 12 HDDs to have a balances node. Don't forget to choose SSD with a high WDPD ratio (>10) The network needs of your node depend of the bandwith of your disks. As explain over, your 12 HDDs can handle 1200 GBps so you need to have a public and a private network that can handle it. In your case, a minimum a two 10 Gbps networks for per node are needed. If you need redondancy, just use two LACP networks with each having two 10 Gbps links. The scrub or deep scrub operations will not have a significant impact on your network but on your disks utilisations. So you need to plan them during low usage by your clients ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
Re: [ceph-users] iSCSI Multipath (Load Balancing) vs RBD Exclusive Lock
Hi Jason, >As discussed in this thread, for active/passive, upon initiator >failover, we used the RBD exclusive-lock feature to blacklist the old >"active" iSCSI target gateway so that it cannot talk w/ the Ceph >cluster before new writes are accepted on the new target gateway. I can get during the new active target gateway was talking to rbd the old active target gateway cannot write because of the RBD exclusive-lock But after the new target gateway done the writes,if the old target gateway had some blocked io during the failover,cant it then get the lock and overwrite the new writes? PS: Petasan say they can do active/active iscsi with patched suse kernel. 2018-03-10 shadowlin 发件人:Jason Dillaman发送时间:2018-03-10 21:40 主题:Re: [ceph-users] iSCSI Multipath (Load Balancing) vs RBD Exclusive Lock 收件人:"shadow_lin" 抄送:"Mike Christie" ,"Lazuardi Nasution" ,"Ceph Users" On Sat, Mar 10, 2018 at 7:42 AM, shadow_lin wrote: > Hi Mike, > So for now only suse kernel with target_rbd_core and tcmu-runner can run > active/passive multipath safely? Negative, the LIO / tcmu-runner implementation documented here [1] is safe for active/passive. > I am a newbie to iscsi. I think the stuck io get excuted cause overwrite > problem can happen with both active/active and active/passive. > What makes the active/passive safer than active/active? As discussed in this thread, for active/passive, upon initiator failover, we used the RBD exclusive-lock feature to blacklist the old "active" iSCSI target gateway so that it cannot talk w/ the Ceph cluster before new writes are accepted on the new target gateway. > What mechanism should be implement to avoid the problem with active/passive > and active/active multipath? Active/passive it solved as discussed above. For active/active, we don't have a solution that is known safe under all failure conditions. If LIO supported MCS (multiple connections per session) instead of just MPIO (multipath IO), the initiator would provide enough context to the target to detect IOs from a failover situation. > 2018-03-10 > > shadowlin > > > > 发件人:Mike Christie > 发送时间:2018-03-09 00:54 > 主题:Re: [ceph-users] iSCSI Multipath (Load Balancing) vs RBD Exclusive Lock > 收件人:"shadow_lin" ,"Lazuardi > Nasution" ,"Ceph Users" > 抄送: > > On 03/07/2018 09:24 AM, shadow_lin wrote: >> Hi Christie, >> Is it safe to use active/passive multipath with krbd with exclusive lock >> for lio/tgt/scst/tcmu? > > No. We tried to use lio and krbd initially, but there is a issue where > IO might get stuck in the target/block layer and get executed after new > IO. So for lio, tgt and tcmu it is not safe as is right now. We could > add some code tcmu's file_example handler which can be used with krbd so > it works like the rbd one. > > I do know enough about SCST right now. > > >> Is it safe to use active/active multipath If use suse kernel with >> target_core_rbd? >> Thanks. >> >> 2018-03-07 >> >> shadowlin >> >> >> >> *发件人:*Mike Christie >> *发送时间:*2018-03-07 03:51 >> *主题:*Re: [ceph-users] iSCSI Multipath (Load Balancing) vs RBD >> Exclusive Lock >> *收件人:*"Lazuardi Nasution" ,"Ceph >> Users" >> *抄送:* >> >> On 03/06/2018 01:17 PM, Lazuardi Nasution wrote: >> > Hi, >> > >> > I want to do load balanced multipathing (multiple iSCSI >> gateway/exporter >> > nodes) of iSCSI backed with RBD images. Should I disable exclusive >> lock >> > feature? What if I don't disable that feature? I'm using TGT (manual >> > way) since I get so many CPU stuck error messages when I was using >> LIO. >> > >> >> You are using LIO/TGT with krbd right? >> >> You cannot or shouldn't do active/active multipathing. If you have the >> lock enabled then it bounces between paths for each IO and will be >> slow. >> If you do not have it enabled then you can end up with stale IO >> overwriting current data. >> >> >> >> > > > ___ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > [1] http://docs.ceph.com/docs/master/rbd/iscsi-overview/ -- Jason ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
Re: [ceph-users] iSCSI Multipath (Load Balancing) vs RBD Exclusive Lock
On Sat, Mar 10, 2018 at 7:42 AM, shadow_linwrote: > Hi Mike, > So for now only suse kernel with target_rbd_core and tcmu-runner can run > active/passive multipath safely? Negative, the LIO / tcmu-runner implementation documented here [1] is safe for active/passive. > I am a newbie to iscsi. I think the stuck io get excuted cause overwrite > problem can happen with both active/active and active/passive. > What makes the active/passive safer than active/active? As discussed in this thread, for active/passive, upon initiator failover, we used the RBD exclusive-lock feature to blacklist the old "active" iSCSI target gateway so that it cannot talk w/ the Ceph cluster before new writes are accepted on the new target gateway. > What mechanism should be implement to avoid the problem with active/passive > and active/active multipath? Active/passive it solved as discussed above. For active/active, we don't have a solution that is known safe under all failure conditions. If LIO supported MCS (multiple connections per session) instead of just MPIO (multipath IO), the initiator would provide enough context to the target to detect IOs from a failover situation. > 2018-03-10 > > shadowlin > > > > 发件人:Mike Christie > 发送时间:2018-03-09 00:54 > 主题:Re: [ceph-users] iSCSI Multipath (Load Balancing) vs RBD Exclusive Lock > 收件人:"shadow_lin" ,"Lazuardi > Nasution" ,"Ceph Users" > 抄送: > > On 03/07/2018 09:24 AM, shadow_lin wrote: >> Hi Christie, >> Is it safe to use active/passive multipath with krbd with exclusive lock >> for lio/tgt/scst/tcmu? > > No. We tried to use lio and krbd initially, but there is a issue where > IO might get stuck in the target/block layer and get executed after new > IO. So for lio, tgt and tcmu it is not safe as is right now. We could > add some code tcmu's file_example handler which can be used with krbd so > it works like the rbd one. > > I do know enough about SCST right now. > > >> Is it safe to use active/active multipath If use suse kernel with >> target_core_rbd? >> Thanks. >> >> 2018-03-07 >> >> shadowlin >> >> >> >> *发件人:*Mike Christie >> *发送时间:*2018-03-07 03:51 >> *主题:*Re: [ceph-users] iSCSI Multipath (Load Balancing) vs RBD >> Exclusive Lock >> *收件人:*"Lazuardi Nasution" ,"Ceph >> Users" >> *抄送:* >> >> On 03/06/2018 01:17 PM, Lazuardi Nasution wrote: >> > Hi, >> > >> > I want to do load balanced multipathing (multiple iSCSI >> gateway/exporter >> > nodes) of iSCSI backed with RBD images. Should I disable exclusive >> lock >> > feature? What if I don't disable that feature? I'm using TGT (manual >> > way) since I get so many CPU stuck error messages when I was using >> LIO. >> > >> >> You are using LIO/TGT with krbd right? >> >> You cannot or shouldn't do active/active multipathing. If you have the >> lock enabled then it bounces between paths for each IO and will be >> slow. >> If you do not have it enabled then you can end up with stale IO >> overwriting current data. >> >> >> >> > > > ___ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > [1] http://docs.ceph.com/docs/master/rbd/iscsi-overview/ -- Jason ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
Re: [ceph-users] iSCSI Multipath (Load Balancing) vs RBD Exclusive Lock
Hi Mike, So for now only suse kernel with target_rbd_core and tcmu-runner can run active/passive multipath safely? I am a newbie to iscsi. I think the stuck io get excuted cause overwrite problem can happen with both active/active and active/passive. What makes the active/passive safer than active/active? What mechanism should be implement to avoid the problem with active/passive and active/active multipath? 2018-03-10 shadowlin 发件人:Mike Christie发送时间:2018-03-09 00:54 主题:Re: [ceph-users] iSCSI Multipath (Load Balancing) vs RBD Exclusive Lock 收件人:"shadow_lin" ,"Lazuardi Nasution" ,"Ceph Users" 抄送: On 03/07/2018 09:24 AM, shadow_lin wrote: > Hi Christie, > Is it safe to use active/passive multipath with krbd with exclusive lock > for lio/tgt/scst/tcmu? No. We tried to use lio and krbd initially, but there is a issue where IO might get stuck in the target/block layer and get executed after new IO. So for lio, tgt and tcmu it is not safe as is right now. We could add some code tcmu's file_example handler which can be used with krbd so it works like the rbd one. I do know enough about SCST right now. > Is it safe to use active/active multipath If use suse kernel with > target_core_rbd? > Thanks. > > 2018-03-07 > > shadowlin > > > > *发件人:*Mike Christie > *发送时间:*2018-03-07 03:51 > *主题:*Re: [ceph-users] iSCSI Multipath (Load Balancing) vs RBD > Exclusive Lock > *收件人:*"Lazuardi Nasution" ,"Ceph > Users" > *抄送:* > > On 03/06/2018 01:17 PM, Lazuardi Nasution wrote: > > Hi, > > > > I want to do load balanced multipathing (multiple iSCSI > gateway/exporter > > nodes) of iSCSI backed with RBD images. Should I disable exclusive lock > > > feature? What if I don't disable that feature? I'm using TGT (manual > > way) since I get so many CPU stuck error messages when I was using LIO. > > > > > You are using LIO/TGT with krbd right? > > You cannot or shouldn't do active/active multipathing. If you have the > lock enabled then it bounces between paths for each IO and will be slow. > If you do not have it enabled then you can end up with stale IO > overwriting current data. > > > > ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
Re: [ceph-users] [jewel] High fs_apply_latency osds
Hi Chris, The osds are running on arm nodes. Every node has a two core 1.5Ghz arm 32bit cpu and 2G ram and runs 2 osds.Hdd is 10TB and journal colocate with data on the same disk. Drives are half full now,but the problem I described also happened when the hdds are empty. Filesystem is ext4 because I have some problem to run xfs for now. I am trying to better balance the pg distrubition now to see if it can ease the high latency problem. 2018-03-10 shadowlin 发件人:Chris Hoy Poy发送时间:2018-03-10 09:44 主题:Re: [ceph-users] [jewel] High fs_apply_latency osds 收件人:"shadow_lin" ,"ceph-users" 抄送: Hi Shadowlin, Can you describe your hardware ? Cpu/ram/harddrives involved etc Also how your drives are set up? How full are the drives? What filesystem is it? Cheers /chris Sent from my SAMSUNG Galaxy S6 on the Telstra Mobile Network Original message From: shadow_lin Date: 10/3/18 1:41 am (GMT+08:00) To: ceph-users Subject: [ceph-users] [jewel] High fs_apply_latency osds Hi list, During my write test,I find there are always some of the osds have high fs_apply_latency(1k-5kms,2-8times more than others). At first I think it is caused by unbalanced pg distribution, but after I reweight the osds the problem hasn't gone away. I looked into the osds with high latency and find one thing in common.There are alot of read activities on these osds.Result of iostat shows there are 400-500 r/s and 2000-3000 rKB/s on these high latnecy osds while the normal osds have around 100 r/s 300-400 rKB/s. I tried to restart osd daemon with high latency.It did became normal sometime but there will be another high latency osds come out.And the new high latency osd has the same high read activities. What are these read activities for? Is there a way to lower the latency? Thanks 2018-03-10 shadowlin Message protected by MailGuard: e-mail anti-virus, anti-spam and content filtering. http://www.mailguard.com.au Report this message as spam Message protected by MailGuard: e-mail anti-virus, anti-spam and content filtering. http://www.mailguard.com.au ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com