Hi,
I have an webcluster setup, where on the loadbalancers the persistence
timeout is 0. To share the sessions I use ceph version 0.56.7, like you
see on the diagram.
+----------------+
| Internet |
+----------------+
|
+-----+-----------------------+
| |
+-----------------+ +-----------------+
| loadbalancer-01 | | loadbalancer-02 |
+-----------------+ +-----------------+
| |
+-----+--192.168.1.0/24-------+
|
+--------+ | +--------+
| web-01 |---+-----| web-02 |
+--------+ | +--------+
|
+--------+ | +--------+
| osd-01 |---+-----| osd-02 |
+--------+ | +--------+
|
+--------+ | +--------+
| mds-01 |---+-----| mds-02 |
+--------+ | +--------+
|
+--------+--------+-------------------+
| | |
+--------+ +--------+ +--------+
| mon-01 | | mon-02 | | mon-03 |
+--------+ +--------+ +--------+
I mount on the web nodes the ceph mds tables:
** /etc/fstab **
mon-01:/ < Session_mountpoint> ceph
defaults,name=admin,secret=<secret_key> 0 0
My probem is that when the sessions gets a frequent update, I sometimes
loss my session data.
I can reproduse my problem with the following PHP script:
<?php
// page2.php
session_save_path('/var/www/storage/sessions/');
session_start();
$_SESSION['count']++;
echo 'count: ';
echo $_SESSION['count'];
?>
When i run the following commands:
michielp@michielp-hp:~$ wget --no-check-certificate
--keep-session-cookies --save-cookies /tmp/cookies.txt
https://sub.domain.nl/page2.php -O -
michielp@michielp-hp:~$ for foo in {1..10000}; do wget
--no-check-certificate --load-cookies /tmp/cookies.txt
"https://sub.domain.nl/page2.php" -O - -o /dev/null; sleep 0.3; done
At 10, 100 and 1000 and further the couter hits to 0. When i use sleep
0.4 everthing works fine.
michielp@michielp-hp:~$ for foo in {1..10000}; do wget
--no-check-certificate --load-cookies /tmp/cookies.txt
"https://sub.domain.nl/page2.php" -O - -o /dev/null; done
count: 1
count: 2
count: 3
count: 4
count: 5
count: 6
count: 7
count: 8
count: 9
count: 10
count: 1
count: 2
count: 1
Also when is switch off one of the webservers the problem disappears.
On mds-01 is see the folllowing message:
root@isp-oscaccstormds-01:/var/log/ceph# tail ceph-mds.5.log
2013-11-13 20:49:00.428592 7f20fca22700 0 mds.0.server
handle_client_file_setlock: start: 0, length: 0, client: 18900, pid:
10032, type: 4
The config of the ceph cluster looks like:
[global]
auth cluster required = cephx
auth service required = cephx
auth client required = cephx
keyring = /etc/ceph/keyring
cluster network = 192.168.1.0/24
public network = 192.168.1.0/24
fsid = 82ecbd50-81ff-4f6c-a009-0bd02a1b4043
[mon]
mon data = /var/lib/ceph/mon/mon.$id
[osd]
osd journal size = 4096
filestore flusher = false
osd data = /var/lib/ceph/osd/osd.$id
osd journal = /var/lib/ceph/osd/osd.$id/journal
osd mkfs type = xfs
keyring = /var/lib/ceph/osd/osd.$id/keyring
[mds]
mds data = /var/lib/ceph/mds/mds.$id
keyring = /var/lib/ceph/mds/mds.$id/keyring
[mon.0]
host = mon-01
mon addr = 192.168.1.56:6789
[mon.1]
host = mon-02
mon addr = 192.168.1.57:6789
[mon.2]
host = mon-03
mon addr = 192.168.1.58:6789
[mds.5]
host = mds-01
[mds.6]
host = mds-02
[osd.0]
host = osd-02
devs = /dev/sdb1
cluster addr = 192.168.1.60
public addr = 192.168.1.60
[osd.1]
host = osd-01
devs = /dev/sdb1
cluster addr = 192.168.1.59
public addr = 192.168.1.59
Kinds regards,
Michiel Piscaer
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com