Hi Paul, 

Thanks for your reply, I am running 3.10.0-957.12.2.el7.x86_64, it is 
from may 2019.



-----Original Message-----
From: Paul Emmerich [mailto:paul.emmer...@croit.io] 
Sent: vrijdag 12 juli 2019 12:34
To: Marc Roos
Cc: ceph-users
Subject: Re: [ceph-users] "session established", "io error", "session 
lost, hunting for new mon" solution/fix



On Thu, Jul 11, 2019 at 11:36 PM Marc Roos <m.r...@f1-outsourcing.eu> 
wrote:


         
        
        Anyone know why I would get these? Is it not strange to get them in 
a 
        'standard' setup?
        


you are probably running on an ancient kernel. this bug has been fixed a 
long time ago.


Paul

 






        -----Original Message-----
        Subject: [ceph-users] "session established", "io error", "session 
lost, 
        hunting for new mon" solution/fix
        
        
        I have on a cephfs client again (luminous cluster, centos7, only 32 

        osds!). Wanted to share the 'fix'
        
        [Thu Jul 11 12:16:09 2019] libceph: mon0 192.168.10.111:6789 
session 
        established
        [Thu Jul 11 12:16:09 2019] libceph: mon0 192.168.10.111:6789 io 
error
        [Thu Jul 11 12:16:09 2019] libceph: mon0 192.168.10.111:6789 
session 
        lost, hunting for new mon
        [Thu Jul 11 12:16:09 2019] libceph: mon2 192.168.10.113:6789 
session 
        established
        [Thu Jul 11 12:16:09 2019] libceph: mon2 192.168.10.113:6789 io 
error
        [Thu Jul 11 12:16:09 2019] libceph: mon2 192.168.10.113:6789 
session 
        lost, hunting for new mon
        [Thu Jul 11 12:16:09 2019] libceph: mon0 192.168.10.111:6789 
session 
        established
        [Thu Jul 11 12:16:09 2019] libceph: mon0 192.168.10.111:6789 io 
error
        [Thu Jul 11 12:16:09 2019] libceph: mon0 192.168.10.111:6789 
session 
        lost, hunting for new mon
        [Thu Jul 11 12:16:09 2019] libceph: mon1 192.168.10.112:6789 
session 
        established
        [Thu Jul 11 12:16:09 2019] libceph: mon1 192.168.10.112:6789 io 
error
        [Thu Jul 11 12:16:09 2019] libceph: mon1 192.168.10.112:6789 
session 
        lost, hunting for new mon
        
        1) I blocked client access to the monitors with
        iptables -I INPUT -p tcp -s 192.168.10.43 --dport 6789 -j REJECT
        Resulting in 
        
        [Thu Jul 11 12:34:16 2019] libceph: mon1 192.168.10.112:6789 socket 

        closed (con state CONNECTING)
        [Thu Jul 11 12:34:18 2019] libceph: mon1 192.168.10.112:6789 socket 

        closed (con state CONNECTING)
        [Thu Jul 11 12:34:22 2019] libceph: mon1 192.168.10.112:6789 socket 

        closed (con state CONNECTING)
        [Thu Jul 11 12:34:26 2019] libceph: mon2 192.168.10.113:6789 socket 

        closed (con state CONNECTING)
        [Thu Jul 11 12:34:27 2019] libceph: mon2 192.168.10.113:6789 socket 

        closed (con state CONNECTING)
        [Thu Jul 11 12:34:28 2019] libceph: mon2 192.168.10.113:6789 socket 

        closed (con state CONNECTING)
        [Thu Jul 11 12:34:30 2019] libceph: mon1 192.168.10.112:6789 socket 

        closed (con state CONNECTING)
        [Thu Jul 11 12:34:30 2019] libceph: mon2 192.168.10.113:6789 socket 

        closed (con state CONNECTING)
        [Thu Jul 11 12:34:34 2019] libceph: mon2 192.168.10.113:6789 socket 

        closed (con state CONNECTING)
        [Thu Jul 11 12:34:42 2019] libceph: mon2 192.168.10.113:6789 socket 

        closed (con state CONNECTING)
        [Thu Jul 11 12:34:44 2019] libceph: mon0 192.168.10.111:6789 socket 

        closed (con state CONNECTING)
        [Thu Jul 11 12:34:45 2019] libceph: mon0 192.168.10.111:6789 socket 

        closed (con state CONNECTING)
        [Thu Jul 11 12:34:46 2019] libceph: mon0 192.168.10.111:6789 socket 

        closed (con state CONNECTING)
        
        2) I applied the suggested changes to the osd map message max, 
mentioned 
        
        in early threads[0]
        ceph tell osd.* injectargs '--osd_map_message_max=10'
        ceph tell mon.* injectargs '--osd_map_message_max=10'
        [@c01 ~]# ceph daemon osd.0 config show|grep message_max
            "osd_map_message_max": "10",
        [@c01 ~]# ceph daemon mon.a config show|grep message_max
            "osd_map_message_max": "10",
        
        [0]
        https://www.mail-archive.com/ceph-users@lists.ceph.com/msg54419.htm
l
        http://tracker.ceph.com/issues/38040
        
        3) Allow access to a monitor with
        iptables -D INPUT -p tcp -s 192.168.10.43 --dport 6789 -j REJECT
        
        Getting 
        [Thu Jul 11 12:39:26 2019] libceph: mon0 192.168.10.111:6789 
session 
        established
        [Thu Jul 11 12:39:26 2019] libceph: osd0 down
        [Thu Jul 11 12:39:26 2019] libceph: osd0 up
        
        Problems solved, in D state hung unmount was released. 
        
        I am not sure if the prolonged disconnection to the monitors was 
the 
        solution or the osd_map_message_max=10, or both. 
        
        
        
        
        
        _______________________________________________
        ceph-users mailing list
        ceph-users@lists.ceph.com
        http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
        
        
        _______________________________________________
        ceph-users mailing list
        ceph-users@lists.ceph.com
        http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
        


_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to