Well, there's no other way to get reliable performance and SLAs compared to 
traditional storage when what you work with is commodity hardware in a mesh-y 
configuration.
And we do like the idea of killing the traditional storage, right? I think 80s 
called already and wanted their SAN back...

Jan

> On 27 Aug 2015, at 21:01, Robert LeBlanc <[email protected]> wrote:
> 
> -----BEGIN PGP SIGNED MESSAGE-----
> Hash: SHA256
> 
> I know writing to min_size as sync and size-min_size as async has been
> discussed before and would help here. From what I understand required
> a lot of code changes and goes against the strong consistency model of
> Ceph. I'm not sure if it will be implemented although I do love this
> idea to help against tail latency.
> - ----------------
> Robert LeBlanc
> PGP Fingerprint 79A2 9CA4 6CC4 45DD A904  C70E E654 3BB2 FA62 B9F1
> 
> 
> On Thu, Aug 27, 2015 at 12:48 PM, Jan Schermer  wrote:
>> Don't kick out the node, just deal with it gracefully and without 
>> interruption... if the IO reached the quorum number of OSDs then there's no 
>> need to block anymore, just queue it. Reads can be mirrored or retried (much 
>> quicker, because making writes idempotent, ordered and async is pretty hard 
>> and expensive).
>> If there's an easy way to detect unreliable OSD that flaps - great, let's 
>> have a warning in ceph health.
>> 
>> Jan
>> 
>>> On 27 Aug 2015, at 20:43, Robert LeBlanc  wrote:
>>> 
>>> -----BEGIN PGP SIGNED MESSAGE-----
>>> Hash: SHA256
>>> 
>>> This has been discussed a few times. The consensus seems to be to make
>>> sure error rates of NICs or other such metrics are included in your
>>> monitoring solution. It would also be good to preform periodic network
>>> tests like a full size ping with nofrag set between all nodes and have
>>> your monitoring solution report that as well.
>>> 
>>> Although I would like to see such a feature in Ceph, the concern is
>>> that such a feature can quickly get out of hand and that something
>>> else that is really designed for it should do it. I can understand
>>> where they are coming from in that regard, but having Ceph kick out a
>>> misbehaving node quickly is appealing as well (there would have to be
>>> a way to specify that only so many nodes could be kicked out).
>>> - ----------------
>>> Robert LeBlanc
>>> PGP Fingerprint 79A2 9CA4 6CC4 45DD A904  C70E E654 3BB2 FA62 B9F1
>>> 
>>> 
>>> On Thu, Aug 27, 2015 at 9:37 AM, Christoph Adomeit  wrote:
>>>> Hello Ceph Users,
>>>> 
>>>> yesterday I had a defective Gbic in 1 node of my 10 node ceph cluster.
>>>> 
>>>> The Gbic was working somehow but had 50% packet-loss. Some packets went 
>>>> through, some did not.
>>>> 
>>>> What happend that the whole cluster did not service requests in time, 
>>>> there were lots of timeouts and so on
>>>> until the problem was isolated. Monitors and osds where asked for data but 
>>>> did dot answer or answer late.
>>>> 
>>>> I am wondering, here we have a highly redundant network setup and a highly 
>>>> redundant piece of software, but a small
>>>> network fault brings down the whole cluster.
>>>> 
>>>> Is there anything that can be configured or changed in ceph so that 
>>>> availability will become better in case of flapping networks ?
>>>> 
>>>> I understand, it is not a ceph problem but a network problem but maybe 
>>>> something can be learned from such incidents  ?
>>>> 
>>>> Thanks
>>>> Christoph
>>>> --
>>>> Christoph Adomeit
>>>> GATWORKS GmbH
>>>> Reststrauch 191
>>>> 41199 Moenchengladbach
>>>> Sitz: Moenchengladbach
>>>> Amtsgericht Moenchengladbach, HRB 6303
>>>> Geschaeftsfuehrer:
>>>> Christoph Adomeit, Hans Wilhelm Terstappen
>>>> 
>>>> [email protected]     Internetloesungen vom Feinsten
>>>> Fon. +49 2166 9149-32                      Fax. +49 2166 9149-10
>>>> _______________________________________________
>>>> ceph-users mailing list
>>>> [email protected]
>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>> 
>>> -----BEGIN PGP SIGNATURE-----
>>> Version: Mailvelope v1.0.2
>>> Comment: https://www.mailvelope.com
>>> 
>>> wsFcBAEBCAAQBQJV31pFCRDmVDuy+mK58QAA7qwQAL0EvbHneC00qhCX/jjT
>>> Xl8whWvQgm/UUDEPAWe2wGkgVZtP3cSAx/p+IkusZuD6NClIiWvazdz5n+vf
>>> cj4Y+S8Zj4Lw7gypHjy5GSCDSbQnEni32QNKp74GM/EZ1331gXuDvP0bS2Sz
>>> 7g5MXu8Vpf0Kdrj8JrOPnHY1PtljxkQXdrEmijDkmnjruO+XGFQrl8l9GFbN
>>> enFZI+PpEAoSEJPZosCnX+ZLM3/ZiwAfAPtvcARyDwdmjV7CjyRjVviloR3K
>>> DV/b+VuWX+NVzTZMKCnILVubt1Khexzk6reU3m7Yjy713dmEehDmKQsESFci
>>> pMi61iEuxje0O+iqOp+mhhYWtv+Iv7bbpHcGv04vfMsl6+ms6v/EHo/Cccoi
>>> ZiOa+xD6l7ZkO+A+2bvunBvC3cjBFXn8yrNpHDj6G+jUWMDuJcs7wAhExhPv
>>> Qicjhzk9AoTFXPIkfkGnuHJ/ngFnswdHeVa1DU7GV+Evh/2BCtoHH7Ur+XQY
>>> u7gL6LXt+2UAB3+ZIEvr2NOAFiIVsPqnGqQqNiNz5XQDFh5bD3e1iScucZbm
>>> VNStBkWDoDwrBYVe74cN55ZXA5auTSDYuYlen+BPbYhAKmpkBp+Suv1H4CFy
>>> 01cnANvJfbaxoBIPLzvhdx4c73Qd+J6ttxi2g8u8EedXDbPIYGFPy2madvtW
>>> JNPc
>>> =3sV8
>>> -----END PGP SIGNATURE-----
>>> _______________________________________________
>>> ceph-users mailing list
>>> [email protected]
>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>> 
> 
> -----BEGIN PGP SIGNATURE-----
> Version: Mailvelope v1.0.2
> Comment: https://www.mailvelope.com
> 
> wsFcBAEBCAAQBQJV316iCRDmVDuy+mK58QAAxogP+QGDVhfGxa4OeIslEsoj
> aW3LY4nzzFP1iDJNjlvPDDTj5AcC56c2QhvshLRy3pYUmwoWqO0gnTOGh/YX
> ma6+hGVJCaBZU5L6rZ0SQfcfo3CNglIzQ2ts07Xb5XPQRrS6/yLsMki+kDf0
> qjCqpZpPTL/d80sBrbCNDoZcnMKBYKwBZbay8RsSBZ0pHmdylfnhvGSxZBEk
> U8ZTrUdsZ9ejzdfh29byR3V/Mz6EkGVnnFPlkIAdkuZJPns+i6NGoe16z3kL
> 3u967qFfFcNGUWCGO0MC/iYT4fcRoramqMWhY5hBUD8DmWXgbKQmutal8vaO
> sDBfKLgmQkBpDgOTng6/uE/BpkROmjsjXuCar/xf+QwXMQJhIWFmBqcaKdQl
> TUjd4QovBJVFPWq9qpyh9ia95cfoFm942LaunaA4chTQnxjTbS/0aajSTM7/
> OxUuMcPCnuAsbHXsj/wkPE1ZTmNU9KPQgo10h8UlhMoQU2fOZB8h/p/0fuqk
> bfBo9k07EkdakFFc/ASLpFqIeV49ZTYjUg/0MPdXW5KnJPb+4OBojIZZF9An
> /20UgUXqqBd2LKFY/bwKqoOw5bKCMxQCJptJXmY6zx7vJ76ahyOuiP/OaQWh
> 8uikTxCUeNdaNzez3s3TUgQSR4y94zT1It5rF16VrFPMik66Yq8x11t6z8eR
> Po34
> =tqy/
> -----END PGP SIGNATURE-----

_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to