On 13/09/2018 08:56, Fam Zheng wrote: >> + /* No need to order poll_disable_cnt writes against other updates; >> + * the counter is only used to avoid wasting time and latency on >> + * iterated polling when the system call will be ultimately necessary. >> + * Changing handlers is a rare event, and a little wasted polling until >> + * the aio_notify below is not an issue. >> + */ >> + atomic_set(&ctx->poll_disable_cnt, >> + atomic_read(&ctx->poll_disable_cnt) + poll_disable_change); > > Why not atomic_add?
This is not lockless, it's protected by list_lock, so there's no race condition involved. I'm just mimicking what is done for other similar cases, for example involving seqlocks. The alternative would be to add a full set of atomic_{add,sub,...}_relaxed atomics. Paolo