I have been doing some latency measurement with different modes of grpc. 
The application that I have is time critical, and I need to ensure that 
requests are being completed in sub millisecond delay. In my test runs, I 
noticed that the average latency that I get for number of calls is around 
300us, but there are number of request that are completed around 10 ms 
which are not acceptable. 
I have been trying to find a way to optimize for latency, and it seems to 
me the source of this jitter is batching that it is done in grpc. Hence, I 
found out in streaming mode you can give WriteOptions().set_write_through() 
when 
making the write call in order to send the packet instantly. But it didn't 
really help, I can still see that packets are sent in batches.

Is set_write_through() the right option to use? or there exist a better way 
to achieve this?
 

-- 
You received this message because you are subscribed to the Google Groups 
"grpc.io" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to grpc-io+unsubscr...@googlegroups.com.
To post to this group, send email to grpc-io@googlegroups.com.
Visit this group at https://groups.google.com/group/grpc-io.
To view this discussion on the web visit 
https://groups.google.com/d/msgid/grpc-io/7d2cc8b9-28b0-4d08-acc7-571bbb11c730%40googlegroups.com.
For more options, visit https://groups.google.com/d/optout.

Reply via email to