Hi Dan
On 20/11/13 15:54, Daniel Kulp wrote:

On Nov 19, 2013, at 10:20 PM, Jason Wang <[email protected]> wrote:

Hi all,

I would like to limit the frequencies our APIs can be called. Given that
they will be public APIs.
The limit will most likely be done on IP addresses.

Is there existing mechanism in CXF for this? Otherwise I will create my own
interceptor to do it.

Currently, no.  I had some similar discussions about this with some folks last 
week related more about throttling per endpoint instead of per IP.  However, 
many of the thoughts are the same.   Kind of came up with this list of thoughts 
to think about:

1) What should happen if more than the needed requests come in?  Should they be 
queued and processed later?   Should a fault be thrown?  Should some number be 
queued and then a fault thrown beyond that?   Lots of possible config options 
here.


May be we can ship a couple of basic interceptors which would return 503 if the rate exceeds. One pair of interceptors would go to the core and would simply check how many concurrent requests are under way, another pair will go to the http module and it will rate the individual client IP addresses, the ideas you suggested below can further be explored to support more advanced options

Thanks, Sergey

2) If you want to do this at an endpoint level via an executor, the CXF schemas 
do have an “executor” element for the jaxws:endpoint element that can be used 
to set a specific executor.  There are a couple of “Executor” things that can 
provide limits that may be able to plug right in here.  That said, I’d 
discourage this.   When using an executor, when a request comes in on a Jetty 
(or other transport thread), we have to place the request on the executor and 
then block the transport thread until the request finishes.  Thus, it ties up 
two threads and jetty cannot process more while it’s waiting.   That said, 
there is definitely a possible enhancement here.  If using a transport that 
supports the CXF continuations, we COULD start a continuation prior to flipping 
to the executor.   Something to think about a bit more.

3) Possibly the more “correct” answer to this is that this is a mediation/Camel 
feature, not a service feature.   CXF is about creating/exposing services.   
Placing quality of services requirements around that service is a mediation 
thing.   That could be considered Camel’s job.   This could be a  
from(“jetty://….”).throttle(…).to(“cxf:…”) type thing.  Not sure if the Camel 
throttling has support for per-ip throttling or not.  Would need to investigate 
more.

4) You likely could implement this as a set of CXF interceptors that could use 
the Continuations to “pause” the request for a few milliseconds or similar if 
the load is too high.   Would require some extra coding.   Contributions back 
could be welcome.

5) Jetty (and likely Tomcat and others) do have some throttling control built 
in at the servlet engine level.  You may want to investigate that.   
Additionally, if you run your web service behind an Apache proxy, I believe the 
mod_proxy stuff in Apache has some settings for this.

Anyway, lots of thoughts, but haven’t had time to really look into any of them 
yet.

Reply via email to