Care to share an example? I'm not picturing it.
On Tue, Aug 16, 2011 at 3:23 PM, Taariq Levack <taar...@gmail.com> wrote: > Hi James > > I did that too for what it's worth. > I send the message to a route that forwards to both the aggregator and to the > socket. > When the response comes in I use an enricher to add the ID to the headers and > then forward to the aggregator. > > Taariq > > On 16 Aug 2011, at 8:55 PM, James Carman <ja...@carmanconsulting.com> wrote: > >> Willem, >> >> Thank you for your help. I don't think this is doing exactly what I >> need, though. The real trick here is the asynchronous nature of the >> "server" on the other end of this situation. I thought about using an >> aggregator to make sure the response gets matched up with the request >> using a correlation id. The aggregator wouldn't aggregate multiple >> responses together into one, it would just make sure it matches the >> correct response with its request. Does this sound like a valid >> approach? If so, how the heck do I go about it? :) >> >> Thanks, >> >> James >> >> On Sun, Aug 7, 2011 at 9:03 PM, Willem Jiang <willem.ji...@gmail.com> wrote: >>> Hi James, >>> >>> Camel async process engine already provides the way that you want. >>> You can take a look at the camel-cxf code[1][2] for some example. >>> >>> [1]http://svn.apache.org/viewvc/camel/trunk/components/camel-cxf/src/main/java/org/apache/camel/component/cxf/CxfConsumer.java?view=markup >>> [2]http://svn.apache.org/viewvc/camel/trunk/components/camel-cxf/src/main/java/org/apache/camel/component/cxf/CxfProducer.java?view=markup >>> >>> On 8/7/11 1:29 AM, James Carman wrote: >>>> >>>> On Sat, Aug 6, 2011 at 10:33 AM, Zbarcea Hadrian<hzbar...@gmail.com> >>>> wrote: >>>>> >>>>> Hi James, >>>>> >>>>> I hope I understand your scenario correctly. Here are a few thoughts. I >>>>> assume want to use camel-netty [1] to send messages to your sever (if you >>>>> have your own code that does that, you can use it too, but you'd have to >>>>> write your own Processor or Component). Iiuic, your scenario is >>>>> converting a >>>>> 2x in-only to a 1x in-out async mep. You should then treat your exchange >>>>> as >>>>> an async in-out and let your framework (Camel) decompose it and compose it >>>>> back again. I would not keep threads blocked so I believe your best bet is >>>>> using the Camel async messaging [2] and Futures (look at the examples >>>>> using >>>>> asyncSend* and asyncCallback*). The issue is that Camel is stateless so >>>>> you'll need a correlationId, which you must have already and something to >>>>> keep your state. A good bet would be jms [3], or you could write your own. >>>>> If you used jms you would need to use both a correlationId and a replyTo >>>>> queue. >>>>> >>>>> from("jms:request-queue").to("netty:output?=correlationId"); >>>>> from("netty:input).to("jms:replyTo-queue") >>>>> >>>> >>>> Perhaps a bit more information might be appropriate here. Eventually, >>>> I'd like to "expose" this route via web services (using CXF of >>>> course). So, I would need to either block the request thread, waiting >>>> for a reply or perhaps check out the new Servlet 3.0 asynchronous >>>> processing stuff (I'm thinking this might help us get more done with >>>> less http request threads) to do more of a continuation thing. >>>> >>>> We already have a correlation id. The "protocol" requires one and the >>>> server process just echos it back in the response message. >>>> >>>>> You may have to play a bit with the correlationId and if you cannot use >>>>> the same you can do a second transformation/correlation using a >>>>> claim-check >>>>> sort of pattern. If you don't want to use jms you can implement your own >>>>> (in >>>>> memory) persistence and correlation. You can also use a resequencer [4] if >>>>> you want to enforce the order. If you use asyncCallback, you get the >>>>> replies >>>>> when they become available, and you can control that. >>>>> >>>> >>>> I don't think a resequencer is necessary. I don't want to guarantee >>>> the ordering. I'm mostly interested in throughput here. So, if a >>>> message comes in after another, but it can be processed faster, so be >>>> it. >>>> >>>>> It's an interesting scenario, I'll definitely give it more thought, but I >>>>> hope this helps. >>>>> Hadrian >>>>> >>>> >>>> You have been very helpful. Thank you for taking the time! >>>> >>> >>> >>> -- >>> Willem >>> ---------------------------------- >>> FuseSource >>> Web: http://www.fusesource.com >>> Blog: http://willemjiang.blogspot.com (English) >>> http://jnn.javaeye.com (Chinese) >>> Twitter: willemjiang >>> Weibo: willemjiang >>> >