So, if I understand things correctly - you do send#1, you don't get response#1, so you do send#2. Since the server is NOT out-of-order, it will respond with response#1 then response#2.

The only thing you're going to save is part of the time to parse response#1 - you're STILL going to have to read all of response#1 and parse it enough to find the message boundary.

Do I understand something wrong? Is the overhead of doing a full parse vs. a skip of a message that high?

If the cost of parsing messages constitutes a significant overhead, have you considered using cthrift? It should have a lower overhead for the parsing, perhaps enough to solve your problems. If want to, I'll be more than happy to tailor cthrift to your needs and work with you on any stability issues.

Mayan

Daniel Kluesing wrote:

I'm a little curious why you would expect to get repsonses out-of-order.


It's case where I might bail out after calling the send_ but before calling 
recv_,
> and I don't want to tear-down/reestablish the TCP connection before I send another > request, so the next call needs to know to ignore. (Basically, some tight time
> constraints where it's better to not have that data than to be slow)

I think the split calls is good enough for me.  An aside, if you ever want a
> tester for your CFG stuff, feel free to ping me off list. Love to take a peak.

-----Original Message-----
From: Mayan Moudgill [mailto:[email protected]] Sent: Wednesday, March 17, 2010 10:39 AM
To: [email protected]; Daniel Kluesing
Subject: Re: async c++ client / THRIFT-1

I'm a little curious why you would expect to get repsonses out-of-order. If you're using TCP/IP (or some other inherently FIFO transport) + a FIFO server, then RPC responses will appear in the order in which the send were sent. Why would you need to reorder cseqids? Are you planning on using UDP or something else? Or do you want a non-order-preserving server? Or are you communicating with multiple servers? In the last case, I'm assuming you'll have different sockets for each server - but since the communication with each server will be FIFO, you still won't need to reorder cseqids.

As for recv being blocking:
1. You can use an O_NONBLOCK socket (maybe not in TSocket.py, but cthrift uses O_NONBLOCK sockets). 2. But that doesn't solve the problem if you're talking to multiple sockets and the size of response is > PIPE_BUF, then it is possible for a RPC response to be split up into multiple packets. So, you'll read the data from the first packet, then get blocked waiting for the second packet to come in - UNLESS you can set up things so that when you get blocked, you switch to a different context (thread/coroutine) to process any other data.

I can't speak for the Thrift team, but I don't think splitting a client call into a send part and a receive part is enough to make a fully non-blocking client, and they may be trying to figure out how to get there.

I'm working on some ideas for non-coroutine/context switch based, partial progress receivers [they involve treating the RPC response specification as a CFG {which it is} derive a table driven parser, and then save the intermediate parse state on a block] - these are for the server, but it should be trivial to add to the client. However, I will only be implementing them in cthrift, not Thrift.

Mayan

Daniel Kluesing wrote:

Well, what I'm really curious about is a fully non-blocking c++ client.

 > What I'm doing now is sending the send_'s, setting cseqids myself, and
 > then sorting out the responses based on the rseqid, that's ok, but the
 > recv call in TSocket is still a blocking call. Since it's been an open
 > issue for so long, I'm curious if there are others working on it, or
 > deep seeded reasons why it's not been done that I just don't know about.

-----Original Message-----
From: Mayan Moudgill [mailto:[email protected]] Sent: Tuesday, March 16, 2010 7:41 PM
To: [email protected]; Daniel Kluesing
Subject: Re: async c++ client / THRIFT-1

I'm guessing you want the abilitiy to separate the calls and receives on a client; is that correct?

Thanks

Mayan

Daniel Kluesing wrote:



I wanted/needed  something approximating async requests on the client - 
directly call send_ stuff, do unrelated stuff that might fail, directly call 
recv_ stuff - so I hacked into the c++ generator on 0.2.0 a bit more support 
for cseqid and rseqid. I did this somewhat quickly and while it works for my 
purposes, it isn't the 'right' thing to do. Is there any news on 
http://issues.apache.org/jira/browse/THRIFT-1 or anyone working on a proper 
nonblocking c++ client? (I saw the spotify ASIO patch)

(and if anyone out there wants similar functionality, I can port my changes to 
trunk and make a patch, I didn't bother since this has lain dormant for quite a 
while, so I'm guessing interest is low)









Reply via email to