Paul Barton-Davis wrote:
> 
> The problem with "delivering individual events as fast as it
> can" is demonstrated rather nicely in Quasimodo. Because of the way it
> implements "patches" between modules, external events (GUI stuff, MIDI
> data, etc.) that alter parameter values happen *too* fast. 

This is a good point, but I don't have any data on the response time 
of my external synths (I suppose I could write a little test program).
During the realtime takes, I want messages to go through the 
flowgraph as fast as possible, but during playback I do want the delay
to be taken into effect.  However, I'm already used to fiddling with
offsets manually to acheive the latter (I'm sure most other folks are
too).

> For point-to-point communication, if the data flow rate is not an
> issue (ie. its not audio or video), than a Unix pipe will probably
> work just fine. If you need to multiplex, then use shared memory
> queues. In short, use the existing IPC mechanisms, since what you 
> want is in fact *IPC* not a plugin API.

Thoroughly agreed, but what happens when I want to talk to a program
designed around LADSPA or MuCoS?  Will either of those offer an easy
(out of the box) way to talk to a pipe, socket, or shared memory 
queue?  If they go out of their way to accomodate me, then I don't
have to go out of my way to accomodate them; I just didn't expect
to get off that easily.  :-)

Div.

-- 
David Slomin, Engineer         mailto:[EMAIL PROTECTED]
AltaVista Business Solutions   http://solutions.altavista.com/
RFC 822 plaintext email strongly preferred except for attachments

Reply via email to