Eddie... thanks.  Yes, that sounds like I would not have the advantage of DUCC 
managing the UIMA pipeline. 

To break it down a little for the uninitiated (me), 

 1. how do I start a DUCC job that stays resident because it has high startup 
cost (e.g. 2 minutes to load all the resources for the UIMA pipeline VS about 2 
seconds to process each request)?

2. once I have a resident job, how do I get the Job Driver to iteratively feed 
references to each next document (as they are received) to the resident Job 
Process?  Because all the input jobs will be archived anyhow, I'm okay with 
passing them through the file system if needed.

Thanks / Dan

-----Original Message-----
From: Eddie Epstein [mailto:eaepst...@gmail.com] 
Sent: Thursday, November 20, 2014 6:06 PM
To: user@uima.apache.org
Subject: Re: DUCC web server interfacing

Ooops, in this case the web server would be feeding the service directly.

On Thu, Nov 20, 2014 at 9:04 PM, Eddie Epstein <eaepst...@gmail.com> wrote:

> The preferred approach is to run the analytics as a DUCC service, and 
> have an application driver that feeds the service instances with incoming 
> data.
> This service would be a scalable UIMA-AS service, which could have as 
> many instances as are needed to keep up with the load. The driver 
> would use the uima-as client API to feed the service. The application 
> driver could itself be another DUCC service.
>
> DUCC manages the life cycle of its services, including restarting them 
> on failure.
>
> Eddie
>
>
> On Thu, Nov 20, 2014 at 6:45 PM, Daniel Heinze <dhei...@san.rr.com> wrote:
>
>> I just installed DUCC this week and can process batch jobs.  I would 
>> like DUCC to initiate/manage one or more copies of the same UIMA 
>> pipeline that has high startup overhead and keep it/them active and 
>> feed it/them with documents that arrive periodically over a web 
>> service.  Any suggestions on the preferred way (if any) to do this in DUCC.
>>
>>
>>
>> Thanks / Dan
>>
>>
>


Reply via email to