Eddie... thanks. Yes, that sounds like I would not have the advantage of DUCC managing the UIMA pipeline.
To break it down a little for the uninitiated (me), 1. how do I start a DUCC job that stays resident because it has high startup cost (e.g. 2 minutes to load all the resources for the UIMA pipeline VS about 2 seconds to process each request)? 2. once I have a resident job, how do I get the Job Driver to iteratively feed references to each next document (as they are received) to the resident Job Process? Because all the input jobs will be archived anyhow, I'm okay with passing them through the file system if needed. Thanks / Dan -----Original Message----- From: Eddie Epstein [mailto:eaepst...@gmail.com] Sent: Thursday, November 20, 2014 6:06 PM To: user@uima.apache.org Subject: Re: DUCC web server interfacing Ooops, in this case the web server would be feeding the service directly. On Thu, Nov 20, 2014 at 9:04 PM, Eddie Epstein <eaepst...@gmail.com> wrote: > The preferred approach is to run the analytics as a DUCC service, and > have an application driver that feeds the service instances with incoming > data. > This service would be a scalable UIMA-AS service, which could have as > many instances as are needed to keep up with the load. The driver > would use the uima-as client API to feed the service. The application > driver could itself be another DUCC service. > > DUCC manages the life cycle of its services, including restarting them > on failure. > > Eddie > > > On Thu, Nov 20, 2014 at 6:45 PM, Daniel Heinze <dhei...@san.rr.com> wrote: > >> I just installed DUCC this week and can process batch jobs. I would >> like DUCC to initiate/manage one or more copies of the same UIMA >> pipeline that has high startup overhead and keep it/them active and >> feed it/them with documents that arrive periodically over a web >> service. Any suggestions on the preferred way (if any) to do this in DUCC. >> >> >> >> Thanks / Dan >> >> >