Hi,

Just some update that might be useful for the upcoming board report.

The work on UIMA CPP is ongoing and with good progress.

A suitable delivery mechanism for releases for the project has been
identified (Docker image) and the hope is to have access to a UIMA CPP
Docker Hub slot whenever the chair files for an INFRA issue for it.

Issues regarding Python multi-interpreter support have been identified and
two possible releases of the Pythonator are available, one fully embedded
and the other with multi-interpreter support. Sadly, NumPy, a popular
numeric processing library (and a requirement to most machine learning
packages in Python) does not support multi-interpreter.

Using the fully embedded library, a proof-of-concept running an annotator
with the popular spaCy NLP library was successfully executed.

Work is still ongoing packaging UIMA CPP to build as a system library
inside the Docker image rather than using custom paths.

I'm currently looking for a good application of UIMA CPP to process large
amounts of text related to LLMs (the topic du jour), to attract more
interest in the new release.

Once this base release is done, the objective is to move to support
aggregates natively in UIMA CPP and to bring back the JNI code that would
allow to run UIMAJ annotators in the pipeline.

Questions? Comments? Would love to hear them.

P

Reply via email to