On Jun 1, 2008, at 10:53 PM, Cloud Zhang [EMAIL PROTECTED] wrote:
Thank a lot for this very detailed guide, I'll forward this to
Chinese Python community, since the first thing a Chinese developer
looking for about Lucene is a tokenizer for Chinese and get stuck
with importing a jar...
One thing I noticed is that your package contains _jcc.so for Python 2.4
and Python 2.5. Is that intentionally?
No, it came for free from the build infrastructure.
That's because the linker can't find libjvm.so (which should be part of Sun's
JDK). You
can set an rpath in setup.py but at least
Jeff Breidenbach schrieb:
That's because the linker can't find libjvm.so (which should be part of Sun's
JDK). You
can set an rpath in setup.py but at least for Fedora this may not be acceptable
[...]
Thanks; that was it. I'll try adding an entry in /etc/ld.so.conf.d
because Debian disallows
On Jun 1, 2008, at 11:35 PM, Cloud Zhang [EMAIL PROTECTED] wrote:
Sure, there are two Chinese analyzer (including the CJKAnalyzer)
bundled with Lucene. But both are character based and far from
acceptable.
A practical Chinese tokenizer should know Chinese words (with one or
several
I've placed an Ubuntu PyLucene package online at just now. I'm also in
progress submitting a package for inclusion with Debian. But that will
almost certainly require some time and back and forth - the last time
I did this, the package was rejected for quality reasons. This
represents some