Hi Dat,
We have an entity called 'record' which contain a record id, a table name
and a set of values. When we insert records to our data layer, we index
those records by the id and the values. Indexing is done in a separate
thread. I ll explain how this done. When we insert records to data layer,
Can you post you scenario in detail along with your modification please?
On 14:09, Sun, 14 Jun 2015 Gimantha Bandara wrote:
> Hi Dat,
>
> I can reproduce this behavior even with like 5 records. Is what you
> said the only reason that make this exception occur?
>
> Thanks,
>
> On Sat, Jun 13,
Hi Dat,
I can reproduce this behavior even with like 5 records. Is what you
said the only reason that make this exception occur?
Thanks,
On Sat, Jun 13, 2015 at 5:40 AM, Đạt Cao Mạnh
wrote:
> Hi, the total number of documents in an index of lucene is
> Integer.MAX_VALUE. So using a single
Hi, the total number of documents in an index of lucene is
Integer.MAX_VALUE. So using a single lucene index to index billions
documents is not a proper ways. You should consider using Solr Cloud or
Elasticsearch to index your documents.
On 19:43, Fri, 12 Jun 2015 Gimantha Bandara wrote:
> Hi al
Hi all,
We are using Lucene 4.10.3 for indexing. Recently we changed our
implementation so that we give data batchwise to lucene to index. Earlier
we just query all the data from the data source and index all data at
once. It works well. But the number of entries can be up to billions. So
getting