Re: Seek past EOF

2011-12-01 Thread Ruben Chadien
We are using ext3 on Debian.

Noticed today that i only need to reload the core to get it working again….


On 30 November 2011 19:59, Simon Willnauer
wrote:

> can you give us some details about what filesystem you are using?
>
> simon
>
> On Wed, Nov 30, 2011 at 3:07 PM, Ruben Chadien 
> wrote:
> > Happened again….
> >
> > I got 3 directories in my index dir
> >
> > 4096 Nov  4 09:31 index.2004083156
> > 4096 Nov 21 10:04 index.2021090440
> > 4096 Nov 30 14:55 index.2029024919
> >
> > as you can se the first two are old and also empty , the last one from
> > today is and containing 9 files none of the are 0 size
> > and total size 7 GB. The size of the index on the master is 14GB.
> >
> > Any ideas on what to look for ?
> >
> > Thanks
> > Ruben Chadien
> >
> >
> >
> >
> > On 29 November 2011 15:58, Mark Miller  wrote:
> >
> >> Hmm...I've seen a bug like this, but I don't think it would be tickled
> if
> >> you are replicating config files...
> >>
> >> It def looks related though ... I'll try to dig around.
> >>
> >> Next time it happens, take a look on the slave for 0 size files - also
> if
> >> the index dir on the slave is plain 'index' or has a timestamp as part
> of
> >> the name (eg timestamp.index).
> >>
> >> On Tue, Nov 29, 2011 at 9:53 AM, Ruben Chadien <
> ruben.chad...@aspiro.com
> >> >wrote:
> >>
> >> > Hi, for the moment there are no 0 sized files, but all indexes are
> >> working
> >> > now. I will have to look next time it breaks.
> >> > Yes, the directory name is "index" and it replicates the schema and a
> >> > synonyms file.
> >> >
> >> > /Ruben Chadien
> >> >
> >> > On 29 November 2011 15:29, Mark Miller  wrote:
> >> >
> >> > > Also, on your master, what is the name of the index directory? Just
> >> > > 'index'?
> >> > >
> >> > > And are you replicating config files as well or no?
> >> > >
> >> > >
> >> > > On Nov 29, 2011, at 9:23 AM, Mark Miller wrote:
> >> > >
> >> > > > Does the problem index have any 0 size files in it?
> >> > > >
> >> > > > On Nov 29, 2011, at 2:54 AM, Ruben Chadien wrote:
> >> > > >
> >> > > >> HI all
> >> > > >>
> >> > > >> After upgrading tol Solr 3.4 we are having trouble with the
> >> > replication.
> >> > > >> The setup is one indexing master with a few slaves that replicate
> >> the
> >> > > >> indexes once every night.
> >> > > >> The largest index is 20 GB and the master and slaves are on the
> same
> >> > > DMZ.
> >> > > >>
> >> > > >> Almost every night one of the indexes (17 in total) fail after
> the
> >> > > >> replication with an EOF file.
> >> > > >>
> >> > > >> SEVERE: Error during auto-warming of
> >> > > >> key:org.apache.solr.search.QueryResultKey@bda006e3
> >> > :java.io.IOException:
> >> > > >> seek past EOF
> >> > > >> at
> >> > > >>
> >> > >
> >> >
> >>
> org.apache.lucene.store.MMapDirectory$MMapIndexInput.seek(MMapDirectory.java:347)
> >> > > >> at
> >> > >
> org.apache.lucene.index.SegmentTermEnum.seek(SegmentTermEnum.java:114)
> >> > > >> at
> >> > > >>
> >> > >
> >> >
> >>
> org.apache.lucene.index.TermInfosReader.seekEnum(TermInfosReader.java:203)
> >> > > >> at
> >> > org.apache.lucene.index.TermInfosReader.get(TermInfosReader.java:273)
> >> > > >> at
> >> > org.apache.lucene.index.TermInfosReader.get(TermInfosReader.java:210)
> >> > > >> at
> >> > org.apache.lucene.index.SegmentReader.docFreq(SegmentReader.java:507)
> >> > > >> at
> >> > >
> >> org.apache.solr.search.SolrIndexReader.docFreq(SolrIndexReader.java:309)
> >> > > >> at
> >> > >
> org.apache.lucene.search.TermQuery$TermWeight$1.add(TermQuery.java:56)
> >> > > >> at
> org.apache.lucene.util.ReaderUtil$Gather.run(Reader

Re: Seek past EOF

2011-11-30 Thread Ruben Chadien
Happened again….

I got 3 directories in my index dir

4096 Nov  4 09:31 index.2004083156
4096 Nov 21 10:04 index.2021090440
4096 Nov 30 14:55 index.2029024919

as you can se the first two are old and also empty , the last one from
today is and containing 9 files none of the are 0 size
and total size 7 GB. The size of the index on the master is 14GB.

Any ideas on what to look for ?

Thanks
Ruben Chadien




On 29 November 2011 15:58, Mark Miller  wrote:

> Hmm...I've seen a bug like this, but I don't think it would be tickled if
> you are replicating config files...
>
> It def looks related though ... I'll try to dig around.
>
> Next time it happens, take a look on the slave for 0 size files - also if
> the index dir on the slave is plain 'index' or has a timestamp as part of
> the name (eg timestamp.index).
>
> On Tue, Nov 29, 2011 at 9:53 AM, Ruben Chadien  >wrote:
>
> > Hi, for the moment there are no 0 sized files, but all indexes are
> working
> > now. I will have to look next time it breaks.
> > Yes, the directory name is "index" and it replicates the schema and a
> > synonyms file.
> >
> > /Ruben Chadien
> >
> > On 29 November 2011 15:29, Mark Miller  wrote:
> >
> > > Also, on your master, what is the name of the index directory? Just
> > > 'index'?
> > >
> > > And are you replicating config files as well or no?
> > >
> > >
> > > On Nov 29, 2011, at 9:23 AM, Mark Miller wrote:
> > >
> > > > Does the problem index have any 0 size files in it?
> > > >
> > > > On Nov 29, 2011, at 2:54 AM, Ruben Chadien wrote:
> > > >
> > > >> HI all
> > > >>
> > > >> After upgrading tol Solr 3.4 we are having trouble with the
> > replication.
> > > >> The setup is one indexing master with a few slaves that replicate
> the
> > > >> indexes once every night.
> > > >> The largest index is 20 GB and the master and slaves are on the same
> > > DMZ.
> > > >>
> > > >> Almost every night one of the indexes (17 in total) fail after the
> > > >> replication with an EOF file.
> > > >>
> > > >> SEVERE: Error during auto-warming of
> > > >> key:org.apache.solr.search.QueryResultKey@bda006e3
> > :java.io.IOException:
> > > >> seek past EOF
> > > >> at
> > > >>
> > >
> >
> org.apache.lucene.store.MMapDirectory$MMapIndexInput.seek(MMapDirectory.java:347)
> > > >> at
> > > org.apache.lucene.index.SegmentTermEnum.seek(SegmentTermEnum.java:114)
> > > >> at
> > > >>
> > >
> >
> org.apache.lucene.index.TermInfosReader.seekEnum(TermInfosReader.java:203)
> > > >> at
> > org.apache.lucene.index.TermInfosReader.get(TermInfosReader.java:273)
> > > >> at
> > org.apache.lucene.index.TermInfosReader.get(TermInfosReader.java:210)
> > > >> at
> > org.apache.lucene.index.SegmentReader.docFreq(SegmentReader.java:507)
> > > >> at
> > >
> org.apache.solr.search.SolrIndexReader.docFreq(SolrIndexReader.java:309)
> > > >> at
> > > org.apache.lucene.search.TermQuery$TermWeight$1.add(TermQuery.java:56)
> > > >> at org.apache.lucene.util.ReaderUtil$Gather.run(ReaderUtil.java:77)
> > > >> at org.apache.lucene.util.ReaderUtil$Gather.run(ReaderUtil.java:82)
> > > >>
> > > >>
> > > >> After a restart the errors are gone, anyone else seen this ?
> > > >>
> > > >> Thanks
> > > >> Ruben Chadien
> > > >
> > > > - Mark Miller
> > > > lucidimagination.com
> > > >
> > > >
> > > >
> > > >
> > > >
> > > >
> > > >
> > > >
> > > >
> > > >
> > > >
> > >
> > > - Mark Miller
> > > lucidimagination.com
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> >
> >
> > --
> > *Ruben Chadien
> > *Senior Developer
> > Mobile +47 900 35 371
> > ruben.chad...@aspiro.com
> > *
> >
> > Aspiro Music AS*
> > Øvre Slottsgate 25, P.O. Box 8710 Youngstorget, N-0028 Oslo
> > Tel +47 452 86 900, fax +47 22 37 36 59
> > www.aspiro.com/music
> >
>
>
>
> --
> - Mark
>
> http://www.lucidimagination.com
>



-- 
*Ruben Chadien
*Senior Developer
Mobile +47 900 35 371
ruben.chad...@aspiro.com
*

Aspiro Music AS*
Øvre Slottsgate 25, P.O. Box 8710 Youngstorget, N-0028 Oslo
Tel +47 452 86 900, fax +47 22 37 36 59
www.aspiro.com/music


Re: Seek past EOF

2011-11-29 Thread Ruben Chadien
Hi, for the moment there are no 0 sized files, but all indexes are working
now. I will have to look next time it breaks.
Yes, the directory name is "index" and it replicates the schema and a
synonyms file.

/Ruben Chadien

On 29 November 2011 15:29, Mark Miller  wrote:

> Also, on your master, what is the name of the index directory? Just
> 'index'?
>
> And are you replicating config files as well or no?
>
>
> On Nov 29, 2011, at 9:23 AM, Mark Miller wrote:
>
> > Does the problem index have any 0 size files in it?
> >
> > On Nov 29, 2011, at 2:54 AM, Ruben Chadien wrote:
> >
> >> HI all
> >>
> >> After upgrading tol Solr 3.4 we are having trouble with the replication.
> >> The setup is one indexing master with a few slaves that replicate the
> >> indexes once every night.
> >> The largest index is 20 GB and the master and slaves are on the same
> DMZ.
> >>
> >> Almost every night one of the indexes (17 in total) fail after the
> >> replication with an EOF file.
> >>
> >> SEVERE: Error during auto-warming of
> >> key:org.apache.solr.search.QueryResultKey@bda006e3:java.io.IOException:
> >> seek past EOF
> >> at
> >>
> org.apache.lucene.store.MMapDirectory$MMapIndexInput.seek(MMapDirectory.java:347)
> >> at
> org.apache.lucene.index.SegmentTermEnum.seek(SegmentTermEnum.java:114)
> >> at
> >>
> org.apache.lucene.index.TermInfosReader.seekEnum(TermInfosReader.java:203)
> >> at org.apache.lucene.index.TermInfosReader.get(TermInfosReader.java:273)
> >> at org.apache.lucene.index.TermInfosReader.get(TermInfosReader.java:210)
> >> at org.apache.lucene.index.SegmentReader.docFreq(SegmentReader.java:507)
> >> at
> org.apache.solr.search.SolrIndexReader.docFreq(SolrIndexReader.java:309)
> >> at
> org.apache.lucene.search.TermQuery$TermWeight$1.add(TermQuery.java:56)
> >> at org.apache.lucene.util.ReaderUtil$Gather.run(ReaderUtil.java:77)
> >> at org.apache.lucene.util.ReaderUtil$Gather.run(ReaderUtil.java:82)
> >>
> >>
> >> After a restart the errors are gone, anyone else seen this ?
> >>
> >> Thanks
> >> Ruben Chadien
> >
> > - Mark Miller
> > lucidimagination.com
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
>
> - Mark Miller
> lucidimagination.com
>
>
>
>
>
>
>
>
>
>
>
>


-- 
*Ruben Chadien
*Senior Developer
Mobile +47 900 35 371
ruben.chad...@aspiro.com
*

Aspiro Music AS*
Øvre Slottsgate 25, P.O. Box 8710 Youngstorget, N-0028 Oslo
Tel +47 452 86 900, fax +47 22 37 36 59
www.aspiro.com/music


Seek past EOF

2011-11-28 Thread Ruben Chadien
HI all

After upgrading tol Solr 3.4 we are having trouble with the replication.
The setup is one indexing master with a few slaves that replicate the
indexes once every night.
The largest index is 20 GB and the master and slaves are on the same DMZ.

Almost every night one of the indexes (17 in total) fail after the
replication with an EOF file.

SEVERE: Error during auto-warming of
key:org.apache.solr.search.QueryResultKey@bda006e3:java.io.IOException:
seek past EOF
 at
org.apache.lucene.store.MMapDirectory$MMapIndexInput.seek(MMapDirectory.java:347)
at org.apache.lucene.index.SegmentTermEnum.seek(SegmentTermEnum.java:114)
 at
org.apache.lucene.index.TermInfosReader.seekEnum(TermInfosReader.java:203)
at org.apache.lucene.index.TermInfosReader.get(TermInfosReader.java:273)
 at org.apache.lucene.index.TermInfosReader.get(TermInfosReader.java:210)
at org.apache.lucene.index.SegmentReader.docFreq(SegmentReader.java:507)
 at org.apache.solr.search.SolrIndexReader.docFreq(SolrIndexReader.java:309)
at org.apache.lucene.search.TermQuery$TermWeight$1.add(TermQuery.java:56)
 at org.apache.lucene.util.ReaderUtil$Gather.run(ReaderUtil.java:77)
at org.apache.lucene.util.ReaderUtil$Gather.run(ReaderUtil.java:82)


After a restart the errors are gone, anyone else seen this ?

Thanks
Ruben Chadien


expungeDeletes on commit in Dataimport

2010-03-25 Thread Ruben Chadien
Hi

I know this has been discussed before, but is there any way do 
expungeDeletes=true when the DataImportHandler does the commit.
I am using the deleteDocByQuery in a Transformer when doing a delta-import and 
as discussed before the documents are not deleted until restart.

Also, how do i know in a Transformer if its running a Delta or Full Import , i 
tries looking at Context. currentProcess() but that gives me "FULL_DUMP" when 
doing a delta import...?

Thanks!
Ruben Chadien

Re: DataImport, remove doc when marked as deleted

2009-04-16 Thread Ruben Chadien

I have now :-)
Thanks , missed that in the Wiki.
Ruben

On Apr 16, 2009, at 7:10 PM, Noble Paul നോബിള്‍  
नोब्ळ् wrote:



did you try the deletedPkQuery?

On Thu, Apr 16, 2009 at 7:49 PM, Ruben Chadien > wrote:

Hi

I am new to Solr, but have been using Lucene for a while. I am  
trying to

rewrite
some old lucene indexing code using the Jdbc DataImport i Solr, my  
problem:


I have Entities that can be marked in the db as "deleted", these i  
don't

want to index
and thats no problem when doing a full-import. When doing a delta- 
import my

deltaQuery will catch
Entities that has been marked as deleted since last index, but how  
do i get

it to delete those from the index ?
I tried making the deltaImportQuery so that in don't return the  
Entity if

its deleted, that didnt help...

Any ideas ?

Thanks
Ruben







--
--Noble Paul




DataImport, remove doc when marked as deleted

2009-04-16 Thread Ruben Chadien

Hi

I am new to Solr, but have been using Lucene for a while. I am trying  
to rewrite
some old lucene indexing code using the Jdbc DataImport i Solr, my  
problem:


I have Entities that can be marked in the db as "deleted", these i  
don't want to index
and thats no problem when doing a full-import. When doing a delta- 
import my deltaQuery will catch
Entities that has been marked as deleted since last index, but how do  
i get it to delete those from the index ?
I tried making the deltaImportQuery so that in don't return the Entity  
if its deleted, that didnt help...


Any ideas ?

Thanks
Ruben