See Paul's e-mail, he's talking about a place I haven't been in Lucene yet.
Other than that, see below.... On 5/19/07, Peter Bloem <[EMAIL PROTECTED]> wrote:
Ah, now we're getting somewhere. So I run the first query on the collection index, get a set of collection id's from that. But how do I use them in the second query on the document index? It should be easy enough to retrieve all documents in the returned collections (which is what I'm after), but then I want to rank them as if they had the collection's term vector as a field. Is there some way to modify a document just prior to processing?
You could think about boosting on the "collectionid" based on the ranking in the first query. NOTE: I really have no idea whether this would be satisfactory or not, so please post what you find out if you try it. Say your first query returned values like collectionid score 88 0.80 63 0.45 100 0.20 You might then build up a query something like... text:the text from the query AND (collectionID:88^8 OR CollectionID:63^4.5 OR CollectionID:100^2).. (NOTE, not very good Lucene query syntax here<G>). Which *should* help weight the results by collectionID by how "well" the query matched the collection vector. You'll have to delve into the scoring and boosting. For instance, if you get your collection IDs from a HitCollector, the scores will be raw. That is, not between 0.0 and 1.0. Also, you may want to boost the text term in the second query depending upon how you want to bias hits in the actual text of the document. You could also do this manually. That is, you've generated a list of collection IDs and scores anyway. Just keep it around, execute your second query without reference to the collectionIDs (i.e. text: text of query), and then use a HitCollector and massage the scores as you see fit. Think about using a FieldSortedHitQueue to sort the results. If you do this, also think about going directly to your index via TermEnum/TermDocs to get the collection IDs associated with each doc rather than using IndexReader.document(ID). See the warning in HitCollector.collect(). I have several thousand collections, but the number of collections
matching a query should remain quite small. The collections contain about as much text as a small webpage, so the chance that one query matches huge amounts of collections is small. If this does become a problem, I could still store the document id's. My data won't change, so there's no danger of the document id's changing. The end result of the project has to look like a production system, but it doesn't have to be one. :)
I don't see how storing the doc IDs makes the problem better. Building up the OR clause would seem to be worse since there's a many-to-one relationship between the document IDs and collection IDs. Your clause would only get bigger. I can see why using Lucene like a database is worrying. There's already
the problem of referential integrity (what if you update document/collection id's), which databases do well, and Lucene doesn't do at all (as there doesn't seem to be a standard mechanism for this sort of thing). On the other hand, I don't think this technique is very new. I think it's a common smoothing method in xml element retrieval (smoothing an element with the contents of its ancestor elements). So surely this sort of thing gets done a lot. I guess there are bound to be some limits to the inverse index that require less pretty tricks like these.
Which is one of the reasons I mentioned collection IDs. Since you control them, they don't need to change ever. You can freely update their content as you add documents without having to regenerate the entire index. But you're right, referential integrity is not part of Lucene. And I doubt if it ever will be, it's just not part of what Lucene is designed to do. There was a very interesting thread a while ago about embedding Lucene in an Oracle database. Since it's not relevant to my problem space, I barely skimmed it, but if you search the archive for posts authored by Marcelo Ochoa, you'll find them. Best Erick regards,
Peter Erick Erickson wrote: > You're right, your index will bloat considerably. In fact, I'm surprised > it's only a factor of 5.... > > The only thing that comes to mind is really a variant on your approach > from your first e-mail. But I wouldn't use document ids because document > IDs can change. So using doc IDs is...er.... fraught. > > So here's the variant. Go ahead and index your "collection vector", > but index it with a second field that is your "collection ID". Then, add > that collection ID to each document in your original index. So, you have > something like > > a: text:{look, a, cat} collectionID:32 > b: text:{my, chimpansee, is , hairy} collectionID:32 > c: text:{dogs, are, playful} collectionID:32 > > Your other index has > collectionID:32 collectionVector:{look, a, cat, my, chimpansee, is , > hairy, > dogs, are, playful} > > Now, you essentially make two queries, one to get a set of > collection IDs from your second index (that is, querying your terms > against collectionVector) and using that set of collectionIDs in a > query against your first index. > > You might be able to do some interesting things with boosts > to score either query more to your liking. > > This will come close to doubling the size of your index, but your > first approach could bloat it by an arbitrary factor depending upon > how many documents were in your largest collection..... > > One thing to note, however, is that there is no need to have > two separate physical indexes. Lucene does not require that > all documents have the same fields. So this could all be in one > big happy index. As long as the fields are different in the two > sets of documents, the queries won't interfere with each other. In > that case, you'd have to name the "foreign key" field differently for > the sets of documents, say collectionID1 and collectionID2. > > > All that said, this approach bothers me because it's mixing > some database ideas with a Lucene index. I suppose in a controlled > situation where you won't be trying to do arbitrary joins it's probably > a misplaced unease. But I'm leery of trying to make Lucene act > like a database. But that may just be a personal problem <G> > > The only other consideration is "how many collections do you have?" > The reason I ask is that in the worst case scenario, you'll have an > OR clause for every collection ID you have. Lucene can easily handle > many thousands of terms in an OR, but your search time will suffer. > And you'll have to take special action (really, just set > MaxBooleanClauses) > if this is over 1024 or you'll get a TooManyClauses exception. > > Best > Erick > > > On 5/19/07, Peter Bloem <[EMAIL PROTECTED]> wrote: >> >> I'm sorry, I should have explained the intended behavior more clearly. >> >> The basic idea (without the collection fields) is that there are very >> simple documents in the index with one content field each. All I do with >> this index is a standard search in this text field. To improve the >> search results, I want to also add the concatenation of all documents in >> a collection as a field to every single document. I then search the >> index using both fields, and diminishing the effect of the collection >> field. This should improve the search results. >> >> As an example, say I have the documents a:"look a cat" b:"my chimpansee >> is hairy" c:"dogs are playful" and many others. These three documents >> are grouped into one collection (of many). The term vectors for the >> documents would then be >> a: {look, a, cat} >> b: {my, chimpansee, is , hairy} >> c: {dogs, are, playful} >> If I create a term vector for the whole collection: {look, a, cat, my, >> chimpansee, is , hairy, dogs, are, playful} and add it to each of the >> documents as a separate field, the query "my hairy cat" scores well >> against document a because of the match on cat, but also because of the >> match on both cat and hairy on the collection field. Documents about the >> linux command 'cat' do not have the word "hairy" in their collection >> field (because they're part of a different collection), and so would not >> get this benefit. It's essentially a smoothing technique, since it >> allows query words that aren't in the document to still have some >> effect. >> >> The problem of course is that storing these collection term vectors for >> each document greatly increases the size of the index and the indexing >> time. It would be alot faster if I could somehow use a second index to >> store the collections as documents, so I would only have to store one >> term vector per collection. (This isn't my own idea btw, I'm trying to >> replicate the results from some other research that used this method). >> >> I hope this is more clear, >> Peter >> >> Erick Erickson wrote: >> > This seems kind of kludgy, but that may just mean I don't understand >> > your problem very well. >> > >> > What is it that you're trying to accomplish? Searching constrained >> > by topic or groups? >> > >> > If you're trying to search by groups, search the archive for the >> > word "facet" or "faceted search". >> > >> > Otherwise, could you describe what behavior you're after and maybe >> > there'd be more ideas.... >> > >> > Best >> > Erick >> > >> > On 5/19/07, Peter Bloem <[EMAIL PROTECTED]> wrote: >> >> >> >> Hi, >> >> >> >> I have the following problem. I'm indexing documents that belong to >> some >> >> collection (ie. the dataset is divided into collections, which are >> >> divided into documents). These documents become my lucene documents, >> >> with some relatively small string that becomes the field I want to >> >> search. However, I would also like to add to document d the >> >> concatenation of all documents in d's collection as a field >> (mainly as >> a >> >> smoothing technique, because documents correspond roughly to topics). >> >> I'm currently doing just that, adding an extra field for the entire >> >> concatenated collection to each document in that collection. Of >> course >> >> this increases the index size and indexing time greatly (about >> >> five-fold). >> >> >> >> There must be a better way to do this. My idea was to create a second >> >> index where the collections are indexed as (lucene) documents. This >> >> index would have the text as a field, and a list of document id's >> >> referring back to the main index. I could then retrieve the term >> vector >> >> for each collection from this second index for each search result >> from >> >> the original index. >> >> >> >> My question is if this is a smart approach. And if it is, which of >> >> Lucene's classes should I use for this. The best I could find was the >> >> FilterIndexReader. If extending the FilterIndexReader is really the >> best >> >> way to go, could I simply override the document(int, FieldSelector) >> >> method, or is there more to it? I doubt I'm the first person that's >> ever >> >> wanted a many to one relation between fields and documents, so I hope >> >> there's a simpler way about this. >> >> >> >> Thank you, >> >> Peter >> >> >> >> --------------------------------------------------------------------- >> >> To unsubscribe, e-mail: [EMAIL PROTECTED] >> >> For additional commands, e-mail: [EMAIL PROTECTED] >> >> >> >> >> > >> >> >> --------------------------------------------------------------------- >> To unsubscribe, e-mail: [EMAIL PROTECTED] >> For additional commands, e-mail: [EMAIL PROTECTED] >> >> > --------------------------------------------------------------------- To unsubscribe, e-mail: [EMAIL PROTECTED] For additional commands, e-mail: [EMAIL PROTECTED]