Question: I need to dynamically data to SOLR, so I do not have a "predefined" list of field names...
so i use the dynamicField option in the schma and match approioate datatype.. in my schema.xml <field name="id" type="string" indexed="true" stored="true" required="true" /> <dynamicField name="*_s" type="string" indexed="true" stored="true"/> Then programatically my code ... document.addField( dynamicFieldName + "_s", dynamicFieldValue, 10 ); facetFieldNames.put( dynamicFieldName + "_s",null);//TODO:use copyField.. server.add( document,true ); server.commit(); when i attempt to graph results, i want to display SolrQuery query = new SolrQuery(); query.setQuery( "*:*" ); query.setFacetLimit(10);//TODO: Iterator facetsIt = facetFieldNames.entrySet().iterator(); while(facetsIt.hasNext()){ Entry<String,String>entry = (Entry)facetsIt.next(); String facetName = (String)entry.getKey(); query.addFacetField(facetName); } QueryResponse rsp; rsp = server.query( query ); List<FacetField> facetFieldList = rsp.getFacetFields(); assertNotNull(facetFieldList); .... my facetFieldList is null, of course if i addFacetField if "id" it works..because i define it in the schema.xml........ is this just a something that is not implemented? or am i missing something... Thanks. Jeryl Cook /^\ Pharaoh /^\ http://pharaohofkush.blogspot.com/ "..Act your age, and not your shoe size.." -Prince(1986) > Date: Fri, 30 Nov 2007 21:23:59 -0500 > From: [EMAIL PROTECTED] > To: solr-user@lucene.apache.org > Subject: Re: Solr Highlighting, word index > > It's good you already have the data because if you somehow got it from > some sort of calculations I'd have to tell my product manager that > the feature he wanted that I told him couldn't be done with our data > was possible after all <G>... > > About page breaks: > > Another approach to paging is to index a special page token with an > increment of 0 from the last word of the page. Say you have the following: > last ctrl-l first. Then index last, $$$$$$$ at an increment of 0 then first. > > You can then quite quickly calculate the pages by using > termdocs/termenum on your special token and count. > > Which approach you use depends upon whether you want span and/or > phrase queries to match across page boundaries. If you use an increment as > Mike suggests, matching "last first"~3 won't work. It just depends upon > whether how you want to match across the page break. > > Best > Erick > > On Nov 30, 2007 4:37 PM, Mike Klaas <[EMAIL PROTECTED]> wrote: > > > On 30-Nov-07, at 1:02 PM, Owens, Martin wrote: > > > > > > > > Hello everyone, > > > > > > We're working to replace the old Linux version of dtSearch with > > > Lucene/Solr, using the http requests for our perl side and java for > > > the indexing. > > > > > > The functionality that is causing the most problems is the > > > highlighting since we're not storing the text in solr (only > > > indexing) and we need to highlight an image file (ocr) so what we > > > really need is to request from solr the word indexes of the > > > matches, we then tie this up to the ocr image and create html boxes > > > to do the highlighting. > > > > This isn't possible with Solr out-of-the-box. Also, the usual > > methods for highlighting won't work because Solr typically re- > > analyzes the raw text to find the appropriate highlighting points. > > However, it shouldn't be too hard to come up with a custom solution. > > You can tell lucene to store token offsets using TermVectors > > (configurable via schema.xml). Then you can customize the request > > handler to return the token offsets (and/or positions) by retrieving > > the TVs. > > > > > The text is also multi page, each page is seperated by Ctrl-L page > > > breaks, should we handle the paging out selves or can Solr tell use > > > which page the match happened on too? > > > > Again, not automatically. However, if you wrote an analyzer that > > bumped up the position increment of tokens every time a new page was > > found (to, say the next multiple of 1000), then you infer the > > matching page by the token position. > > > > cheers, > > -Mike > >