Not Able to Build Spellcheck index - SpellCheckComponent.prepare 500 Error
Hi I am trying to use spellcheck in solr with below config but it throwing with error while using spellcheck build or reload it works fine otherwise for indexed search, can someone please help implementing spellcheck corectly schema.xml: // fieldType declaration fieldType name=textSpell class=solr.TextField positionIncrementGap=100 analyzer tokenizer class=solr.StandardTokenizerFactory/ filter class=solr.LowerCaseFilterFactory/ filter class=solr.StopFilterFactory ignoreCase=true words=misspelled_words.txt/ filter class=solr.PatternReplaceFilterFactory pattern=([^a-z]) replacement= replace=all/ filter class=solr.LengthFilterFactory min=2 max=50/ /analyzer /fieldType //field name field name=spell type=textSpell indexed=true stored=true multiValued=true/ //copyFields copyField source=dealName dest=spell/ copyField source=dealUri dest=spell/ copyField source=merchantName dest=spell/ copyField source=categoryName dest=spell/ copyField source=dealTitle dest=spell/ copyField source=highlights dest=spell/ copyField source=seoTags dest=spell/ copyField source=description dest=spell/ copyField source=text dest=spell/ solrconfig.xml: //searchComponent searchComponent name=spellcheck class=solr.SpellCheckComponent str name=queryAnalyzerFieldTypetextSpell/str lst name=spellchecker str name=classnamesolr.IndexBasedSpellChecker/str str name=namedefault/str str name=spellcheckIndexDir./spellchecker/str str name=fieldcategoryName,dealName,seoTags,description,dealTitle,merchantName,dealUri,highlights/str str name=buildOnCommittrue/str str name=accuracy0.9/str /lst /searchComponent //default requestHandler requestHandler name=standard class=solr.StandardRequestHandler default=true lst name=defaults str name=echoParamsexplicit/str str name=spellchecktrue/str str name=spellcheck.dictionarydirect/str str name=spellcheckon/str str name=spellcheck.extendedResultstrue/str str name=spellcheck.count5/str str name=spellcheck.collatetrue/str str name=spellcheck.collateExtendedResultstrue/str /lst arr name=last-components strspellcheck/str /arr /requestHandler // URL params select?q=*%3A*wt=phpindent=truespellcheck=truespellcheck.build=true //output array( 'responseHeader'=array( 'status'=500, 'QTime'=4, 'params'=array( 'spellcheck'='true', 'indent'='true', 'q'='*:*', '_'='1396684768649', 'wt'='php', 'spellcheck.build'='true')), 'error'=array( 'trace'='java.lang.NullPointerException at org.apache.solr.handler.component.SpellCheckComponent.prepare(SpellCheckComponent.java:125) at org.apache.solr.handler.component.SearchHandler.handleRequestBody(SearchHandler.java:187) at org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:135) at org.apache.solr.core.SolrCore.execute(SolrCore.java:1859) at org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:710) at org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:413) at org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:197) at org.mortbay.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1212) at org.mortbay.jetty.servlet.ServletHandler.handle(ServletHandler.java:399) at org.mortbay.jetty.security.SecurityHandler.handle(SecurityHandler.java:216) at org.mortbay.jetty.servlet.SessionHandler.handle(SessionHandler.java:182) at org.mortbay.jetty.handler.ContextHandler.handle(ContextHandler.java:767) at org.mortbay.jetty.webapp.WebAppContext.handle(WebAppContext.java:450) at org.mortbay.jetty.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:230) at org.mortbay.jetty.handler.HandlerCollection.handle(HandlerCollection.java:114) at org.mortbay.jetty.handler.HandlerWrapper.handle(HandlerWrapper.java:152) at org.mortbay.jetty.handler.DebugHandler.handle(DebugHandler.java:77) at org.mortbay.jetty.handler.HandlerWrapper.handle(HandlerWrapper.java:152) at org.mortbay.jetty.Server.handle(Server.java:326) at org.mortbay.jetty.HttpConnection.handleRequest(HttpConnection.java:542) at org.mortbay.jetty.HttpConnection$RequestHandler.headerComplete(HttpConnection.java:928) at org.mortbay.jetty.HttpParser.parseNext(HttpParser.java:549) at org.mortbay.jetty.HttpParser.parseAvailable(HttpParser.java:212) at org.mortbay.jetty.HttpConnection.handle(HttpConnection.java:404) at org.mortbay.io.nio.SelectChannelEndPoint.run(SelectChannelEndPoint.java:410) at org.mortbay.thread.QueuedThreadPool$PoolThread.run(QueuedThreadPool.java:582) ', 'code'=500)) -- View this message in context:
Re: Not Able to Build Spellcheck index - SpellCheckComponent.prepare 500 Error
It's a null pointer exception. Either something is not defined correctly or you are hitting a funny unexpected. Which version of Solr is it? Regards, Alex. Personal website: http://www.outerthoughts.com/ Current project: http://www.solr-start.com/ - Accelerating your Solr proficiency On Sat, Apr 5, 2014 at 3:24 PM, sameer sameer.sha...@trideal.in wrote: Hi I am trying to use spellcheck in solr with below config but it throwing with error while using spellcheck build or reload
Re: Full Indexing is Causing a Java Heap Out of Memory Exception
Hi, Now we have a more informative error : org.apache.solr.handler.dataimport.DataImportHandlerException: java.lang.OutOfMemoryError: Java heap space Caused by: org.apache.solr.handler.dataimport.DataImportHandlerException: java.lang.OutOfMemoryError: Java heap space at org.apache.solr.handler.dataimport.DocBuilder.buildDocument(DocBuilder.java:535) at org.apache.solr.handler.dataimport.DocBuilder.buildDocument(DocBuilder.java:404) 1) Does this happen when you increase -Xmx64m -Xms64m ? 2) I see you use custom jars called MDSolrDIHTransformer JARs inside But I don't see any Transformers used in database.xm, why is that. I would remove them just to be sure. 3) I see you have org.apache.solr.core.StandardDirectoryFactory declared in sorlconfig. Assuming you are using, 64 bit windows, it is recommended to use MMap http://blog.thetaphi.de/2012/07/use-lucenes-mmapdirectory-on-64bit.html 4) In your previous mail you had batch size set, now there is not batchSize defined in database.xml. For MySQL it is recommended to use -1. Not sure about oracle, I personally used 10,000 once for Oracle. http://wiki.apache.org/solr/DataImportHandlerFaq#I.27m_using_DataImportHandler_with_a_MySQL_database._My_table_is_huge_and_DataImportHandler_is_going_out_of_memory._Why_does_DataImportHandler_bring_everything_to_memory.3F You have a lot of copyFields defined. There could be some gotchas when handling unusually much copy fields. I would really try CSV option here. Given that you have only full import SQL defined and it is not a complex one. It queries only one table. I believe Oracle has some tool to export a table to CSV file efficiently. On Saturday, April 5, 2014 3:05 AM, Candygram For Mongo candygram.for.mo...@gmail.com wrote: Does this user list allow attachments? I have four files attached (database.xml, error.txt, schema.xml, solrconfig.xml). We just ran the process again using the parameters you suggested, but not to a csv file. It errored out quickly. We are working on the csv file run. Removed both autoCommit and autoSoftCommit parts/definitions from solrconfig.xml Disabled tlog by removing updateLog str name=dir${solr.ulog.dir:}/str /updateLog from solrconfig.xml Used commit=true parameter. ?commit=truecommand=full-import On Fri, Apr 4, 2014 at 3:29 PM, Ahmet Arslan iori...@yahoo.com wrote: Hi, This may not solve your problem but generally it is recommended to disable auto commit and transaction logs for bulk indexing. And issue one commit at the very end. Do you tlogs enabled? I see commit failed in the error message thats why I am offering this. And regarding comma separated values, with this approach you focus on just solr importing process. You separate data acquisition phrase. And it is very fast load even big csv files http://wiki.apache.org/solr/UpdateCSV I have never experienced OOM during indexing, I suspect data acquisition has role in it. Ahmet On Saturday, April 5, 2014 1:18 AM, Candygram For Mongo candygram.for.mo...@gmail.com wrote: We would be happy to try that. That sounds counter intuitive for the high volume of records we have. Can you help me understand how that might solve our problem? On Fri, Apr 4, 2014 at 2:34 PM, Ahmet Arslan iori...@yahoo.com wrote: Hi, Can you remove auto commit for bulk import. Commit at the very end? Ahmet On Saturday, April 5, 2014 12:16 AM, Candygram For Mongo candygram.for.mo...@gmail.com wrote: In case the attached database.xml file didn't show up, I have pasted in the contents below: dataConfig dataSource name=org_only type=JdbcDataSource driver=oracle.jdbc.OracleDriver url=jdbc:oracle:thin:@test2.abc.com:1521:ORCL user=admin password=admin readOnly=false batchSize=100 / document entity name=full-index query= select NVL(cast(ORCL.ADDRESS_ACCT_ALL.RECORD_ID as varchar2(100)), 'null') as SOLR_ID, 'ORCL.ADDRESS_ACCT_ALL' as SOLR_CATEGORY, NVL(cast(ORCL.ADDRESS_ACCT_ALL.RECORD_ID as varchar2(255)), ' ') as ADDRESSALLROWID, NVL(cast(ORCL.ADDRESS_ACCT_ALL.ADDR_TYPE_CD as varchar2(255)), ' ') as ADDRESSALLADDRTYPECD, NVL(cast(ORCL.ADDRESS_ACCT_ALL.LONGITUDE as varchar2(255)), ' ') as ADDRESSALLLONGITUDE, NVL(cast(ORCL.ADDRESS_ACCT_ALL.LATITUDE as varchar2(255)), ' ') as ADDRESSALLLATITUDE, NVL(cast(ORCL.ADDRESS_ACCT_ALL.ADDR_NAME as varchar2(255)), ' ') as ADDRESSALLADDRNAME, NVL(cast(ORCL.ADDRESS_ACCT_ALL.CITY as varchar2(255)), ' ') as ADDRESSALLCITY, NVL(cast(ORCL.ADDRESS_ACCT_ALL.STATE as varchar2(255)), ' ') as ADDRESSALLSTATE, NVL(cast(ORCL.ADDRESS_ACCT_ALL.EMAIL_ADDR as varchar2(255)), ' ') as ADDRESSALLEMAILADDR from ORCL.ADDRESS_ACCT_ALL field column=SOLR_ID name=id / field column=SOLR_CATEGORY name=category / field column=ADDRESSALLROWID name=ADDRESS_ACCT_ALL.RECORD_ID_abc / field column=ADDRESSALLADDRTYPECD name=ADDRESS_ACCT_ALL.ADDR_TYPE_CD_abc / field column=ADDRESSALLLONGITUDE name=ADDRESS_ACCT_ALL.LONGITUDE_abc / field column=ADDRESSALLLATITUDE
Re: Not Able to Build Spellcheck index - SpellCheckComponent.prepare 500 Error
its solr-4.6.0 -- View this message in context: http://lucene.472066.n3.nabble.com/Not-Able-to-Build-Spellcheck-index-SpellCheckComponent-prepare-500-Error-tp4129368p4129392.html Sent from the Solr - User mailing list archive at Nabble.com.
Combining eDismax and SpellChecker
Hi All, I want to suggest the correct phrase if a typo is made while searching and then search it using eDismax parser(pf,pf2,pf3), if no typo is made then search it using eDismax parser alone. Is there a way I can combine these two components , I have seen examples for eDismax and also for SpellChecker , but nothing that combines these two together. Can you please let me know ? Thanks.
Re: How to see the value of long type (solr) ?
One of the Lucene guys is going to need to address this question. I d know that Trie fields index additional values to support fast range queries, so maybe you are merely seeing some of those generated values, and if you look further you should see your actual indexed value. What exactly are you seeing in Luke for this TrieLongField? -- Jack Krupansky -Original Message- From: Lisheng Zhang Sent: Friday, April 4, 2014 12:44 PM To: solr-user@lucene.apache.org Subject: How to see the value of long type (solr) ? Hi, We use solr 3.6 to index a field of long type: fieldType name=long class=solr.TrieLongField ... Now for debugging purpose we need to see the original value (the field is not stored), but in luke we cannot see. 1/ is there a way to see original long type value (using luke or not) ? 2/ if we need to use lucene to search this field, what analyzer should we use ? Thanks very much for helps, Lisheng
Re: Combining eDismax and SpellChecker
There is one commercial solution http://www.sematext.com/products/dym-researcher/index.html On Saturday, April 5, 2014 4:07 PM, S.L simpleliving...@gmail.com wrote: Hi All, I want to suggest the correct phrase if a typo is made while searching and then search it using eDismax parser(pf,pf2,pf3), if no typo is made then search it using eDismax parser alone. Is there a way I can combine these two components , I have seen examples for eDismax and also for SpellChecker , but nothing that combines these two together. Can you please let me know ? Thanks.
Re: How to see the value of long type (solr) ?
You can use faceting to human readable values. On Saturday, April 5, 2014 7:08 PM, Jack Krupansky j...@basetechnology.com wrote: One of the Lucene guys is going to need to address this question. I d know that Trie fields index additional values to support fast range queries, so maybe you are merely seeing some of those generated values, and if you look further you should see your actual indexed value. What exactly are you seeing in Luke for this TrieLongField? -- Jack Krupansky -Original Message- From: Lisheng Zhang Sent: Friday, April 4, 2014 12:44 PM To: solr-user@lucene.apache.org Subject: How to see the value of long type (solr) ? Hi, We use solr 3.6 to index a field of long type: fieldType name=long class=solr.TrieLongField ... Now for debugging purpose we need to see the original value (the field is not stored), but in luke we cannot see. 1/ is there a way to see original long type value (using luke or not) ? 2/ if we need to use lucene to search this field, what analyzer should we use ? Thanks very much for helps, Lisheng
Re: Searching multivalue fields.
Hi, Did restart solr and you re-index after schema change? On Saturday, April 5, 2014 2:39 AM, Vijay Kokatnur kokatnur.vi...@gmail.com wrote: I had already tested with omitTermFreqAndPositions=false . I still got the same error. Is there something that I am overlooking? On Fri, Apr 4, 2014 at 2:45 PM, Ahmet Arslan iori...@yahoo.com wrote: Hi Vijay, Add omitTermFreqAndPositions=false attribute to fieldType definitions. fieldType name=string class=solr.StrField omitTermFreqAndPositions=false sortMissingLast=true / fieldType name=int class=solr.TrieIntField omitTermFreqAndPositions=false precisionStep=0 positionIncrementGap=0/ You don't need termVectors for this. 1.2: omitTermFreqAndPositions attribute introduced, true by default except for text fields. And please reply to solr user mail, so others can use the threat later on. Ahmet On Saturday, April 5, 2014 12:18 AM, Vijay Kokatnur kokatnur.vi...@gmail.com wrote: Hey Ahmet, Sorry it took some time to test this. But schema definition seem to conflict with SpanQuery. I get following error when I use Spans field OrderLineType was indexed without position data; cannot run SpanTermQuery (term=11) I changed field definition in the schema but can't find the right attribute to set this. My last attempt was with following definition field name=OrderLineType type=string indexed=true stored=true multiValued=true termVectors=true termPositions=true termOffsets=true/ Any ideas what I am doing wrong? Thanks, -Vijay On Wed, Mar 26, 2014 at 1:54 PM, Ahmet Arslan iori...@yahoo.com wrote: Hi Vijay, After reading the documentation it seems that following query is what you are after. It will return OrderId:345 without matching OrderId:123 SpanQuery q1 = new SpanTermQuery(new Term(BookingRecordId, 234)); SpanQuery q2 = new SpanTermQuery(new Term(OrderLineType, 11)); SpanQuery q2m new FieldMaskingSpanQuery(q2, BookingRecordId); Query q = new SpanNearQuery(new SpanQuery[]{q1, q2m}, -1, false); Ahmet On Wednesday, March 26, 2014 10:39 PM, Ahmet Arslan iori...@yahoo.com wrote: Hi Vijay, I personally don't understand joins very well. Just a guess may be FieldMaskingSpanQuery could be used? http://blog.griddynamics.com/2011/07/solr-experience-search-parent-child.html Ahmet On Wednesday, March 26, 2014 9:46 PM, Vijay Kokatnur kokatnur.vi...@gmail.com wrote: Hi, I am bumping this thread again one last time to see if anyone has a solution. In it's current state, our application is storing child items as multivalue fields. Consider some orders, for example - { OrderId:123 BookingRecordId : [145, 987, *234*] OrderLineType : [11, 12, *13*] . } { OrderId:345 BookingRecordId : [945, 882, *234*] OrderLineType : [1, 12, *11*] . } { OrderId:678 BookingRecordId : [444] OrderLineType : [11] . } Here, If you look up for an Order with BookingRecordId: 234 And OrderLineType:11. You will get two orders with orderId : 123 and 345, which is correct. You have two arrays in both the orders that satisfy this condition. However, for OrderId:123, the value at 3rd index of OrderLineType array is 13 and not 11( this is for OrderId:345). So orderId 123 should be excluded. This is what I am trying to achieve. I got some suggestions from a solr-user to use FieldsCollapsing, Join, Block-join or string concatenation. None of these approaches can be used without re-indexing schema. Has anyone found a non-invasive solution for this? Thanks, -Vijay
Re: Does sorting skip everything having to do with relevancy?
Shawn, I suppose e yields syntax error. Therefore, this case doesn't prove anything yet. Haven't you tried sqrt(-1) or log(-1) ? On Sat, Apr 5, 2014 at 1:47 AM, Shawn Heisey s...@elyograg.org wrote: On 4/4/2014 3:13 PM, Mikhail Khludnev wrote: I suppose SolrIndexSearcher.buildTopDocsCollector() doesn't create a Collector which calls score() in this case. Hence, it shouldn't waste CPU. Just my impression. Haven't you tried to check it supplying some weird formula, which throws exception? I didn't think of that. That's a good idea -- as long as there's not independent code that checks the function in addition to the code that actually runs it. With the following parameters added to an edismax query that otherwise works, I get an exception. It works if I change the e to 5. sort=registered_date ascboost=sum(5,e) I will take Alvaro's suggestion and add boost=1 to queries that use a sort parameter. It's probably a good idea to file that Jira. Thanks, Shawn -- Sincerely yours Mikhail Khludnev Principal Engineer, Grid Dynamics http://www.griddynamics.com mkhlud...@griddynamics.com
Re: SOLR Jetty Server on Windows 2003
Is the URL for the Solr request absolutely 100% identical in both cases? By not getting a response, do you mean it hangs andtimes out or that the response is empty? -- Jack Krupansky -Original Message- From: EXTERNAL Taminidi Ravi (ETI, Automotive-Service-Solutions) Sent: Friday, April 4, 2014 11:49 AM To: solr-user@lucene.apache.org Subject: SOLR Jetty Server on Windows 2003 Hi , I am trying to install solr on the Windows 2003 with Jetty server. Form browser everything works , but when I try to acesss from another javascript Code in other machine I am not getting reponse. I am using Xmlhttprequest to get the response from server using javascript. Any Help...? --Ravi
Re: Strange behavior of edismax and mm=0 with long queries (bug?)
Set the q.op parameter to OR and set mm=10% or something like that. The idea is to not excessively restrict the documents that will match, but weight the matched results based on how many word pairs and triples do match. In addition, use the pf parameter to provide extra weight when the full query term phrase matches exactly. -- Jack Krupansky From: Nils Kaiser Sent: Friday, April 4, 2014 10:10 AM To: solr-user@lucene.apache.org Subject: Strange behavior of edismax and mm=0 with long queries (bug?) Hey, I am currently using solr to recognize songs and people from a list of user comments. My index stores the titles of the songs. At the moment my application builds word ngrams and fires a search with that query, which works well but is quite inefficient. So my thought was to simply use the collated comments as query. So it is a case where the query is much longer. I need to use mm=0 or mm=1. My plan was to use edismax as the pf2 and pf3 parameters should work well for my usecase. However when using longer queries, I get a strange behavior which can be seen in debugQuery. Here is an example: Collated Comments (used as query) I love Henry so much. It is hard to tear your eyes away from Maria, but watch just his feet. You'll be amazed. sometimes pure skill can will a comp, sometimes pure joy can win... put them both together and there is no competition This video clip makes me smile. Pure joy! so good! Who's the person that gave this a thumbs down?!? This is one of the best routines I've ever seen. Period. And it's a competitionl! How is that possible? They're so good it boggles my mind. It's gorgeous. Flawless victory. Great number! Does anybody know the name of the piece? I believe it's called Sunny side of the street Maria is like, the best 'follow' I've ever seen. She's so amazing. Thanks so much Johnathan! Song name in Index Louis Armstrong - Sunny Side of The Street parsedquery_toString: +(((text:I) (text:love) (text:Henry) (text:so) (text:much.) (text:It) (text:is) (text:hard) (text:to) (text:tear) (text:your) (text:eyes) (text:away) (text:from) (text:Maria,) (text:but) (text:watch) (text:just) (text:his) (text:feet.) (text:You'll) (text:be) (text:amazed.) (text:sometimes) (text:pure) (text:skill) (text:can) (text:will) (text:a) (text:comp,) (text:sometimes) (text:pure) (text:joy) (text:can) (text:win...) (text:put) (text:them) (text:both) +(text:together) +(text:there) (text:is) (text:no) (text:competition) (text:This) (text:video) (text:clip) (text:makes) (text:me) (text:smile.) (text:Pure) (text:joy!) (text:so) (text:good!) (text:Who's) (text:the) (text:person) (text:that) (text:gave) (text:this) (text:a) (text:thumbs) (text:down?!?) (text:This) (text:is) (text:one) (text:of) (text:the) (text:best) (text:routines) (text:I've) (text:ever) (text:seen.) +(text:Period.) +(text:it's) (text:a) (text:competitionl!) (text:How) (text:is) (text:that) (text:possible?) (text:They're) (text:so) (text:good) (text:it) (text:boggles) (text:my) (text:mind.) (text:It's) (text:gorgeous.) (text:Flawless) (text:victory.) (text:Great) (text:number!) (text:Does) (text:anybody) (text:know) (text:the) (text:name) (text:of) (text:the) (text:piece?) (text:I) (text:believe) (text:it's) (text:called) (text:Sunny) (text:side) (text:of) (text:the) (text:street) (text:Maria) (text:is) (text:like,) (text:the) (text:best) (text:'follow') (text:I've) (text:ever) (text:seen.) (text:She's) (text:so) (text:amazing.) (text:Thanks) (text:so) (text:much) (text:Johnathan!))~1)/str This query generates 0 results. The reason is it expects terms together, there, Period., it's to be part of the document (see parsedquery above, all other terms are optional, those terms are must). Is there any reason for this behavior? If I use shorter queries it works flawlessly and returns the document. I've appended the whole query. Best, Nils
Re: Query and field name with wildcard
The LucidWorks Search query parser lets you use the all pseudo-field to search across all fields. See: http://docs.lucidworks.com/display/lweug/Field+Queries For example: q = all:some_word -- Jack Krupansky -Original Message- From: Ahmet Arslan Sent: Friday, April 4, 2014 8:13 AM To: solr-user@lucene.apache.org Subject: Re: Query and field name with wildcard Hi, bq. possible to search a word over the entire index. You can a get list of all searchable fields (indexed=true) programmatically by https://wiki.apache.org/solr/LukeRequestHandler And then you can fed this list to qf parameter of (e)dismax. This could be implemented as a custom query parser plugin that searches a word over the entire index. Ahmet On Friday, April 4, 2014 12:08 PM, Alexandre Rafalovitch arafa...@gmail.com wrote: Are you using eDisMax. That gives a lot of options, including field aliasing, including a single name to multiple fields: http://wiki.apache.org/solr/ExtendedDisMax#Field_aliasing_.2F_renaming (with example on p77 of my book http://www.packtpub.com/apache-solr-for-indexing-data/book :-) Regards, Alex. Personal website: http://www.outerthoughts.com/ Current project: http://www.solr-start.com/ - Accelerating your Solr proficiency On Fri, Apr 4, 2014 at 3:52 PM, Croci Francesco Luigi (ID SWS) fcr...@id.ethz.ch wrote: In my index I have some fields which have the same prefix(rmDocumentTitle, rmDocumentClass, rmDocumentSubclass, rmDocumentArt). Apparently it is not possible to specify a query like this: q = rm* : some_word Is there a way to do this without having to write a long list of ORs? Another question is if it is really not possible to search a word over the entire index. Something like this: q = * : some_word Thank you Francesco
Re: Does sorting skip everything having to do with relevancy?
On 4/5/2014 1:21 PM, Mikhail Khludnev wrote: I suppose e yields syntax error. Therefore, this case doesn't prove anything yet. Haven't you tried sqrt(-1) or log(-1) ? Using boost=sqrt(-1) is error-free whether I include the sort parameter or not. That seems like a bug. Thanks, Shawn
Re: Using Sentence Information For Snippet Generation
Hi Dmitry; I think that such kind of hacking may reduce the search speed. I think that it should be done with boundary scanner isn't it? I think that bs.type=LINE is what I am looking for? There is one more point. I want to do that for Turkish language and I think that I should customize it or if I put special characters to point boundaries I can use simple boundary scanner? Thanks; Furkan KAMACI 2014-03-24 21:14 GMT+02:00 Dmitry Kan solrexp...@gmail.com: Hi Furkan, I have done an implementation with a custom filler (special character) sequence in between sentences. A better solution I landed at was increasing the position of each sentence's first token by a large number, like 1 (perhaps, a smaller number could be used too). Then a user search can be conducted with a proximity query: some tokens ~5000 (the recently committed complexphrase parser supports rich phrase syntax, for example). This of course expects that a sentence fits the 5000 window size and the total number of sentences in the field * 10k does not exceed Integer.MAX_VALUE. Then on the highlighter side you'd get the hits within sentences naturally. Is this something you are looking for? Dmitry On Mon, Mar 24, 2014 at 5:43 PM, Furkan KAMACI furkankam...@gmail.com wrote: Hi; When I generate snippet via Solr I do not want to remove beginning of any sentence at the snippet. So I need to do a sentence detection. I think that I can do it before I send documents into Solr. I can put some special characters that signs beginning or end of a sentence. Then I can use that information when generating snippet. On the other hand I should not show that special character to the user. What do you think that how can I do it or do you have any other ideas for my purpose? PS: I do not do it for English sentences. Thanks; Furkan KAMACI -- Dmitry Blog: http://dmitrykan.blogspot.com Twitter: http://twitter.com/dmitrykan
Re: Combining eDismax and SpellChecker
Yes, I saw that earlier in one of your other postings. Is it the case that we cannot use the SpellChecker with a parser like edismax by making a configuration change without having to go thru this commercial product? Sent from my HTC - Reply message - From: Ahmet Arslan iori...@yahoo.com To: solr-user@lucene.apache.org solr-user@lucene.apache.org Subject: Combining eDismax and SpellChecker Date: Sat, Apr 5, 2014 12:11 PM There is one commercial solution http://www.sematext.com/products/dym-researcher/index.html On Saturday, April 5, 2014 4:07 PM, S.L simpleliving...@gmail.com wrote: Hi All, I want to suggest the correct phrase if a typo is made while searching and then search it using eDismax parser(pf,pf2,pf3), if no typo is made then search it using eDismax parser alone. Is there a way I can combine these two components , I have seen examples for eDismax and also for SpellChecker , but nothing that combines these two together. Can you please let me know ? Thanks.
Re: Solr Search on Fields name
One technique is to add a copyField directive to your schema, which can use a wildcard to copy a bunch of fields to a single, combinaed field that you can query directly, such as rullAll:key. Or, consider using a multivalued field. -- Jack Krupansky -Original Message- From: anuragwalia Sent: Friday, April 4, 2014 7:15 AM To: solr-user@lucene.apache.org Subject: Solr Search on Fields name Hi, Thank for giving your important time. Problem : I am unable to find a way how can I search Key with OR operator like if I search Items having RuleA OR RuleE. Format of Indexed Data: result name=response numFound=27 start=0 maxScore=1.0 doc float name=score1.0/float . int name=RuleA4/int int name=RuleD2/int int name=RuleE2/int int name=RuleF2/int /doc Can any one help me out how can prepare SearchQuery for key search. Regards Anurag -- View this message in context: http://lucene.472066.n3.nabble.com/Solr-Search-on-Fields-name-tp4129119.html Sent from the Solr - User mailing list archive at Nabble.com.
maxDistErr 0.000009
As we all know, maxDistErr=0.09 is approx 1 meter. If I increase it to maxDistErr=0.9 then it would be 10 meters. Still really good for most usages (finding a house, etc). What would be the index size improvement on a million rows? And what would the anticipated performance gain be? In theory. Of coruce I would have to run performance tests to know for sure. Thoughts? -- Bill Bell billnb...@gmail.com cell 720-256-8076
Anyone going to ApacheCon in Denver next week?
Thoughts on getting together for breakfast? a little Solr meet up? -- Bill Bell billnb...@gmail.com cell 720-256-8076
RE: Anyone going to ApacheCon in Denver next week?
I'll be there. I'd love to meet up. Let me know! Sent from my Windows Phone From: William Bell Sent: 4/5/2014 10:40 PM To: solr-user@lucene.apache.org Subject: Anyone going to ApacheCon in Denver next week? Thoughts on getting together for breakfast? a little Solr meet up? -- Bill Bell billnb...@gmail.com cell 720-256-8076
Piggy backing off Bloomberg :)
Healthgrades is also hiring for a Linux/SOLR Admin. Ability to: - Manage production, and development SOLR machines using Debian Linux - Knowledge of Jetty, Java 7 - 1+ years Solr experience Downtown Denver, CO location. Contact me or see me at ApacheCon... Or ghay...@healthgrades.com -- Bill Bell billnb...@gmail.com cell 720-256-8076