RE: Solr 1.4.1 compatible with Lucene 3.0.1?
Hi RichSimon, In short: Solr is only compatible to the version it is shipped with (or the other minor patch levels). So Solr 1.4 works with Lucene 2.9, so you must have any Lucene 2.9.x in your classpath. Lucene 3.0 is a major release of Lucene and had removed all backwards compatibility layers, that are still existent in Lucene 2.9 (functionality is identical in 2.9 and 3.0). Unfortunately, Solr 1.4 still used some of the deprecated code, so it's not compatible. About one year ago, Lucene and Solr merged to one Apache project and now Lucene and Solr is released together with same version number. The latest Lucene and Solr release in 3.1. But again, Solr should only be used with the version it is shipped with. There was never a separate release of Solr for Lucene 3.0, you can only upgrade or downgrade. matchVersion parameters have less to do with API compatibility (sometimes it is used also for that to keep backwards), but it's more used to control behavior of some analyzers (means how text tokenization is done to make analysis compatible between different versions). Uwe - Uwe Schindler H.-H.-Meier-Allee 63, D-28213 Bremen http://www.thetaphi.de eMail: u...@thetaphi.de -Original Message- From: RichSimon [mailto:richard_si...@hms.harvard.edu] Sent: Monday, April 11, 2011 4:37 PM To: solr-...@lucene.apache.org Subject: Solr 1.4.1 compatible with Lucene 3.0.1? Short story: I am using Lucene 3.0.1, and I'm trying to run Solr 1.4.1. I get an error starting the embedded Solr server that says it cannot find the method FSDirectory.getDirectory. The release notes for Solr 1.4.1 says it is compatible with Lucene 2.9.3, and I see that Lucene 3.0.1 does not have the FSDirectory.getDirectory method any more. Dorwngrading Lucene to 2.9.x is not an option for me. What version of Solr should I use for Lucene 3.0.1? (We're just starting with Solr, so changing that version is not hard.) Or, do I have to upgrade both Solr and Lucene? Thanks, -Rich Here's the long story: I am using Lucene 3.0.1, and I'm trying to run Solr 1.4.1. I have not used any other version of Lucene. We have an existing project using Lucene 3.0.1, and we want to start using Solr. When I try to initialize an embedded Solr server, like so: String solrHome = PATH_TO_SOLR_HOME; File home = new File(solrHome); File solrXML = new File(home, solr.xml); coreContainer = new CoreContainer(); coreContainer.load(solrHome, solrXML); embeddedSolr = new EmbeddedSolrServer(coreContainer, SOLR_CORE); [04-08 11:48:39] ERROR CoreContainer [main]: java.lang.NoSuchMethodError: org.apache.lucene.store.FSDirectory.getDirectory(Ljava/lang/String;)Lorg/ap ache/lucene/store/FSDirectory; at org.apache.solr.spelling.AbstractLuceneSpellChecker.initIndex(AbstractLuce neSpellChecker.java:186) at org.apache.solr.spelling.AbstractLuceneSpellChecker.init(AbstractLuceneSpe llChecker.java:101) at org.apache.solr.spelling.IndexBasedSpellChecker.init(IndexBasedSpellCheck er.java:56) at org.apache.solr.handler.component.SpellCheckComponent.inform(SpellChe ckComponent.java:274) at org.apache.solr.core.SolrResourceLoader.inform(SolrResourceLoader.java:50 8) at org.apache.solr.core.SolrCore.(SolrCore.java:588) at org.apache.solr.core.CoreContainer.create(CoreContainer.java:428) at org.apache.solr.core.CoreContainer.load(CoreContainer.java:278) Looking at Google posts about this, it seemed that this can be caused by a version mismatch between the Lucene version in use and the one Solr tries to use. I noticed a Lucene version tag in the example solrconfig.xml that I’m modifying: LUCENE_40 I changing it to LUCENE_301, changing it to LUCENE_30, and commenting it out, but I still get the same error. Using LucenePackage.get().getImplementationVersion() shows this as the Lucene version: Lucene version: 3.0.1 912433 - 2010-02-21 23:51:22 I also printed my classpath and found the following lucene jars: lucene-analyzers-3.0.1.jar lucene-core-3.0.1.jar lucene-highlighter-3.0.1.jar lucene-memory-3.0.1.jar lucene-misc-2.9.3.jar lucene-queries-2.9.3.jar lucene-snowball-2.9.3.jar lucene-spellchecker-2.9.3.jar The FSDirectory class is in lucene-core. I decompiled the class file in the jar, and did not see a getDirectory method. Also, I used a ClassLoader statement to get an instance of the FSDirectory class my code is using, and printed out the methods; no getDirectory method. I gather from the Lucene Javadoc that the getDirectory method is in FSDirectory for 2.4.0 and for 2.9.0, but is gone in 3.0.1 (the version I'm using). Is Lucene 3.0.1 completely incompatible with Solr 1.4.1? Is there some way to use the luceneMatchVersion tag to tell Solr what version I want to use? -- View this message in context:
[jira] [Commented] (LUCENE-2956) Support updateDocument() with DWPTs
[ https://issues.apache.org/jira/browse/LUCENE-2956?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13018743#comment-13018743 ] Simon Willnauer commented on LUCENE-2956: - bq. I think I have an idea, however can you explain the ticketQueue? Sure, since with DWPT the flush process is concurrent and several DWPT could flush at the same time we must maintain the order of the flushes before we can apply the flushed segment and the frozen global deletes it is buffering. The reason for this is that the global deletes mark a certain point in time where we took a DWPT out of rotation and freeze the global deletes. Example: A DWPT 'A' starts flushing and freezes the global deletes, then DWPT 'B' starts flushing and freezes all deletes occurred since 'A' has started. if 'B' finishes before 'A' we need to wait until 'A' is done otherwise the deletes frozen by 'B' are not applied to 'A' and we might miss to deletes documents in 'A'. The Ticket queue simply ensures that we push the frozen deletes and the new created segment in the same order as the corresponding DWPT have started flushing. If a DWPT finishes flushing we update its Ticket and then check the head of the queue if we can remove / publish the ticket. if so we continue publishing until the head of the queue can not be published yet or the queue is empty. Support updateDocument() with DWPTs --- Key: LUCENE-2956 URL: https://issues.apache.org/jira/browse/LUCENE-2956 Project: Lucene - Java Issue Type: Bug Components: Index Affects Versions: Realtime Branch Reporter: Michael Busch Assignee: Simon Willnauer Priority: Minor Fix For: Realtime Branch Attachments: LUCENE-2956.patch With separate DocumentsWriterPerThreads (DWPT) it can currently happen that the delete part of an updateDocument() is flushed and committed separately from the corresponding new document. We need to make sure that updateDocument() is always an atomic operation from a IW.commit() and IW.getReader() perspective. See LUCENE-2324 for more details. -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
Re: termInfosIndexDivisor typo in the Solr-UIMA config?
Hi Otis, you're right, that can be safely changed to setTermIndexDivisor (note that line's commented so it's just for the matter of consistency I think). Tommaso 2011/4/11 Otis Gospodnetic otis_gospodne...@yahoo.com Hi, I was looking at term index divisor and spotted this: .../lucene-solr-3.1$ ffxg -i IndexDivisor ./solr/src/test-files/solr/conf/solrconfig-termindex.xml:int name=setTermIndexDivisor12/int ./solr/src/test-files/solr/conf/solrconfig-xinclude.xml:int name=setTermIndexDivisor12/int ./solr/contrib/uima/src/test/resources/solr-uima/conf/solrconfig.xml: !-- To set the termInfosIndexDivisor, do this: -- ./solr/contrib/uima/src/test/resources/solr-uima/conf/solrconfig.xml: name=termInfosIndexDivisor12/int /indexReaderFactory HERE ./solr/example/solr/conf/solrconfig.xml: !-- By explicitly declaring the Factory, the termIndexDivisor can ./solr/example/solr/conf/solrconfig.xml: int name=setTermIndexDivisor12/int Is that termInfosIndexDivisor a typo in there? Should it be setTermIndexDivisor like in the other configs? Otis Sematext :: http://sematext.com/ :: Solr - Lucene - Nutch Lucene ecosystem search :: http://search-lucene.com/ - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-3018) Lucene Native Directory implementation need automated build
[ https://issues.apache.org/jira/browse/LUCENE-3018?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13018762#comment-13018762 ] Simon Willnauer commented on LUCENE-3018: - Varun, are you getting along here so far? Lucene Native Directory implementation need automated build --- Key: LUCENE-3018 URL: https://issues.apache.org/jira/browse/LUCENE-3018 Project: Lucene - Java Issue Type: Wish Components: Build Affects Versions: 4.0 Reporter: Simon Willnauer Assignee: Varun Thacker Priority: Minor Fix For: 4.0 Currently the native directory impl in contrib/misc require manual action to compile the c code (partially) documented in https://svn.apache.org/repos/asf/lucene/dev/trunk/lucene/contrib/misc/src/java/overview.html yet it would be nice if we had an ant task and documentation for all platforms how to compile them and set up the prerequisites. -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
Patch for http_proxy support in solr-ruby client
Hi, I have a patch for adding http_proxy support to the solr-ruby client. I thought the project was managed via Github, but this turns out not to be the case. It the process the same as for Solr itself? https://github.com/bbcrd/solr-ruby/compare/5b06e66f4e%5E...a76aee983e Best, Duncan http://www.bbc.co.uk/ This e-mail (and any attachments) is confidential and may contain personal views which are not the views of the BBC unless specifically stated. If you have received it in error, please delete it from your system. Do not use, copy or disclose the information in any way nor act in reliance on it and notify the sender immediately. Please note that the BBC monitors e-mails sent or received. Further communication will signify your consent to this. - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (LUCENE-3022) DictionaryCompoundWordTokenFilter Flag onlyLongestMatch has no affect
DictionaryCompoundWordTokenFilter Flag onlyLongestMatch has no affect - Key: LUCENE-3022 URL: https://issues.apache.org/jira/browse/LUCENE-3022 Project: Lucene - Java Issue Type: Bug Components: contrib/analyzers Affects Versions: 3.1, 2.9.4 Reporter: Johann Höchtl Priority: Minor When using the DictionaryCompoundWordTokenFilter with a german dictionary, I got a strange behaviour: The german word streifenbluse (blouse with stripes) was decompounded to streifen (stripe),reifen(tire) which makes no sense at all. I thought the flag onlyLongestMatch would fix this, because streifen is longer than reifen, but it had no effect. So I reviewed the sourcecode and found the problem: [code] protected void decomposeInternal(final Token token) { // Only words longer than minWordSize get processed if (token.length() this.minWordSize) { return; } char[] lowerCaseTermBuffer=makeLowerCaseCopy(token.buffer()); for (int i=0;itoken.length()-this.minSubwordSize;++i) { Token longestMatchToken=null; for (int j=this.minSubwordSize-1;jthis.maxSubwordSize;++j) { if(i+jtoken.length()) { break; } if(dictionary.contains(lowerCaseTermBuffer, i, j)) { if (this.onlyLongestMatch) { if (longestMatchToken!=null) { if (longestMatchToken.length()j) { longestMatchToken=createToken(i,j,token); } } else { longestMatchToken=createToken(i,j,token); } } else { tokens.add(createToken(i,j,token)); } } } if (this.onlyLongestMatch longestMatchToken!=null) { tokens.add(longestMatchToken); } } } [/code] should be changed to [code] protected void decomposeInternal(final Token token) { // Only words longer than minWordSize get processed if (token.termLength() this.minWordSize) { return; } char[] lowerCaseTermBuffer=makeLowerCaseCopy(token.termBuffer()); Token longestMatchToken=null; for (int i=0;itoken.termLength()-this.minSubwordSize;++i) { for (int j=this.minSubwordSize-1;jthis.maxSubwordSize;++j) { if(i+jtoken.termLength()) { break; } if(dictionary.contains(lowerCaseTermBuffer, i, j)) { if (this.onlyLongestMatch) { if (longestMatchToken!=null) { if (longestMatchToken.termLength()j) { longestMatchToken=createToken(i,j,token); } } else { longestMatchToken=createToken(i,j,token); } } else { tokens.add(createToken(i,j,token)); } } } } if (this.onlyLongestMatch longestMatchToken!=null) { tokens.add(longestMatchToken); } } [/code] So, that only the longest token is really indexed and the onlyLongestMatch Flag makes sense. -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
Numerical ids for terms?
Hi -- has there been any effort to create a numerical representation of Lucene indices. That is, to use the Lucene Directory backend as a large term-document matrix at index level. As this would require bijective mapping between terms (per-field, as customary in Lucene) and a numerical index (integer, monotonous from 0 to numTerms()-1), I guess this requires some some special modifications to the Lucene core. Another interesting feature would be to use Lucene's Directory backend for storage of large dense matrices, for instance to data-mining tasks from within Lucene. Any suggestions? Best regards and thanks gregor - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
Re: Numerical ids for terms?
On Tue, Apr 12, 2011 at 13:41, Gregor Heinrich gre...@arbylon.net wrote: Hi -- has there been any effort to create a numerical representation of Lucene indices. That is, to use the Lucene Directory backend as a large term-document matrix at index level. As this would require bijective mapping between terms (per-field, as customary in Lucene) and a numerical index (integer, monotonous from 0 to numTerms()-1), I guess this requires some some special modifications to the Lucene core. Lucene index already provides term - id mapping in some form. Another interesting feature would be to use Lucene's Directory backend for storage of large dense matrices, for instance to data-mining tasks from within Lucene. Lucene's Directory is a dumb abstraction for random-access named write-once byte streams. It doesn't add /any/ value over mmap. Any suggestions? *troll mode on* Use numpy/scipy? :) -- Kirill Zakharenko/Кирилл Захаренко E-Mail/Jabber: ear...@gmail.com Phone: +7 (495) 683-567-4 ICQ: 104465785 - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-3017) FST should differentiate between final vs non-final stop nodes
[ https://issues.apache.org/jira/browse/LUCENE-3017?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13018819#comment-13018819 ] Michael McCandless commented on LUCENE-3017: I hear you :) I think Lucene's needs put pressure on the traditional FST bounds... so we need to stretch things a bit. FST should differentiate between final vs non-final stop nodes -- Key: LUCENE-3017 URL: https://issues.apache.org/jira/browse/LUCENE-3017 Project: Lucene - Java Issue Type: Improvement Reporter: Michael McCandless Assignee: Michael McCandless Priority: Minor Fix For: 4.0 Attachments: LUCENE-3017.patch I'm breaking out this one improvement from LUCENE-2948... Currently, if a node has no outgoing edges (a stop node) the FST forcefully marks this as a final node, but it need not do this. Ie, whether that node is final or not should be orthogonal to whether it has arcs leaving or not. -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (LUCENE-3023) Land DWPT on trunk
Land DWPT on trunk -- Key: LUCENE-3023 URL: https://issues.apache.org/jira/browse/LUCENE-3023 Project: Lucene - Java Issue Type: Task Affects Versions: CSF branch, 4.0 Reporter: Simon Willnauer Fix For: 4.0 With LUCENE-2956 we have resolved the last remaining issue for LUCENE-2324 so we can proceed landing the DWPT development on trunk soon. I think one of the bigger issues here is to make sure that all JavaDocs for IW etc. are still correct though. I will start going through that first. -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Assigned] (LUCENE-3023) Land DWPT on trunk
[ https://issues.apache.org/jira/browse/LUCENE-3023?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Simon Willnauer reassigned LUCENE-3023: --- Assignee: Simon Willnauer Land DWPT on trunk -- Key: LUCENE-3023 URL: https://issues.apache.org/jira/browse/LUCENE-3023 Project: Lucene - Java Issue Type: Task Affects Versions: CSF branch, 4.0 Reporter: Simon Willnauer Assignee: Simon Willnauer Fix For: 4.0 With LUCENE-2956 we have resolved the last remaining issue for LUCENE-2324 so we can proceed landing the DWPT development on trunk soon. I think one of the bigger issues here is to make sure that all JavaDocs for IW etc. are still correct though. I will start going through that first. -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Resolved] (LUCENE-3021) randomize skipInterval in tests
[ https://issues.apache.org/jira/browse/LUCENE-3021?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Robert Muir resolved LUCENE-3021. - Resolution: Fixed Fix Version/s: 4.0 Committed revision 1091408. randomize skipInterval in tests --- Key: LUCENE-3021 URL: https://issues.apache.org/jira/browse/LUCENE-3021 Project: Lucene - Java Issue Type: Test Reporter: Robert Muir Fix For: 4.0 Attachments: LUCENE-3021.patch, LUCENE-3021.patch we probably don't test the multi-level skipping very well, but skipInterval etc is now private to the codec, so for better test coverage we should parameterize it to the postings writers, and randomize it via mockrandomcodec. -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (LUCENE-3024) If index has more than Integer.MAX_VALUE terms, seeking can it AIOOBE due to long/int overflow
If index has more than Integer.MAX_VALUE terms, seeking can it AIOOBE due to long/int overflow -- Key: LUCENE-3024 URL: https://issues.apache.org/jira/browse/LUCENE-3024 Project: Lucene - Java Issue Type: Bug Reporter: Michael McCandless Assignee: Michael McCandless Fix For: 3.1.1, 3.2, 4.0 Tom hit a new long/int overflow case: http://markmail.org/thread/toyl2ujcl4suqvf3 This is a regression, in 3.1, introduced with LUCENE-2075. Worse, our Test2BTerms failed to catch this, so I've fixed that test to show the failure. -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (LUCENE-3025) TestIndexWriterExceptions fails on windows (2)
TestIndexWriterExceptions fails on windows (2) -- Key: LUCENE-3025 URL: https://issues.apache.org/jira/browse/LUCENE-3025 Project: Lucene - Java Issue Type: Bug Reporter: Robert Muir Note: this is a different problem than LUCENE-2991 (I disabled the assert for that problem). -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-3025) TestIndexWriterExceptions fails on windows (2)
[ https://issues.apache.org/jira/browse/LUCENE-3025?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13018849#comment-13018849 ] Robert Muir commented on LUCENE-3025: - junit-sequential: [junit] Testsuite: org.apache.lucene.index.TestIndexWriterExceptions [junit] Tests run: 1, Failures: 1, Errors: 0, Time elapsed: 0.681 sec [junit] [junit] - Standard Error - [junit] NOTE: reproduce with: ant test -Dtestcase=TestIndexWriterExceptions -Dtestmethod=testExceptionsDuringCommit -Dtests.seed=-2008642232753917842:-38943 12978511486973 -Dtests.codec=MockRandom [junit] NOTE: test params are: codec=MockRandom, locale=sl, timezone=Asia/Pyongyang [junit] NOTE: all tests run in this JVM: [junit] [TestIndexWriterExceptions] [junit] NOTE: Windows Vista 6.0 x86/Sun Microsystems Inc. 1.6.0_23 (32-bit)/cpus=4,threads=1,free=14526984,total=32202752 [junit] - --- [junit] Testcase: testExceptionsDuringCommit(org.apache.lucene.index.TestIndexWriterExceptions): FAILED [junit] [junit] junit.framework.AssertionFailedError: [junit] at org.apache.lucene.util.LuceneTestCase$LuceneTestCaseRunner.runChild(LuceneTestCase.java:1232) [junit] at org.apache.lucene.util.LuceneTestCase$LuceneTestCaseRunner.runChild(LuceneTestCase.java:1160) [junit] at org.apache.lucene.index.TestIndexWriterExceptions.testExceptionsDuringCommit(TestIndexWriterExceptions.java:867) [junit] [junit] [junit] Test org.apache.lucene.index.TestIndexWriterExceptions FAILED TestIndexWriterExceptions fails on windows (2) -- Key: LUCENE-3025 URL: https://issues.apache.org/jira/browse/LUCENE-3025 Project: Lucene - Java Issue Type: Bug Reporter: Robert Muir Note: this is a different problem than LUCENE-2991 (I disabled the assert for that problem). -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-3025) TestIndexWriterExceptions fails on windows (2)
[ https://issues.apache.org/jira/browse/LUCENE-3025?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13018864#comment-13018864 ] Simon Willnauer commented on LUCENE-3025: - bq. Note: this is a different problem than LUCENE-2991 (I disabled the assert for that problem). Robert, I think this is the same problem as LUCENE-2991 has. I actually wonder if that problem still occurs if we close the directory first and then reopen another one to see if the files are still there. I don't have a windows ready to try it myself but what I would be interested in is if this situation occurs are there still open files around? I mean if we are holding on to the files and this test doesn't fail on unix then we have a problem with MockDirectoryWrapper since it is supposed to 'act' like windows in that respect. When I look at w.rollback() I see a possibility that this test can fail under windows since it is calling closeInternal(false) that does not wait for merges. So some of the files can still be referenced by the merger. I wonder if it makes sense to call w.close() to make sure all files are done before checking if the files have been removed. TestIndexWriterExceptions fails on windows (2) -- Key: LUCENE-3025 URL: https://issues.apache.org/jira/browse/LUCENE-3025 Project: Lucene - Java Issue Type: Bug Reporter: Robert Muir Note: this is a different problem than LUCENE-2991 (I disabled the assert for that problem). -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
Re: Numerical ids for terms?
Thanks for the quick response. Please be a bit more concrete than some form of term--id mapping: Do you refer to subclassing SegmentReader with the appropriate Map implementation or is there a tested structure in the existing API that I've overseen? Regarding a Directory abstraction backed by a memory mapping API, my question refers to using Lucene API because even if may be perceived dumb, it hides a lot of boilerplate code. Are there any efforts going on regarding this? Cheers gregor On 4/12/11 1:21 PM, Earwin Burrfoot wrote: On Tue, Apr 12, 2011 at 13:41, Gregor Heinrichgre...@arbylon.net wrote: Hi -- has there been any effort to create a numerical representation of Lucene indices. That is, to use the Lucene Directory backend as a large term-document matrix at index level. As this would require bijective mapping between terms (per-field, as customary in Lucene) and a numerical index (integer, monotonous from 0 to numTerms()-1), I guess this requires some some special modifications to the Lucene core. Lucene index already provides term- id mapping in some form. Another interesting feature would be to use Lucene's Directory backend for storage of large dense matrices, for instance to data-mining tasks from within Lucene. Lucene's Directory is a dumb abstraction for random-access named write-once byte streams. It doesn't add /any/ value over mmap. Any suggestions? *troll mode on* Use numpy/scipy? :) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Resolved] (SOLR-2335) FunctionQParser can't handle fieldnames containing whitespace
[ https://issues.apache.org/jira/browse/SOLR-2335?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Hoss Man resolved SOLR-2335. Resolution: Fixed Fix Version/s: 4.0 Assignee: Hoss Man Committed revision 1091516. FunctionQParser can't handle fieldnames containing whitespace - Key: SOLR-2335 URL: https://issues.apache.org/jira/browse/SOLR-2335 Project: Solr Issue Type: Bug Affects Versions: 1.4.1 Reporter: Miriam Doelle Assignee: Hoss Man Priority: Minor Fix For: 4.0 Attachments: SOLR-2335.patch, SOLR-2335.patch FunctionQParser has some simplistic assumptions about what types of field names it will deal with, in particular it can't deal with field names containing whitespaces. -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-3018) Lucene Native Directory implementation need automated build
[ https://issues.apache.org/jira/browse/LUCENE-3018?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13018951#comment-13018951 ] Varun Thacker commented on LUCENE-3018: --- This is a small ant task I wrote to build a shared library from a .cpp file. http://pastebin.com/diTXru9w I will update the code to add options for command line parameters which are required to build NativePosixUtil.cpp Lucene Native Directory implementation need automated build --- Key: LUCENE-3018 URL: https://issues.apache.org/jira/browse/LUCENE-3018 Project: Lucene - Java Issue Type: Wish Components: Build Affects Versions: 4.0 Reporter: Simon Willnauer Assignee: Varun Thacker Priority: Minor Fix For: 4.0 Currently the native directory impl in contrib/misc require manual action to compile the c code (partially) documented in https://svn.apache.org/repos/asf/lucene/dev/trunk/lucene/contrib/misc/src/java/overview.html yet it would be nice if we had an ant task and documentation for all platforms how to compile them and set up the prerequisites. -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Resolved] (LUCENE-3017) FST should differentiate between final vs non-final stop nodes
[ https://issues.apache.org/jira/browse/LUCENE-3017?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Michael McCandless resolved LUCENE-3017. Resolution: Fixed FST should differentiate between final vs non-final stop nodes -- Key: LUCENE-3017 URL: https://issues.apache.org/jira/browse/LUCENE-3017 Project: Lucene - Java Issue Type: Improvement Reporter: Michael McCandless Assignee: Michael McCandless Priority: Minor Fix For: 4.0 Attachments: LUCENE-3017.patch I'm breaking out this one improvement from LUCENE-2948... Currently, if a node has no outgoing edges (a stop node) the FST forcefully marks this as a final node, but it need not do this. Ie, whether that node is final or not should be orthogonal to whether it has arcs leaving or not. -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (SOLR-2466) CommonsHttpSolrServer will retry a query even if _maxRetries is 0
CommonsHttpSolrServer will retry a query even if _maxRetries is 0 - Key: SOLR-2466 URL: https://issues.apache.org/jira/browse/SOLR-2466 Project: Solr Issue Type: Bug Components: clients - java Affects Versions: 3.1, 1.4.1, 4.0 Reporter: Tomás Fernández Löbbe The HttpClient library used by CommonsHttpSolrServer will retry by default 3 times a request that failed on the server side, even if the _maxRetries field of CommonsHttpSolrServer is set to 0. The retry count should be managed in just one place and CommonsHttpSolrServer seems to be the right one. CommonsHttpSolrServer should override that HttpClient default to 0 retries, and manage the retry count with the value of the field _maxRetries. -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-3024) If index has more than Integer.MAX_VALUE terms, seeking can it AIOOBE due to long/int overflow
[ https://issues.apache.org/jira/browse/LUCENE-3024?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13019003#comment-13019003 ] Michael McCandless commented on LUCENE-3024: Fixed in 3.2, 4.0. I'm leaving this open in case we ever release 3.1.1. If index has more than Integer.MAX_VALUE terms, seeking can it AIOOBE due to long/int overflow -- Key: LUCENE-3024 URL: https://issues.apache.org/jira/browse/LUCENE-3024 Project: Lucene - Java Issue Type: Bug Reporter: Michael McCandless Assignee: Michael McCandless Fix For: 3.1.1, 3.2, 4.0 Attachments: LUCENE-3024.patch Tom hit a new long/int overflow case: http://markmail.org/thread/toyl2ujcl4suqvf3 This is a regression, in 3.1, introduced with LUCENE-2075. Worse, our Test2BTerms failed to catch this, so I've fixed that test to show the failure. -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-2956) Support updateDocument() with DWPTs
[ https://issues.apache.org/jira/browse/LUCENE-2956?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13019019#comment-13019019 ] Michael Busch commented on LUCENE-2956: --- Cool patch! :) Though it worries me a little how complex the whole delete/update logic is becoming (not only the part this patch adds). Originally we decided to not go with sequenceIDs partly because we thought the implementation might be too complex, but I think it'd be simpler than the current approach that uses bits. The sequenceIDs approach we had in the beginning was also completely lockless in a very very simple way. Anyway, I have yet to take a closer look here. Just something that might be worth discussing. Support updateDocument() with DWPTs --- Key: LUCENE-2956 URL: https://issues.apache.org/jira/browse/LUCENE-2956 Project: Lucene - Java Issue Type: Bug Components: Index Affects Versions: Realtime Branch Reporter: Michael Busch Assignee: Simon Willnauer Priority: Minor Fix For: Realtime Branch Attachments: LUCENE-2956.patch With separate DocumentsWriterPerThreads (DWPT) it can currently happen that the delete part of an updateDocument() is flushed and committed separately from the corresponding new document. We need to make sure that updateDocument() is always an atomic operation from a IW.commit() and IW.getReader() perspective. See LUCENE-2324 for more details. -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-2466) CommonsHttpSolrServer will retry a query even if _maxRetries is 0
[ https://issues.apache.org/jira/browse/SOLR-2466?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13019021#comment-13019021 ] Yonik Seeley commented on SOLR-2466: Hmmm, that's interesting. Does anyone know why we (CommonsHttpSolrServer) do retries when HttpClient already does them? Is there an advantage to doing it ourselves? CommonsHttpSolrServer will retry a query even if _maxRetries is 0 - Key: SOLR-2466 URL: https://issues.apache.org/jira/browse/SOLR-2466 Project: Solr Issue Type: Bug Components: clients - java Affects Versions: 1.4.1, 3.1, 4.0 Reporter: Tomás Fernández Löbbe The HttpClient library used by CommonsHttpSolrServer will retry by default 3 times a request that failed on the server side, even if the _maxRetries field of CommonsHttpSolrServer is set to 0. The retry count should be managed in just one place and CommonsHttpSolrServer seems to be the right one. CommonsHttpSolrServer should override that HttpClient default to 0 retries, and manage the retry count with the value of the field _maxRetries. -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-2466) CommonsHttpSolrServer will retry a query even if _maxRetries is 0
[ https://issues.apache.org/jira/browse/SOLR-2466?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13019028#comment-13019028 ] Tomás Fernández Löbbe commented on SOLR-2466: - Not sure why Solr does it on CommonsHttpSolrServer. I do think is important to be able to specify the exact number of retries. CommonsHttpSolrServer will retry a query even if _maxRetries is 0 - Key: SOLR-2466 URL: https://issues.apache.org/jira/browse/SOLR-2466 Project: Solr Issue Type: Bug Components: clients - java Affects Versions: 1.4.1, 3.1, 4.0 Reporter: Tomás Fernández Löbbe The HttpClient library used by CommonsHttpSolrServer will retry by default 3 times a request that failed on the server side, even if the _maxRetries field of CommonsHttpSolrServer is set to 0. The retry count should be managed in just one place and CommonsHttpSolrServer seems to be the right one. CommonsHttpSolrServer should override that HttpClient default to 0 retries, and manage the retry count with the value of the field _maxRetries. -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-2466) CommonsHttpSolrServer will retry a query even if _maxRetries is 0
[ https://issues.apache.org/jira/browse/SOLR-2466?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13019030#comment-13019030 ] Hoss Man commented on SOLR-2466: I haven't checked hte code but if i remember correctly (from another project) HttpClient and it's RetryHandler hook are only used when dealing with *network* failures -- ie: connection refused, connection timeout, connection aborted. If a request is a success at the TCP layer, but a failure at the HTTP layer (ie: 500) then you need your own retry logic external to the HttpClient that may be what SolrJ is doing, to account for transient errors (ie: trying to add during a blocking commit or something like that) CommonsHttpSolrServer will retry a query even if _maxRetries is 0 - Key: SOLR-2466 URL: https://issues.apache.org/jira/browse/SOLR-2466 Project: Solr Issue Type: Bug Components: clients - java Affects Versions: 1.4.1, 3.1, 4.0 Reporter: Tomás Fernández Löbbe The HttpClient library used by CommonsHttpSolrServer will retry by default 3 times a request that failed on the server side, even if the _maxRetries field of CommonsHttpSolrServer is set to 0. The retry count should be managed in just one place and CommonsHttpSolrServer seems to be the right one. CommonsHttpSolrServer should override that HttpClient default to 0 retries, and manage the retry count with the value of the field _maxRetries. -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
An IDF variation with penalty for very rare terms
Excuse me for somewhat of an offtopic, but have anybody ever seen/used -subj- ? Something that looks like like http://dl.dropbox.com/u/920413/IDFplusplus.png Traditional log(N/x) tail, but when nearing zero freq, instead of going to +inf you do a nice round bump (with controlled height/location/sharpness) and drop down to -inf (or zero). Should be cool when doing cosine-measure(or something comparable)-based document comparisons (eg. in a more like this query, to mention Lucene at least once :) ), over dirty data. Rationale is that - most good, discriminating terms are found in at least a certain percentage of your documents, but there are lots of mostly unique crapterms, which at some collection sizes stop being strictly unique and with IDF's help explode your scores. -- Kirill Zakharenko/Кирилл Захаренко E-Mail/Jabber: ear...@gmail.com Phone: +7 (495) 683-567-4 ICQ: 104465785 - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[HUDSON] Lucene-Solr-tests-only-trunk - Build # 7045 - Failure
Build: https://hudson.apache.org/hudson/job/Lucene-Solr-tests-only-trunk/7045/ 1 tests failed. REGRESSION: org.apache.solr.cloud.CloudStateUpdateTest.testCoreRegistration Error Message: null Stack Trace: junit.framework.AssertionFailedError: at org.apache.lucene.util.LuceneTestCase$LuceneTestCaseRunner.runChild(LuceneTestCase.java:1232) at org.apache.lucene.util.LuceneTestCase$LuceneTestCaseRunner.runChild(LuceneTestCase.java:1160) at org.apache.solr.cloud.CloudStateUpdateTest.testCoreRegistration(CloudStateUpdateTest.java:227) Build Log (for compile errors): [...truncated 8822 lines...] - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (SOLR-1566) Allow components to add fields to outgoing documents
[ https://issues.apache.org/jira/browse/SOLR-1566?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jan Høydahl updated SOLR-1566: -- Attachment: SOLR-1566-PageTool.patch Also encountered the Velocity bug. Fixed it by patching PageTool (attached). Allow components to add fields to outgoing documents Key: SOLR-1566 URL: https://issues.apache.org/jira/browse/SOLR-1566 Project: Solr Issue Type: New Feature Components: search Reporter: Noble Paul Fix For: 4.0 Attachments: SOLR-1566-DocTransformer.patch, SOLR-1566-DocTransformer.patch, SOLR-1566-DocTransformer.patch, SOLR-1566-DocTransformer.patch, SOLR-1566-DocTransformer.patch, SOLR-1566-DocTransformer.patch, SOLR-1566-PageTool.patch, SOLR-1566-gsi.patch, SOLR-1566-rm.patch, SOLR-1566-rm.patch, SOLR-1566-rm.patch, SOLR-1566-rm.patch, SOLR-1566-rm.patch, SOLR-1566.patch, SOLR-1566.patch, SOLR-1566.patch, SOLR-1566.patch, SOLR-1566_parsing.patch Currently it is not possible for components to add fields to outgoing documents which are not in the the stored fields of the document. This makes it cumbersome to add computed fields/metadata . -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[HUDSON] Lucene-Solr-tests-only-trunk - Build # 7051 - Failure
Build: https://hudson.apache.org/hudson/job/Lucene-Solr-tests-only-trunk/7051/ 9 tests failed. REGRESSION: org.apache.solr.cloud.CloudStateUpdateTest.testCoreRegistration Error Message: null Stack Trace: org.apache.solr.common.cloud.ZooKeeperException: at org.apache.solr.cloud.ZkController.init(ZkController.java:280) at org.apache.solr.cloud.ZkController.init(ZkController.java:133) at org.apache.solr.core.CoreContainer.initZooKeeper(CoreContainer.java:164) at org.apache.solr.core.CoreContainer.load(CoreContainer.java:333) at org.apache.solr.core.CoreContainer$Initializer.initialize(CoreContainer.java:242) at org.apache.solr.cloud.CloudStateUpdateTest.setUp(CloudStateUpdateTest.java:122) at org.apache.lucene.util.LuceneTestCase$LuceneTestCaseRunner.runChild(LuceneTestCase.java:1232) at org.apache.lucene.util.LuceneTestCase$LuceneTestCaseRunner.runChild(LuceneTestCase.java:1160) Caused by: org.apache.zookeeper.KeeperException$ConnectionLossException: KeeperErrorCode = ConnectionLoss for /live_nodes at org.apache.zookeeper.KeeperException.create(KeeperException.java:90) at org.apache.zookeeper.KeeperException.create(KeeperException.java:42) at org.apache.zookeeper.ZooKeeper.setData(ZooKeeper.java:1038) at org.apache.solr.common.cloud.SolrZkClient.setData(SolrZkClient.java:224) at org.apache.solr.common.cloud.SolrZkClient.makePath(SolrZkClient.java:354) at org.apache.solr.common.cloud.SolrZkClient.makePath(SolrZkClient.java:308) at org.apache.solr.common.cloud.SolrZkClient.makePath(SolrZkClient.java:290) at org.apache.solr.common.cloud.SolrZkClient.makePath(SolrZkClient.java:255) at org.apache.solr.cloud.ZkController.init(ZkController.java:275) REGRESSION: org.apache.solr.core.TestXIncludeConfig.testXInclude Error Message: org.apache.solr.common.cloud.ZooKeeperException: Stack Trace: java.lang.RuntimeException: org.apache.solr.common.cloud.ZooKeeperException: at org.apache.solr.util.TestHarness.init(TestHarness.java:153) at org.apache.solr.util.TestHarness.init(TestHarness.java:135) at org.apache.solr.util.TestHarness.init(TestHarness.java:125) at org.apache.solr.util.AbstractSolrTestCase.setUp(AbstractSolrTestCase.java:132) at org.apache.solr.core.TestXIncludeConfig.setUp(TestXIncludeConfig.java:55) at org.apache.lucene.util.LuceneTestCase$LuceneTestCaseRunner.runChild(LuceneTestCase.java:1232) at org.apache.lucene.util.LuceneTestCase$LuceneTestCaseRunner.runChild(LuceneTestCase.java:1160) Caused by: org.apache.solr.common.cloud.ZooKeeperException: at org.apache.solr.core.CoreContainer.initZooKeeper(CoreContainer.java:183) at org.apache.solr.util.TestHarness$Initializer$1.init(TestHarness.java:184) at org.apache.solr.util.TestHarness$Initializer.initialize(TestHarness.java:179) at org.apache.solr.util.TestHarness.init(TestHarness.java:140) Caused by: java.util.concurrent.TimeoutException: Could not connect to ZooKeeper 127.0.0.1:25804/solr within 5000 ms at org.apache.solr.common.cloud.ConnectionManager.waitForConnected(ConnectionManager.java:124) at org.apache.solr.common.cloud.SolrZkClient.init(SolrZkClient.java:121) at org.apache.solr.common.cloud.SolrZkClient.init(SolrZkClient.java:69) at org.apache.solr.cloud.ZkController.init(ZkController.java:104) at org.apache.solr.core.CoreContainer.initZooKeeper(CoreContainer.java:164) FAILED: junit.framework.TestSuite.org.apache.solr.handler.component.QueryElevationComponentTest Error Message: org.apache.solr.common.cloud.ZooKeeperException: Stack Trace: java.lang.RuntimeException: org.apache.solr.common.cloud.ZooKeeperException: at org.apache.solr.util.TestHarness.init(TestHarness.java:153) at org.apache.solr.util.TestHarness.init(TestHarness.java:135) at org.apache.solr.util.TestHarness.init(TestHarness.java:125) at org.apache.solr.SolrTestCaseJ4.initCore(SolrTestCaseJ4.java:238) at org.apache.solr.SolrTestCaseJ4.initCore(SolrTestCaseJ4.java:101) at org.apache.solr.SolrTestCaseJ4.initCore(SolrTestCaseJ4.java:89) at org.apache.solr.handler.component.QueryElevationComponentTest.beforeClass(QueryElevationComponentTest.java:48) Caused by: org.apache.solr.common.cloud.ZooKeeperException: at org.apache.solr.core.CoreContainer.initZooKeeper(CoreContainer.java:183) at org.apache.solr.util.TestHarness$Initializer$1.init(TestHarness.java:184) at org.apache.solr.util.TestHarness$Initializer.initialize(TestHarness.java:179) at org.apache.solr.util.TestHarness.init(TestHarness.java:140) Caused by: java.util.concurrent.TimeoutException: Could not connect to ZooKeeper 127.0.0.1:25804/solr within 5000 ms at
[HUDSON] Lucene-trunk - Build # 1528 - Still Failing
Build: https://hudson.apache.org/hudson/job/Lucene-trunk/1528/ 1 tests failed. REGRESSION: org.apache.lucene.index.TestNRTThreads.testNRTThreads Error Message: Some threads threw uncaught exceptions! Stack Trace: junit.framework.AssertionFailedError: Some threads threw uncaught exceptions! at org.apache.lucene.util.LuceneTestCase$LuceneTestCaseRunner.runChild(LuceneTestCase.java:1232) at org.apache.lucene.util.LuceneTestCase$LuceneTestCaseRunner.runChild(LuceneTestCase.java:1160) at org.apache.lucene.util.LuceneTestCase.tearDown(LuceneTestCase.java:521) Build Log (for compile errors): [...truncated 11900 lines...] - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[HUDSON] Lucene-Solr-tests-only-3.x - Build # 7043 - Failure
Build: https://hudson.apache.org/hudson/job/Lucene-Solr-tests-only-3.x/7043/ 1 tests failed. REGRESSION: org.apache.solr.client.solrj.TestLBHttpSolrServer.testReliability Error Message: No live SolrServers available to handle this request Stack Trace: org.apache.solr.client.solrj.SolrServerException: No live SolrServers available to handle this request at org.apache.solr.client.solrj.impl.LBHttpSolrServer.request(LBHttpSolrServer.java:222) at org.apache.solr.client.solrj.request.QueryRequest.process(QueryRequest.java:89) at org.apache.solr.client.solrj.SolrServer.query(SolrServer.java:118) at org.apache.solr.client.solrj.TestLBHttpSolrServer.waitForServer(TestLBHttpSolrServer.java:189) at org.apache.solr.client.solrj.TestLBHttpSolrServer.testReliability(TestLBHttpSolrServer.java:182) at org.apache.lucene.util.LuceneTestCase$LuceneTestCaseRunner.runChild(LuceneTestCase.java:1082) at org.apache.lucene.util.LuceneTestCase$LuceneTestCaseRunner.runChild(LuceneTestCase.java:1010) Caused by: org.apache.solr.client.solrj.SolrServerException: java.net.SocketTimeoutException: Read timed out at org.apache.solr.client.solrj.impl.CommonsHttpSolrServer.request(CommonsHttpSolrServer.java:484) at org.apache.solr.client.solrj.impl.CommonsHttpSolrServer.request(CommonsHttpSolrServer.java:245) at org.apache.solr.client.solrj.impl.LBHttpSolrServer.request(LBHttpSolrServer.java:206) Caused by: java.net.SocketTimeoutException: Read timed out at java.net.SocketInputStream.socketRead0(Native Method) at java.net.SocketInputStream.read(SocketInputStream.java:146) at java.io.BufferedInputStream.fill(BufferedInputStream.java:235) at java.io.BufferedInputStream.read(BufferedInputStream.java:254) at org.apache.commons.httpclient.HttpParser.readRawLine(HttpParser.java:78) at org.apache.commons.httpclient.HttpParser.readLine(HttpParser.java:106) at org.apache.commons.httpclient.HttpConnection.readLine(HttpConnection.java:1116) at org.apache.commons.httpclient.MultiThreadedHttpConnectionManager$HttpConnectionAdapter.readLine(MultiThreadedHttpConnectionManager.java:1413) at org.apache.commons.httpclient.HttpMethodBase.readStatusLine(HttpMethodBase.java:1973) at org.apache.commons.httpclient.HttpMethodBase.readResponse(HttpMethodBase.java:1735) at org.apache.commons.httpclient.HttpMethodBase.execute(HttpMethodBase.java:1098) at org.apache.commons.httpclient.HttpMethodDirector.executeWithRetry(HttpMethodDirector.java:398) at org.apache.commons.httpclient.HttpMethodDirector.executeMethod(HttpMethodDirector.java:171) at org.apache.commons.httpclient.HttpClient.executeMethod(HttpClient.java:397) at org.apache.commons.httpclient.HttpClient.executeMethod(HttpClient.java:323) at org.apache.solr.client.solrj.impl.CommonsHttpSolrServer.request(CommonsHttpSolrServer.java:428) Build Log (for compile errors): [...truncated 10823 lines...] - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
Re: Patch for http_proxy support in solr-ruby client
Hi, Hm, maybe you are asking where solr-ruby actually lives and is being developed? I'm not sure. I see it under solr/client/ruby/solr-ruby (no new development in ages?), but I also see an *active* solr-ruby fork over on https://github.com/bbcrd/solr-ruby . So if you want to contribute to solr-ruby on Github, get yourself a Github account, fork that solr-ruby, make your change, and submit it via the pull request. This is separate from Solr @ Apache. Otis Sematext :: http://sematext.com/ :: Solr - Lucene - Nutch Lucene ecosystem search :: http://search-lucene.com/ - Original Message From: Duncan Robertson duncan.robert...@bbc.co.uk To: dev@lucene.apache.org Sent: Tue, April 12, 2011 4:36:17 AM Subject: Patch for http_proxy support in solr-ruby client Hi, I have a patch for adding http_proxy support to the solr-ruby client. I thought the project was managed via Github, but this turns out not to be the case. It the process the same as for Solr itself? https://github.com/bbcrd/solr-ruby/compare/5b06e66f4e%5E...a76aee983e Best, Duncan http://www.bbc.co.uk/ This e-mail (and any attachments) is confidential and may contain personal views which are not the views of the BBC unless specifically stated. If you have received it in error, please delete it from your system. Do not use, copy or disclose the information in any way nor act in reliance on it and notify the sender immediately. Please note that the BBC monitors e-mails sent or received. Further communication will signify your consent to this. - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org