Re: Revisions to Incubator proposal
On Thu, Jul 1, 2010 at 10:08 AM, Marvin Humphrey mar...@rectangular.com wrote: There is a possibility that this proposal won't be accepted. But if that happens, we have a contingency plan: leave Apache, yet continue to operate according to many of the same procedures and values, possibly returning with a new proposal at some later date. We should seek to exploit any critiques that we receive during the proposal process to help guide us, regardless of the outcome -- just as we have attempted to make best use of the Lucene PMC's advice. Having slept on the matter, here are some thoughts on the current proposal. The first half of the Rationale seems fine... There is great hunger for a search engine library in the mode of Lucene which is accessible from various dynamic languages, and for one accessible from pure C. Individuals naturally wish to code in their language of choice. Organizations which do not have significant Java expertise may not want to support Java strictly for the sake of running a Lucene instance. Native applications can be launched much more quickly than JVMs. Lucy will meet all these demands. I like this, but might stop after the first two sentences. I think they're the stronger reasons. If you wanted to go further, I'd mention the ways in which C has traditionally been the language of Unix server programming, and allows one to take full advantage of the machine's potential. I think this is essentially what made Apache a great web server: a brilliant architecture combined with the clarity of C. Can you picture writing mod_php, mod_perl, or mod_python in Java? It's certainly not the only way to go, but fills a niche that Lucene never will. We acknowledge that Apache seems like a natural home for Lucy given that it is also the home of Lucene, and speculate that this may have been on the minds of the Lucene PMC when Lucy was green-lighted as a sub-project. More importantly, though, the Lucy development community strongly believes that The Apache Way is right for Lucy. I agree that this part is somewhat weak. And while I really like Apache, I'm not sure it's the only way for Lucy. For example, I also really like the SQLite way. While Lucy certainly can take advantage of the Apache infrastructure, I'm not sure it really needs it. What it needs is more people writing code for it, and the best way to achieve this is probably to get more technically proficient users. If the code is clear (which it is) and the architecture is graspable (needs improvement?) some significant percentage of these users will contribute. The one thing I don't think we've done well (and this is my fault) is handle releases and backwards compatibility. Father Chrysostomos put in an awful lot of work creating a subclassable Highlighter, but later releases of KS broke back compat on him. Nevertheless, I think we have learned from how that played out, and that the backwards compatibility policy we arrived at last year goes a long way towards solving those problems. I think you're doing fine on backwards compatibility, and if anything you're spending too much time on it. Instead of worrying about not breaking things that are old, spend the effort making it easier to write things that are new. Let's make the dogfood tasty, and start writing 'extensions' instead of trying to include everything in core. If it's easy enough and useful enough, someone will port the old to the new. In fact, that would be an excellent project for a beginner to familiarize themselves with the code base. Lastly, it would be nice to cover our contingency plan of growing the community and coming back with a bigger committer list at some later date. However, I think that may arise naturally during the discussion, and it's probably too big a topic to squeeze in. I don't know if you need to discuss this. And once we have more developers, do we really need to come back? I mean, I really like Apache, but Trac plus taking the best points of the Incubator approach might offer 90% of the benefit with a lot less overhead. Which doesn't mean Lucy wouldn't benefit from being included, but I wouldn't precommit to returning at a later date if they don't want us. --nate
[Lucy Wiki] Update of LucyIncubatorProposal by Nathan Kurz
Dear Wiki user, You have subscribed to a wiki page or wiki category on Lucy Wiki for change notification. The LucyIncubatorProposal page has been changed by NathanKurz. http://wiki.apache.org/lucy/LucyIncubatorProposal?action=diffrev1=35rev2=36 -- === Core Developers === * Marvin Humphrey is the project founder of KinoSearch, and co-founded the existing Lucy sub-project. He is presently employed by Eventful, Inc. * Peter Karman has contributed to several open source projects since 2001, including being a committer at http://swish-e.org/ (a search engine), http://code.google.com/p/rose/ (an ORM) and http://catalyst.perl.org/ (web framework). He is employed by American Public Media. - * Nathan Kurz has participated in numerous open source projects and has been a KinoSearch committer since 2007. He is currently Chief Flavor Engineer of Scream Sorbet, and writes software in his copious free time. + * Nathan Kurz is excited by the intersection of search and recommendations, and has been a KinoSearch committer since 2007. As the owner of Scream Sorbet (http://screamsorbet.com), he divides his time between code and fruit. === Alignment === One Apache value which is particularly cherished by the Lucy community is codebase transparency. We have developed institutions which enable us to measure and maximize usability (see [http://wiki.apache.org/lucy/BrainLog]), and we feel strongly that the bindings for Lucy must present APIs and documentation which are idiomatic to the host language culture so that end users can consume our work as easily as possible.
[jira] Created: (LUCENE-2523) if index is too old you should hit an exception saying so
if index is too old you should hit an exception saying so - Key: LUCENE-2523 URL: https://issues.apache.org/jira/browse/LUCENE-2523 Project: Lucene - Java Issue Type: Bug Components: Index Reporter: Michael McCandless Priority: Minor Fix For: 4.0 If you create an index in 2.3.x (I used demo's IndexFiles) and then try to read it in 4.0.x (I used CheckIndex), you hit a confusing exception like this: {noformat} java.io.IOException: read past EOF at org.apache.lucene.store.BufferedIndexInput.refill(BufferedIndexInput.java:154) at org.apache.lucene.store.BufferedIndexInput.readByte(BufferedIndexInput.java:39) at org.apache.lucene.store.ChecksumIndexInput.readByte(ChecksumIndexInput.java:40) at org.apache.lucene.store.DataInput.readInt(DataInput.java:76) at org.apache.lucene.index.SegmentInfo.init(SegmentInfo.java:171) at org.apache.lucene.index.SegmentInfos.read(SegmentInfos.java:230) at org.apache.lucene.index.SegmentInfos$1.doBody(SegmentInfos.java:269) at org.apache.lucene.index.SegmentInfos$FindSegmentsFile.run(SegmentInfos.java:649) at org.apache.lucene.index.SegmentInfos$FindSegmentsFile.run(SegmentInfos.java:484) at org.apache.lucene.index.SegmentInfos.read(SegmentInfos.java:265) at org.apache.lucene.index.CheckIndex.checkIndex(CheckIndex.java:308) at org.apache.lucene.index.CheckIndex.checkIndex(CheckIndex.java:287) at org.apache.lucene.index.CheckIndex.main(CheckIndex.java:930) {noformat} I think instead we should throw an IndexTooOldException or something like that? -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online. - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
Build failed in Hudson: Solr-trunk #1195
See http://hudson.zones.apache.org/hudson/job/Solr-trunk/1195/changes Changes: [yonik] formatting: make consistent with other clauses [gsingers] fix typo in javadoc [gsingers] LUCENE-1810: added LATENT field selector option -- [...truncated 9858 lines...] [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 9.208 sec [junit] [junit] Testsuite: org.apache.solr.handler.SpellCheckerRequestHandlerTest [junit] Tests run: 5, Failures: 0, Errors: 0, Time elapsed: 66.886 sec [junit] [junit] Testsuite: org.apache.solr.handler.StandardRequestHandlerTest [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 7.217 sec [junit] [junit] Testsuite: org.apache.solr.handler.TestCSVLoader [junit] Tests run: 4, Failures: 0, Errors: 0, Time elapsed: 13.609 sec [junit] [junit] Testsuite: org.apache.solr.handler.TestReplicationHandler [junit] Tests run: 7, Failures: 0, Errors: 0, Time elapsed: 98.945 sec [junit] [junit] - Standard Error - [junit] Jul 2, 2010 9:29:39 AM org.apache.solr.handler.SnapPuller fetchLatestIndex [junit] SEVERE: Master at: http://localhost:55735/solr/replication is not available. Index fetch failed. Exception: Connection refused [junit] Jul 2, 2010 9:29:40 AM org.apache.solr.handler.SnapPuller fetchLatestIndex [junit] SEVERE: Master at: http://localhost:55735/solr/replication is not available. Index fetch failed. Exception: Connection refused [junit] Jul 2, 2010 9:29:57 AM org.apache.solr.handler.SnapPuller fetchLatestIndex [junit] SEVERE: Master at: http://localhost:55757/solr/replication is not available. Index fetch failed. Exception: Connection refused [junit] Jul 2, 2010 9:30:41 AM org.apache.solr.handler.SnapPuller fetchLatestIndex [junit] SEVERE: Master at: http://localhost:55809/solr/replication is not available. Index fetch failed. Exception: Connection refused [junit] Jul 2, 2010 9:30:51 AM org.apache.solr.handler.SnapPuller fetchLatestIndex [junit] SEVERE: Master at: http://localhost:55831/solr/replication is not available. Index fetch failed. Exception: Connection refused [junit] Jul 2, 2010 9:30:52 AM org.apache.solr.handler.SnapPuller fetchLatestIndex [junit] SEVERE: Master at: http://localhost:55831/solr/replication is not available. Index fetch failed. Exception: Connection refused [junit] Jul 2, 2010 9:30:52 AM org.apache.solr.handler.SnapPuller fetchLatestIndex [junit] SEVERE: Master at: http://localhost:55822/solr/replication is not available. Index fetch failed. Exception: Connection refused [junit] Jul 2, 2010 9:30:53 AM org.apache.solr.handler.SnapPuller fetchLatestIndex [junit] SEVERE: Master at: http://localhost:55831/solr/replication is not available. Index fetch failed. Exception: Connection refused [junit] Jul 2, 2010 9:30:54 AM org.apache.solr.handler.SnapPuller fetchLatestIndex [junit] SEVERE: Master at: http://localhost:55831/solr/replication is not available. Index fetch failed. Exception: Connection refused [junit] Jul 2, 2010 9:30:55 AM org.apache.solr.handler.SnapPuller fetchLatestIndex [junit] SEVERE: Master at: http://localhost:55831/solr/replication is not available. Index fetch failed. Exception: Connection refused [junit] Jul 2, 2010 9:30:56 AM org.apache.solr.handler.SnapPuller fetchLatestIndex [junit] SEVERE: Master at: http://localhost:55831/solr/replication is not available. Index fetch failed. Exception: Connection refused [junit] Jul 2, 2010 9:30:57 AM org.apache.solr.handler.SnapPuller fetchLatestIndex [junit] SEVERE: Master at: http://localhost:55831/solr/replication is not available. Index fetch failed. Exception: Connection refused [junit] Jul 2, 2010 9:30:59 AM org.apache.solr.handler.SnapPuller fetchLatestIndex [junit] SEVERE: Master at: http://localhost:55831/solr/replication is not available. Index fetch failed. Exception: Connection refused [junit] - --- [junit] Testsuite: org.apache.solr.handler.XmlUpdateRequestHandlerTest [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 6 sec [junit] [junit] Testsuite: org.apache.solr.handler.admin.LukeRequestHandlerTest [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 12.173 sec [junit] [junit] Testsuite: org.apache.solr.handler.admin.SystemInfoHandlerTest [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 0.5 sec [junit] [junit] Testsuite: org.apache.solr.handler.component.DistributedSpellCheckComponentTest [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 58.226 sec [junit] [junit] Testsuite: org.apache.solr.handler.component.DistributedTermsComponentTest [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 49.52 sec [junit] [junit] Testsuite:
[jira] Commented: (SOLR-1144) replication hang
[ https://issues.apache.org/jira/browse/SOLR-1144?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=12884644#action_12884644 ] Toby Cole commented on SOLR-1144: - Just over a year since it was first spotted, I'm consistently getting the same symptoms as this bug. We've got a single master, with two slaves polling it, both slaves have stalled at exactly the same point in the replication. Here's the relevent section of the replication handler's 'details' response: Node A {code:xml} str name=numFilesDownloaded18/str str name=replicationStartTimeFri Jul 02 10:40:00 BST 2010/str str name=timeElapsed6683s/str str name=currentFile_9du.prx/str str name=currentFileSize8.17 MB/str str name=currentFileSizeDownloaded8.17 MB/str str name=currentFileSizePercent100.0/str str name=bytesDownloaded40.55 MB/str str name=totalPercent0.0/str str name=timeRemaining8290722s/str str name=downloadSpeed6.21 KB/str {code} Node B {code:xml} str name=numFilesDownloaded18/str str name=replicationStartTimeFri Jul 02 10:40:00 BST 2010/str str name=timeElapsed6752s/str str name=currentFile_9du.prx/str str name=currentFileSize8.17 MB/str str name=currentFileSizeDownloaded8.17 MB/str str name=currentFileSizePercent100.0/str str name=bytesDownloaded40.55 MB/str str name=totalPercent0.0/str str name=timeRemaining8376322s/str str name=downloadSpeed6.15 KB/str {code} replication hang Key: SOLR-1144 URL: https://issues.apache.org/jira/browse/SOLR-1144 Project: Solr Issue Type: Bug Reporter: Yonik Seeley Assignee: Noble Paul Fix For: 1.4 It seems that replication can sometimes hang. http://www.lucidimagination.com/search/document/403305a3fda18599 -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online. - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
Re: [Lucy] Revisions to Incubator proposal
Marvin Humphrey wrote on 7/1/10 12:08 PM: ... but I'm dissatisfied with the second half: We acknowledge that Apache seems like a natural home for Lucy given that it is also the home of Lucene, and speculate that this may have been on the minds of the Lucene PMC when Lucy was green-lighted as a sub-project. More importantly, though, the Lucy development community strongly believes that The Apache Way is right for Lucy. First, this passage only asserts that we believe in The Apache Way rather than demonstrating our understanding of it. We should show, not tell. Second, we should purge the PMC mind reading. Who knows what they were thinking! :) Third, I don't want to leave in any mention of Lucy belonging at Apache because Lucene is there, too. That's Lucy sponging off the Lucene brand, and it's not a benefit to Apache. We should just leave that unstated and stand on our merits. +1 to all your points. Nuke that second half. The Community section does a fine job of identifying our challenges and presenting a plan: Lucy currently has a small community, most members of which originated in the KinoSearch community. Lucy's chief challenge is growing its community, which it hopes to achieve through efforts in two areas: reaching a 1.0 release, and actively reaching out to its target audience, users and developers in the dynamic language communities who want a fast, scalable full-text search solution in their native language. Still, I think we deserve a little more credit. We've taken a lot of flak regarding the size of the Lucy community, but you know, if you consider how the *KinoSearch* community has operated over the years, we haven't done so bad. +1 for mentioning KS, since *that* is the code that is being donated. The one thing I don't think we've done well (and this is my fault) is handle releases and backwards compatibility. I agree with what Nate said on this. The perfect is the enemy of the good. Lastly, it would be nice to cover our contingency plan of growing the community and coming back with a bigger committer list at some later date. However, I think that may arise naturally during the discussion, and it's probably too big a topic to squeeze in. +1 on Nate's comments here. I'm afk most of the day today (Friday), Marvin. I'm fine with the proposal as it written at the moment; it looks like you've already addressed most of the points above in your edits from last night. cheers on a hard week's work! pek -- Peter Karman . http://peknet.com/ . pe...@peknet.com
[jira] Updated: (SOLR-1144) replication hang
[ https://issues.apache.org/jira/browse/SOLR-1144?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Toby Cole updated SOLR-1144: Attachment: stacktrace-master.txt stacktrace-slave-1.txt stacktrace-slave-2.txt Adding stacktraces for both slave instances and the master instance. replication hang Key: SOLR-1144 URL: https://issues.apache.org/jira/browse/SOLR-1144 Project: Solr Issue Type: Bug Reporter: Yonik Seeley Assignee: Noble Paul Fix For: 1.4 Attachments: stacktrace-master.txt, stacktrace-slave-1.txt, stacktrace-slave-2.txt It seems that replication can sometimes hang. http://www.lucidimagination.com/search/document/403305a3fda18599 -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online. - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] Issue Comment Edited: (SOLR-1144) replication hang
[ https://issues.apache.org/jira/browse/SOLR-1144?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=12884672#action_12884672 ] Toby Cole edited comment on SOLR-1144 at 7/2/10 9:37 AM: - Adding stacktraces for both slave instances and the master instance. These stack traces are from a reproduction of the original problem, so the timestamps will not matchup with the XML from the replication-handler previously posted. was (Author: tub): Adding stacktraces for both slave instances and the master instance. replication hang Key: SOLR-1144 URL: https://issues.apache.org/jira/browse/SOLR-1144 Project: Solr Issue Type: Bug Reporter: Yonik Seeley Assignee: Noble Paul Fix For: 1.4 Attachments: stacktrace-master.txt, stacktrace-slave-1.txt, stacktrace-slave-2.txt It seems that replication can sometimes hang. http://www.lucidimagination.com/search/document/403305a3fda18599 -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online. - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] Commented: (SOLR-1144) replication hang
[ https://issues.apache.org/jira/browse/SOLR-1144?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=12884690#action_12884690 ] Yonik Seeley commented on SOLR-1144: Thanks for the stack traces Toby! Interesting... seems like the commit in the slave blocked... {code} at org.apache.solr.common.util.ConcurrentLRUCache.getLatestAccessedItems(ConcurrentLRUCache.java:276) {code} So perhaps another thread locked, but didn't unlock the lock? SOLR-1538 did fix something that could possibly lead to a deadlock, but it's super unlikely (a very small object allocation would have to fail at just the right spot). Still, if this is easy enough to reproduce, could you try Solr 1.4.1 and see if it's fixed? (and if it hangs again, be sure to get stack traces... they are super helpful!) replication hang Key: SOLR-1144 URL: https://issues.apache.org/jira/browse/SOLR-1144 Project: Solr Issue Type: Bug Reporter: Yonik Seeley Assignee: Noble Paul Fix For: 1.4 Attachments: stacktrace-master.txt, stacktrace-slave-1.txt, stacktrace-slave-2.txt It seems that replication can sometimes hang. http://www.lucidimagination.com/search/document/403305a3fda18599 -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online. - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] Commented: (SOLR-1144) replication hang
[ https://issues.apache.org/jira/browse/SOLR-1144?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=12884693#action_12884693 ] Toby Cole commented on SOLR-1144: - Oh yes, should have mentioned... we're already on Solr 1.4.1 in production as of yesterday (we don't hang about y'know ;) ). replication hang Key: SOLR-1144 URL: https://issues.apache.org/jira/browse/SOLR-1144 Project: Solr Issue Type: Bug Reporter: Yonik Seeley Assignee: Noble Paul Fix For: 1.4 Attachments: stacktrace-master.txt, stacktrace-slave-1.txt, stacktrace-slave-2.txt It seems that replication can sometimes hang. http://www.lucidimagination.com/search/document/403305a3fda18599 -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online. - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
Re: Core Developer bios
On Wed, Jun 30, 2010 at 05:01:45PM -0700, Marvin Humphrey wrote: On Wed, Jun 30, 2010 at 11:47:42PM -, Apache Wiki wrote: + * Nathan Kurz has participated in numerous open source projects and has been a KinoSearch committer since 2007. He is currently Chief Flavor Engineer of Scream Sorbet, and writes software in his copious free time. Nate, I didn't know what to write to describe your crazy bio, and I don't think this does enough to convey your serious C and comp-sci chops. I also don't think Chief Flavor Engineer is your title at Scream, though IMO it ought to be. Can you please edit? Heh, I hear from Nate via private email that he likes the existing bio. So be it! Marvin Humphrey
[Lucy Wiki] Update of LucyIncubatorProposal by Marvin Humphrey
Dear Wiki user, You have subscribed to a wiki page or wiki category on Lucy Wiki for change notification. The LucyIncubatorProposal page has been changed by MarvinHumphrey. The comment on this change is: Change Rationale to mention C's interoperability, fine-grained control.. http://wiki.apache.org/lucy/LucyIncubatorProposal?action=diffrev1=33rev2=34 -- Since the Lucene PMC will not be responsible for Lucy much longer, it is more appropriate for the software grant to take place within the context of the Incubator than the Lucene TLP. As none of our current members have Apache PMC experience, we also seek to take advantage of the Incubator environment to prepare ourselves for responsible self-governance. == Rationale == - There is great hunger for a search engine library in the mode of Lucene which is accessible from various dynamic languages, and for one accessible from pure C. Individuals naturally wish to code in their language of choice. Organizations which do not have significant Java expertise may not want to support Java strictly for the sake of running a Lucene installation. Native applications may be launched much more quickly than JVMs. Lucy will meet all these demands. + There is great hunger for a search engine library in the mode of Lucene which is accessible from various dynamic languages, and for one accessible from pure C. Individuals naturally wish to code in their language of choice. Organizations which do not have significant Java expertise may not want to support Java strictly for the sake of running a Lucene installation. Developers may want to take advantage of C's interoperability and fine-grained control. Lucy will meet all these demands. Apache is a natural home for our project given the way it has always operated: user-driven innovation, security as a requirement, lively and amiable mailing list discussions, strength through diversity, and so on. We feel comfortable here, and we believe that we will become exemplary Apache citizens.
[jira] Commented: (SOLR-1144) replication hang
[ https://issues.apache.org/jira/browse/SOLR-1144?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=12884717#action_12884717 ] Yonik Seeley commented on SOLR-1144: The odd thing is that the line numbers in the stack traces don't match up for either 1.4.0 or 1.4.1 Specifically ConcurrentLRUCache.java:276 is in the middle of markAndSweep() in both versions (as opposed to getLatestAccessedItems() which your stack trace would suggest). Are these stack traces from stock 1.4.0 or 1.41? If so, does anyone have a clue why the line numbers would be off? replication hang Key: SOLR-1144 URL: https://issues.apache.org/jira/browse/SOLR-1144 Project: Solr Issue Type: Bug Reporter: Yonik Seeley Assignee: Noble Paul Fix For: 1.4 Attachments: stacktrace-master.txt, stacktrace-slave-1.txt, stacktrace-slave-2.txt It seems that replication can sometimes hang. http://www.lucidimagination.com/search/document/403305a3fda18599 -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online. - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] Commented: (SOLR-1144) replication hang
[ https://issues.apache.org/jira/browse/SOLR-1144?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=12884719#action_12884719 ] Toby Cole commented on SOLR-1144: - I know exactly why the line numbers would be off. I just remembered we're using a custom war package so we can add our own plugins in (yes, I know we can use solr.home/lib, but we've not got round to that yet). The only classes we're overriding from solr are ConcurrentLRUCache and FastLRUCache. This was from pre solr 1.4, when the cache implementations were slowing faceting right down. I have a feeling if I remove those overridden classes and use the new (bug-free) ones, the hang may stop. I'll give it a go now, sorry in advance if it was my oversight that is causing this bug to re-appear. T replication hang Key: SOLR-1144 URL: https://issues.apache.org/jira/browse/SOLR-1144 Project: Solr Issue Type: Bug Reporter: Yonik Seeley Assignee: Noble Paul Fix For: 1.4 Attachments: stacktrace-master.txt, stacktrace-slave-1.txt, stacktrace-slave-2.txt It seems that replication can sometimes hang. http://www.lucidimagination.com/search/document/403305a3fda18599 -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online. - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
Re: Revisions to Incubator proposal
On Thu, Jul 01, 2010 at 11:44:46PM -0700, Nathan Kurz wrote: There is great hunger for a search engine library in the mode of Lucene which is accessible from various dynamic languages, and for one accessible from pure C. Individuals naturally wish to code in their language of choice. Organizations which do not have significant Java expertise may not want to support Java strictly for the sake of running a Lucene instance. Native applications can be launched much more quickly than JVMs. Lucy will meet all these demands. I like this, but might stop after the first two sentences. I think they're the stronger reasons. If you wanted to go further, I'd mention the ways in which C has traditionally been the language of Unix server programming, and allows one to take full advantage of the machine's potential. OK, I went with this for today: Developers may want to take advantage of C's interoperability and fine-grained control. I do think it's important to at least touch on the advantages of natively compiled code in the Rationale, and not limit ourselves to the language-loyalty angle. While Lucy certainly can take advantage of the Apache infrastructure, I'm not sure it really needs it. Acknowledged. However, the more successful Lucy gets, the more useful it is to have the support systems and institutions of Apache protecting it and instilling confidence. And it's not just about backups, uptime, and fending off crackers -- it's things like stable governance, legal protection mechanisms, encouraging contributions from developers sponsored by major corporations, and so on. What it needs is more people writing code for it, and the best way to achieve this is probably to get more technically proficient users. I agree. That is our most pressing need, so we should prioritize accordingly. I think you're doing fine on backwards compatibility, and if anything you're spending too much time on it. Instead of worrying about not breaking things that are old, spend the effort making it easier to write things that are new. I believe that the new backwards compatibility policy properly privileges active developers, while still offering strong stability guarantees to those who require them. It's not perfect, but I think it strikes a good balance. Let's make the dogfood tasty, and start writing 'extensions' instead of trying to include everything in core. A public C API will facilitate extensions. Had one been available, ProximityQuery would have been written to be a separate CPAN distro. I think separating ProximityQuery from core would be a good concrete goal to help us focus while we design the C API. I don't know if you need to discuss this. And once we have more developers, do we really need to come back? I mean, I really like Apache, but Trac plus taking the best points of the Incubator approach might offer 90% of the benefit with a lot less overhead. Which doesn't mean Lucy wouldn't benefit from being included, but I wouldn't precommit to returning at a later date if they don't want us. Your response helps to clarify the point that I really wanted to get in there, which is that a long overdue release is imminent. That's now covered in the revised Community section. OK, I only made that one minor content change, so I think we have adequate consensus. I'm going to perform a QA pass and maybe tweak some phrasings so that the word community doesn't appear eleventy-billion times. Then I'll send it off to gene...@lucene. Thanks, folks -- this has worked out well! Marvin Humphrey
Re: Trouble updating Solr website
: Like I said, on Mahout we went w/ a simple landing page and everything else in Confluence, as inspired by OFBiz. : : : Sounds good to me any objections? Any tips on how to get us started? FWIW: Ryan actually setup a SOLR Confluence wiki a long time ago with the plan of it replacing the mail solr site... https://cwiki.apache.org/SOLRxSITE/ While i have mo particular love for forrest, i would like to suggest the following things be considered before moving forward... 1) now may be a good time to consider consolidating the entire Lucene TLP website into a single update model (ie: instead of having /java, /solr, /pylucene, /lucene.net, etc... all be managed with distinct wikis or publishing flows, lets just have one site with one edit model. 2) docs that ship as part of the releases can (and probably should) be seperated out and managed independently -- either as distinct wikis, or via some other means -- and still available online linked to from the main site. /java already has this working well for the per release lucene-java documentation where both teh site and the docs are in forrest -- we can follow this model for all of the individual releases independently of how the files them selves are generated. 3) if we do move to confluence we should think carefully in advance about how we want to deal with the ACLs ... what should be editable to anyone with an wiki acount, vs. what should be editable only to people with a CLA on file, vs what should be editable to only committers -- the first two in particular will heavily impact what can be packaged up in releases. -Hoss - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] Commented: (SOLR-461) Highlighting TokenStream Truncation capability
[ https://issues.apache.org/jira/browse/SOLR-461?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=12884806#action_12884806 ] Uwe Schindler commented on SOLR-461: The TokenFilter for limiting Token count is already in Lucene 3.x and trunk: LimitTokenCountFilter and a Wrapping Analyzer that adds this filter on top of any existing Analyzer. I think somebody already added it to Solr, too, dont know the issue, but have seen it. Highlighting TokenStream Truncation capability -- Key: SOLR-461 URL: https://issues.apache.org/jira/browse/SOLR-461 Project: Solr Issue Type: Improvement Components: highlighter Reporter: Grant Ingersoll Assignee: Grant Ingersoll Priority: Minor Attachments: SOLR-461.patch It is sometimes the case when generating snippets that one need not fragment/analyze the whole document (especially for large documents) in order to show meaningful snippet highlights. Patch to follow that adds a counting TokenFilter that returns null after X number of Tokens have been seen. This filter will then be hooked into the SolrHighlighter and configurable via solrconfig.xml. The default value will be Integer.MAX_VALUE or, I suppose, it could be set to whatever Max Field Length is set to, as well. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online. - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] Created: (SOLR-1981) solr.xml should fail to load if multiple cores with the same name
solr.xml should fail to load if multiple cores with the same name - Key: SOLR-1981 URL: https://issues.apache.org/jira/browse/SOLR-1981 Project: Solr Issue Type: Improvement Reporter: Hoss Man Attachments: SOLR-1981.patch As noted in email a while back... http://search.lucidimagination.com/search/document/674bf5dfbbb349bc/multiple_cores_w_same_name_in_solr_xml {quote} but there is currently no assertion that every core have a name, or that the names be unique before the SolrCore is constructed ... it's not until the core is registered that an error will be generated if the core name is null, or that the previous core with an identicle name will be close()ed. {quote} I think we should fail fast if solr.xml specifies the same name more then once -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online. - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] Updated: (LUCENE-2522) add simple japanese tokenizer, based on tinysegmenter
[ https://issues.apache.org/jira/browse/LUCENE-2522?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Robert Muir updated LUCENE-2522: Attachment: LUCENE-2522.patch i refactored the TinySegmenterConstants to use ints/switch statements instead of all the hashmaps. this creates a larger .java file, but its a smaller .class, and scoring no longer has to create 24 strings per character add simple japanese tokenizer, based on tinysegmenter - Key: LUCENE-2522 URL: https://issues.apache.org/jira/browse/LUCENE-2522 Project: Lucene - Java Issue Type: New Feature Components: contrib/analyzers Reporter: Robert Muir Priority: Minor Attachments: LUCENE-2522.patch, LUCENE-2522.patch TinySegmenter (http://www.chasen.org/~taku/software/TinySegmenter/) is a tiny japanese segmenter. It was ported to java/lucene by Kohei TAKETA k-...@void.in, and is under friendly license terms (BSD, some files explicitly disclaim copyright to the source code, giving a blessing instead) Koji knows the author, and already contacted about incorporating into lucene: {noformat} I've contacted Takeda-san who is the creater of Java version of TinySegmenter. He said he is happy if his program is part of Lucene. He is a co-author of my book about Solr published in Japan, BTW. ;-) {noformat} -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online. - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] Resolved: (SOLR-1949) overwrite document fails if Solr index is not optimized
[ https://issues.apache.org/jira/browse/SOLR-1949?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Hoss Man resolved SOLR-1949. Resolution: Not A Problem overwrite document fails if Solr index is not optimized --- Key: SOLR-1949 URL: https://issues.apache.org/jira/browse/SOLR-1949 Project: Solr Issue Type: Bug Components: update Affects Versions: 1.4 Environment: linux centos Reporter: Miguel B. Scenario: - Solr 1.4 with multicore - We have a set of 5.000 source documents that we want to index. - We send these set to Solr by SolrJ API and they are added correctly. We have field ID as string and uniqueKey, so the update operation overwite documents with the same ID. The result is 4500 unique documents in Solr. Also all documents have an index field that contains the source repository of each document, we need it because we want to index another sources. - After add operation, we send optimization. All works fine at this point. Solr have 4.500 documents at Solr core (and 4.500 max documents too). Now these 5.000 sources documents are updated by users, and a set of them are deleted (supose, 1000). So, now we want to update our Solr index with these change (unfortunately our repository doesn't support an incremental approach), the operations are: - At index Solr, delete documents by query (by the field that contains document source repository). We use deleteByQuery and commit SolrJ operations. - At this point Solr core have 0 documents (but 4.500 max documents, important!!!) - Now we add to Solr the new version of source documents (4000). Remember that documents don't have unique identifiers, supose that unique items are 3000. So when add operation finish (after commit sended) Solr index must have 3.000 unique items. But the result isn't 3.000 unique items, we obtains a random results: 3000, 2980, 2976, etc. It's a serious problem because we lost documents. We have a workaround. At these operations just after delete operation, we send an optimization to Solr (maxDocuments are updated). After this, we send new documents. By this way the result is always fine. In our tests, we can see that this issue is only when the new documents overwrites documents that existed in solr. Thanks!! -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online. - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] Commented: (SOLR-1949) overwrite document fails if Solr index is not optimized
[ https://issues.apache.org/jira/browse/SOLR-1949?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=12884849#action_12884849 ] Hoss Man commented on SOLR-1949: 1) In the future, please don't use jira to ask questions about odd behavior you are seeing -- that is what the solr-user mailing list is for. as a general rule you should only open a Bug in Jira issue if you have already asked a question on the solr-user mailing list and have verified that there isn't a mistake/missunderstanding in your config. 2) your initial report is inconsistent and makes no sense... bq. We have field ID as string and uniqueKey, so the update operation overwite documents with the same ID. ... bq. Remember that documents don't have unique identifiers ...if you do follow up on solr-user, please be more explicit, and clarify exactly what you are doing (showing us your schema.xml, and some sample documents is pretty much a neccdessity to make sense of problems like this that don't produce an error message) overwrite document fails if Solr index is not optimized --- Key: SOLR-1949 URL: https://issues.apache.org/jira/browse/SOLR-1949 Project: Solr Issue Type: Bug Components: update Affects Versions: 1.4 Environment: linux centos Reporter: Miguel B. Scenario: - Solr 1.4 with multicore - We have a set of 5.000 source documents that we want to index. - We send these set to Solr by SolrJ API and they are added correctly. We have field ID as string and uniqueKey, so the update operation overwite documents with the same ID. The result is 4500 unique documents in Solr. Also all documents have an index field that contains the source repository of each document, we need it because we want to index another sources. - After add operation, we send optimization. All works fine at this point. Solr have 4.500 documents at Solr core (and 4.500 max documents too). Now these 5.000 sources documents are updated by users, and a set of them are deleted (supose, 1000). So, now we want to update our Solr index with these change (unfortunately our repository doesn't support an incremental approach), the operations are: - At index Solr, delete documents by query (by the field that contains document source repository). We use deleteByQuery and commit SolrJ operations. - At this point Solr core have 0 documents (but 4.500 max documents, important!!!) - Now we add to Solr the new version of source documents (4000). Remember that documents don't have unique identifiers, supose that unique items are 3000. So when add operation finish (after commit sended) Solr index must have 3.000 unique items. But the result isn't 3.000 unique items, we obtains a random results: 3000, 2980, 2976, etc. It's a serious problem because we lost documents. We have a workaround. At these operations just after delete operation, we send an optimization to Solr (maxDocuments are updated). After this, we send new documents. By this way the result is always fine. In our tests, we can see that this issue is only when the new documents overwrites documents that existed in solr. Thanks!! -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online. - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] Commented: (SOLR-1982) Leading wildcard queries work for all fields if ReversedWildcardFilterFactory is used for any field
[ https://issues.apache.org/jira/browse/SOLR-1982?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=12884847#action_12884847 ] Hoss Man commented on SOLR-1982: The behavior comes from the fact that during initialization, SolrQueryParser.checkAllowLeadingWildcards calls setAllowLeadingWildcard(true); if any field type uses ReversedWildcardFilterFactory. Then in getWildcardQuery, it the specific field type before calling ReverseStringFilter.reverse, but then, regardless of field type, delegates to super.getWildcardQuery which will allow the leading wildcard for *all* fields based on the previous call to setAllowLeadingWildcard(true). I'm not really sure what the intention was for fields that don't use ReversedWildcardFilterFactory, but the current behavior makes no sense at all. Either leading wildcards should *only* be allowed for fieldtypes that use ReversedWildcardFilterFactory, or the QParser should have a config option to control it for other fields -- but as it stands it makes no sense what so ever. Leading wildcard queries work for all fields if ReversedWildcardFilterFactory is used for any field --- Key: SOLR-1982 URL: https://issues.apache.org/jira/browse/SOLR-1982 Project: Solr Issue Type: Bug Affects Versions: 1.4, 1.4.1 Reporter: Hoss Man As noted on the mailing list... http://search.lucidimagination.com/search/document/8064e6877f49e4c4/leading_wildcard_query_strangeness ...SolrQueryParse supports leading wild card queries for *any* field as long as at least one field type exists in the schema.xml which uses ReversedWildcardFilterFactory -- even if that field type is never used. This is extremely confusing, and ost likely indicates a bug in how SolrQueryParser deals with ReversedWildcardFilterFactory -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online. - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] Updated: (SOLR-1283) Mark Invalid error on indexing
[ https://issues.apache.org/jira/browse/SOLR-1283?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Hoss Man updated SOLR-1283: --- Fix Version/s: 3.1 4.0 we have a patch that *seems* to work, so we should dfinitely try to get this into the next release ... i'm hoping someone more familiar with the code can sanity check it soon. Mark Invalid error on indexing -- Key: SOLR-1283 URL: https://issues.apache.org/jira/browse/SOLR-1283 Project: Solr Issue Type: Bug Affects Versions: 1.3 Environment: Ubuntu 8.04, Sun Java 6 Reporter: solrize Fix For: 3.1, 4.0 Attachments: SOLR-1283.modules.patch, SOLR-1283.patch When indexing large (1 megabyte) documents I get a lot of exceptions with stack traces like the below. It happens both in the Solr 1.3 release and in the July 9 1.4 nightly. I believe this to NOT be the same issue as SOLR-42. I found some further discussion on solr-user: http://www.nabble.com/IOException:-Mark-invalid-while-analyzing-HTML-td17052153.html In that discussion, Grant asked the original poster to open a Jira issue, but I didn't see one so I'm opening one; please feel free to merge or close if it's redundant. My stack trace follows. Jul 15, 2009 8:36:42 AM org.apache.solr.core.SolrCore execute INFO: [] webapp=/solr path=/update params={} status=500 QTime=3 Jul 15, 2009 8:36:42 AM org.apache.solr.common.SolrException log SEVERE: java.io.IOException: Mark invalid at java.io.BufferedReader.reset(BufferedReader.java:485) at org.apache.solr.analysis.HTMLStripReader.restoreState(HTMLStripReader.java:171) at org.apache.solr.analysis.HTMLStripReader.read(HTMLStripReader.java:728) at org.apache.solr.analysis.HTMLStripReader.read(HTMLStripReader.java:742) at java.io.Reader.read(Reader.java:123) at org.apache.lucene.analysis.CharTokenizer.next(CharTokenizer.java:108) at org.apache.lucene.analysis.StopFilter.next(StopFilter.java:178) at org.apache.lucene.analysis.standard.StandardFilter.next(StandardFilter.java:84) at org.apache.lucene.analysis.LowerCaseFilter.next(LowerCaseFilter.java:53) at org.apache.solr.analysis.WordDelimiterFilter.next(WordDelimiterFilter.java:347) at org.apache.lucene.index.DocInverterPerField.processFields(DocInverterPerField.java:159) at org.apache.lucene.index.DocFieldConsumersPerField.processFields(DocFieldConsumersPerField.java:36) at org.apache.lucene.index.DocFieldProcessorPerThread.processDocument(DocFieldProcessorPerThread.java:234) at org.apache.lucene.index.DocumentsWriter.updateDocument(DocumentsWriter.java:765) at org.apache.lucene.index.DocumentsWriter.updateDocument(DocumentsWriter.java:748) at org.apache.lucene.index.IndexWriter.updateDocument(IndexWriter.java:2512) at org.apache.lucene.index.IndexWriter.updateDocument(IndexWriter.java:2484) at org.apache.solr.update.DirectUpdateHandler2.addDoc(DirectUpdateHandler2.java:240) at org.apache.solr.update.processor.RunUpdateProcessor.processAdd(RunUpdateProcessorFactory.java:61) at org.apache.solr.handler.XMLLoader.processUpdate(XMLLoader.java:140) at org.apache.solr.handler.XMLLoader.load(XMLLoader.java:69) at org.apache.solr.handler.ContentStreamHandlerBase.handleRequestBody(ContentStreamHandlerBase.java:54) at org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:131) at org.apache.solr.core.SolrCore.execute(SolrCore.java:1292) at org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:338) at org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:241) at org.mortbay.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1089) at org.mortbay.jetty.servlet.ServletHandler.handle(ServletHandler.java:365) at org.mortbay.jetty.security.SecurityHandler.handle(SecurityHandler.java:216) at org.mortbay.jetty.servlet.SessionHandler.handle(SessionHandler.java:181) at org.mortbay.jetty.handler.ContextHandler.handle(ContextHandler.java:712) at org.mortbay.jetty.webapp.WebAppContext.handle(WebAppContext.java:405) at org.mortbay.jetty.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:211) at org.mortbay.jetty.handler.HandlerCollection.handle(HandlerCollection.java:114) at org.mortbay.jetty.handler.HandlerWrapper.handle(HandlerWrapper.java:139) at org.mortbay.jetty.Server.handle(Server.java:285) at org.mortbay.jetty.HttpConnection.handleRequest(HttpConnection.java:502) at
Re: [VOTE] RC2 Release Solr 1.4.1
: Is there any chance for the patch posted on issue SOLR-1283 ( : https://issues.apache.org/jira/browse/SOLR-1283) to be integrated in the : forthcoming release ? : It appears this old issue is still present in 1.4.1 rc2. Julien: I'm afraid it's definitely too late for it to make it into 1.4.1 -- by the looks of it it was already too late when you sent your email (i think timing wise the release was already official, there just hadn't been an announcement yet while miller was waiting for it to hit the mirrors) Sometimes patches slip through the cracks, and that seems to be what happened with your patch in SOLR-1283 -- i'm not familiar enough with that code to commit to feel comfortable committing it myself -- but i've flagged it in Jira in hopes that someone who is might take a look. -Hoss - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] Updated: (SOLR-1283) Mark Invalid error on indexing
[ https://issues.apache.org/jira/browse/SOLR-1283?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Hoss Man updated SOLR-1283: --- Attachment: SOLR-1283.modules.patch Updates patch to trunk (where the charfilter stuff has been refactored into the new top level modules directory) I'm not familiar with the HTMLStripCharFilter stuff, so i can't say whether the fix is correct (no idea if peek should be incrementing that counter -- that's why even private methods should have javadocs), but the test certainly looks valid to me Mark Invalid error on indexing -- Key: SOLR-1283 URL: https://issues.apache.org/jira/browse/SOLR-1283 Project: Solr Issue Type: Bug Affects Versions: 1.3 Environment: Ubuntu 8.04, Sun Java 6 Reporter: solrize Attachments: SOLR-1283.modules.patch, SOLR-1283.patch When indexing large (1 megabyte) documents I get a lot of exceptions with stack traces like the below. It happens both in the Solr 1.3 release and in the July 9 1.4 nightly. I believe this to NOT be the same issue as SOLR-42. I found some further discussion on solr-user: http://www.nabble.com/IOException:-Mark-invalid-while-analyzing-HTML-td17052153.html In that discussion, Grant asked the original poster to open a Jira issue, but I didn't see one so I'm opening one; please feel free to merge or close if it's redundant. My stack trace follows. Jul 15, 2009 8:36:42 AM org.apache.solr.core.SolrCore execute INFO: [] webapp=/solr path=/update params={} status=500 QTime=3 Jul 15, 2009 8:36:42 AM org.apache.solr.common.SolrException log SEVERE: java.io.IOException: Mark invalid at java.io.BufferedReader.reset(BufferedReader.java:485) at org.apache.solr.analysis.HTMLStripReader.restoreState(HTMLStripReader.java:171) at org.apache.solr.analysis.HTMLStripReader.read(HTMLStripReader.java:728) at org.apache.solr.analysis.HTMLStripReader.read(HTMLStripReader.java:742) at java.io.Reader.read(Reader.java:123) at org.apache.lucene.analysis.CharTokenizer.next(CharTokenizer.java:108) at org.apache.lucene.analysis.StopFilter.next(StopFilter.java:178) at org.apache.lucene.analysis.standard.StandardFilter.next(StandardFilter.java:84) at org.apache.lucene.analysis.LowerCaseFilter.next(LowerCaseFilter.java:53) at org.apache.solr.analysis.WordDelimiterFilter.next(WordDelimiterFilter.java:347) at org.apache.lucene.index.DocInverterPerField.processFields(DocInverterPerField.java:159) at org.apache.lucene.index.DocFieldConsumersPerField.processFields(DocFieldConsumersPerField.java:36) at org.apache.lucene.index.DocFieldProcessorPerThread.processDocument(DocFieldProcessorPerThread.java:234) at org.apache.lucene.index.DocumentsWriter.updateDocument(DocumentsWriter.java:765) at org.apache.lucene.index.DocumentsWriter.updateDocument(DocumentsWriter.java:748) at org.apache.lucene.index.IndexWriter.updateDocument(IndexWriter.java:2512) at org.apache.lucene.index.IndexWriter.updateDocument(IndexWriter.java:2484) at org.apache.solr.update.DirectUpdateHandler2.addDoc(DirectUpdateHandler2.java:240) at org.apache.solr.update.processor.RunUpdateProcessor.processAdd(RunUpdateProcessorFactory.java:61) at org.apache.solr.handler.XMLLoader.processUpdate(XMLLoader.java:140) at org.apache.solr.handler.XMLLoader.load(XMLLoader.java:69) at org.apache.solr.handler.ContentStreamHandlerBase.handleRequestBody(ContentStreamHandlerBase.java:54) at org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:131) at org.apache.solr.core.SolrCore.execute(SolrCore.java:1292) at org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:338) at org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:241) at org.mortbay.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1089) at org.mortbay.jetty.servlet.ServletHandler.handle(ServletHandler.java:365) at org.mortbay.jetty.security.SecurityHandler.handle(SecurityHandler.java:216) at org.mortbay.jetty.servlet.SessionHandler.handle(SessionHandler.java:181) at org.mortbay.jetty.handler.ContextHandler.handle(ContextHandler.java:712) at org.mortbay.jetty.webapp.WebAppContext.handle(WebAppContext.java:405) at org.mortbay.jetty.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:211) at org.mortbay.jetty.handler.HandlerCollection.handle(HandlerCollection.java:114) at org.mortbay.jetty.handler.HandlerWrapper.handle(HandlerWrapper.java:139) at org.mortbay.jetty.Server.handle(Server.java:285) at
[jira] Created: (SOLR-1983) snappuller fails when modifiedConfFiles is not empty and full copy of index is needed
snappuller fails when modifiedConfFiles is not empty and full copy of index is needed - Key: SOLR-1983 URL: https://issues.apache.org/jira/browse/SOLR-1983 Project: Solr Issue Type: Bug Components: replication (java) Reporter: Noble Paul Assignee: Noble Paul Fix For: 1.4 see the mail thread. http://markmail.org/thread/orxyqftqrsqvrt5w -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online. - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] Commented: (SOLR-1983) snappuller fails when modifiedConfFiles is not empty and full copy of index is needed
[ https://issues.apache.org/jira/browse/SOLR-1983?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=12884867#action_12884867 ] Hoss Man commented on SOLR-1983: cloned from SOLR-1264 where a new attachment was recently added addressing a continuation of the original problem in a specific situation. Cloned rather then reopend because SOLR-1264 is already closed and part of Solr 1.4 -- for tracking in CAHNGES.txt we need a new issue number -- see comments in SOLR-1264 for more details snappuller fails when modifiedConfFiles is not empty and full copy of index is needed - Key: SOLR-1983 URL: https://issues.apache.org/jira/browse/SOLR-1983 Project: Solr Issue Type: Bug Components: replication (java) Reporter: Noble Paul Assignee: Noble Paul Fix For: 1.4 see the mail thread. http://markmail.org/thread/orxyqftqrsqvrt5w -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online. - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] Updated: (SOLR-1556) TermVectorComponents should provide good error messages when fieldtype isn't compatible with requested options
[ https://issues.apache.org/jira/browse/SOLR-1556?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Grant Ingersoll updated SOLR-1556: -- Attachment: SOLR-1556.patch Ready to go. Going to commit this weekend. TermVectorComponents should provide good error messages when fieldtype isn't compatible with requested options -- Key: SOLR-1556 URL: https://issues.apache.org/jira/browse/SOLR-1556 Project: Solr Issue Type: Bug Components: search Affects Versions: 1.4 Reporter: Hoss Man Assignee: Grant Ingersoll Fix For: Next Attachments: SOLR-1556.patch, SOLR-1556.patch As noted by grant on the email list, asking TermVectorComponent for things like termVectors, positions, and offsets can't produce meaningful results unless the field in question has the corrisponding schema option set to true -- but the behavior of TVC when they not true is confusing to users. We should make TVC return a meaningful error if it's asked to return a certain type of info for field that it can't deal with - something making it clear what properties of hte schema need to be changed to make it work... http://old.nabble.com/Re%3A-TermVectorComponent-%3A-Required---Optional-Parameters-p26181454.html -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online. - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
Hudson build is back to normal : Lucene-3.x #59
See http://hudson.zones.apache.org/hudson/job/Lucene-3.x/59/ - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
Hudson build is back to normal : Lucene-trunk #1232
See http://hudson.zones.apache.org/hudson/job/Lucene-trunk/1232/changes - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org