Re: Can we add isDaemon check to LTC.threadCleanup?

2012-02-22 Thread Shai Erera
I use a Jetty server instance in my tests, and LTC almost always reports
that the 'qtp' threads are left open. 'qtp' are the prefix name of Jetty's
QueuedThreadPool threads. QueuedThreadPool has two defaults:

1) It is not daemon, therefore its PoolThreads are not set daemon too
2) It defaults to keep idle threads for 60 seconds around

In real-life these two settings make sense, but in tests, when I know that
I call server.stop only after all my tasks have completed, it's annoying.

When you call server.stop() followed by server.join(), it calls qtp.stop()
which interrupts all running PoolThreads. I'm not sure if there's a bug in
QueuedThreadPool's impl (or maybe PoolThread's impl), but I can tell for
sure that the JVM process keeps running for ~60 seconds, even though the
Socket and port have been released.

So I wrote this code in my tests that start a Server instance:

Server s = new Server();
QueuedThreadPool qtp = new QueuedThreadPool();
qtp.setDaemon(true); // I don't mind it'll be daemon in tests.
qtp.setMaxIdleTimeMs(0); // when a thread becomes idle, kill it
s.setThreadPool(qtp);

Yet, sometimes even that's not enough and LTC's afterClass is called before
the PoolThread dies. So I'm just thinking that if LTC would ignore daemon
threads, the annoying messages will be gone :).

It's not a big deal, just annoying to see these prints, and wasting so much
time trying to make them go away :).

Shai

On Thu, Feb 23, 2012 at 9:14 AM, Dawid Weiss
wrote:

> I think it should be closed gracefully, even if it's a daemon thread.
> Which thread do you have in mind?
>
> Dawid
>
> On Thu, Feb 23, 2012 at 7:33 AM, Shai Erera  wrote:
> > Hi
> >
> > I think that if a thread is daemon, but still running, LTC.threadCleanup
> > should not report it. What do you think?
> >
> > Shai
>
> -
> To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
> For additional commands, e-mail: dev-h...@lucene.apache.org
>
>


[JENKINS] Lucene-Solr-tests-only-3.x - Build # 12514 - Failure

2012-02-22 Thread Apache Jenkins Server
Build: https://builds.apache.org/job/Lucene-Solr-tests-only-3.x/12514/

1 tests failed.
REGRESSION:  org.apache.solr.client.solrj.embedded.JettyWebappTest.testJSP

Error Message:
Server returned HTTP response code: 500 for URL: 
http://localhost:56349/test/admin/threaddump.jsp

Stack Trace:
java.io.IOException: Server returned HTTP response code: 500 for URL: 
http://localhost:56349/test/admin/threaddump.jsp
at 
sun.net.www.protocol.http.HttpURLConnection.getInputStream(HttpURLConnection.java:1403)
at java.net.URL.openStream(URL.java:1029)
at 
org.apache.solr.client.solrj.embedded.JettyWebappTest.testJSP(JettyWebappTest.java:106)
at 
org.apache.lucene.util.LuceneTestCase$SubclassSetupTeardownRule$1.evaluate(LuceneTestCase.java:606)
at 
org.apache.lucene.util.LuceneTestCase$InternalSetupTeardownRule$1.evaluate(LuceneTestCase.java:495)
at 
org.apache.lucene.util.LuceneTestCase$TestResultInterceptorRule$1.evaluate(LuceneTestCase.java:400)
at 
org.apache.lucene.util.LuceneTestCase$RememberThreadRule$1.evaluate(LuceneTestCase.java:458)
at 
org.apache.lucene.util.LuceneTestCaseRunner.runChild(LuceneTestCaseRunner.java:147)
at 
org.apache.lucene.util.LuceneTestCaseRunner.runChild(LuceneTestCaseRunner.java:50)




Build Log (for compile errors):
[...truncated 16485 lines...]



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org

[jira] [Closed] (LUCENE-3810) Text Alignment in the "status-item" tab of the Website needs to be word wrapped

2012-02-22 Thread Varun Thacker (Closed) (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-3810?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Varun Thacker closed LUCENE-3810.
-


> Text Alignment in the "status-item" tab of the Website needs to be word 
> wrapped
> ---
>
> Key: LUCENE-3810
> URL: https://issues.apache.org/jira/browse/LUCENE-3810
> Project: Lucene - Java
>  Issue Type: Improvement
>Reporter: Varun Thacker
>Assignee: Varun Thacker
>Priority: Minor
> Attachments: LUCENE-3810.patch, lucene.png
>
>
> If we go to http://lucene.apache.org/core/ or http://lucene.apache.org/solr/ 
> under the Latest Dev section the text is exceeding the css boundary. This is 
> happening on both Firefox and Chrome. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Resolved] (LUCENE-3810) Text Alignment in the "status-item" tab of the Website needs to be word wrapped

2012-02-22 Thread Varun Thacker (Resolved) (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-3810?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Varun Thacker resolved LUCENE-3810.
---

   Resolution: Duplicate
Lucene Fields:   (was: New)

> Text Alignment in the "status-item" tab of the Website needs to be word 
> wrapped
> ---
>
> Key: LUCENE-3810
> URL: https://issues.apache.org/jira/browse/LUCENE-3810
> Project: Lucene - Java
>  Issue Type: Improvement
>Reporter: Varun Thacker
>Assignee: Varun Thacker
>Priority: Minor
> Attachments: LUCENE-3810.patch, lucene.png
>
>
> If we go to http://lucene.apache.org/core/ or http://lucene.apache.org/solr/ 
> under the Latest Dev section the text is exceeding the css boundary. This is 
> happening on both Firefox and Chrome. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-3820) Wrong trailing index calculation in PatternReplaceCharFilter

2012-02-22 Thread Dawid Weiss (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3820?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13214412#comment-13214412
 ] 

Dawid Weiss commented on LUCENE-3820:
-

Thanks for looking at this, Robert. I'll go back to this later today, but I can 
tell you right now that from my paper considerations negative indexes make 
logical sense in case of "prepended" characters. So:

PATTERN: A
INPUT: ABCDEF
REPLACEMENT: XYZ
OUTPUT:XYZBCDEF

then (in my patch) X and Y would have negative offsets. It's a matter of 
agreement I guess. Negative indexes are consistent with something like this:

PATTERN: ^
INPUT: ABC
REPLACEMENT: XYZ
OUTPUT:XYZABC

then all three characters (XYZ) have a negative index to indicate they're 
prepended. Thoughts?


> Wrong trailing index calculation in PatternReplaceCharFilter
> 
>
> Key: LUCENE-3820
> URL: https://issues.apache.org/jira/browse/LUCENE-3820
> Project: Lucene - Java
>  Issue Type: Bug
>Reporter: Dawid Weiss
>Assignee: Dawid Weiss
>Priority: Minor
> Fix For: 4.0
>
> Attachments: LUCENE-3820.patch, LUCENE-3820_test.patch, 
> LUCENE-3820_test.patch
>
>
> I need to use PatternReplaceCharFilter's index corrections directly and it 
> fails for me -- the trailing index is not mapped correctly for a pattern 
> "\\.[\\s]*" and replacement ".", input "A. .B.".
> I tried to understand the logic in getReplaceBlock but I eventually failed 
> and simply rewrote it from scratch. After my changes a few tests don't pass 
> but I don't know if it's the tests that are screwed up or my logic. In 
> essence, the difference between the previous implementation and my 
> implementation is how indexes are mapped for shorter replacements. I shift 
> indexes of shorter regions to the "right" of the original index pool and the 
> previous patch seems to squeeze them to the left (don't know why though).
> If anybody remembers how it's supposed to work, feel free to correct me?

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-3810) Text Alignment in the "status-item" tab of the Website needs to be word wrapped

2012-02-22 Thread Varun Thacker (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3810?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13214411#comment-13214411
 ] 

Varun Thacker commented on LUCENE-3810:
---

Hi Mark,

I just saw LUCENE-3819. All the problems stated here have been addressed there. 
So closing this issue.

> Text Alignment in the "status-item" tab of the Website needs to be word 
> wrapped
> ---
>
> Key: LUCENE-3810
> URL: https://issues.apache.org/jira/browse/LUCENE-3810
> Project: Lucene - Java
>  Issue Type: Improvement
>Reporter: Varun Thacker
>Assignee: Varun Thacker
>Priority: Minor
> Attachments: LUCENE-3810.patch, lucene.png
>
>
> If we go to http://lucene.apache.org/core/ or http://lucene.apache.org/solr/ 
> under the Latest Dev section the text is exceeding the css boundary. This is 
> happening on both Firefox and Chrome. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: Can we add isDaemon check to LTC.threadCleanup?

2012-02-22 Thread Dawid Weiss
I think it should be closed gracefully, even if it's a daemon thread.
Which thread do you have in mind?

Dawid

On Thu, Feb 23, 2012 at 7:33 AM, Shai Erera  wrote:
> Hi
>
> I think that if a thread is daemon, but still running, LTC.threadCleanup
> should not report it. What do you think?
>
> Shai

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-3144) Parameter marshalling is broken in Solrj JavaBinUpdateRequestCodec

2012-02-22 Thread Sami Siren (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-3144?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13214401#comment-13214401
 ] 

Sami Siren commented on SOLR-3144:
--

b1. However it might be a good idea to replace the toNamedList method in the 
Codec by the toNamedList method from SolrParams.

Yeah, I agree.

> Parameter marshalling is broken in Solrj JavaBinUpdateRequestCodec 
> ---
>
> Key: SOLR-3144
> URL: https://issues.apache.org/jira/browse/SOLR-3144
> Project: Solr
>  Issue Type: Bug
>  Components: clients - java
>Affects Versions: 4.0
>Reporter: Jörg Maier
> Fix For: 4.0
>
> Attachments: JavaBinUpdateRequestCodec.patch
>
>
> The parameter marshalling de-marshalling is broken in Solrj's 
> JavaBinUpdateRequestCodec. 
> The bug can be reproduced by adding a parameter e.g. overwrite=false as 
> parameter to the UpdateRequest. After desiarilizing on the backend side the 
> value will be not "false" it will be "[false]" which results in an Exception 
> in the backend and documents will not be imported.
> This issue can easily be fixed by replacing the serialization method with the 
> correct one in SolrParams. See also this gist for a working version: 
> https://gist.github.com/1853544

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Closed] (SOLR-3144) Parameter marshalling is broken in Solrj JavaBinUpdateRequestCodec

2012-02-22 Thread Closed

 [ 
https://issues.apache.org/jira/browse/SOLR-3144?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jörg Maier closed SOLR-3144.



> Parameter marshalling is broken in Solrj JavaBinUpdateRequestCodec 
> ---
>
> Key: SOLR-3144
> URL: https://issues.apache.org/jira/browse/SOLR-3144
> Project: Solr
>  Issue Type: Bug
>  Components: clients - java
>Affects Versions: 4.0
>Reporter: Jörg Maier
> Fix For: 4.0
>
> Attachments: JavaBinUpdateRequestCodec.patch
>
>
> The parameter marshalling de-marshalling is broken in Solrj's 
> JavaBinUpdateRequestCodec. 
> The bug can be reproduced by adding a parameter e.g. overwrite=false as 
> parameter to the UpdateRequest. After desiarilizing on the backend side the 
> value will be not "false" it will be "[false]" which results in an Exception 
> in the backend and documents will not be imported.
> This issue can easily be fixed by replacing the serialization method with the 
> correct one in SolrParams. See also this gist for a working version: 
> https://gist.github.com/1853544

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Resolved] (SOLR-3144) Parameter marshalling is broken in Solrj JavaBinUpdateRequestCodec

2012-02-22 Thread Resolved

 [ 
https://issues.apache.org/jira/browse/SOLR-3144?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jörg Maier resolved SOLR-3144.
--

Resolution: Duplicate

Hi Sami,

i was working on an earlier build than the current head. Obviously previous to 
your fix. I couldnt verify the existence of this bug on trunk anymore and 
everything works fine for me. I missed SOLR-3037.

However it might be a good idea to replace the toNamedList method in the Codec 
by the toNamedList method from SolrParams.

I close this as duplicate to SOLR-3037.


> Parameter marshalling is broken in Solrj JavaBinUpdateRequestCodec 
> ---
>
> Key: SOLR-3144
> URL: https://issues.apache.org/jira/browse/SOLR-3144
> Project: Solr
>  Issue Type: Bug
>  Components: clients - java
>Affects Versions: 4.0
>Reporter: Jörg Maier
> Fix For: 4.0
>
> Attachments: JavaBinUpdateRequestCodec.patch
>
>
> The parameter marshalling de-marshalling is broken in Solrj's 
> JavaBinUpdateRequestCodec. 
> The bug can be reproduced by adding a parameter e.g. overwrite=false as 
> parameter to the UpdateRequest. After desiarilizing on the backend side the 
> value will be not "false" it will be "[false]" which results in an Exception 
> in the backend and documents will not be imported.
> This issue can easily be fixed by replacing the serialization method with the 
> correct one in SolrParams. See also this gist for a working version: 
> https://gist.github.com/1853544

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-3155) Zookeeper info servlet should use JSON library

2012-02-22 Thread Ryan McKinley (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-3155?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ryan McKinley updated SOLR-3155:


Attachment: ZookeeperInfoServlet.java

Here is an updated version that uses the noggit library to write JSON

I don't have a good zookeeper setup... stefan, can you test this?

> Zookeeper info servlet should use JSON library
> --
>
> Key: SOLR-3155
> URL: https://issues.apache.org/jira/browse/SOLR-3155
> Project: Solr
>  Issue Type: Improvement
>Reporter: Ryan McKinley
> Attachments: ZookeeperInfoServlet.java
>
>
> Some of the JSON that the zookeeper info servlet spits out is not valid.  
> Rather then try to fix it, I think we should just use an existing (tested!) 
> json framework.  the noggit writer is available, or maybe expose the solr 
> JSONWriter

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Can we add isDaemon check to LTC.threadCleanup?

2012-02-22 Thread Shai Erera
Hi

I think that if a thread is daemon, but still running, LTC.threadCleanup
should not report it. What do you think?

Shai


[jira] [Resolved] (SOLR-3153) When a leader goes down he should ask replicas to sync in parallel rather than serially.

2012-02-22 Thread Mark Miller (Resolved) (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-3153?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mark Miller resolved SOLR-3153.
---

Resolution: Fixed

Committed.

> When a leader goes down he should ask replicas to sync in parallel rather 
> than serially.
> 
>
> Key: SOLR-3153
> URL: https://issues.apache.org/jira/browse/SOLR-3153
> Project: Solr
>  Issue Type: Improvement
>  Components: SolrCloud
>Reporter: Mark Miller
>Assignee: Mark Miller
>Priority: Minor
> Fix For: 4.0
>
> Attachments: SOLR-3153.patch
>
>
> Need to finish this todo.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-3040) Avoid use of qt param for the DIH in the admin UI

2012-02-22 Thread David Smiley (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-3040?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13214321#comment-13214321
 ] 

David Smiley commented on SOLR-3040:


I'm going to commit this within 24 hours unless someone says otherwise.  

> Avoid use of qt param for the DIH in the admin UI
> -
>
> Key: SOLR-3040
> URL: https://issues.apache.org/jira/browse/SOLR-3040
> Project: Solr
>  Issue Type: Improvement
>  Components: web gui
>Reporter: David Smiley
>Priority: Minor
> Attachments: SOLR-3040_avoid_qt_for_DIH_in_admin_ui.patch
>
>
> I really, really dislike that the qt parameter can used to refer to request 
> handlers starting with a '/', _especially_ for non-search handlers.  The 
> admin UI has one place I am aware of that attempts to do this, which is the 
> DIH's admin page.  Since we have two UIs in trunk, the new and old, there are 
> actually two UIs where this occurs, and the old UI has two related files that 
> need updating, in order to address this issue.
> An example URL generated by the UI today is this:
>   http://localhost:8983/solr/rss/select?qt=/dataimport&command=show-config
> And here it is without using qt:
>   http://localhost:8983/solr/rss/dataimport?command=show-config
> I do realize that fixing this issue as I do in my patch will make the UI not 
> work if the DIH isn't registered with a leading '/', but honestly I wonder 
> who if anyone out there does that.  And even then, it's easily rectified.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: [JENKINS] Lucene-Solr-tests-only-trunk - Build # 12502 - Failure

2012-02-22 Thread Ryan McKinley
should be fixed in revision: 1292642


On Wed, Feb 22, 2012 at 7:15 PM, Apache Jenkins Server
 wrote:
> Build: https://builds.apache.org/job/Lucene-Solr-tests-only-trunk/12502/
>
> 1 tests failed.
> FAILED:  
> org.apache.solr.handler.admin.LogLevelHandlerTest.testLogLevelHandlerOutput
>
> Error Message:
> Exception during query
>
> Stack Trace:
> java.lang.RuntimeException: Exception during query
>        at org.apache.solr.SolrTestCaseJ4.assertQ(SolrTestCaseJ4.java:423)
>        at 
> org.apache.solr.handler.admin.LogLevelHandlerTest.testLogLevelHandlerOutput(LogLevelHandlerTest.java:35)
>        at 
> org.apache.lucene.util.LuceneTestCase$SubclassSetupTeardownRule$1.evaluate(LuceneTestCase.java:707)
>        at 
> org.apache.lucene.util.LuceneTestCase$InternalSetupTeardownRule$1.evaluate(LuceneTestCase.java:606)
>        at 
> org.apache.lucene.util.LuceneTestCase$TestResultInterceptorRule$1.evaluate(LuceneTestCase.java:511)
>        at 
> org.apache.lucene.util.LuceneTestCase$RememberThreadRule$1.evaluate(LuceneTestCase.java:569)
>        at 
> org.apache.lucene.util.LuceneTestCaseRunner.runChild(LuceneTestCaseRunner.java:165)
>        at 
> org.apache.lucene.util.LuceneTestCaseRunner.runChild(LuceneTestCaseRunner.java:57)
> Caused by: java.lang.RuntimeException: REQUEST FAILED: 
> xpath=//arr[@name='loggers']/lst/str[.='global']/../str[@name='level'][.='INFO']
>        xml response was: 
> 
> 0 name="QTime">5java.util.logging name="levels">FINESTFINECONFIGINFOWARNINGSEVEREOFF  name="loggers">root name="level">SEVEREtrue name="name">/solrSEVERE name="set">falseglobal name="level">SEVEREfalse name="name">httpclient name="set">falsehttpclient.wire name="level"/>false name="name">httpclient.wire.contentSEVERE name="set">false name="name">httpclient.wire.headerSEVERE name="set">falsejavax name="level"/>false name="name">javax.managementSEVERE name="set">false name="name">javax.management.mbeanserver name="level">SEVEREfalse name="name">javax.management.miscSEVERE name="set">false name="name">javax.management.mletSEVERE name="set">false name="name">javax.management.modelmbean name="level">SEVEREfalse name="name">javax.management.monitorSEVERE name="set">false name="name">javax.management.notification name="level">SEVEREfalse name="name">javax.management.relation name="level">SEVEREfalse name="name">javax.management.remote name="set">false name="name">javax.management.remote.misc name="level">SEVEREfalse name="name">javax.management.remote.rmi name="level">SEVEREfalse name="name">javax.management.remote.timeout name="level">SEVEREfalse name="name">javax.management.snmpSEVERE name="set">false name="name">javax.management.snmp.daemon name="level">SEVEREfalse name="name">javax.management.timerSEVERE name="set">falseorg name="level"/>false name="name">org.apache name="set">false name="name">org.apache.commons name="set">false name="name">org.apache.commons.httpclient name="set">false name="name">org.apache.commons.httpclient.ChunkedInputStream name="level">SEVEREfalse name="name">org.apache.commons.httpclient.HeaderElement name="level">SEVEREfalse name="name">org.apache.commons.httpclient.HttpClient name="level">SEVEREfalse name="name">org.apache.commons.httpclient.HttpConnection name="level">SEVEREfalse name="name">org.apache.commons.httpclient.HttpMethodBase name="level">SEVEREfalse name="name">org.apache.commons.httpclient.HttpMethodDirector name="level">SEVEREfalse name="name">org.apache.commons.httpclient.HttpParser name="level">SEVEREfalse name="name">org.apache.commons.httpclient.HttpState name="level">SEVEREfalse name="name">org.apache.commons.httpclient.MultiThreadedHttpConnectionManager  name="level">SEVEREfalse name="name">org.apache.commons.httpclient.SimpleHttpConnectionManager  name="level">SEVEREfalse name="name">org.apache.commons.httpclient.auth name="set">false name="name">org.apache.commons.httpclient.auth.AuthChallengeProcessor  name="level">SEVEREfalse name="name">org.apache.commons.httpclient.cookie name="level"/>false name="name">org.apache.commons.httpclient.cookie.CookiePolicy name="level">SEVEREfalse name="name">org.apache.commons.httpclient.cookie.CookieSpec name="level">SEVEREfalse name="name">org.apache.commons.httpclient.methods name="level"/>false name="name">org.apache.commons.httpclient.methods.EntityEnclosingMethod  name="level">SEVEREfalse name="name">org.apache.commons.httpclient.methods.ExpectContinueMethod  name="level">SEVEREfalse name="name">org.apache.commons.httpclient.methods.GetMethod name="level">SEVEREfalse name="name">org.apache.commons.httpclient.methods.InputStreamRequestEntity  name="level">SEVEREfalse name="name">org.apache.commons.httpclient.methods.PostMethod name="level">SEVEREfalse name="name">org.apache.commons.httpclient.params name="level"/>false name="name">org.apache.commons.httpclient.params.DefaultHttpParams name="level">SEVEREfalse name="name">org.apache.commons.httpclient.params.HttpMethodParams name="level">SEVEREfalse name="name">org.apache.common

[JENKINS] Lucene-Solr-tests-only-trunk-java7 - Build # 1794 - Failure

2012-02-22 Thread Apache Jenkins Server
Build: https://builds.apache.org/job/Lucene-Solr-tests-only-trunk-java7/1794/

1 tests failed.
FAILED:  
org.apache.solr.handler.admin.LogLevelHandlerTest.testLogLevelHandlerOutput

Error Message:
Exception during query

Stack Trace:
java.lang.RuntimeException: Exception during query
at org.apache.solr.SolrTestCaseJ4.assertQ(SolrTestCaseJ4.java:423)
at 
org.apache.solr.handler.admin.LogLevelHandlerTest.testLogLevelHandlerOutput(LogLevelHandlerTest.java:35)
at 
org.apache.lucene.util.LuceneTestCase$SubclassSetupTeardownRule$1.evaluate(LuceneTestCase.java:707)
at 
org.apache.lucene.util.LuceneTestCase$InternalSetupTeardownRule$1.evaluate(LuceneTestCase.java:606)
at 
org.apache.lucene.util.LuceneTestCase$TestResultInterceptorRule$1.evaluate(LuceneTestCase.java:511)
at 
org.apache.lucene.util.LuceneTestCase$RememberThreadRule$1.evaluate(LuceneTestCase.java:569)
at 
org.apache.lucene.util.LuceneTestCaseRunner.runChild(LuceneTestCaseRunner.java:165)
at 
org.apache.lucene.util.LuceneTestCaseRunner.runChild(LuceneTestCaseRunner.java:57)
Caused by: java.lang.RuntimeException: REQUEST FAILED: 
xpath=//arr[@name='loggers']/lst/str[.='global']/../str[@name='level'][.='INFO']
xml response was: 

08java.util.loggingFINESTFINECONFIGINFOWARNINGSEVEREOFFrootSEVEREtrue/solrSEVEREfalseglobalSEVEREfalsehttpclientfalsehttpclient.wirefalsehttpclient.wire.contentSEVEREfalsehttpclient.wire.headerSEVEREfalsejavaxfalsejavax.managementSEVEREfalsejavax.management.mbeanserverSEVEREfalsejavax.management.miscSEVEREfalsejavax.management.mletSEVEREfalsejavax.management.modelmbeanSEVEREfalsejavax.management.monitorSEVEREfalsejavax.management.notificationSEVEREfalsejavax.management.relationSEVEREfalsejavax.management.snmpSEVEREfalsejavax.management.snmp.daemonSEVEREfalsejavax.management.timerSEVEREfalseorgfalseorg.apachefalseorg.apache.commonsfalseorg.apache.commons.httpclientfalseorg.apache.commons.httpclient.ChunkedInputStreamSEVEREfalseorg.apache.commons.httpclient.HeaderElementSEVEREfalseorg.apache.commons.httpclient.HttpClientSEVEREfalseorg.apache.commons.httpclient.HttpConnectionSEVEREfalseorg.apache.commons.httpclient.HttpMethodBaseSEVEREfalseorg.apache.commons.httpclient.HttpMethodDirectorSEVEREfalseorg.apache.commons.httpclient.HttpParserSEVEREfalseorg.apache.commons.httpclient.HttpStateSEVEREfalseorg.apache.commons.httpclient.MultiThreadedHttpConnectionManagerSEVEREfalseorg.apache.commons.httpclient.SimpleHttpConnectionManagerSEVEREfalseorg.apache.commons.httpclient.authfalseorg.apache.commons.httpclient.auth.AuthChallengeProcessorSEVEREfalseorg.apache.commons.httpclient.cookiefalseorg.apache.commons.httpclient.cookie.CookiePolicySEVEREfalseorg.apache.commons.httpclient.cookie.CookieSpecSEVEREfalseorg.apache.commons.httpclient.methodsfalseorg.apache.commons.httpclient.methods.EntityEnclosingMethodSEVEREfalseorg.apache.commons.httpclient.methods.ExpectContinueMethodSEVEREfalseorg.apache.commons.httpclient.methods.GetMethodSEVEREfalseorg.apache.commons.httpclient.methods.InputStreamRequestEntitySEVEREfalseorg.apache.commons.httpclient.methods.PostMethodSEVEREfalseorg.apache.commons.httpclient.paramsfalseorg.apache.commons.httpclient.params.DefaultHttpParamsSEVEREfalseorg.apache.commons.httpclient.params.HttpMethodParamsSEVEREfalseorg.apache.commons.httpclient.utilfalseorg.apache.commons.httpclient.util.EncodingUtilSEVEREfalseorg.apache.commons.httpclient.util.ExceptionUtilSEVEREfalseorg.apache.commons.httpclient.util.IdleConnectionHandlerSEVEREfalseorg.apache.solrfalseorg.apache.solr.BaseDistributedSearchTestCaseSEVEREfalseorg.apache.solr.SolrTestCaseJ4SEVEREfalseorg.apache.solr.analysisfalseorg.apache.solr.analysis.BaseCharFilterFactorySEVEREfalseorg.apache.solr.analysis.BaseTokenFilterFactorySEVEREfalseorg.apache.solr.analysis.BaseTokenStreamFactorySEVEREfalseorg.apache.solr.analysis.BaseTokenizerFactorySEVEREfalseorg.apache.solr.clientfalseorg.apache.solr.client.solrjfalseorg.apache.solr.client.solrj.implfalseorg.apache.solr.client.solrj.impl.CommonsHttpSolrServerSEVEREfalseorg.apache.solr.cloudfalseorg.apache.solr.cloud.AbstractZkTestCaseSEVEREfalseorg.apache.solr.cloud.LeaderElectorSEVEREfalseorg.apache.solr.cloud.NodeStateWatcherSEVEREfalseorg.apache.solr.cloud.OverseerSEVEREfalseorg.apache.solr.cloud.RecoveryStrategySEVEREfalseorg.apache.solr.cloud.ShardLeaderWatcherSEVEREfalseorg.apache.solr.cloud.SyncStrategySEVEREfalseorg.apache.solr.cloud.ZkControllerSEVEREfalseorg.apache.solr.cloud.ZkTestServerSEVEREfalseorg.apache.solr.commonfalseorg.apache.solr.common.cloudfalseorg.apache.solr.common.cloud.ConnectionManagerSEVEREfalseorg.apache.solr.common.cloud.DefaultConnectionStrategySEVEREfalseorg.apache.solr.common.cloud.SolrZkClientSEVEREfalseorg.apache.solr.common.cloud.ZkStateReaderSEVEREfalseorg.apache.solr.common.utilfalseorg.apache.solr.common.util.SystemIdResolverSEVEREfalseorg.apache.solr.corefalseorg.apache.solr.core.CachingDirectoryFactorySEVEREfalse

[JENKINS] Lucene-Solr-tests-only-trunk - Build # 12502 - Failure

2012-02-22 Thread Apache Jenkins Server
Build: https://builds.apache.org/job/Lucene-Solr-tests-only-trunk/12502/

1 tests failed.
FAILED:  
org.apache.solr.handler.admin.LogLevelHandlerTest.testLogLevelHandlerOutput

Error Message:
Exception during query

Stack Trace:
java.lang.RuntimeException: Exception during query
at org.apache.solr.SolrTestCaseJ4.assertQ(SolrTestCaseJ4.java:423)
at 
org.apache.solr.handler.admin.LogLevelHandlerTest.testLogLevelHandlerOutput(LogLevelHandlerTest.java:35)
at 
org.apache.lucene.util.LuceneTestCase$SubclassSetupTeardownRule$1.evaluate(LuceneTestCase.java:707)
at 
org.apache.lucene.util.LuceneTestCase$InternalSetupTeardownRule$1.evaluate(LuceneTestCase.java:606)
at 
org.apache.lucene.util.LuceneTestCase$TestResultInterceptorRule$1.evaluate(LuceneTestCase.java:511)
at 
org.apache.lucene.util.LuceneTestCase$RememberThreadRule$1.evaluate(LuceneTestCase.java:569)
at 
org.apache.lucene.util.LuceneTestCaseRunner.runChild(LuceneTestCaseRunner.java:165)
at 
org.apache.lucene.util.LuceneTestCaseRunner.runChild(LuceneTestCaseRunner.java:57)
Caused by: java.lang.RuntimeException: REQUEST FAILED: 
xpath=//arr[@name='loggers']/lst/str[.='global']/../str[@name='level'][.='INFO']
xml response was: 

05java.util.loggingFINESTFINECONFIGINFOWARNINGSEVEREOFFrootSEVEREtrue/solrSEVEREfalseglobalSEVEREfalsehttpclientfalsehttpclient.wirefalsehttpclient.wire.contentSEVEREfalsehttpclient.wire.headerSEVEREfalsejavaxfalsejavax.managementSEVEREfalsejavax.management.mbeanserverSEVEREfalsejavax.management.miscSEVEREfalsejavax.management.mletSEVEREfalsejavax.management.modelmbeanSEVEREfalsejavax.management.monitorSEVEREfalsejavax.management.notificationSEVEREfalsejavax.management.relationSEVEREfalsejavax.management.remotefalsejavax.management.remote.miscSEVEREfalsejavax.management.remote.rmiSEVEREfalsejavax.management.remote.timeoutSEVEREfalsejavax.management.snmpSEVEREfalsejavax.management.snmp.daemonSEVEREfalsejavax.management.timerSEVEREfalseorgfalseorg.apachefalseorg.apache.commonsfalseorg.apache.commons.httpclientfalseorg.apache.commons.httpclient.ChunkedInputStreamSEVEREfalseorg.apache.commons.httpclient.HeaderElementSEVEREfalseorg.apache.commons.httpclient.HttpClientSEVEREfalseorg.apache.commons.httpclient.HttpConnectionSEVEREfalseorg.apache.commons.httpclient.HttpMethodBaseSEVEREfalseorg.apache.commons.httpclient.HttpMethodDirectorSEVEREfalseorg.apache.commons.httpclient.HttpParserSEVEREfalseorg.apache.commons.httpclient.HttpStateSEVEREfalseorg.apache.commons.httpclient.MultiThreadedHttpConnectionManagerSEVEREfalseorg.apache.commons.httpclient.SimpleHttpConnectionManagerSEVEREfalseorg.apache.commons.httpclient.authfalseorg.apache.commons.httpclient.auth.AuthChallengeProcessorSEVEREfalseorg.apache.commons.httpclient.cookiefalseorg.apache.commons.httpclient.cookie.CookiePolicySEVEREfalseorg.apache.commons.httpclient.cookie.CookieSpecSEVEREfalseorg.apache.commons.httpclient.methodsfalseorg.apache.commons.httpclient.methods.EntityEnclosingMethodSEVEREfalseorg.apache.commons.httpclient.methods.ExpectContinueMethodSEVEREfalseorg.apache.commons.httpclient.methods.GetMethodSEVEREfalseorg.apache.commons.httpclient.methods.InputStreamRequestEntitySEVEREfalseorg.apache.commons.httpclient.methods.PostMethodSEVEREfalseorg.apache.commons.httpclient.paramsfalseorg.apache.commons.httpclient.params.DefaultHttpParamsSEVEREfalseorg.apache.commons.httpclient.params.HttpMethodParamsSEVEREfalseorg.apache.commons.httpclient.utilfalseorg.apache.commons.httpclient.util.EncodingUtilSEVEREfalseorg.apache.commons.httpclient.util.ExceptionUtilSEVEREfalseorg.apache.commons.httpclient.util.IdleConnectionHandlerSEVEREfalseorg.apache.solrfalseorg.apache.solr.BaseDistributedSearchTestCaseSEVEREfalseorg.apache.solr.SolrTestCaseJ4SEVEREfalseorg.apache.solr.analysisfalseorg.apache.solr.analysis.BaseCharFilterFactorySEVEREfalseorg.apache.solr.analysis.BaseTokenFilterFactorySEVEREfalseorg.apache.solr.analysis.BaseTokenStreamFactorySEVEREfalseorg.apache.solr.analysis.BaseTokenizerFactorySEVEREfalseorg.apache.solr.clientfalseorg.apache.solr.client.solrjfalseorg.apache.solr.client.solrj.implfalseorg.apache.solr.client.solrj.impl.CommonsHttpSolrServerSEVEREfalseorg.apache.solr.client.solrj.impl.StreamingUpdateSolrServerSEVEREfalseorg.apache.solr.cloudfalseorg.apache.solr.cloud.AbstractZkTestCaseSEVEREfalseorg.apache.solr.cloud.LeaderElectorSEVEREfalseorg.apache.solr.cloud.NodeStateWatcherSEVEREfalseorg.apache.solr.cloud.OverseerSEVEREfalseorg.apache.solr.cloud.RecoveryStrategySEVEREfalseorg.apache.solr.cloud.ShardLeaderWatcherSEVEREfalseorg.apache.solr.cloud.SyncStrategySEVEREfalseorg.apache.solr.cloud.ZkControllerSEVEREfalseorg.apache.solr.cloud.ZkTestServerSEVEREfalseorg.apache.solr.commonfalseorg.apache.solr.common.cloudfalseorg.apache.solr.common.cloud.ConnectionManagerSEVEREfalseorg.apache.solr.common.cloud.DefaultConnectionStrategySEVEREfalseorg.apache.solr.common.cloud.SolrZkClientSEVEREfalseorg.apac

[jira] [Updated] (LUCENE-3820) Wrong trailing index calculation in PatternReplaceCharFilter

2012-02-22 Thread Robert Muir (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-3820?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Robert Muir updated LUCENE-3820:


Attachment: LUCENE-3820_test.patch

updated patch, this tests only ascii (to avoid stupid problems in outdated 
regex support).

But there are a lot of offset problems (perhaps this corresponds to the warning 
in the class's javadocs?), including things like offsets being corrected to 
negative numbers...

> Wrong trailing index calculation in PatternReplaceCharFilter
> 
>
> Key: LUCENE-3820
> URL: https://issues.apache.org/jira/browse/LUCENE-3820
> Project: Lucene - Java
>  Issue Type: Bug
>Reporter: Dawid Weiss
>Assignee: Dawid Weiss
>Priority: Minor
> Fix For: 4.0
>
> Attachments: LUCENE-3820.patch, LUCENE-3820_test.patch, 
> LUCENE-3820_test.patch
>
>
> I need to use PatternReplaceCharFilter's index corrections directly and it 
> fails for me -- the trailing index is not mapped correctly for a pattern 
> "\\.[\\s]*" and replacement ".", input "A. .B.".
> I tried to understand the logic in getReplaceBlock but I eventually failed 
> and simply rewrote it from scratch. After my changes a few tests don't pass 
> but I don't know if it's the tests that are screwed up or my logic. In 
> essence, the difference between the previous implementation and my 
> implementation is how indexes are mapped for shorter replacements. I shift 
> indexes of shorter regions to the "right" of the original index pool and the 
> previous patch seems to squeeze them to the left (don't know why though).
> If anybody remembers how it's supposed to work, feel free to correct me?

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-3820) Wrong trailing index calculation in PatternReplaceCharFilter

2012-02-22 Thread Robert Muir (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3820?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13214235#comment-13214235
 ] 

Robert Muir commented on LUCENE-3820:
-

to fix dawid's problem we can probably modify this test only for ascii, i 
suspect the unicode "problems"
are going to be impossible to fix given java's regex library (i think it does 
not treat "." as codepoint
but code unit). I'll take another stab at that just to tackle the offsets issue 
he is seeing.

> Wrong trailing index calculation in PatternReplaceCharFilter
> 
>
> Key: LUCENE-3820
> URL: https://issues.apache.org/jira/browse/LUCENE-3820
> Project: Lucene - Java
>  Issue Type: Bug
>Reporter: Dawid Weiss
>Assignee: Dawid Weiss
>Priority: Minor
> Fix For: 4.0
>
> Attachments: LUCENE-3820.patch, LUCENE-3820_test.patch
>
>
> I need to use PatternReplaceCharFilter's index corrections directly and it 
> fails for me -- the trailing index is not mapped correctly for a pattern 
> "\\.[\\s]*" and replacement ".", input "A. .B.".
> I tried to understand the logic in getReplaceBlock but I eventually failed 
> and simply rewrote it from scratch. After my changes a few tests don't pass 
> but I don't know if it's the tests that are screwed up or my logic. In 
> essence, the difference between the previous implementation and my 
> implementation is how indexes are mapped for shorter replacements. I shift 
> indexes of shorter regions to the "right" of the original index pool and the 
> previous patch seems to squeeze them to the left (don't know why though).
> If anybody remembers how it's supposed to work, feel free to correct me?

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-3767) Explore streaming Viterbi search in Kuromoji

2012-02-22 Thread Robert Muir (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3767?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13214232#comment-13214232
 ] 

Robert Muir commented on LUCENE-3767:
-

{quote}
Mike, could you share some details on any query-time trickiness here? Are you 
thinking about composing a query with both the compound and its 
parts/decompounds? Thanks.
{quote}

I don't understand Mike's comment there either. positionIncrement=0 is really 
going to be treated like synonyms by the QP,
so it shouldn't involve any trickiness.

> Explore streaming Viterbi search in Kuromoji
> 
>
> Key: LUCENE-3767
> URL: https://issues.apache.org/jira/browse/LUCENE-3767
> Project: Lucene - Java
>  Issue Type: Improvement
>  Components: modules/analysis
>Reporter: Michael McCandless
>Assignee: Michael McCandless
> Fix For: 3.6, 4.0
>
> Attachments: LUCENE-3767.patch, LUCENE-3767.patch, LUCENE-3767.patch, 
> compound_diffs.txt
>
>
> I've been playing with the idea of changing the Kuromoji viterbi
> search to be 2 passes (intersect, backtrace) instead of 4 passes
> (break into sentences, intersect, score, backtrace)... this is very
> much a work in progress, so I'm just getting my current state up.
> It's got tons of nocommits, doesn't properly handle the user dict nor
> extended modes yet, etc.
> One thing I'm playing with is to add a double backtrace for the long
> compound tokens, ie, instead of penalizing these tokens so that
> shorter tokens are picked, leave the scores unchanged but on backtrace
> take that penalty and use it as a threshold for a 2nd best
> segmentation...

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-3767) Explore streaming Viterbi search in Kuromoji

2012-02-22 Thread Christian Moen (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3767?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13214227#comment-13214227
 ] 

Christian Moen commented on LUCENE-3767:


{quote}
I left the default mode as Mode.SEARCH... maybe if we can somehow run some 
relevance tests we can make the default SEARCH_WITH_COMPOUNDS.
But it'd also be tricky at query time...
{quote}

Mike, could you share some details on any query-time trickiness here?  Are you 
thinking about composing a query with both the compound and its 
parts/decompounds?  Thanks.


> Explore streaming Viterbi search in Kuromoji
> 
>
> Key: LUCENE-3767
> URL: https://issues.apache.org/jira/browse/LUCENE-3767
> Project: Lucene - Java
>  Issue Type: Improvement
>  Components: modules/analysis
>Reporter: Michael McCandless
>Assignee: Michael McCandless
> Fix For: 3.6, 4.0
>
> Attachments: LUCENE-3767.patch, LUCENE-3767.patch, LUCENE-3767.patch, 
> compound_diffs.txt
>
>
> I've been playing with the idea of changing the Kuromoji viterbi
> search to be 2 passes (intersect, backtrace) instead of 4 passes
> (break into sentences, intersect, score, backtrace)... this is very
> much a work in progress, so I'm just getting my current state up.
> It's got tons of nocommits, doesn't properly handle the user dict nor
> extended modes yet, etc.
> One thing I'm playing with is to add a double backtrace for the long
> compound tokens, ie, instead of penalizing these tokens so that
> shorter tokens are picked, leave the scores unchanged but on backtrace
> take that penalty and use it as a threshold for a 2nd best
> segmentation...

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (LUCENE-3820) Wrong trailing index calculation in PatternReplaceCharFilter

2012-02-22 Thread Robert Muir (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-3820?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Robert Muir updated LUCENE-3820:


Attachment: LUCENE-3820_test.patch

Here's a simple random test showing some existing bugs in the filter... 
* there are offsets problems as dawid notices...
* blockbuffer should always oversize by 1 character, if a block ends on a high 
surrogate (rare) it should do one additional read() so it doesnt create invalid 
unicode


> Wrong trailing index calculation in PatternReplaceCharFilter
> 
>
> Key: LUCENE-3820
> URL: https://issues.apache.org/jira/browse/LUCENE-3820
> Project: Lucene - Java
>  Issue Type: Bug
>Reporter: Dawid Weiss
>Assignee: Dawid Weiss
>Priority: Minor
> Fix For: 4.0
>
> Attachments: LUCENE-3820.patch, LUCENE-3820_test.patch
>
>
> I need to use PatternReplaceCharFilter's index corrections directly and it 
> fails for me -- the trailing index is not mapped correctly for a pattern 
> "\\.[\\s]*" and replacement ".", input "A. .B.".
> I tried to understand the logic in getReplaceBlock but I eventually failed 
> and simply rewrote it from scratch. After my changes a few tests don't pass 
> but I don't know if it's the tests that are screwed up or my logic. In 
> essence, the difference between the previous implementation and my 
> implementation is how indexes are mapped for shorter replacements. I shift 
> indexes of shorter regions to the "right" of the original index pool and the 
> previous patch seems to squeeze them to the left (don't know why though).
> If anybody remembers how it's supposed to work, feel free to correct me?

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-3767) Explore streaming Viterbi search in Kuromoji

2012-02-22 Thread Christian Moen (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3767?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13214217#comment-13214217
 ] 

Christian Moen commented on LUCENE-3767:


Robert, some comments are below.

{quote}
And I tend to like Mike's improvements from a relevance perspective for these 
reasons:

# keeping the original compound term for improved precision
# preventing compound decomposition from having any unrelated negative impact 
on the rest of the tokenization
{quote}

Very good points.

{quote}
So I think we should pursue this change, even if we want to separately train a 
dictionary in the future, because in that case, we would just disable the kanji 
decomposition heuristic but keep the heuristic (obviously re-tuned!) for 
katakana?
{quote}

I agree completely.

{quote}
The dictionary documentation for the original ipadic has the ability to hold 
compound data (not in mecab-ipadic though, so maybe it was never 
implemented?!), but I don't actually see it in any implementations. So yeah, we 
would need to find a corpus containing compound information (and of course 
extend the file format and add support to kuromoji) to support that.

However, would this really solve the total issue? Wouldn't that really only 
help for known kanji compounds... whereas most katakana compounds (e.g. the 
software engineer example) are expected to be OOV anyway? So it seems like, 
even if we ensured the dictionary was annotated for long kanji such that we 
always used decompounded forms, we need a 'heuristical' decomposition like 
search-mode either way, at least for the unknown katakana case?
{quote}

I've made an inquiry to a friend who did his PhD work at Prof. Matsumoto's lab 
at NAIST (where ChaSen was made) and I've made en inquiry regarding compound 
information and the Kyoto Corpus.


You are perfectly right that this doesn't solve the complete problem as unknown 
words can actually be compounds -- unknown compounds.  The approach used today 
is basically adding all the potential decompounds the model knows about to the 
lattice and see if a short path can be found often in combination with an 
unknown word.

We get errors such as クイーンズコミックス (Queen's Comics) becoming クイーン ズコミックス (Queen 
Zukuomikkusu) because クイーン (Queen) is known.

I'll open up a separate JIRA for discussing search-mode improvements. :)


> Explore streaming Viterbi search in Kuromoji
> 
>
> Key: LUCENE-3767
> URL: https://issues.apache.org/jira/browse/LUCENE-3767
> Project: Lucene - Java
>  Issue Type: Improvement
>  Components: modules/analysis
>Reporter: Michael McCandless
>Assignee: Michael McCandless
> Fix For: 3.6, 4.0
>
> Attachments: LUCENE-3767.patch, LUCENE-3767.patch, LUCENE-3767.patch, 
> compound_diffs.txt
>
>
> I've been playing with the idea of changing the Kuromoji viterbi
> search to be 2 passes (intersect, backtrace) instead of 4 passes
> (break into sentences, intersect, score, backtrace)... this is very
> much a work in progress, so I'm just getting my current state up.
> It's got tons of nocommits, doesn't properly handle the user dict nor
> extended modes yet, etc.
> One thing I'm playing with is to add a double backtrace for the long
> compound tokens, ie, instead of penalizing these tokens so that
> shorter tokens are picked, leave the scores unchanged but on backtrace
> take that penalty and use it as a threshold for a 2nd best
> segmentation...

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Resolved] (SOLR-3151) Replace zookeeper.jsp with a servlet

2012-02-22 Thread Ryan McKinley (Resolved) (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-3151?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ryan McKinley resolved SOLR-3151.
-

Resolution: Fixed

This is now a servlet, and future improvements can be tracked in SOLR-3155

> Replace zookeeper.jsp with a servlet
> 
>
> Key: SOLR-3151
> URL: https://issues.apache.org/jira/browse/SOLR-3151
> Project: Solr
>  Issue Type: Improvement
>Reporter: Ryan McKinley
> Fix For: 4.0
>
> Attachments: SOLR-3151-invalid-json-for-solrconfig.json, 
> SOLR-3151-zookeeper-servlet.patch
>
>
> The zookeeper info is currently generated with a jsp file -- making it harder 
> to maintain then it should be, and harder to include in other applications

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (SOLR-3155) Zookeeper info servlet should use JSON library

2012-02-22 Thread Ryan McKinley (Created) (JIRA)
Zookeeper info servlet should use JSON library
--

 Key: SOLR-3155
 URL: https://issues.apache.org/jira/browse/SOLR-3155
 Project: Solr
  Issue Type: Improvement
Reporter: Ryan McKinley


Some of the JSON that the zookeeper info servlet spits out is not valid.  
Rather then try to fix it, I think we should just use an existing (tested!) 
json framework.  the noggit writer is available, or maybe expose the solr 
JSONWriter

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-3820) Wrong trailing index calculation in PatternReplaceCharFilter

2012-02-22 Thread Robert Muir (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3820?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13214203#comment-13214203
 ] 

Robert Muir commented on LUCENE-3820:
-

Unrelated to this change: we should be using StringBuilder here.

> Wrong trailing index calculation in PatternReplaceCharFilter
> 
>
> Key: LUCENE-3820
> URL: https://issues.apache.org/jira/browse/LUCENE-3820
> Project: Lucene - Java
>  Issue Type: Bug
>Reporter: Dawid Weiss
>Assignee: Dawid Weiss
>Priority: Minor
> Fix For: 4.0
>
> Attachments: LUCENE-3820.patch
>
>
> I need to use PatternReplaceCharFilter's index corrections directly and it 
> fails for me -- the trailing index is not mapped correctly for a pattern 
> "\\.[\\s]*" and replacement ".", input "A. .B.".
> I tried to understand the logic in getReplaceBlock but I eventually failed 
> and simply rewrote it from scratch. After my changes a few tests don't pass 
> but I don't know if it's the tests that are screwed up or my logic. In 
> essence, the difference between the previous implementation and my 
> implementation is how indexes are mapped for shorter replacements. I shift 
> indexes of shorter regions to the "right" of the original index pool and the 
> previous patch seems to squeeze them to the left (don't know why though).
> If anybody remembers how it's supposed to work, feel free to correct me?

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-3767) Explore streaming Viterbi search in Kuromoji

2012-02-22 Thread Robert Muir (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3767?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13214201#comment-13214201
 ] 

Robert Muir commented on LUCENE-3767:
-

Well my comments are only questions based on your previous JIRA comments,
I don't really know anything about decompounding Japanese and my suggestions
are only intuition, mostly based on patterns that have worked for other
languages... so just consider it thinking out loud.

I think your points on search mode are actually related here: though we can't 
really have a plan its good to think about possible paths we might take in the 
future so that we don't lock ourselves out.


> Explore streaming Viterbi search in Kuromoji
> 
>
> Key: LUCENE-3767
> URL: https://issues.apache.org/jira/browse/LUCENE-3767
> Project: Lucene - Java
>  Issue Type: Improvement
>  Components: modules/analysis
>Reporter: Michael McCandless
>Assignee: Michael McCandless
> Fix For: 3.6, 4.0
>
> Attachments: LUCENE-3767.patch, LUCENE-3767.patch, LUCENE-3767.patch, 
> compound_diffs.txt
>
>
> I've been playing with the idea of changing the Kuromoji viterbi
> search to be 2 passes (intersect, backtrace) instead of 4 passes
> (break into sentences, intersect, score, backtrace)... this is very
> much a work in progress, so I'm just getting my current state up.
> It's got tons of nocommits, doesn't properly handle the user dict nor
> extended modes yet, etc.
> One thing I'm playing with is to add a double backtrace for the long
> compound tokens, ie, instead of penalizing these tokens so that
> shorter tokens are picked, leave the scores unchanged but on backtrace
> take that penalty and use it as a threshold for a 2nd best
> segmentation...

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Issue Comment Edited] (SOLR-1301) Solr + Hadoop

2012-02-22 Thread Alexander Kanarsky (Issue Comment Edited) (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-1301?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13214102#comment-13214102
 ] 

Alexander Kanarsky edited comment on SOLR-1301 at 2/23/12 2:09 AM:
---

Note, the hadoop-core-0.20.2-cdh3u3.jar is a part of Cloudera's CDH3 Hadoop 
distribution and is licensed under Apache License v. 2.0.


  was (Author: kanarsky):
Note, the hadoop-core-0.20.2-cdh3u3.jar is part of the Coludera's CDH3u3 
Hadoop distribution and licensed under Apache License v. 2.0.
  
> Solr + Hadoop
> -
>
> Key: SOLR-1301
> URL: https://issues.apache.org/jira/browse/SOLR-1301
> Project: Solr
>  Issue Type: Improvement
>Affects Versions: 1.4
>Reporter: Andrzej Bialecki 
> Fix For: 3.6, 4.0
>
> Attachments: README.txt, SOLR-1301-hadoop-0-20.patch, 
> SOLR-1301-hadoop-0-20.patch, SOLR-1301.patch, SOLR-1301.patch, 
> SOLR-1301.patch, SOLR-1301.patch, SOLR-1301.patch, SOLR-1301.patch, 
> SOLR-1301.patch, SOLR-1301.patch, SOLR-1301.patch, SOLR-1301.patch, 
> SolrRecordWriter.java, commons-logging-1.0.4.jar, 
> commons-logging-api-1.0.4.jar, hadoop-0.19.1-core.jar, 
> hadoop-0.20.1-core.jar, hadoop-core-0.20.2-cdh3u3.jar, hadoop.patch, 
> log4j-1.2.15.jar
>
>
> This patch contains  a contrib module that provides distributed indexing 
> (using Hadoop) to Solr EmbeddedSolrServer. The idea behind this module is 
> twofold:
> * provide an API that is familiar to Hadoop developers, i.e. that of 
> OutputFormat
> * avoid unnecessary export and (de)serialization of data maintained on HDFS. 
> SolrOutputFormat consumes data produced by reduce tasks directly, without 
> storing it in intermediate files. Furthermore, by using an 
> EmbeddedSolrServer, the indexing task is split into as many parts as there 
> are reducers, and the data to be indexed is not sent over the network.
> Design
> --
> Key/value pairs produced by reduce tasks are passed to SolrOutputFormat, 
> which in turn uses SolrRecordWriter to write this data. SolrRecordWriter 
> instantiates an EmbeddedSolrServer, and it also instantiates an 
> implementation of SolrDocumentConverter, which is responsible for turning 
> Hadoop (key, value) into a SolrInputDocument. This data is then added to a 
> batch, which is periodically submitted to EmbeddedSolrServer. When reduce 
> task completes, and the OutputFormat is closed, SolrRecordWriter calls 
> commit() and optimize() on the EmbeddedSolrServer.
> The API provides facilities to specify an arbitrary existing solr.home 
> directory, from which the conf/ and lib/ files will be taken.
> This process results in the creation of as many partial Solr home directories 
> as there were reduce tasks. The output shards are placed in the output 
> directory on the default filesystem (e.g. HDFS). Such part-N directories 
> can be used to run N shard servers. Additionally, users can specify the 
> number of reduce tasks, in particular 1 reduce task, in which case the output 
> will consist of a single shard.
> An example application is provided that processes large CSV files and uses 
> this API. It uses a custom CSV processing to avoid (de)serialization overhead.
> This patch relies on hadoop-core-0.19.1.jar - I attached the jar to this 
> issue, you should put it in contrib/hadoop/lib.
> Note: the development of this patch was sponsored by an anonymous contributor 
> and approved for release under Apache License.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (LUCENE-3820) Wrong trailing index calculation in PatternReplaceCharFilter

2012-02-22 Thread Dawid Weiss (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-3820?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dawid Weiss updated LUCENE-3820:


Attachment: LUCENE-3820.patch

A patch with reimplementation of getReplaceBlock and a test case that is 
failing with AIOOB (apply test changes without modifying 
PatternReplaceCharFilter to get the error).

> Wrong trailing index calculation in PatternReplaceCharFilter
> 
>
> Key: LUCENE-3820
> URL: https://issues.apache.org/jira/browse/LUCENE-3820
> Project: Lucene - Java
>  Issue Type: Bug
>Reporter: Dawid Weiss
>Assignee: Dawid Weiss
>Priority: Minor
> Fix For: 4.0
>
> Attachments: LUCENE-3820.patch
>
>
> I need to use PatternReplaceCharFilter's index corrections directly and it 
> fails for me -- the trailing index is not mapped correctly for a pattern 
> "\\.[\\s]*" and replacement ".", input "A. .B.".
> I tried to understand the logic in getReplaceBlock but I eventually failed 
> and simply rewrote it from scratch. After my changes a few tests don't pass 
> but I don't know if it's the tests that are screwed up or my logic. In 
> essence, the difference between the previous implementation and my 
> implementation is how indexes are mapped for shorter replacements. I shift 
> indexes of shorter regions to the "right" of the original index pool and the 
> previous patch seems to squeeze them to the left (don't know why though).
> If anybody remembers how it's supposed to work, feel free to correct me?

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-3767) Explore streaming Viterbi search in Kuromoji

2012-02-22 Thread Christian Moen (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3767?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13214196#comment-13214196
 ] 

Christian Moen commented on LUCENE-3767:


I agree completely; we should definitely proceed with Mike's improvements.  A 
big +1 from me.

I'm sorry if this wasn't clear.  My comments on search mode are unrelated and I 
didn't mean to confuse these with the improvements made.  None of this is 
necessary for Mike's improvements to be used, of course.


> Explore streaming Viterbi search in Kuromoji
> 
>
> Key: LUCENE-3767
> URL: https://issues.apache.org/jira/browse/LUCENE-3767
> Project: Lucene - Java
>  Issue Type: Improvement
>  Components: modules/analysis
>Reporter: Michael McCandless
>Assignee: Michael McCandless
> Fix For: 3.6, 4.0
>
> Attachments: LUCENE-3767.patch, LUCENE-3767.patch, LUCENE-3767.patch, 
> compound_diffs.txt
>
>
> I've been playing with the idea of changing the Kuromoji viterbi
> search to be 2 passes (intersect, backtrace) instead of 4 passes
> (break into sentences, intersect, score, backtrace)... this is very
> much a work in progress, so I'm just getting my current state up.
> It's got tons of nocommits, doesn't properly handle the user dict nor
> extended modes yet, etc.
> One thing I'm playing with is to add a double backtrace for the long
> compound tokens, ie, instead of penalizing these tokens so that
> shorter tokens are picked, leave the scores unchanged but on backtrace
> take that penalty and use it as a threshold for a 2nd best
> segmentation...

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (LUCENE-3820) Wrong trailing index calculation in PatternReplaceCharFilter

2012-02-22 Thread Dawid Weiss (Created) (JIRA)
Wrong trailing index calculation in PatternReplaceCharFilter


 Key: LUCENE-3820
 URL: https://issues.apache.org/jira/browse/LUCENE-3820
 Project: Lucene - Java
  Issue Type: Bug
Reporter: Dawid Weiss
Assignee: Dawid Weiss
Priority: Minor
 Fix For: 4.0


I need to use PatternReplaceCharFilter's index corrections directly and it 
fails for me -- the trailing index is not mapped correctly for a pattern 
"\\.[\\s]*" and replacement ".", input "A. .B.".

I tried to understand the logic in getReplaceBlock but I eventually failed and 
simply rewrote it from scratch. After my changes a few tests don't pass but I 
don't know if it's the tests that are screwed up or my logic. In essence, the 
difference between the previous implementation and my implementation is how 
indexes are mapped for shorter replacements. I shift indexes of shorter regions 
to the "right" of the original index pool and the previous patch seems to 
squeeze them to the left (don't know why though).

If anybody remembers how it's supposed to work, feel free to correct me?

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-2459) Implement LogLevelSelection as a RequestHandler

2012-02-22 Thread Ryan McKinley (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-2459?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13214193#comment-13214193
 ] 

Ryan McKinley commented on SOLR-2459:
-

Added in revision: 1292617

Lets keep the existing servlet in place until the admin UI works

> Implement LogLevelSelection as a RequestHandler
> ---
>
> Key: SOLR-2459
> URL: https://issues.apache.org/jira/browse/SOLR-2459
> Project: Solr
>  Issue Type: Wish
>  Components: web gui
>Reporter: Stefan Matheis (steffkes)
>Priority: Trivial
> Attachments: LogLevelHandler.patch, SOLR-2459-LogLevel.patch, 
> SOLR-2459-LogLevel.patch, sample-output.json, sample-output.xml
>
>
> The current available Output of the LogLevelSelection Servlet is plain HTML, 
> which made it unpossible, to integrate the Logging-Information in the new 
> Admin-UI. Format-Agnostic Output (like every [?] other Servlet offers) would 
> be really nice!
> Just as an Idea for a future structure, the new admin-ui is [actually based 
> on that 
> json-structure|https://github.com/steffkes/solr-admin/blob/master/logging.json]
>  :)

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-2459) Implement LogLevelSelection as a RequestHandler

2012-02-22 Thread Ryan McKinley (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-2459?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ryan McKinley updated SOLR-2459:


Attachment: SOLR-2459-LogLevel.patch

An updated/simplified generic version that uses the SolrTestCaseJ4

The format is different, but seems to match the JS needs better.

I will commit this soon so the UI stuff can continue in a different issue

> Implement LogLevelSelection as a RequestHandler
> ---
>
> Key: SOLR-2459
> URL: https://issues.apache.org/jira/browse/SOLR-2459
> Project: Solr
>  Issue Type: Wish
>  Components: web gui
>Reporter: Stefan Matheis (steffkes)
>Priority: Trivial
> Attachments: LogLevelHandler.patch, SOLR-2459-LogLevel.patch, 
> SOLR-2459-LogLevel.patch, sample-output.json, sample-output.xml
>
>
> The current available Output of the LogLevelSelection Servlet is plain HTML, 
> which made it unpossible, to integrate the Logging-Information in the new 
> Admin-UI. Format-Agnostic Output (like every [?] other Servlet offers) would 
> be really nice!
> Just as an Idea for a future structure, the new admin-ui is [actually based 
> on that 
> json-structure|https://github.com/steffkes/solr-admin/blob/master/logging.json]
>  :)

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-3767) Explore streaming Viterbi search in Kuromoji

2012-02-22 Thread Robert Muir (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3767?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13214159#comment-13214159
 ] 

Robert Muir commented on LUCENE-3767:
-

{quote}
I've done some analysis of these cases and it's possible to add more heuristics 
to deal with some that are obviouslt wrong, such a word starting with a long 
vowel sound in katakana.  This is a slippery slope that I'm reluctant to 
pursue...
{quote}

I've wondered about that also at least for the unknown katakana case: (though I 
don't know all the rules that could be applied).

Adding such heuristics isn't really unprecedented, in a way its very similar to 
the compounds/ package (geared towards german, etc)
using TeX hyphenation rules to restrict word splits to hyphenation breaks; and 
similar to DictionaryBasedBreakIterators in
ICU/your JRE that use orthographic rules in combination with a dictionary to 
segment southeast asian languages like Thai, and 
not too far from simple rules like "don't separate a base character from any 
combining characters that follow it", or "don't
separate a lead surrogate from a trail surrogate" that you would generally use 
across all languages.


> Explore streaming Viterbi search in Kuromoji
> 
>
> Key: LUCENE-3767
> URL: https://issues.apache.org/jira/browse/LUCENE-3767
> Project: Lucene - Java
>  Issue Type: Improvement
>  Components: modules/analysis
>Reporter: Michael McCandless
>Assignee: Michael McCandless
> Fix For: 3.6, 4.0
>
> Attachments: LUCENE-3767.patch, LUCENE-3767.patch, LUCENE-3767.patch, 
> compound_diffs.txt
>
>
> I've been playing with the idea of changing the Kuromoji viterbi
> search to be 2 passes (intersect, backtrace) instead of 4 passes
> (break into sentences, intersect, score, backtrace)... this is very
> much a work in progress, so I'm just getting my current state up.
> It's got tons of nocommits, doesn't properly handle the user dict nor
> extended modes yet, etc.
> One thing I'm playing with is to add a double backtrace for the long
> compound tokens, ie, instead of penalizing these tokens so that
> shorter tokens are picked, leave the scores unchanged but on backtrace
> take that penalty and use it as a threshold for a 2nd best
> segmentation...

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-1301) Solr + Hadoop

2012-02-22 Thread Alexander Kanarsky (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-1301?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13214152#comment-13214152
 ] 

Alexander Kanarsky commented on SOLR-1301:
--

OK, so I changed the patch to work with 3.5 ant build and re-tested it with 
Solr 3.5 and Cloudera's CDH3u3 (both the build and csv test run in 
pseudo-distributed mode). Still no unit tests but I am working on this :)

No changes compared to previous version except that I had to comment out the 
code that sets the debug level dynamically in SolrRecordWriter - because of the 
conflics with slf4j parts in current Solr; I think it is minor but if not 
please feel free to resolve this and update the patch. With this done, no need 
to put the log4j and commons-logging jars in the hadoop/lib at a compile time 
anymore, only the hadoop jar. I provided the hadoop-core-0.20.2-cdh3u3.jar used 
for testing as a part of the patch but you can use the other versions of 0.20.x 
if you'd like; it also should work with hadoop 0.21.x. Note that you still need 
to make the other related jars (solr, solrj, lucene, commons etc) available 
while you running your job; one way to do this is to put all the needed jars 
into the lib subfolder of apache-solr-hadoop jar, another ways are described 
here: 
http://www.cloudera.com/blog/2011/01/how-to-include-third-party-libraries-in-your-map-reduce-job/.
 

Finally, the quick steps to get the patch compiled (on linux):
1.  get the solr source tarball (apache-solr-3.5.0-src.tgz in this example), 
put it into some folder, cd there
2.  tar -xzf apache-solr-3.5.0-src.tgz
3.  cd apache-solr-3.5.0/solr
4.  wget 
https://issues.apache.org/jira/secure/attachment/12515662/SOLR-1301.patch
5.  patch -p0 -i SOLR-1301.patch
6.  mkdir contrib/hadoop/lib
7.  cd contrib/hadoop/lib
8.  wget 
https://issues.apache.org/jira/secure/attachment/12515663/hadoop-core-0.20.2-cdh3u3.jar
9.  cd ../../..
10. ant dist

and you should have the apache-solr-hadoop-3.5-SNAPSHOT.jar in solr/dist folder.

> Solr + Hadoop
> -
>
> Key: SOLR-1301
> URL: https://issues.apache.org/jira/browse/SOLR-1301
> Project: Solr
>  Issue Type: Improvement
>Affects Versions: 1.4
>Reporter: Andrzej Bialecki 
> Fix For: 3.6, 4.0
>
> Attachments: README.txt, SOLR-1301-hadoop-0-20.patch, 
> SOLR-1301-hadoop-0-20.patch, SOLR-1301.patch, SOLR-1301.patch, 
> SOLR-1301.patch, SOLR-1301.patch, SOLR-1301.patch, SOLR-1301.patch, 
> SOLR-1301.patch, SOLR-1301.patch, SOLR-1301.patch, SOLR-1301.patch, 
> SolrRecordWriter.java, commons-logging-1.0.4.jar, 
> commons-logging-api-1.0.4.jar, hadoop-0.19.1-core.jar, 
> hadoop-0.20.1-core.jar, hadoop-core-0.20.2-cdh3u3.jar, hadoop.patch, 
> log4j-1.2.15.jar
>
>
> This patch contains  a contrib module that provides distributed indexing 
> (using Hadoop) to Solr EmbeddedSolrServer. The idea behind this module is 
> twofold:
> * provide an API that is familiar to Hadoop developers, i.e. that of 
> OutputFormat
> * avoid unnecessary export and (de)serialization of data maintained on HDFS. 
> SolrOutputFormat consumes data produced by reduce tasks directly, without 
> storing it in intermediate files. Furthermore, by using an 
> EmbeddedSolrServer, the indexing task is split into as many parts as there 
> are reducers, and the data to be indexed is not sent over the network.
> Design
> --
> Key/value pairs produced by reduce tasks are passed to SolrOutputFormat, 
> which in turn uses SolrRecordWriter to write this data. SolrRecordWriter 
> instantiates an EmbeddedSolrServer, and it also instantiates an 
> implementation of SolrDocumentConverter, which is responsible for turning 
> Hadoop (key, value) into a SolrInputDocument. This data is then added to a 
> batch, which is periodically submitted to EmbeddedSolrServer. When reduce 
> task completes, and the OutputFormat is closed, SolrRecordWriter calls 
> commit() and optimize() on the EmbeddedSolrServer.
> The API provides facilities to specify an arbitrary existing solr.home 
> directory, from which the conf/ and lib/ files will be taken.
> This process results in the creation of as many partial Solr home directories 
> as there were reduce tasks. The output shards are placed in the output 
> directory on the default filesystem (e.g. HDFS). Such part-N directories 
> can be used to run N shard servers. Additionally, users can specify the 
> number of reduce tasks, in particular 1 reduce task, in which case the output 
> will consist of a single shard.
> An example application is provided that processes large CSV files and uses 
> this API. It uses a custom CSV processing to avoid (de)serialization overhead.
> This patch relies on hadoop-core-0.19.1.jar - I attached the jar to this 
> issue, you should put it in contrib/hadoop/lib.
> Note: the development of this patch was spons

[jira] [Commented] (LUCENE-3767) Explore streaming Viterbi search in Kuromoji

2012-02-22 Thread Robert Muir (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3767?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13214148#comment-13214148
 ] 

Robert Muir commented on LUCENE-3767:
-

{quote}
Robert mentioned earlier that he believes IPADIC could have been annotated with 
compounds as the documentation mentions them, but they're not part of the 
IPADIC model we are using. If it is possible to get the decompounds from the 
training data (Kyoto Corpus), a better overall approach is then to do regular 
segmentation (normal mode) and then provide the decompounds directly from the 
token info for the compounds. We might need to retrain the model and preserving 
the decompounds in order for this to work, but I think it is worth 
investigating.
{quote}

The dictionary documentation for the original ipadic has the ability to hold 
compound data (not in mecab-ipadic though, so maybe it was never 
implemented?!), 
but I don't actually see it in any implementations. So yeah, we would need to 
find a corpus containing compound information (and of course extend the file 
format
and add support to kuromoji) to support that.

However, would this really solve the total issue? Wouldn't that really only 
help for known kanji compounds... whereas most katakana compounds 
(e.g. the software engineer example) are expected to be OOV anyway? So it seems 
like, even if we ensured the dictionary was annotated for 
long kanji such that we always used decompounded forms, we need a 'heuristical' 
decomposition like search-mode either way, at least for 
the unknown katakana case?

And I tend to like Mike's improvements from a relevance perspective for these 
reasons:
# keeping the original compound term for improved precision
# preventing compound decomposition from having any unrelated negative impact 
on the rest of the tokenization

So I think we should pursue this change, even if we want to separately train a 
dictionary in the future, because in that case, 
we would just disable the kanji decomposition heuristic but keep the heuristic 
(obviously re-tuned!) for katakana?


> Explore streaming Viterbi search in Kuromoji
> 
>
> Key: LUCENE-3767
> URL: https://issues.apache.org/jira/browse/LUCENE-3767
> Project: Lucene - Java
>  Issue Type: Improvement
>  Components: modules/analysis
>Reporter: Michael McCandless
>Assignee: Michael McCandless
> Fix For: 3.6, 4.0
>
> Attachments: LUCENE-3767.patch, LUCENE-3767.patch, LUCENE-3767.patch, 
> compound_diffs.txt
>
>
> I've been playing with the idea of changing the Kuromoji viterbi
> search to be 2 passes (intersect, backtrace) instead of 4 passes
> (break into sentences, intersect, score, backtrace)... this is very
> much a work in progress, so I'm just getting my current state up.
> It's got tons of nocommits, doesn't properly handle the user dict nor
> extended modes yet, etc.
> One thing I'm playing with is to add a double backtrace for the long
> compound tokens, ie, instead of penalizing these tokens so that
> shorter tokens are picked, leave the scores unchanged but on backtrace
> take that penalty and use it as a threshold for a 2nd best
> segmentation...

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Issue Comment Edited] (SOLR-1301) Solr + Hadoop

2012-02-22 Thread Alexander Kanarsky (Issue Comment Edited) (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-1301?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13214102#comment-13214102
 ] 

Alexander Kanarsky edited comment on SOLR-1301 at 2/22/12 11:33 PM:


Note, the hadoop-core-0.20.2-cdh3u3.jar is part of the Coludera's CDH3u3 Hadoop 
distribution and licensed under Apache License v. 2.0.

  was (Author: kanarsky):
part of the Coludera's CDH3u3 distribution (licensed under Apache License 
v. 2.0)
  
> Solr + Hadoop
> -
>
> Key: SOLR-1301
> URL: https://issues.apache.org/jira/browse/SOLR-1301
> Project: Solr
>  Issue Type: Improvement
>Affects Versions: 1.4
>Reporter: Andrzej Bialecki 
> Fix For: 3.6, 4.0
>
> Attachments: README.txt, SOLR-1301-hadoop-0-20.patch, 
> SOLR-1301-hadoop-0-20.patch, SOLR-1301.patch, SOLR-1301.patch, 
> SOLR-1301.patch, SOLR-1301.patch, SOLR-1301.patch, SOLR-1301.patch, 
> SOLR-1301.patch, SOLR-1301.patch, SOLR-1301.patch, SOLR-1301.patch, 
> SolrRecordWriter.java, commons-logging-1.0.4.jar, 
> commons-logging-api-1.0.4.jar, hadoop-0.19.1-core.jar, 
> hadoop-0.20.1-core.jar, hadoop-core-0.20.2-cdh3u3.jar, hadoop.patch, 
> log4j-1.2.15.jar
>
>
> This patch contains  a contrib module that provides distributed indexing 
> (using Hadoop) to Solr EmbeddedSolrServer. The idea behind this module is 
> twofold:
> * provide an API that is familiar to Hadoop developers, i.e. that of 
> OutputFormat
> * avoid unnecessary export and (de)serialization of data maintained on HDFS. 
> SolrOutputFormat consumes data produced by reduce tasks directly, without 
> storing it in intermediate files. Furthermore, by using an 
> EmbeddedSolrServer, the indexing task is split into as many parts as there 
> are reducers, and the data to be indexed is not sent over the network.
> Design
> --
> Key/value pairs produced by reduce tasks are passed to SolrOutputFormat, 
> which in turn uses SolrRecordWriter to write this data. SolrRecordWriter 
> instantiates an EmbeddedSolrServer, and it also instantiates an 
> implementation of SolrDocumentConverter, which is responsible for turning 
> Hadoop (key, value) into a SolrInputDocument. This data is then added to a 
> batch, which is periodically submitted to EmbeddedSolrServer. When reduce 
> task completes, and the OutputFormat is closed, SolrRecordWriter calls 
> commit() and optimize() on the EmbeddedSolrServer.
> The API provides facilities to specify an arbitrary existing solr.home 
> directory, from which the conf/ and lib/ files will be taken.
> This process results in the creation of as many partial Solr home directories 
> as there were reduce tasks. The output shards are placed in the output 
> directory on the default filesystem (e.g. HDFS). Such part-N directories 
> can be used to run N shard servers. Additionally, users can specify the 
> number of reduce tasks, in particular 1 reduce task, in which case the output 
> will consist of a single shard.
> An example application is provided that processes large CSV files and uses 
> this API. It uses a custom CSV processing to avoid (de)serialization overhead.
> This patch relies on hadoop-core-0.19.1.jar - I attached the jar to this 
> issue, you should put it in contrib/hadoop/lib.
> Note: the development of this patch was sponsored by an anonymous contributor 
> and approved for release under Apache License.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-1301) Solr + Hadoop

2012-02-22 Thread Alexander Kanarsky (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-1301?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alexander Kanarsky updated SOLR-1301:
-

Attachment: hadoop-core-0.20.2-cdh3u3.jar

part of the Coludera's CDH3u3 distribution (licensed under Apache License v. 
2.0)

> Solr + Hadoop
> -
>
> Key: SOLR-1301
> URL: https://issues.apache.org/jira/browse/SOLR-1301
> Project: Solr
>  Issue Type: Improvement
>Affects Versions: 1.4
>Reporter: Andrzej Bialecki 
> Fix For: 3.6, 4.0
>
> Attachments: README.txt, SOLR-1301-hadoop-0-20.patch, 
> SOLR-1301-hadoop-0-20.patch, SOLR-1301.patch, SOLR-1301.patch, 
> SOLR-1301.patch, SOLR-1301.patch, SOLR-1301.patch, SOLR-1301.patch, 
> SOLR-1301.patch, SOLR-1301.patch, SOLR-1301.patch, SOLR-1301.patch, 
> SolrRecordWriter.java, commons-logging-1.0.4.jar, 
> commons-logging-api-1.0.4.jar, hadoop-0.19.1-core.jar, 
> hadoop-0.20.1-core.jar, hadoop-core-0.20.2-cdh3u3.jar, hadoop.patch, 
> log4j-1.2.15.jar
>
>
> This patch contains  a contrib module that provides distributed indexing 
> (using Hadoop) to Solr EmbeddedSolrServer. The idea behind this module is 
> twofold:
> * provide an API that is familiar to Hadoop developers, i.e. that of 
> OutputFormat
> * avoid unnecessary export and (de)serialization of data maintained on HDFS. 
> SolrOutputFormat consumes data produced by reduce tasks directly, without 
> storing it in intermediate files. Furthermore, by using an 
> EmbeddedSolrServer, the indexing task is split into as many parts as there 
> are reducers, and the data to be indexed is not sent over the network.
> Design
> --
> Key/value pairs produced by reduce tasks are passed to SolrOutputFormat, 
> which in turn uses SolrRecordWriter to write this data. SolrRecordWriter 
> instantiates an EmbeddedSolrServer, and it also instantiates an 
> implementation of SolrDocumentConverter, which is responsible for turning 
> Hadoop (key, value) into a SolrInputDocument. This data is then added to a 
> batch, which is periodically submitted to EmbeddedSolrServer. When reduce 
> task completes, and the OutputFormat is closed, SolrRecordWriter calls 
> commit() and optimize() on the EmbeddedSolrServer.
> The API provides facilities to specify an arbitrary existing solr.home 
> directory, from which the conf/ and lib/ files will be taken.
> This process results in the creation of as many partial Solr home directories 
> as there were reduce tasks. The output shards are placed in the output 
> directory on the default filesystem (e.g. HDFS). Such part-N directories 
> can be used to run N shard servers. Additionally, users can specify the 
> number of reduce tasks, in particular 1 reduce task, in which case the output 
> will consist of a single shard.
> An example application is provided that processes large CSV files and uses 
> this API. It uses a custom CSV processing to avoid (de)serialization overhead.
> This patch relies on hadoop-core-0.19.1.jar - I attached the jar to this 
> issue, you should put it in contrib/hadoop/lib.
> Note: the development of this patch was sponsored by an anonymous contributor 
> and approved for release under Apache License.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-1301) Solr + Hadoop

2012-02-22 Thread Randy Prager (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-1301?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13214101#comment-13214101
 ] 

Randy Prager commented on SOLR-1301:


I will be out of the office from 2/21 to 2/27 with limited email access.

For immediate inquiries please contact supp...@fixflyer.com or 888-349-3593.


> Solr + Hadoop
> -
>
> Key: SOLR-1301
> URL: https://issues.apache.org/jira/browse/SOLR-1301
> Project: Solr
>  Issue Type: Improvement
>Affects Versions: 1.4
>Reporter: Andrzej Bialecki 
> Fix For: 3.6, 4.0
>
> Attachments: README.txt, SOLR-1301-hadoop-0-20.patch, 
> SOLR-1301-hadoop-0-20.patch, SOLR-1301.patch, SOLR-1301.patch, 
> SOLR-1301.patch, SOLR-1301.patch, SOLR-1301.patch, SOLR-1301.patch, 
> SOLR-1301.patch, SOLR-1301.patch, SOLR-1301.patch, SOLR-1301.patch, 
> SolrRecordWriter.java, commons-logging-1.0.4.jar, 
> commons-logging-api-1.0.4.jar, hadoop-0.19.1-core.jar, 
> hadoop-0.20.1-core.jar, hadoop.patch, log4j-1.2.15.jar
>
>
> This patch contains  a contrib module that provides distributed indexing 
> (using Hadoop) to Solr EmbeddedSolrServer. The idea behind this module is 
> twofold:
> * provide an API that is familiar to Hadoop developers, i.e. that of 
> OutputFormat
> * avoid unnecessary export and (de)serialization of data maintained on HDFS. 
> SolrOutputFormat consumes data produced by reduce tasks directly, without 
> storing it in intermediate files. Furthermore, by using an 
> EmbeddedSolrServer, the indexing task is split into as many parts as there 
> are reducers, and the data to be indexed is not sent over the network.
> Design
> --
> Key/value pairs produced by reduce tasks are passed to SolrOutputFormat, 
> which in turn uses SolrRecordWriter to write this data. SolrRecordWriter 
> instantiates an EmbeddedSolrServer, and it also instantiates an 
> implementation of SolrDocumentConverter, which is responsible for turning 
> Hadoop (key, value) into a SolrInputDocument. This data is then added to a 
> batch, which is periodically submitted to EmbeddedSolrServer. When reduce 
> task completes, and the OutputFormat is closed, SolrRecordWriter calls 
> commit() and optimize() on the EmbeddedSolrServer.
> The API provides facilities to specify an arbitrary existing solr.home 
> directory, from which the conf/ and lib/ files will be taken.
> This process results in the creation of as many partial Solr home directories 
> as there were reduce tasks. The output shards are placed in the output 
> directory on the default filesystem (e.g. HDFS). Such part-N directories 
> can be used to run N shard servers. Additionally, users can specify the 
> number of reduce tasks, in particular 1 reduce task, in which case the output 
> will consist of a single shard.
> An example application is provided that processes large CSV files and uses 
> this API. It uses a custom CSV processing to avoid (de)serialization overhead.
> This patch relies on hadoop-core-0.19.1.jar - I attached the jar to this 
> issue, you should put it in contrib/hadoop/lib.
> Note: the development of this patch was sponsored by an anonymous contributor 
> and approved for release under Apache License.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-1301) Solr + Hadoop

2012-02-22 Thread Alexander Kanarsky (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-1301?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alexander Kanarsky updated SOLR-1301:
-

Attachment: SOLR-1301.patch

SOLR-1301 patch modified to work with Solr 3.x ant build. tested with Solr 
3.5.0 and Cloudera CDH3u3 (also attached) 

> Solr + Hadoop
> -
>
> Key: SOLR-1301
> URL: https://issues.apache.org/jira/browse/SOLR-1301
> Project: Solr
>  Issue Type: Improvement
>Affects Versions: 1.4
>Reporter: Andrzej Bialecki 
> Fix For: 3.6, 4.0
>
> Attachments: README.txt, SOLR-1301-hadoop-0-20.patch, 
> SOLR-1301-hadoop-0-20.patch, SOLR-1301.patch, SOLR-1301.patch, 
> SOLR-1301.patch, SOLR-1301.patch, SOLR-1301.patch, SOLR-1301.patch, 
> SOLR-1301.patch, SOLR-1301.patch, SOLR-1301.patch, SOLR-1301.patch, 
> SolrRecordWriter.java, commons-logging-1.0.4.jar, 
> commons-logging-api-1.0.4.jar, hadoop-0.19.1-core.jar, 
> hadoop-0.20.1-core.jar, hadoop.patch, log4j-1.2.15.jar
>
>
> This patch contains  a contrib module that provides distributed indexing 
> (using Hadoop) to Solr EmbeddedSolrServer. The idea behind this module is 
> twofold:
> * provide an API that is familiar to Hadoop developers, i.e. that of 
> OutputFormat
> * avoid unnecessary export and (de)serialization of data maintained on HDFS. 
> SolrOutputFormat consumes data produced by reduce tasks directly, without 
> storing it in intermediate files. Furthermore, by using an 
> EmbeddedSolrServer, the indexing task is split into as many parts as there 
> are reducers, and the data to be indexed is not sent over the network.
> Design
> --
> Key/value pairs produced by reduce tasks are passed to SolrOutputFormat, 
> which in turn uses SolrRecordWriter to write this data. SolrRecordWriter 
> instantiates an EmbeddedSolrServer, and it also instantiates an 
> implementation of SolrDocumentConverter, which is responsible for turning 
> Hadoop (key, value) into a SolrInputDocument. This data is then added to a 
> batch, which is periodically submitted to EmbeddedSolrServer. When reduce 
> task completes, and the OutputFormat is closed, SolrRecordWriter calls 
> commit() and optimize() on the EmbeddedSolrServer.
> The API provides facilities to specify an arbitrary existing solr.home 
> directory, from which the conf/ and lib/ files will be taken.
> This process results in the creation of as many partial Solr home directories 
> as there were reduce tasks. The output shards are placed in the output 
> directory on the default filesystem (e.g. HDFS). Such part-N directories 
> can be used to run N shard servers. Additionally, users can specify the 
> number of reduce tasks, in particular 1 reduce task, in which case the output 
> will consist of a single shard.
> An example application is provided that processes large CSV files and uses 
> this API. It uses a custom CSV processing to avoid (de)serialization overhead.
> This patch relies on hadoop-core-0.19.1.jar - I attached the jar to this 
> issue, you should put it in contrib/hadoop/lib.
> Note: the development of this patch was sponsored by an anonymous contributor 
> and approved for release under Apache License.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-3731) Create a analysis/uima module for UIMA based tokenizers/analyzers

2012-02-22 Thread Tommaso Teofili (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3731?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13214093#comment-13214093
 ] 

Tommaso Teofili commented on LUCENE-3731:
-

After some more testing I think the CasPool is good just for scenarios where 
the pool serves different CAS to different clients (the tokenizers), so not 
really helpful in the current implementation, however it may be useful if we 
abstract the operation of obtaining and releasing a CAS outside the 
BaseTokenizer.

In the meantime I noticed the AEProviderFactory getAEProvider() methods have a 
keyPrefix parameter that came from Solr implementation and was intended to hold 
the core name, so, at the moment I think it'd be better to have (also) methods 
which don't need that paramater for the Lucene uses.

> Create a analysis/uima module for UIMA based tokenizers/analyzers
> -
>
> Key: LUCENE-3731
> URL: https://issues.apache.org/jira/browse/LUCENE-3731
> Project: Lucene - Java
>  Issue Type: Improvement
>  Components: modules/analysis
>Reporter: Tommaso Teofili
>Assignee: Tommaso Teofili
> Fix For: 3.6, 4.0
>
> Attachments: LUCENE-3731.patch, LUCENE-3731_2.patch, 
> LUCENE-3731_3.patch, LUCENE-3731_4.patch, LUCENE-3731_rsrel.patch, 
> LUCENE-3731_speed.patch, LUCENE-3731_speed.patch, LUCENE-3731_speed.patch
>
>
> As discussed in SOLR-3013 the UIMA Tokenizers/Analyzer should be refactored 
> out in a separate module (modules/analysis/uima) as they can be used in plain 
> Lucene. Then the solr/contrib/uima will contain only the related factories.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Resolved] (LUCENE-3819) Clean up what we show in right side bar of website.

2012-02-22 Thread Mark Miller (Resolved) (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-3819?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mark Miller resolved LUCENE-3819.
-

Resolution: Fixed

> Clean up what we show in right side bar of website.
> ---
>
> Key: LUCENE-3819
> URL: https://issues.apache.org/jira/browse/LUCENE-3819
> Project: Lucene - Java
>  Issue Type: Improvement
>Reporter: Mark Miller
>Assignee: Mark Miller
>Priority: Minor
>
> I'd love to remove a couple things - it's pretty crowded on the right side 
> bar. I find the latest JIRA and email displays are hard to read, tend to 
> format badly, and don't offer much value.
> I'd like to remove them and just leave svn commits and twitter mentions 
> (which are much easier to read and format better). Will help with some info 
> overload on each page.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-3819) Clean up what we show in right side bar of website.

2012-02-22 Thread Mark Miller (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3819?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13214009#comment-13214009
 ] 

Mark Miller commented on LUCENE-3819:
-

So I'm done here for now - just made my last tweak. Thank you everyone for your 
suggestions and input!

> Clean up what we show in right side bar of website.
> ---
>
> Key: LUCENE-3819
> URL: https://issues.apache.org/jira/browse/LUCENE-3819
> Project: Lucene - Java
>  Issue Type: Improvement
>Reporter: Mark Miller
>Assignee: Mark Miller
>Priority: Minor
>
> I'd love to remove a couple things - it's pretty crowded on the right side 
> bar. I find the latest JIRA and email displays are hard to read, tend to 
> format badly, and don't offer much value.
> I'd like to remove them and just leave svn commits and twitter mentions 
> (which are much easier to read and format better). Will help with some info 
> overload on each page.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-3810) Text Alignment in the "status-item" tab of the Website needs to be word wrapped

2012-02-22 Thread Mark Miller (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3810?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13213986#comment-13213986
 ] 

Mark Miller commented on LUCENE-3810:
-

Thanks Varun! Hand't seen this issue - I'll take a look at making this change 
(though we have trimmed some from the sidebar recently).

> Text Alignment in the "status-item" tab of the Website needs to be word 
> wrapped
> ---
>
> Key: LUCENE-3810
> URL: https://issues.apache.org/jira/browse/LUCENE-3810
> Project: Lucene - Java
>  Issue Type: Improvement
>Reporter: Varun Thacker
>Assignee: Varun Thacker
>Priority: Minor
> Attachments: LUCENE-3810.patch, lucene.png
>
>
> If we go to http://lucene.apache.org/core/ or http://lucene.apache.org/solr/ 
> under the Latest Dev section the text is exceeding the css boundary. This is 
> happening on both Firefox and Chrome. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-3819) Clean up what we show in right side bar of website.

2012-02-22 Thread Mark Miller (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3819?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13213983#comment-13213983
 ] 

Mark Miller commented on LUCENE-3819:
-

Explicitly, based on some informal feedback and my thoughts:

1. Changed color of Solr download button to orange.
2. On the main page, put small gray text indicating what each download button 
points to (lucene/solr)
3. Added new redirects so that Download buttons now points to the latest 
release explicitly (3.5)

> Clean up what we show in right side bar of website.
> ---
>
> Key: LUCENE-3819
> URL: https://issues.apache.org/jira/browse/LUCENE-3819
> Project: Lucene - Java
>  Issue Type: Improvement
>Reporter: Mark Miller
>Assignee: Mark Miller
>Priority: Minor
>
> I'd love to remove a couple things - it's pretty crowded on the right side 
> bar. I find the latest JIRA and email displays are hard to read, tend to 
> format badly, and don't offer much value.
> I'd like to remove them and just leave svn commits and twitter mentions 
> (which are much easier to read and format better). Will help with some info 
> overload on each page.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-3804) Swap Features and News on the website.

2012-02-22 Thread Mark Miller (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3804?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13213976#comment-13213976
 ] 

Mark Miller commented on LUCENE-3804:
-

Thanks hossman - ill get to some of this shortly.

> Swap Features and News on the website.
> --
>
> Key: LUCENE-3804
> URL: https://issues.apache.org/jira/browse/LUCENE-3804
> Project: Lucene - Java
>  Issue Type: Improvement
>Reporter: Mark Miller
>Assignee: Mark Miller
>
> I think we can do even better, but that is a nice, easy incremental 
> improvement.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-3819) Clean up what we show in right side bar of website.

2012-02-22 Thread Mark Miller (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3819?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13213966#comment-13213966
 ] 

Mark Miller commented on LUCENE-3819:
-

I just made some more improvements as well.

> Clean up what we show in right side bar of website.
> ---
>
> Key: LUCENE-3819
> URL: https://issues.apache.org/jira/browse/LUCENE-3819
> Project: Lucene - Java
>  Issue Type: Improvement
>Reporter: Mark Miller
>Assignee: Mark Miller
>Priority: Minor
>
> I'd love to remove a couple things - it's pretty crowded on the right side 
> bar. I find the latest JIRA and email displays are hard to read, tend to 
> format badly, and don't offer much value.
> I'd like to remove them and just leave svn commits and twitter mentions 
> (which are much easier to read and format better). Will help with some info 
> overload on each page.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (SOLR-3154) SolrJ CloudServer should be leader aware when adding docs

2012-02-22 Thread Grant Ingersoll (Created) (JIRA)
SolrJ CloudServer should be leader aware when adding docs
-

 Key: SOLR-3154
 URL: https://issues.apache.org/jira/browse/SOLR-3154
 Project: Solr
  Issue Type: Improvement
  Components: SolrCloud
Affects Versions: 4.0
Reporter: Grant Ingersoll
Priority: Minor
 Fix For: 4.0


It would be good when indexing if the SolrJ CloudServer was leader aware so 
that we could avoid doing an extra hop for the data.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-tests-only-3.x-java7 - Build # 1821 - Failure

2012-02-22 Thread Apache Jenkins Server
Build: https://builds.apache.org/job/Lucene-Solr-tests-only-3.x-java7/1821/

1 tests failed.
REGRESSION:  org.apache.solr.TestDistributedGrouping.testDistribSearch

Error Message:
java.lang.AssertionError: Some threads threw uncaught exceptions!

Stack Trace:
java.lang.RuntimeException: java.lang.AssertionError: Some threads threw 
uncaught exceptions!
at 
org.apache.lucene.util.LuceneTestCase.tearDownInternal(LuceneTestCase.java:679)
at 
org.apache.lucene.util.LuceneTestCase.access$1000(LuceneTestCase.java:116)
at 
org.apache.lucene.util.LuceneTestCase$InternalSetupTeardownRule$1.evaluate(LuceneTestCase.java:496)
at 
org.apache.lucene.util.LuceneTestCase$TestResultInterceptorRule$1.evaluate(LuceneTestCase.java:400)
at 
org.apache.lucene.util.LuceneTestCase$RememberThreadRule$1.evaluate(LuceneTestCase.java:458)
at 
org.apache.lucene.util.LuceneTestCaseRunner.runChild(LuceneTestCaseRunner.java:147)
at 
org.apache.lucene.util.LuceneTestCaseRunner.runChild(LuceneTestCaseRunner.java:50)
at 
org.apache.lucene.util.LuceneTestCase.checkUncaughtExceptionsAfter(LuceneTestCase.java:707)
at 
org.apache.lucene.util.LuceneTestCase.tearDownInternal(LuceneTestCase.java:651)




Build Log (for compile errors):
[...truncated 16072 lines...]



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org

[jira] [Commented] (LUCENE-3819) Clean up what we show in right side bar of website.

2012-02-22 Thread Mark Miller (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3819?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13213848#comment-13213848
 ] 

Mark Miller commented on LUCENE-3819:
-

bq. Thoughts?

Sounds good - I've made a first pass at this.

> Clean up what we show in right side bar of website.
> ---
>
> Key: LUCENE-3819
> URL: https://issues.apache.org/jira/browse/LUCENE-3819
> Project: Lucene - Java
>  Issue Type: Improvement
>Reporter: Mark Miller
>Assignee: Mark Miller
>Priority: Minor
>
> I'd love to remove a couple things - it's pretty crowded on the right side 
> bar. I find the latest JIRA and email displays are hard to read, tend to 
> format badly, and don't offer much value.
> I'd like to remove them and just leave svn commits and twitter mentions 
> (which are much easier to read and format better). Will help with some info 
> overload on each page.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-2604) add regexpquery to queryparser

2012-02-22 Thread Robert Muir (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-2604?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13213847#comment-13213847
 ] 

Robert Muir commented on LUCENE-2604:
-

Its no bug: the regexp syntax is optionalField:/stuff/



> add regexpquery to queryparser
> --
>
> Key: LUCENE-2604
> URL: https://issues.apache.org/jira/browse/LUCENE-2604
> Project: Lucene - Java
>  Issue Type: New Feature
>  Components: core/queryparser
>Affects Versions: 4.0
>Reporter: Robert Muir
>Assignee: Simon Willnauer
> Fix For: 4.0
>
> Attachments: LUCENE-2604.patch, LUCENE-2604.patch, LUCENE-2604.patch
>
>
> patch that adds RegexpQuery if you /enter an expression between slashes like 
> this/
> i didnt do the contrib ones but could add it there too if it seems like a 
> good idea.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-2604) add regexpquery to queryparser

2012-02-22 Thread Yury Kats (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-2604?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13213841#comment-13213841
 ] 

Yury Kats commented on LUCENE-2604:
---

I just came across a problem with this, where regular expressions are being 
parsed across fields. Feels like a bug to me.

Queries like these fail to parse:
fieldName:/a fieldName:/a
SEVERE: org.apache.solr.common.SolrException: no field name specified in query 
and no defaultSearchField defined in schema.xml
at 
org.apache.solr.search.SolrQueryParser.checkNullField(SolrQueryParser.java:106)
at 
org.apache.solr.search.SolrQueryParser.getFieldQuery(SolrQueryParser.java:124)
at 
org.apache.lucene.queryparser.classic.QueryParserBase.handleBareTokenQuery(QueryParserBase.java:1058)
at 
org.apache.lucene.queryparser.classic.QueryParser.Term(QueryParser.java:358)
at 
org.apache.lucene.queryparser.classic.QueryParser.Clause(QueryParser.java:257)
at 
org.apache.lucene.queryparser.classic.QueryParser.Query(QueryParser.java:212)
at 
org.apache.lucene.queryparser.classic.QueryParser.TopLevelQuery(QueryParser.java:170)
at 
org.apache.lucene.queryparser.classic.QueryParserBase.parse(QueryParserBase.java:118)
at 
org.apache.solr.search.LuceneQParser.parse(LuceneQParserPlugin.java:74)
at org.apache.solr.search.QParser.getQuery(QParser.java:143)

fieldName:/* fieldName:/*
java.lang.NullPointerException
at 
org.apache.solr.schema.IndexSchema$DynamicReplacement.matches(IndexSchema.java:747)
at 
org.apache.solr.schema.IndexSchema.getDynamicFieldType(IndexSchema.java:1026)
at org.apache.solr.schema.IndexSchema.getFieldType(IndexSchema.java:980)
at 
org.apache.solr.search.SolrQueryParser.getWildcardQuery(SolrQueryParser.java:172)
at 
org.apache.lucene.queryparser.classic.QueryParserBase.handleBareTokenQuery(QueryParserBase.java:1039)
at 
org.apache.lucene.queryparser.classic.QueryParser.Term(QueryParser.java:358)
at 
org.apache.lucene.queryparser.classic.QueryParser.Clause(QueryParser.java:257)
at 
org.apache.lucene.queryparser.classic.QueryParser.Query(QueryParser.java:212)
at 
org.apache.lucene.queryparser.classic.QueryParser.TopLevelQuery(QueryParser.java:170)
at 
org.apache.lucene.queryparser.classic.QueryParserBase.parse(QueryParserBase.java:118)
at 
org.apache.solr.search.LuceneQParser.parse(LuceneQParserPlugin.java:74)
at org.apache.solr.search.QParser.getQuery(QParser.java:143)


> add regexpquery to queryparser
> --
>
> Key: LUCENE-2604
> URL: https://issues.apache.org/jira/browse/LUCENE-2604
> Project: Lucene - Java
>  Issue Type: New Feature
>  Components: core/queryparser
>Affects Versions: 4.0
>Reporter: Robert Muir
>Assignee: Simon Willnauer
> Fix For: 4.0
>
> Attachments: LUCENE-2604.patch, LUCENE-2604.patch, LUCENE-2604.patch
>
>
> patch that adds RegexpQuery if you /enter an expression between slashes like 
> this/
> i didnt do the contrib ones but could add it there too if it seems like a 
> good idea.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-3819) Clean up what we show in right side bar of website.

2012-02-22 Thread Steven Rowe (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3819?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13213787#comment-13213787
 ] 

Steven Rowe commented on LUCENE-3819:
-

+1 to all of Christian's suggestions

> Clean up what we show in right side bar of website.
> ---
>
> Key: LUCENE-3819
> URL: https://issues.apache.org/jira/browse/LUCENE-3819
> Project: Lucene - Java
>  Issue Type: Improvement
>Reporter: Mark Miller
>Assignee: Mark Miller
>Priority: Minor
>
> I'd love to remove a couple things - it's pretty crowded on the right side 
> bar. I find the latest JIRA and email displays are hard to read, tend to 
> format badly, and don't offer much value.
> I'd like to remove them and just leave svn commits and twitter mentions 
> (which are much easier to read and format better). Will help with some info 
> overload on each page.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-3819) Clean up what we show in right side bar of website.

2012-02-22 Thread Christian Moen (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3819?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13213776#comment-13213776
 ] 

Christian Moen commented on LUCENE-3819:


Thanks, Mark.

I like your idea of renaming "User Resources" to "Projects" on the top level 
page. "Lucene Projects" is perhaps also possible, but I think I like "Projects" 
better.

Regarding downloads, the idea was to make convenience links to the most common 
downloads -- Lucene Core and Solr -- from the top page, but I realize this 
breaks the layering structure somewhat.  Perhaps "Downloads", "Common 
Downloads" or "Download Shortcuts" makes sense?

Alternatively, we could put download buttons for Lucene Core and Solr in the 
sidebar. If we follow this approach perhaps we might want to move the buttons 
to the sidebar for the Lucene Core and Solr page as well.

To sum up, I'm proposing:

# We rename "User Resources" to "Projects" as you suggested (on the top page)
# We add two download buttons for Solr and Lucene to the sidebar (on the top 
page)
# We move the download button to the sidebar on the Lucene Core and Solr pages.

(Lucene gets a green button and Solr gets an orange button on the relevant 
pages.)

Thoughts?


> Clean up what we show in right side bar of website.
> ---
>
> Key: LUCENE-3819
> URL: https://issues.apache.org/jira/browse/LUCENE-3819
> Project: Lucene - Java
>  Issue Type: Improvement
>Reporter: Mark Miller
>Assignee: Mark Miller
>Priority: Minor
>
> I'd love to remove a couple things - it's pretty crowded on the right side 
> bar. I find the latest JIRA and email displays are hard to read, tend to 
> format badly, and don't offer much value.
> I'd like to remove them and just leave svn commits and twitter mentions 
> (which are much easier to read and format better). Will help with some info 
> overload on each page.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-3144) Parameter marshalling is broken in Solrj JavaBinUpdateRequestCodec

2012-02-22 Thread Sami Siren (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-3144?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13213758#comment-13213758
 ] 

Sami Siren commented on SOLR-3144:
--

Hi Jörg,

What version did you see this problem with? I tried to fix this already in 
SOLR-3037 but perhaps I missed something...

> Parameter marshalling is broken in Solrj JavaBinUpdateRequestCodec 
> ---
>
> Key: SOLR-3144
> URL: https://issues.apache.org/jira/browse/SOLR-3144
> Project: Solr
>  Issue Type: Bug
>  Components: clients - java
>Affects Versions: 4.0
>Reporter: Jörg Maier
> Fix For: 4.0
>
> Attachments: JavaBinUpdateRequestCodec.patch
>
>
> The parameter marshalling de-marshalling is broken in Solrj's 
> JavaBinUpdateRequestCodec. 
> The bug can be reproduced by adding a parameter e.g. overwrite=false as 
> parameter to the UpdateRequest. After desiarilizing on the backend side the 
> value will be not "false" it will be "[false]" which results in an Exception 
> in the backend and documents will not be imported.
> This issue can easily be fixed by replacing the serialization method with the 
> correct one in SolrParams. See also this gist for a working version: 
> https://gist.github.com/1853544

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-3819) Clean up what we show in right side bar of website.

2012-02-22 Thread Mark Miller (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3819?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13213753#comment-13213753
 ] 

Mark Miller commented on LUCENE-3819:
-

bq. I also think it might be useful to have download shortcuts to Lucene Core 
(Java) and Solr available in the sidebar from http://lucene.apache.org/. 

How do you propose we do this?

Right now we have links to each subproject called User Resources (could use a 
better name) - Should we add another section below that called Project 
Downloads? And rename User Resources to Projects or something?

> Clean up what we show in right side bar of website.
> ---
>
> Key: LUCENE-3819
> URL: https://issues.apache.org/jira/browse/LUCENE-3819
> Project: Lucene - Java
>  Issue Type: Improvement
>Reporter: Mark Miller
>Assignee: Mark Miller
>Priority: Minor
>
> I'd love to remove a couple things - it's pretty crowded on the right side 
> bar. I find the latest JIRA and email displays are hard to read, tend to 
> format badly, and don't offer much value.
> I'd like to remove them and just leave svn commits and twitter mentions 
> (which are much easier to read and format better). Will help with some info 
> overload on each page.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Issue Comment Edited] (LUCENE-3819) Clean up what we show in right side bar of website.

2012-02-22 Thread Christian Moen (Issue Comment Edited) (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3819?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13213740#comment-13213740
 ] 

Christian Moen edited comment on LUCENE-3819 at 2/22/12 4:21 PM:
-

+1, Mark.  +1, Yonik.  

I also think it might be useful to have download shortcuts to Lucene Core 
(Java) and Solr available in the sidebar from http://lucene.apache.org/.  
Perhaps "Download" could be considered becoming a standard sidebar item for the 
subprojects?

 

  was (Author: cm):
+1, Mark.  +1, Yonik.  

I also think it might be useful to have download shortcuts to Lucene Core 
(Java) and Solr available in the sidebar from http://lucene.apache.org/.  
Perhaps "Download" could be considered to be a standard sidebar item.  (I quite 
like the red download button, though! :))

 
  
> Clean up what we show in right side bar of website.
> ---
>
> Key: LUCENE-3819
> URL: https://issues.apache.org/jira/browse/LUCENE-3819
> Project: Lucene - Java
>  Issue Type: Improvement
>Reporter: Mark Miller
>Assignee: Mark Miller
>Priority: Minor
>
> I'd love to remove a couple things - it's pretty crowded on the right side 
> bar. I find the latest JIRA and email displays are hard to read, tend to 
> format badly, and don't offer much value.
> I'd like to remove them and just leave svn commits and twitter mentions 
> (which are much easier to read and format better). Will help with some info 
> overload on each page.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-3819) Clean up what we show in right side bar of website.

2012-02-22 Thread Christian Moen (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3819?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13213740#comment-13213740
 ] 

Christian Moen commented on LUCENE-3819:


+1, Mark.  +1, Yonik.  

I also think it might be useful to have download shortcuts to Lucene Core 
(Java) and Solr available in the sidebar from http://lucene.apache.org/.  
Perhaps "Download" could be considered to be a standard sidebar item.  (I quite 
like the red download button, though! :))

 

> Clean up what we show in right side bar of website.
> ---
>
> Key: LUCENE-3819
> URL: https://issues.apache.org/jira/browse/LUCENE-3819
> Project: Lucene - Java
>  Issue Type: Improvement
>Reporter: Mark Miller
>Assignee: Mark Miller
>Priority: Minor
>
> I'd love to remove a couple things - it's pretty crowded on the right side 
> bar. I find the latest JIRA and email displays are hard to read, tend to 
> format badly, and don't offer much value.
> I'd like to remove them and just leave svn commits and twitter mentions 
> (which are much easier to read and format better). Will help with some info 
> overload on each page.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-3819) Clean up what we show in right side bar of website.

2012-02-22 Thread Yonik Seeley (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3819?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13213713#comment-13213713
 ] 

Yonik Seeley commented on LUCENE-3819:
--

+1

And make the sidebar take less horizontal space too (currently ~40%)

> Clean up what we show in right side bar of website.
> ---
>
> Key: LUCENE-3819
> URL: https://issues.apache.org/jira/browse/LUCENE-3819
> Project: Lucene - Java
>  Issue Type: Improvement
>Reporter: Mark Miller
>Assignee: Mark Miller
>Priority: Minor
>
> I'd love to remove a couple things - it's pretty crowded on the right side 
> bar. I find the latest JIRA and email displays are hard to read, tend to 
> format badly, and don't offer much value.
> I'd like to remove them and just leave svn commits and twitter mentions 
> (which are much easier to read and format better). Will help with some info 
> overload on each page.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (LUCENE-3819) Clean up what we show in right side bar of website.

2012-02-22 Thread Mark Miller (Created) (JIRA)
Clean up what we show in right side bar of website.
---

 Key: LUCENE-3819
 URL: https://issues.apache.org/jira/browse/LUCENE-3819
 Project: Lucene - Java
  Issue Type: Improvement
Reporter: Mark Miller
Assignee: Mark Miller
Priority: Minor


I'd love to remove a couple things - it's pretty crowded on the right side bar. 
I find the latest JIRA and email displays are hard to read, tend to format 
badly, and don't offer much value.

I'd like to remove them and just leave svn commits and twitter mentions (which 
are much easier to read and format better). Will help with some info overload 
on each page.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-3144) Parameter marshalling is broken in Solrj JavaBinUpdateRequestCodec

2012-02-22 Thread Updated

 [ 
https://issues.apache.org/jira/browse/SOLR-3144?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jörg Maier updated SOLR-3144:
-

Attachment: JavaBinUpdateRequestCodec.patch

Attached patch.

> Parameter marshalling is broken in Solrj JavaBinUpdateRequestCodec 
> ---
>
> Key: SOLR-3144
> URL: https://issues.apache.org/jira/browse/SOLR-3144
> Project: Solr
>  Issue Type: Bug
>  Components: clients - java
>Affects Versions: 4.0
>Reporter: Jörg Maier
> Fix For: 4.0
>
> Attachments: JavaBinUpdateRequestCodec.patch
>
>
> The parameter marshalling de-marshalling is broken in Solrj's 
> JavaBinUpdateRequestCodec. 
> The bug can be reproduced by adding a parameter e.g. overwrite=false as 
> parameter to the UpdateRequest. After desiarilizing on the backend side the 
> value will be not "false" it will be "[false]" which results in an Exception 
> in the backend and documents will not be imported.
> This issue can easily be fixed by replacing the serialization method with the 
> correct one in SolrParams. See also this gist for a working version: 
> https://gist.github.com/1853544

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-3074) SimpleTextCodec needs SimpleText DocValues impl

2012-02-22 Thread Robert Muir (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3074?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13213672#comment-13213672
 ] 

Robert Muir commented on LUCENE-3074:
-

first thoughts: looks nice! Thanks for working on this!

I will try to take a look later, I noticed a few imports from lucene40 codec 
into simpletext (which i think we should avoid),
but I think these were just javadocs relics!

> SimpleTextCodec needs SimpleText DocValues impl
> ---
>
> Key: LUCENE-3074
> URL: https://issues.apache.org/jira/browse/LUCENE-3074
> Project: Lucene - Java
>  Issue Type: Task
>  Components: core/index, core/search
>Affects Versions: 4.0
>Reporter: Simon Willnauer
>Assignee: Simon Willnauer
>Priority: Blocker
> Fix For: 4.0
>
> Attachments: LUCENE-3074.patch
>
>
> currently SimpleTextCodec uses binary docValues we should move that to a 
> simple text impl.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (LUCENE-3818) TestIndexWriterNRTIsCurrent failure

2012-02-22 Thread Robert Muir (Created) (JIRA)
TestIndexWriterNRTIsCurrent failure
---

 Key: LUCENE-3818
 URL: https://issues.apache.org/jira/browse/LUCENE-3818
 Project: Lucene - Java
  Issue Type: Bug
Reporter: Robert Muir
 Fix For: 4.0


found by jenkins: 

https://builds.apache.org/job/Lucene-Solr-tests-only-trunk/12492/

make your computer busy (e.g. run tests in another checkout) then,

ant test-core -Dtests.iter=100 -Dtestcase=TestIndexWriterNRTIsCurrent 
-Dtestmethod=testIsCurrentWithThreads 
-Dtests.seed=-78f6fa16b849cf27:382126da79c1e146:-d2cdec79e86e1b3 
-Dtests.multiplier=3 -Dargs="-Dfile.encoding=ISO8859-1"

takes a few tries till it pops...

{noformat}
junit-sequential:
[junit] Testsuite: org.apache.lucene.index.TestIndexWriterNRTIsCurrent
[junit] Tests run: 100, Failures: 1, Errors: 1, Time elapsed: 277.818 sec
[junit] 
[junit] - Standard Output ---
[junit] WARNING: you are using -Dtests.iter=n where n > 1, not all tests 
support this option.
[junit] Some may crash or fail: this is not a bug.
[junit] -  ---
[junit] - Standard Error -
[junit] NOTE: reproduce with: ant test 
-Dtestcase=TestIndexWriterNRTIsCurrent -Dtestmethod=testIsCurrentWithThreads 
-Dtests.seed=-78f6fa16b849cf27:382126da79c1e146:-d2cdec79e86e1b3 
-Dtests.multiplier=3 -Dargs="-Dfile.encoding=ISO8859-1"
[junit] The following exceptions were thrown by threads:
[junit] *** Thread: Lucene Merge Thread #17 ***
[junit] org.apache.lucene.index.MergePolicy$MergeException: 
java.lang.AssertionError
[junit] at 
org.apache.lucene.index.ConcurrentMergeScheduler.handleMergeException(ConcurrentMergeScheduler.java:520)
[junit] at 
org.apache.lucene.index.ConcurrentMergeScheduler$MergeThread.run(ConcurrentMergeScheduler.java:480)
[junit] Caused by: java.lang.AssertionError
[junit] at 
org.apache.lucene.index.IndexWriter$ReadersAndLiveDocs.initWritableLiveDocs(IndexWriter.java:580)
[junit] at 
org.apache.lucene.index.IndexWriter.commitMergedDeletes(IndexWriter.java:3061)
[junit] at 
org.apache.lucene.index.IndexWriter.commitMerge(IndexWriter.java:3137)
[junit] at 
org.apache.lucene.index.IndexWriter.mergeMiddle(IndexWriter.java:3718)
[junit] at 
org.apache.lucene.index.IndexWriter.merge(IndexWriter.java:3257)
[junit] at 
org.apache.lucene.index.ConcurrentMergeScheduler.doMerge(ConcurrentMergeScheduler.java:382)
[junit] at 
org.apache.lucene.index.ConcurrentMergeScheduler$MergeThread.run(ConcurrentMergeScheduler.java:451)
[junit] NOTE: reproduce with: ant test 
-Dtestcase=TestIndexWriterNRTIsCurrent -Dtestmethod=testIsCurrentWithThreads 
-Dtests.seed=-78f6fa16b849cf27:382126da79c1e146:-d2cdec79e86e1b3 
-Dtests.multiplier=3 -Dargs="-Dfile.encoding=ISO8859-1"
[junit] NOTE: reproduce with: ant test 
-Dtestcase=TestIndexWriterNRTIsCurrent -Dtestmethod=testIsCurrentWithThreads 
-Dtests.seed=-78f6fa16b849cf27:382126da79c1e146:-d2cdec79e86e1b3 
-Dtests.multiplier=3 -Dargs="-Dfile.encoding=ISO8859-1"
[junit] NOTE: test params are: codec=Lucene40: 
{id=MockFixedIntBlock(blockSize=525)}, sim=DefaultSimilarity, locale=es_PY, 
timezone=Africa/Luanda
[junit] NOTE: all tests run in this JVM:
[junit] [TestIndexWriterNRTIsCurrent]
[junit] NOTE: Linux 3.0.0-14-generic amd64/Sun Microsystems Inc. 1.6.0_24 
(64-bit)/cpus=8,threads=1,free=74907448,total=255787008
[junit] -  ---
[junit] Testcase: 
testIsCurrentWithThreads(org.apache.lucene.index.TestIndexWriterNRTIsCurrent):  
  FAILED
[junit] info=_qx(4.0):C1/1 isn't live
[junit] junit.framework.AssertionFailedError: info=_qx(4.0):C1/1 isn't live
[junit] at 
org.apache.lucene.util.LuceneTestCaseRunner.runChild(LuceneTestCaseRunner.java:165)
[junit] at 
org.apache.lucene.util.LuceneTestCaseRunner.runChild(LuceneTestCaseRunner.java:57)
[junit] at 
org.apache.lucene.index.IndexWriter$ReaderPool.infoIsLive(IndexWriter.java:663)
[junit] at 
org.apache.lucene.index.IndexWriter$ReaderPool.dropAll(IndexWriter.java:717)
[junit] at 
org.apache.lucene.index.IndexWriter.closeInternal(IndexWriter.java:1136)
[junit] at 
org.apache.lucene.index.IndexWriter.close(IndexWriter.java:1069)
[junit] at 
org.apache.lucene.index.IndexWriter.close(IndexWriter.java:1033)
[junit] at 
org.apache.lucene.index.TestIndexWriterNRTIsCurrent.testIsCurrentWithThreads(TestIndexWriterNRTIsCurrent.java:68)
[junit] at 
org.apache.lucene.util.LuceneTestCase$SubclassSetupTeardownRule$1.evaluate(LuceneTestCase.java:707)
[junit] at 
org.apache.lucene.util.LuceneTestCase$InternalSetupTeardownRule$1.evaluate(LuceneTestCase.java:606)
[junit] at 
org.apache.lucene.util.LuceneTestCase$TestResultInterceptorRule$1.evaluate(

[jira] [Updated] (SOLR-3153) When a leader goes down he should ask replicas to sync in parallel rather than serially.

2012-02-22 Thread Mark Miller (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-3153?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mark Miller updated SOLR-3153:
--

Attachment: SOLR-3153.patch

> When a leader goes down he should ask replicas to sync in parallel rather 
> than serially.
> 
>
> Key: SOLR-3153
> URL: https://issues.apache.org/jira/browse/SOLR-3153
> Project: Solr
>  Issue Type: Improvement
>  Components: SolrCloud
>Reporter: Mark Miller
>Assignee: Mark Miller
>Priority: Minor
> Fix For: 4.0
>
> Attachments: SOLR-3153.patch
>
>
> Need to finish this todo.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (LUCENE-3074) SimpleTextCodec needs SimpleText DocValues impl

2012-02-22 Thread Simon Willnauer (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-3074?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Simon Willnauer updated LUCENE-3074:


Attachment: LUCENE-3074.patch

here is a first patch adding SimpleTextDV and replacing SimpleTextNorms with it 
directly. 
I had to change some upstream classes and especially the merging done in the 
DocValuesConsumer which used the "wrong" type  for merging. In general we 
should use the target type instead of the source type and sources need to 
implement getBytes and do auto conversion otherwise type promotion doesn't work.

this patch writes individual files per field like sep codec which made things a 
lot easier and is maybe better suited for SimpleText

comments welcome

> SimpleTextCodec needs SimpleText DocValues impl
> ---
>
> Key: LUCENE-3074
> URL: https://issues.apache.org/jira/browse/LUCENE-3074
> Project: Lucene - Java
>  Issue Type: Task
>  Components: core/index, core/search
>Affects Versions: 4.0
>Reporter: Simon Willnauer
>Assignee: Simon Willnauer
>Priority: Blocker
> Fix For: 4.0
>
> Attachments: LUCENE-3074.patch
>
>
> currently SimpleTextCodec uses binary docValues we should move that to a 
> simple text impl.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (SOLR-3153) When a leader goes down he should ask replicas to sync in parallel rather than serially.

2012-02-22 Thread Mark Miller (Created) (JIRA)
When a leader goes down he should ask replicas to sync in parallel rather than 
serially.


 Key: SOLR-3153
 URL: https://issues.apache.org/jira/browse/SOLR-3153
 Project: Solr
  Issue Type: Improvement
  Components: SolrCloud
Reporter: Mark Miller
Assignee: Mark Miller
Priority: Minor
 Fix For: 4.0


Need to finish this todo.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Issue Comment Edited] (LUCENE-3800) Readers wrapping other readers don't prevent usage if any of their subreaders was closed

2012-02-22 Thread Uwe Schindler (Issue Comment Edited) (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3800?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13213642#comment-13213642
 ] 

Uwe Schindler edited comment on LUCENE-3800 at 2/22/12 2:18 PM:


Yes! By the way, the same trick is used for MMapDirectory to keep track of its 
clones IndexInputs.

  was (Author: thetaphi):
Yes!
  
> Readers wrapping other readers don't prevent usage if any of their subreaders 
> was closed
> 
>
> Key: LUCENE-3800
> URL: https://issues.apache.org/jira/browse/LUCENE-3800
> Project: Lucene - Java
>  Issue Type: Bug
>  Components: core/index
>Affects Versions: 4.0
>Reporter: Uwe Schindler
>Assignee: Uwe Schindler
> Fix For: 4.0
>
> Attachments: LUCENE-3800.patch, LUCENE-3800.patch, LUCENE-3800.patch
>
>
> On recent trunk test we got this problem:
> org.apache.lucene.index.TestReaderClosed.test
> fails because the inner reader is closed but the wrapped outer ones are still 
> open.
> I fixed the issue partially for SlowCompositeReaderWrapper and 
> ParallelAtomicReader but it failed again. The cool thing with this test is 
> the following:
> The test opens an DirectoryReader and then creates a searcher, closes the 
> reader and executes a search. This is not an issue, if the reader is closed 
> that the search is running on. This test uses LTC.newSearcher(wrap=true), 
> which randomly wraps the passed Reader with SlowComposite or ParallelReader - 
> or with both!!! If you then close the original inner reader, the close is not 
> detected when excuting search. This can cause SIGSEGV when MMAP is used.
> The problem in (in Slow* and Parallel*) is, that both have their own Fields 
> instances thats are kept alive until the reader itsself is closed. If the 
> child reader is closed, the wrapping reader does not know and still uses its 
> own Fields instance that delegates to the inner readers. On this step no more 
> ensureOpen checks are done, causing the failures.
> The first fix done in Slow and Parallel was to call ensureOpen() on the 
> subReader, too when requesting fields(). This works fine until you wrap two 
> times: 
> ParallelAtomicReader(SlowCompositeReaderWrapper(StandardDirectoryReader(segments_1:3:nrt
>  _0(4.0):C42)))
> One solution would be to make ensureOpen also check all subreaders, but that 
> would do the volatile checks way too often (with n is the total number of 
> subreaders and m is the number of hierarchical levels this is n^m) - we 
> cannot do this. Currently we only have n*m which is fine.
> The proposal how to solve this (closing subreaders under the hood of parent 
> readers is to use the readerClosedListeners. Whenever a composite or slow 
> reader wraps another readers, it registers itself as interested in 
> readerClosed events. When a subreader is then forcefully closed (e.g by a 
> programming error or this crazy test), we automatically close the parents, 
> too.
> We should also fix this in 3.x, if we have similar problems there (needs 
> investigation).

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Resolved] (SOLR-3009) hitGrouped.vm isn't shipped with 3.x

2012-02-22 Thread Resolved

 [ 
https://issues.apache.org/jira/browse/SOLR-3009?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jan Høydahl resolved SOLR-3009.
---

Resolution: Fixed

> hitGrouped.vm isn't shipped with 3.x
> 
>
> Key: SOLR-3009
> URL: https://issues.apache.org/jira/browse/SOLR-3009
> Project: Solr
>  Issue Type: Bug
>Affects Versions: 3.5
>Reporter: Erik Hatcher
>Assignee: Jan Høydahl
> Fix For: 3.6
>
> Attachments: SOLR-3009.patch
>
>
> A user reported getting "Can't find resource 'hitGrouped.vm'" when trying 
> /browse?q=Titel%3Anike&group=true&group.field=EAN on a 3.x version.  Looking 
> at svn, we don't ship hitGrouped.vm on 3.x, but we do on trunk/4.0.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Assigned] (SOLR-3009) hitGrouped.vm isn't shipped with 3.x

2012-02-22 Thread Assigned

 [ 
https://issues.apache.org/jira/browse/SOLR-3009?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jan Høydahl reassigned SOLR-3009:
-

Assignee: Jan Høydahl  (was: Erik Hatcher)

> hitGrouped.vm isn't shipped with 3.x
> 
>
> Key: SOLR-3009
> URL: https://issues.apache.org/jira/browse/SOLR-3009
> Project: Solr
>  Issue Type: Bug
>Affects Versions: 3.5
>Reporter: Erik Hatcher
>Assignee: Jan Høydahl
> Fix For: 3.6
>
> Attachments: SOLR-3009.patch
>
>
> A user reported getting "Can't find resource 'hitGrouped.vm'" when trying 
> /browse?q=Titel%3Anike&group=true&group.field=EAN on a 3.x version.  Looking 
> at svn, we don't ship hitGrouped.vm on 3.x, but we do on trunk/4.0.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-3800) Readers wrapping other readers don't prevent usage if any of their subreaders was closed

2012-02-22 Thread Uwe Schindler (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3800?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13213642#comment-13213642
 ] 

Uwe Schindler commented on LUCENE-3800:
---

Yes!

> Readers wrapping other readers don't prevent usage if any of their subreaders 
> was closed
> 
>
> Key: LUCENE-3800
> URL: https://issues.apache.org/jira/browse/LUCENE-3800
> Project: Lucene - Java
>  Issue Type: Bug
>  Components: core/index
>Affects Versions: 4.0
>Reporter: Uwe Schindler
>Assignee: Uwe Schindler
> Fix For: 4.0
>
> Attachments: LUCENE-3800.patch, LUCENE-3800.patch, LUCENE-3800.patch
>
>
> On recent trunk test we got this problem:
> org.apache.lucene.index.TestReaderClosed.test
> fails because the inner reader is closed but the wrapped outer ones are still 
> open.
> I fixed the issue partially for SlowCompositeReaderWrapper and 
> ParallelAtomicReader but it failed again. The cool thing with this test is 
> the following:
> The test opens an DirectoryReader and then creates a searcher, closes the 
> reader and executes a search. This is not an issue, if the reader is closed 
> that the search is running on. This test uses LTC.newSearcher(wrap=true), 
> which randomly wraps the passed Reader with SlowComposite or ParallelReader - 
> or with both!!! If you then close the original inner reader, the close is not 
> detected when excuting search. This can cause SIGSEGV when MMAP is used.
> The problem in (in Slow* and Parallel*) is, that both have their own Fields 
> instances thats are kept alive until the reader itsself is closed. If the 
> child reader is closed, the wrapping reader does not know and still uses its 
> own Fields instance that delegates to the inner readers. On this step no more 
> ensureOpen checks are done, causing the failures.
> The first fix done in Slow and Parallel was to call ensureOpen() on the 
> subReader, too when requesting fields(). This works fine until you wrap two 
> times: 
> ParallelAtomicReader(SlowCompositeReaderWrapper(StandardDirectoryReader(segments_1:3:nrt
>  _0(4.0):C42)))
> One solution would be to make ensureOpen also check all subreaders, but that 
> would do the volatile checks way too often (with n is the total number of 
> subreaders and m is the number of hierarchical levels this is n^m) - we 
> cannot do this. Currently we only have n*m which is fine.
> The proposal how to solve this (closing subreaders under the hood of parent 
> readers is to use the readerClosedListeners. Whenever a composite or slow 
> reader wraps another readers, it registers itself as interested in 
> readerClosed events. When a subreader is then forcefully closed (e.g by a 
> programming error or this crazy test), we automatically close the parents, 
> too.
> We should also fix this in 3.x, if we have similar problems there (needs 
> investigation).

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-3800) Readers wrapping other readers don't prevent usage if any of their subreaders was closed

2012-02-22 Thread Yonik Seeley (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3800?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13213638#comment-13213638
 ] 

Yonik Seeley commented on LUCENE-3800:
--

Ouch - more weak references.  I was hoping we could reduce the number of those 
(I've seen them cause worse GC problems for a number of people).
But if I understand the description correctly, then without this patch things 
could core dump when using mmap directory?

> Readers wrapping other readers don't prevent usage if any of their subreaders 
> was closed
> 
>
> Key: LUCENE-3800
> URL: https://issues.apache.org/jira/browse/LUCENE-3800
> Project: Lucene - Java
>  Issue Type: Bug
>  Components: core/index
>Affects Versions: 4.0
>Reporter: Uwe Schindler
>Assignee: Uwe Schindler
> Fix For: 4.0
>
> Attachments: LUCENE-3800.patch, LUCENE-3800.patch, LUCENE-3800.patch
>
>
> On recent trunk test we got this problem:
> org.apache.lucene.index.TestReaderClosed.test
> fails because the inner reader is closed but the wrapped outer ones are still 
> open.
> I fixed the issue partially for SlowCompositeReaderWrapper and 
> ParallelAtomicReader but it failed again. The cool thing with this test is 
> the following:
> The test opens an DirectoryReader and then creates a searcher, closes the 
> reader and executes a search. This is not an issue, if the reader is closed 
> that the search is running on. This test uses LTC.newSearcher(wrap=true), 
> which randomly wraps the passed Reader with SlowComposite or ParallelReader - 
> or with both!!! If you then close the original inner reader, the close is not 
> detected when excuting search. This can cause SIGSEGV when MMAP is used.
> The problem in (in Slow* and Parallel*) is, that both have their own Fields 
> instances thats are kept alive until the reader itsself is closed. If the 
> child reader is closed, the wrapping reader does not know and still uses its 
> own Fields instance that delegates to the inner readers. On this step no more 
> ensureOpen checks are done, causing the failures.
> The first fix done in Slow and Parallel was to call ensureOpen() on the 
> subReader, too when requesting fields(). This works fine until you wrap two 
> times: 
> ParallelAtomicReader(SlowCompositeReaderWrapper(StandardDirectoryReader(segments_1:3:nrt
>  _0(4.0):C42)))
> One solution would be to make ensureOpen also check all subreaders, but that 
> would do the volatile checks way too often (with n is the total number of 
> subreaders and m is the number of hierarchical levels this is n^m) - we 
> cannot do this. Currently we only have n*m which is fine.
> The proposal how to solve this (closing subreaders under the hood of parent 
> readers is to use the readerClosedListeners. Whenever a composite or slow 
> reader wraps another readers, it registers itself as interested in 
> readerClosed events. When a subreader is then forcefully closed (e.g by a 
> programming error or this crazy test), we automatically close the parents, 
> too.
> We should also fix this in 3.x, if we have similar problems there (needs 
> investigation).

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (LUCENE-3817) TestThreadedForceMerge failure

2012-02-22 Thread Robert Muir (Created) (JIRA)
TestThreadedForceMerge failure
--

 Key: LUCENE-3817
 URL: https://issues.apache.org/jira/browse/LUCENE-3817
 Project: Lucene - Java
  Issue Type: Bug
Affects Versions: 4.0
Reporter: Robert Muir


leaks a file handle

was running tests a lot to keep my computer busy while debugging something else.
{noformat}
[junit] Testsuite: org.apache.lucene.index.TestThreadedForceMerge
[junit] Testcase: 
testThreadedForceMerge(org.apache.lucene.index.TestThreadedForceMerge):   
Caused an ERROR
[junit] MockDirectoryWrapper: cannot close: there are still open files: 
{_58_1.skp=1, _58_1.tib=1, _58_1.tii=1, _58_0.frq=1, _58_0.tim=1, _58.tvf=1, 
_58.tvd=1, _58_1.doc=1, _58.tvx=1, _58.fdt=1, _58.fdx=1}
[junit] java.lang.RuntimeException: MockDirectoryWrapper: cannot close: 
there are still open files: {_58_1.skp=1, _58_1.tib=1, _58_1.tii=1, 
_58_0.frq=1, _58_0.tim=1, _58.tvf=1, _58.tvd=1, _58_1.doc=1, _58.tvx=1, 
_58.fdt=1, _58.fdx=1}
[junit] at 
org.apache.lucene.store.MockDirectoryWrapper.close(MockDirectoryWrapper.java:555)
[junit] at 
org.apache.lucene.index.TestThreadedForceMerge.testThreadedForceMerge(TestThreadedForceMerge.java:141)
[junit] at 
org.apache.lucene.util.LuceneTestCase$SubclassSetupTeardownRule$1.evaluate(LuceneTestCase.java:707)
[junit] at 
org.apache.lucene.util.LuceneTestCase$InternalSetupTeardownRule$1.evaluate(LuceneTestCase.java:606)
[junit] at 
org.apache.lucene.util.LuceneTestCase$TestResultInterceptorRule$1.evaluate(LuceneTestCase.java:511)
[junit] at 
org.apache.lucene.util.LuceneTestCase$RememberThreadRule$1.evaluate(LuceneTestCase.java:569)
[junit] at 
org.apache.lucene.util.LuceneTestCaseRunner.runChild(LuceneTestCaseRunner.java:165)
[junit] at 
org.apache.lucene.util.LuceneTestCaseRunner.runChild(LuceneTestCaseRunner.java:57)
[junit] Caused by: java.lang.RuntimeException: unclosed IndexInput: 
_58_1.doc
[junit] at 
org.apache.lucene.store.MockDirectoryWrapper.addFileHandle(MockDirectoryWrapper.java:479)
[junit] at 
org.apache.lucene.store.MockDirectoryWrapper.openInput(MockDirectoryWrapper.java:504)
[junit] at 
org.apache.lucene.codecs.mocksep.MockSingleIntIndexInput.(MockSingleIntIndexInput.java:41)
[junit] at 
org.apache.lucene.codecs.mocksep.MockSingleIntFactory.openInput(MockSingleIntFactory.java:32)
[junit] at 
org.apache.lucene.codecs.sep.SepPostingsReader.(SepPostingsReader.java:68)
[junit] at 
org.apache.lucene.codecs.mocksep.MockSepPostingsFormat.fieldsProducer(MockSepPostingsFormat.java:89)
[junit] at 
org.apache.lucene.codecs.perfield.PerFieldPostingsFormat$FieldsReader$1.visitOneFormat(PerFieldPostingsFormat.java:189)
[junit] at 
org.apache.lucene.codecs.perfield.PerFieldPostingsFormat$VisitPerFieldFile.(PerFieldPostingsFormat.java:280)
[junit] at 
org.apache.lucene.codecs.perfield.PerFieldPostingsFormat$FieldsReader$1.(PerFieldPostingsFormat.java:186)
[junit] at 
org.apache.lucene.codecs.perfield.PerFieldPostingsFormat$FieldsReader.(PerFieldPostingsFormat.java:186)
[junit] at 
org.apache.lucene.codecs.perfield.PerFieldPostingsFormat.fieldsProducer(PerFieldPostingsFormat.java:256)
[junit] at 
org.apache.lucene.index.SegmentCoreReaders.(SegmentCoreReaders.java:108)
[junit] at 
org.apache.lucene.index.SegmentReader.(SegmentReader.java:51)
[junit] at 
org.apache.lucene.index.IndexWriter$ReadersAndLiveDocs.getMergeReader(IndexWriter.java:521)
[junit] at 
org.apache.lucene.index.IndexWriter.mergeMiddle(IndexWriter.java:3586)
[junit] at 
org.apache.lucene.index.IndexWriter.merge(IndexWriter.java:3256)
[junit] at 
org.apache.lucene.index.ConcurrentMergeScheduler.doMerge(ConcurrentMergeScheduler.java:382)
[junit] at 
org.apache.lucene.index.ConcurrentMergeScheduler$MergeThread.run(ConcurrentMergeScheduler.java:451)
[junit] 
[junit] 
[junit] Tests run: 1, Failures: 0, Errors: 1, Time elapsed: 0.704 sec
[junit] 
[junit] - Standard Error -
[junit] NOTE: reproduce with: ant test -Dtestcase=TestThreadedForceMerge 
-Dtestmethod=testThreadedForceMerge 
-Dtests.seed=-36d98ee49f3f6205:-602830ebcb0e2f6d:9429978db44e64b 
-Dargs="-Dfile.encoding=UTF-8"
[junit] NOTE: test params are: codec=Lucene40: 
{id=PostingsFormat(name=MockSep), contents=PostingsFormat(name=NestedPulsing)}, 
sim=RandomSimilarityProvider(queryNorm=false,coord=true): {}, locale=ar_JO, 
timezone=Asia/Pontianak
[junit] NOTE: all tests run in this JVM:
[junit] [TestSearchForDuplicates, TestMockAnalyzer, TestToken, 
TestIntBlockCodec, TestImpersonation, TestBitVector, TestAtomicUpdate, 
TestBinaryTerms, TestCheckIndex, TestCrashCausesCorruptIndex, 
TestDirectoryReaderReopen, TestDocCount, TestDocValuesIndexing, 
TestForTooMuchCloning, TestIndexInput

[jira] [Commented] (LUCENE-3800) Readers wrapping other readers don't prevent usage if any of their subreaders was closed

2012-02-22 Thread Uwe Schindler (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3800?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13213626#comment-13213626
 ] 

Uwe Schindler commented on LUCENE-3800:
---

If this would be good for 3.x, too -> reopen. 3.x is more safe as if a child 
reader is closed also parent readers have mostly no chance to do anything.

> Readers wrapping other readers don't prevent usage if any of their subreaders 
> was closed
> 
>
> Key: LUCENE-3800
> URL: https://issues.apache.org/jira/browse/LUCENE-3800
> Project: Lucene - Java
>  Issue Type: Bug
>  Components: core/index
>Affects Versions: 4.0
>Reporter: Uwe Schindler
>Assignee: Uwe Schindler
> Fix For: 4.0
>
> Attachments: LUCENE-3800.patch, LUCENE-3800.patch, LUCENE-3800.patch
>
>
> On recent trunk test we got this problem:
> org.apache.lucene.index.TestReaderClosed.test
> fails because the inner reader is closed but the wrapped outer ones are still 
> open.
> I fixed the issue partially for SlowCompositeReaderWrapper and 
> ParallelAtomicReader but it failed again. The cool thing with this test is 
> the following:
> The test opens an DirectoryReader and then creates a searcher, closes the 
> reader and executes a search. This is not an issue, if the reader is closed 
> that the search is running on. This test uses LTC.newSearcher(wrap=true), 
> which randomly wraps the passed Reader with SlowComposite or ParallelReader - 
> or with both!!! If you then close the original inner reader, the close is not 
> detected when excuting search. This can cause SIGSEGV when MMAP is used.
> The problem in (in Slow* and Parallel*) is, that both have their own Fields 
> instances thats are kept alive until the reader itsself is closed. If the 
> child reader is closed, the wrapping reader does not know and still uses its 
> own Fields instance that delegates to the inner readers. On this step no more 
> ensureOpen checks are done, causing the failures.
> The first fix done in Slow and Parallel was to call ensureOpen() on the 
> subReader, too when requesting fields(). This works fine until you wrap two 
> times: 
> ParallelAtomicReader(SlowCompositeReaderWrapper(StandardDirectoryReader(segments_1:3:nrt
>  _0(4.0):C42)))
> One solution would be to make ensureOpen also check all subreaders, but that 
> would do the volatile checks way too often (with n is the total number of 
> subreaders and m is the number of hierarchical levels this is n^m) - we 
> cannot do this. Currently we only have n*m which is fine.
> The proposal how to solve this (closing subreaders under the hood of parent 
> readers is to use the readerClosedListeners. Whenever a composite or slow 
> reader wraps another readers, it registers itself as interested in 
> readerClosed events. When a subreader is then forcefully closed (e.g by a 
> programming error or this crazy test), we automatically close the parents, 
> too.
> We should also fix this in 3.x, if we have similar problems there (needs 
> investigation).

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Resolved] (LUCENE-3800) Readers wrapping other readers don't prevent usage if any of their subreaders was closed

2012-02-22 Thread Uwe Schindler (Resolved) (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-3800?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Uwe Schindler resolved LUCENE-3800.
---

Resolution: Fixed

Committed trunk revision: 1292293

> Readers wrapping other readers don't prevent usage if any of their subreaders 
> was closed
> 
>
> Key: LUCENE-3800
> URL: https://issues.apache.org/jira/browse/LUCENE-3800
> Project: Lucene - Java
>  Issue Type: Bug
>  Components: core/index
>Affects Versions: 4.0
>Reporter: Uwe Schindler
>Assignee: Uwe Schindler
> Fix For: 4.0
>
> Attachments: LUCENE-3800.patch, LUCENE-3800.patch, LUCENE-3800.patch
>
>
> On recent trunk test we got this problem:
> org.apache.lucene.index.TestReaderClosed.test
> fails because the inner reader is closed but the wrapped outer ones are still 
> open.
> I fixed the issue partially for SlowCompositeReaderWrapper and 
> ParallelAtomicReader but it failed again. The cool thing with this test is 
> the following:
> The test opens an DirectoryReader and then creates a searcher, closes the 
> reader and executes a search. This is not an issue, if the reader is closed 
> that the search is running on. This test uses LTC.newSearcher(wrap=true), 
> which randomly wraps the passed Reader with SlowComposite or ParallelReader - 
> or with both!!! If you then close the original inner reader, the close is not 
> detected when excuting search. This can cause SIGSEGV when MMAP is used.
> The problem in (in Slow* and Parallel*) is, that both have their own Fields 
> instances thats are kept alive until the reader itsself is closed. If the 
> child reader is closed, the wrapping reader does not know and still uses its 
> own Fields instance that delegates to the inner readers. On this step no more 
> ensureOpen checks are done, causing the failures.
> The first fix done in Slow and Parallel was to call ensureOpen() on the 
> subReader, too when requesting fields(). This works fine until you wrap two 
> times: 
> ParallelAtomicReader(SlowCompositeReaderWrapper(StandardDirectoryReader(segments_1:3:nrt
>  _0(4.0):C42)))
> One solution would be to make ensureOpen also check all subreaders, but that 
> would do the volatile checks way too often (with n is the total number of 
> subreaders and m is the number of hierarchical levels this is n^m) - we 
> cannot do this. Currently we only have n*m which is fine.
> The proposal how to solve this (closing subreaders under the hood of parent 
> readers is to use the readerClosedListeners. Whenever a composite or slow 
> reader wraps another readers, it registers itself as interested in 
> readerClosed events. When a subreader is then forcefully closed (e.g by a 
> programming error or this crazy test), we automatically close the parents, 
> too.
> We should also fix this in 3.x, if we have similar problems there (needs 
> investigation).

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Resolved] (LUCENE-3816) FilteredDocIdSet does not handle a case where the inner set iterator is null

2012-02-22 Thread Uwe Schindler (Resolved) (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-3816?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Uwe Schindler resolved LUCENE-3816.
---

Resolution: Fixed

Committed trunk revision: 1292282
Committed 3.x revision: 1292288

Thanks Shay!

> FilteredDocIdSet does not handle a case where the inner set iterator is null
> 
>
> Key: LUCENE-3816
> URL: https://issues.apache.org/jira/browse/LUCENE-3816
> Project: Lucene - Java
>  Issue Type: Bug
>  Components: core/search
>Affects Versions: 3.5
>Reporter: Shay Banon
>Assignee: Uwe Schindler
> Fix For: 3.6, 4.0
>
> Attachments: LUCENE-3816.patch, LUCENE-3816.patch
>
>
> DocIdSet#iterator is allowed to return null, when used in FilteredDocIdSet, 
> if null is returned from the inner set, the FilteredDocIdSetIterator fails 
> since it does not allow for nulls to be passed to it.
> The fix is simple, return null in FilteredDocIdSet in the iterator method is 
> the iterator is null.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (LUCENE-3816) FilteredDocIdSet does not handle a case where the inner set iterator is null

2012-02-22 Thread Uwe Schindler (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-3816?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Uwe Schindler updated LUCENE-3816:
--

Attachment: LUCENE-3816.patch

Patch with a new test that verifies this works. Fails without Shay's patch.

Will commit now.

> FilteredDocIdSet does not handle a case where the inner set iterator is null
> 
>
> Key: LUCENE-3816
> URL: https://issues.apache.org/jira/browse/LUCENE-3816
> Project: Lucene - Java
>  Issue Type: Bug
>  Components: core/search
>Affects Versions: 3.5
>Reporter: Shay Banon
>Assignee: Uwe Schindler
> Fix For: 3.6, 4.0
>
> Attachments: LUCENE-3816.patch, LUCENE-3816.patch
>
>
> DocIdSet#iterator is allowed to return null, when used in FilteredDocIdSet, 
> if null is returned from the inner set, the FilteredDocIdSetIterator fails 
> since it does not allow for nulls to be passed to it.
> The fix is simple, return null in FilteredDocIdSet in the iterator method is 
> the iterator is null.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Resolved] (SOLR-2607) Clean up /client directory

2012-02-22 Thread Erik Hatcher (Resolved) (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-2607?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Erik Hatcher resolved SOLR-2607.


Resolution: Fixed

Thanks Jan, committed.

> Clean up /client directory
> --
>
> Key: SOLR-2607
> URL: https://issues.apache.org/jira/browse/SOLR-2607
> Project: Solr
>  Issue Type: Task
>  Components: clients - java, clients - ruby - flare
>Affects Versions: 4.0
>Reporter: Eric Pugh
>Assignee: Erik Hatcher
>Priority: Minor
> Fix For: 4.0
>
> Attachments: SOLR-2607.patch
>
>
> The /clients directory is a bit of a mess.  The only actively maintained 
> client SolrJ is actually in the /dist directory!   The other clients that 
> used to be in here, /php and /javascript (I think!) have been moved.  The 
> only one is /ruby, and it isn't actively maintained the way other ruby 
> clients are.
> I'd recommend just removing the /clients directory since it's very confusing 
> to a new user who would logically go here to find clients, and only find a 
> ruby one!  It would also let us slim down the size of the download.
> Alterntively if we want the /clients directory, then lets copy over the solrj 
> lib to this dir instead of /dist
> I am happy to submit a patch if this makes sense.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-2607) Clean up /client directory

2012-02-22 Thread Erik Hatcher (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-2607?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Erik Hatcher updated SOLR-2607:
---

Summary: Clean up /client directory  (was: Clean up /clients directory)

> Clean up /client directory
> --
>
> Key: SOLR-2607
> URL: https://issues.apache.org/jira/browse/SOLR-2607
> Project: Solr
>  Issue Type: Task
>  Components: clients - java, clients - ruby - flare
>Affects Versions: 4.0
>Reporter: Eric Pugh
>Assignee: Erik Hatcher
>Priority: Minor
> Fix For: 4.0
>
> Attachments: SOLR-2607.patch
>
>
> The /clients directory is a bit of a mess.  The only actively maintained 
> client SolrJ is actually in the /dist directory!   The other clients that 
> used to be in here, /php and /javascript (I think!) have been moved.  The 
> only one is /ruby, and it isn't actively maintained the way other ruby 
> clients are.
> I'd recommend just removing the /clients directory since it's very confusing 
> to a new user who would logically go here to find clients, and only find a 
> ruby one!  It would also let us slim down the size of the download.
> Alterntively if we want the /clients directory, then lets copy over the solrj 
> lib to this dir instead of /dist
> I am happy to submit a patch if this makes sense.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (LUCENE-3816) FilteredDocIdSet does not handle a case where the inner set iterator is null

2012-02-22 Thread Uwe Schindler (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-3816?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Uwe Schindler updated LUCENE-3816:
--

Fix Version/s: 4.0
   3.6

> FilteredDocIdSet does not handle a case where the inner set iterator is null
> 
>
> Key: LUCENE-3816
> URL: https://issues.apache.org/jira/browse/LUCENE-3816
> Project: Lucene - Java
>  Issue Type: Bug
>  Components: core/search
>Affects Versions: 3.5
>Reporter: Shay Banon
>Assignee: Uwe Schindler
> Fix For: 3.6, 4.0
>
> Attachments: LUCENE-3816.patch
>
>
> DocIdSet#iterator is allowed to return null, when used in FilteredDocIdSet, 
> if null is returned from the inner set, the FilteredDocIdSetIterator fails 
> since it does not allow for nulls to be passed to it.
> The fix is simple, return null in FilteredDocIdSet in the iterator method is 
> the iterator is null.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-3816) FilteredDocIdSet does not handle a case where the inner set iterator is null

2012-02-22 Thread Uwe Schindler (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3816?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13213592#comment-13213592
 ] 

Uwe Schindler commented on LUCENE-3816:
---

Tha nks for reporting this, I will check and commit trunk and 3.x branch!

> FilteredDocIdSet does not handle a case where the inner set iterator is null
> 
>
> Key: LUCENE-3816
> URL: https://issues.apache.org/jira/browse/LUCENE-3816
> Project: Lucene - Java
>  Issue Type: Bug
>  Components: core/search
>Affects Versions: 3.5
>Reporter: Shay Banon
>Assignee: Uwe Schindler
> Attachments: LUCENE-3816.patch
>
>
> DocIdSet#iterator is allowed to return null, when used in FilteredDocIdSet, 
> if null is returned from the inner set, the FilteredDocIdSetIterator fails 
> since it does not allow for nulls to be passed to it.
> The fix is simple, return null in FilteredDocIdSet in the iterator method is 
> the iterator is null.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Assigned] (LUCENE-3816) FilteredDocIdSet does not handle a case where the inner set iterator is null

2012-02-22 Thread Uwe Schindler (Assigned) (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-3816?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Uwe Schindler reassigned LUCENE-3816:
-

Assignee: Uwe Schindler

> FilteredDocIdSet does not handle a case where the inner set iterator is null
> 
>
> Key: LUCENE-3816
> URL: https://issues.apache.org/jira/browse/LUCENE-3816
> Project: Lucene - Java
>  Issue Type: Bug
>  Components: core/search
>Affects Versions: 3.5
>Reporter: Shay Banon
>Assignee: Uwe Schindler
> Attachments: LUCENE-3816.patch
>
>
> DocIdSet#iterator is allowed to return null, when used in FilteredDocIdSet, 
> if null is returned from the inner set, the FilteredDocIdSetIterator fails 
> since it does not allow for nulls to be passed to it.
> The fix is simple, return null in FilteredDocIdSet in the iterator method is 
> the iterator is null.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: [JENKINS] Lucene-Solr-tests-only-trunk - Build # 12492 - Failure

2012-02-22 Thread Robert Muir
OK for the TestIndexWriterNRTisCurrent, it reproduces better if you
make your computer busy (run it with -Dtests.iter=100 or so at the
same time as running 'ant test' in another checkout)

improving the assert a bit:
[junit] Caused by: java.lang.AssertionError: out of bounds:
docid=0,liveDocsLength=0
[junit] at
org.apache.lucene.index.IndexWriter$ReadersAndLiveDocs.delete(IndexWriter.java:535)
[junit] at
org.apache.lucene.index.IndexWriter.commitMergedDeletes(IndexWriter.java:3062)
[junit] at
org.apache.lucene.index.IndexWriter.commitMerge(IndexWriter.java:3136)

seems bogus that we have a 0-length bitvector...

On Wed, Feb 22, 2012 at 6:15 AM, Robert Muir  wrote:
> There are actually 2 failures here... TestDeletionPolicy was a bug in
> LuceneTestCase.
> TestIndexWriterNRTIsCurrent is something else, and it doesnt reproduce
> easily yet
>
> On Wed, Feb 22, 2012 at 4:11 AM, Apache Jenkins Server
>  wrote:
>> Build: https://builds.apache.org/job/Lucene-Solr-tests-only-trunk/12492/
>>
>> 2 tests failed.
>> REGRESSION:  
>> org.apache.lucene.index.TestDeletionPolicy.testKeepLastNDeletionPolicyWithCreates
>>
>> Error Message:
>> null
>>
>> Stack Trace:
>> java.lang.NullPointerException
>>        at 
>> org.apache.lucene.index.FilterAtomicReader$FilterFields.terms(FilterAtomicReader.java:53)
>>        at org.apache.lucene.util.TermContext.build(TermContext.java:94)
>>        at org.apache.lucene.search.TermQuery.createWeight(TermQuery.java:180)
>>        at 
>> org.apache.lucene.search.IndexSearcher.createNormalizedWeight(IndexSearcher.java:584)
>>        at 
>> org.apache.lucene.search.AssertingIndexSearcher.createNormalizedWeight(AssertingIndexSearcher.java:59)
>>        at 
>> org.apache.lucene.search.IndexSearcher.search(IndexSearcher.java:258)
>>        at 
>> org.apache.lucene.index.TestDeletionPolicy.testKeepLastNDeletionPolicyWithCreates(TestDeletionPolicy.java:701)
>>        at 
>> org.apache.lucene.util.LuceneTestCase$SubclassSetupTeardownRule$1.evaluate(LuceneTestCase.java:707)
>>        at 
>> org.apache.lucene.util.LuceneTestCase$InternalSetupTeardownRule$1.evaluate(LuceneTestCase.java:606)
>>        at 
>> org.apache.lucene.util.LuceneTestCase$TestResultInterceptorRule$1.evaluate(LuceneTestCase.java:511)
>>        at 
>> org.apache.lucene.util.LuceneTestCase$RememberThreadRule$1.evaluate(LuceneTestCase.java:569)
>>        at 
>> org.apache.lucene.util.LuceneTestCaseRunner.runChild(LuceneTestCaseRunner.java:165)
>>        at 
>> org.apache.lucene.util.LuceneTestCaseRunner.runChild(LuceneTestCaseRunner.java:57)
>>
>>
>> REGRESSION:  
>> org.apache.lucene.index.TestIndexWriterNRTIsCurrent.testIsCurrentWithThreads
>>
>> Error Message:
>> java.lang.AssertionError: Some threads threw uncaught exceptions!
>>
>> Stack Trace:
>> java.lang.RuntimeException: java.lang.AssertionError: Some threads threw 
>> uncaught exceptions!
>>        at 
>> org.apache.lucene.util.LuceneTestCase.tearDownInternal(LuceneTestCase.java:780)
>>        at 
>> org.apache.lucene.util.LuceneTestCase.access$1000(LuceneTestCase.java:138)
>>        at 
>> org.apache.lucene.util.LuceneTestCase$InternalSetupTeardownRule$1.evaluate(LuceneTestCase.java:607)
>>        at 
>> org.apache.lucene.util.LuceneTestCase$TestResultInterceptorRule$1.evaluate(LuceneTestCase.java:511)
>>        at 
>> org.apache.lucene.util.LuceneTestCase$RememberThreadRule$1.evaluate(LuceneTestCase.java:569)
>>        at 
>> org.apache.lucene.util.LuceneTestCaseRunner.runChild(LuceneTestCaseRunner.java:165)
>>        at 
>> org.apache.lucene.util.LuceneTestCaseRunner.runChild(LuceneTestCaseRunner.java:57)
>>        at 
>> org.apache.lucene.util.LuceneTestCase.checkUncaughtExceptionsAfter(LuceneTestCase.java:808)
>>        at 
>> org.apache.lucene.util.LuceneTestCase.tearDownInternal(LuceneTestCase.java:752)
>>
>>
>>
>>
>> Build Log (for compile errors):
>> [...truncated 1547 lines...]
>>
>>
>>
>>
>> -
>> To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
>> For additional commands, e-mail: dev-h...@lucene.apache.org
>
>
>
> --
> lucidimagination.com



-- 
lucidimagination.com

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-3151) Replace zookeeper.jsp with a servlet

2012-02-22 Thread Erik Hatcher (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-3151?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13213568#comment-13213568
 ] 

Erik Hatcher commented on SOLR-3151:


Ok, so definitely not appropriate for this to be a request handler.  And yeah, 
that core admin is crazy stuff - wouldn't allow for a custom (or non-default, 
thus not Velocity) response writing.  Carry on...

> Replace zookeeper.jsp with a servlet
> 
>
> Key: SOLR-3151
> URL: https://issues.apache.org/jira/browse/SOLR-3151
> Project: Solr
>  Issue Type: Improvement
>Reporter: Ryan McKinley
> Fix For: 4.0
>
> Attachments: SOLR-3151-invalid-json-for-solrconfig.json, 
> SOLR-3151-zookeeper-servlet.patch
>
>
> The zookeeper info is currently generated with a jsp file -- making it harder 
> to maintain then it should be, and harder to include in other applications

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-tests-only-trunk - Build # 12494 - Still Failing

2012-02-22 Thread Apache Jenkins Server
Build: https://builds.apache.org/job/Lucene-Solr-tests-only-trunk/12494/

2 tests failed.
FAILED:  org.apache.solr.cloud.BasicDistributedZkTest.testDistribSearch

Error Message:
There are still nodes recoverying

Stack Trace:
junit.framework.AssertionFailedError: There are still nodes recoverying
at 
org.apache.lucene.util.LuceneTestCaseRunner.runChild(LuceneTestCaseRunner.java:165)
at 
org.apache.lucene.util.LuceneTestCaseRunner.runChild(LuceneTestCaseRunner.java:57)
at 
org.apache.solr.cloud.AbstractDistributedZkTestCase.waitForRecoveriesToFinish(AbstractDistributedZkTestCase.java:124)
at 
org.apache.solr.cloud.BasicDistributedZkTest.testANewCollectionInOneInstanceWithManualShardAssignement(BasicDistributedZkTest.java:306)
at 
org.apache.solr.cloud.BasicDistributedZkTest.doTest(BasicDistributedZkTest.java:276)
at 
org.apache.solr.BaseDistributedSearchTestCase.testDistribSearch(BaseDistributedSearchTestCase.java:664)
at 
org.apache.lucene.util.LuceneTestCase$SubclassSetupTeardownRule$1.evaluate(LuceneTestCase.java:707)
at 
org.apache.lucene.util.LuceneTestCase$InternalSetupTeardownRule$1.evaluate(LuceneTestCase.java:606)
at 
org.apache.lucene.util.LuceneTestCase$TestResultInterceptorRule$1.evaluate(LuceneTestCase.java:511)
at 
org.apache.lucene.util.LuceneTestCase$RememberThreadRule$1.evaluate(LuceneTestCase.java:569)


FAILED:  junit.framework.TestSuite.org.apache.solr.cloud.BasicDistributedZkTest

Error Message:
ERROR: SolrIndexSearcher opens=86 closes=85

Stack Trace:
junit.framework.AssertionFailedError: ERROR: SolrIndexSearcher opens=86 
closes=85
at 
org.apache.solr.SolrTestCaseJ4.endTrackingSearchers(SolrTestCaseJ4.java:152)
at 
org.apache.solr.SolrTestCaseJ4.afterClassSolrTestCase(SolrTestCaseJ4.java:76)




Build Log (for compile errors):
[...truncated 7939 lines...]



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org

[JENKINS] Lucene-Solr-tests-only-trunk-java7 - Build # 1788 - Still Failing

2012-02-22 Thread Apache Jenkins Server
Build: https://builds.apache.org/job/Lucene-Solr-tests-only-trunk-java7/1788/

2 tests failed.
REGRESSION:  org.apache.solr.cloud.BasicDistributedZkTest.testDistribSearch

Error Message:
There are still nodes recoverying

Stack Trace:
junit.framework.AssertionFailedError: There are still nodes recoverying
at 
org.apache.lucene.util.LuceneTestCaseRunner.runChild(LuceneTestCaseRunner.java:165)
at 
org.apache.lucene.util.LuceneTestCaseRunner.runChild(LuceneTestCaseRunner.java:57)
at 
org.apache.solr.cloud.AbstractDistributedZkTestCase.waitForRecoveriesToFinish(AbstractDistributedZkTestCase.java:124)
at 
org.apache.solr.cloud.BasicDistributedZkTest.testANewCollectionInOneInstanceWithManualShardAssignement(BasicDistributedZkTest.java:306)
at 
org.apache.solr.cloud.BasicDistributedZkTest.doTest(BasicDistributedZkTest.java:276)
at 
org.apache.solr.BaseDistributedSearchTestCase.testDistribSearch(BaseDistributedSearchTestCase.java:664)
at 
org.apache.lucene.util.LuceneTestCase$SubclassSetupTeardownRule$1.evaluate(LuceneTestCase.java:707)
at 
org.apache.lucene.util.LuceneTestCase$InternalSetupTeardownRule$1.evaluate(LuceneTestCase.java:606)
at 
org.apache.lucene.util.LuceneTestCase$TestResultInterceptorRule$1.evaluate(LuceneTestCase.java:511)
at 
org.apache.lucene.util.LuceneTestCase$RememberThreadRule$1.evaluate(LuceneTestCase.java:569)


FAILED:  junit.framework.TestSuite.org.apache.solr.cloud.BasicDistributedZkTest

Error Message:
ERROR: SolrIndexSearcher opens=88 closes=87

Stack Trace:
junit.framework.AssertionFailedError: ERROR: SolrIndexSearcher opens=88 
closes=87
at 
org.apache.solr.SolrTestCaseJ4.endTrackingSearchers(SolrTestCaseJ4.java:152)
at 
org.apache.solr.SolrTestCaseJ4.afterClassSolrTestCase(SolrTestCaseJ4.java:76)




Build Log (for compile errors):
[...truncated 9741 lines...]



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org

[jira] [Updated] (LUCENE-3688) Bucketing of association value of a certain category

2012-02-22 Thread Sivan Yogev (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-3688?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sivan Yogev updated LUCENE-3688:


Attachment: LUCENE-association-buckets.r1292224.patch

New patch

> Bucketing of association value of a certain category
> 
>
> Key: LUCENE-3688
> URL: https://issues.apache.org/jira/browse/LUCENE-3688
> Project: Lucene - Java
>  Issue Type: Improvement
>  Components: modules/facet
>Reporter: Sivan Yogev
>Priority: Minor
>  Labels: bucket, facet
> Attachments: LUCENE-association-buckets.patch, 
> LUCENE-association-buckets.r1292224.patch
>
>
> Add facet requests which collect associations of a certain category into 
> buckets, and returns each bucket as a facet result node. Association type is 
> either int or float, and there are two methods to define buckets. The first 
> by providing buckets which contain pre-defined ranges. The other is by 
> declaring the required number of buckets, where the ranges of the different 
> buckets are dynamicly set to create balanced bucket sizes.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: [JENKINS] Lucene-Solr-tests-only-trunk - Build # 12492 - Failure

2012-02-22 Thread Robert Muir
There are actually 2 failures here... TestDeletionPolicy was a bug in
LuceneTestCase.
TestIndexWriterNRTIsCurrent is something else, and it doesnt reproduce
easily yet

On Wed, Feb 22, 2012 at 4:11 AM, Apache Jenkins Server
 wrote:
> Build: https://builds.apache.org/job/Lucene-Solr-tests-only-trunk/12492/
>
> 2 tests failed.
> REGRESSION:  
> org.apache.lucene.index.TestDeletionPolicy.testKeepLastNDeletionPolicyWithCreates
>
> Error Message:
> null
>
> Stack Trace:
> java.lang.NullPointerException
>        at 
> org.apache.lucene.index.FilterAtomicReader$FilterFields.terms(FilterAtomicReader.java:53)
>        at org.apache.lucene.util.TermContext.build(TermContext.java:94)
>        at org.apache.lucene.search.TermQuery.createWeight(TermQuery.java:180)
>        at 
> org.apache.lucene.search.IndexSearcher.createNormalizedWeight(IndexSearcher.java:584)
>        at 
> org.apache.lucene.search.AssertingIndexSearcher.createNormalizedWeight(AssertingIndexSearcher.java:59)
>        at 
> org.apache.lucene.search.IndexSearcher.search(IndexSearcher.java:258)
>        at 
> org.apache.lucene.index.TestDeletionPolicy.testKeepLastNDeletionPolicyWithCreates(TestDeletionPolicy.java:701)
>        at 
> org.apache.lucene.util.LuceneTestCase$SubclassSetupTeardownRule$1.evaluate(LuceneTestCase.java:707)
>        at 
> org.apache.lucene.util.LuceneTestCase$InternalSetupTeardownRule$1.evaluate(LuceneTestCase.java:606)
>        at 
> org.apache.lucene.util.LuceneTestCase$TestResultInterceptorRule$1.evaluate(LuceneTestCase.java:511)
>        at 
> org.apache.lucene.util.LuceneTestCase$RememberThreadRule$1.evaluate(LuceneTestCase.java:569)
>        at 
> org.apache.lucene.util.LuceneTestCaseRunner.runChild(LuceneTestCaseRunner.java:165)
>        at 
> org.apache.lucene.util.LuceneTestCaseRunner.runChild(LuceneTestCaseRunner.java:57)
>
>
> REGRESSION:  
> org.apache.lucene.index.TestIndexWriterNRTIsCurrent.testIsCurrentWithThreads
>
> Error Message:
> java.lang.AssertionError: Some threads threw uncaught exceptions!
>
> Stack Trace:
> java.lang.RuntimeException: java.lang.AssertionError: Some threads threw 
> uncaught exceptions!
>        at 
> org.apache.lucene.util.LuceneTestCase.tearDownInternal(LuceneTestCase.java:780)
>        at 
> org.apache.lucene.util.LuceneTestCase.access$1000(LuceneTestCase.java:138)
>        at 
> org.apache.lucene.util.LuceneTestCase$InternalSetupTeardownRule$1.evaluate(LuceneTestCase.java:607)
>        at 
> org.apache.lucene.util.LuceneTestCase$TestResultInterceptorRule$1.evaluate(LuceneTestCase.java:511)
>        at 
> org.apache.lucene.util.LuceneTestCase$RememberThreadRule$1.evaluate(LuceneTestCase.java:569)
>        at 
> org.apache.lucene.util.LuceneTestCaseRunner.runChild(LuceneTestCaseRunner.java:165)
>        at 
> org.apache.lucene.util.LuceneTestCaseRunner.runChild(LuceneTestCaseRunner.java:57)
>        at 
> org.apache.lucene.util.LuceneTestCase.checkUncaughtExceptionsAfter(LuceneTestCase.java:808)
>        at 
> org.apache.lucene.util.LuceneTestCase.tearDownInternal(LuceneTestCase.java:752)
>
>
>
>
> Build Log (for compile errors):
> [...truncated 1547 lines...]
>
>
>
>
> -
> To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
> For additional commands, e-mail: dev-h...@lucene.apache.org



-- 
lucidimagination.com

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (LUCENE-3807) Cleanup suggester API

2012-02-22 Thread Simon Willnauer (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-3807?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Simon Willnauer updated LUCENE-3807:


Attachment: LUCENE-3807.patch

updated patch with davids suggestions

> Cleanup suggester API
> -
>
> Key: LUCENE-3807
> URL: https://issues.apache.org/jira/browse/LUCENE-3807
> Project: Lucene - Java
>  Issue Type: Improvement
>  Components: modules/other
>Affects Versions: 3.6, 4.0
>Reporter: Simon Willnauer
> Fix For: 4.0
>
> Attachments: LUCENE-3807.patch, LUCENE-3807.patch, LUCENE-3807.patch, 
> LUCENE-3807.patch
>
>
> Currently the suggester api and especially TermFreqIterator don't play that 
> nice with BytesRef and other paradigms we use in lucene, further the java 
> iterator pattern isn't that useful when it gets to work with TermsEnum, 
> BytesRef etc. We should try to clean up this api step by step moving over to 
> BytesRef including the Lookup class and its interface...

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-3807) Cleanup suggester API

2012-02-22 Thread Simon Willnauer (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3807?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13213519#comment-13213519
 ] 

Simon Willnauer commented on LUCENE-3807:
-

bq. This shows up in a number of places. I have mixed feelings about certain 
types having a comparator and others not having it, but it's minor.
well this an indicator if we know something about the order or not. if you get 
null there is not order specified...


bq. BufferingTermFreqIteratorWrapper is a nuisance (buffers in memory). It 
would be nicer to have a sort on disk if something doesn't support sorted 
iteration order.
this is the ultimate goal... see my comment above ({noformat} next steps would 
be cleaning up the in-memory sorting stuff and use the sorter which goes to 
disk to do the actual sorting internally if needed {noformat})


bq. I also wonder float -> long = 4 -> 8 bytes... would this count as an 
incompatible API change (because what used to work for a given amount of RAM 
won't work anymore – BufferingTermFreqIteratorWrapper again)?

see above

bq. if I remember correctly Math.min/max are intrinsics, so you can afford to 
be explicit 

I will upload a new patch - we use this in BytesRefComp too, I think its safe 
to fix

bq. This doesn't seem necessary (lookup accepts a CharSequence?).
right - leftover from an old iteration

bq. Why not a specific covariant here?
that would be Float get(String) or Float get(CharSequence) or... ;)



> Cleanup suggester API
> -
>
> Key: LUCENE-3807
> URL: https://issues.apache.org/jira/browse/LUCENE-3807
> Project: Lucene - Java
>  Issue Type: Improvement
>  Components: modules/other
>Affects Versions: 3.6, 4.0
>Reporter: Simon Willnauer
> Fix For: 4.0
>
> Attachments: LUCENE-3807.patch, LUCENE-3807.patch, LUCENE-3807.patch
>
>
> Currently the suggester api and especially TermFreqIterator don't play that 
> nice with BytesRef and other paradigms we use in lucene, further the java 
> iterator pattern isn't that useful when it gets to work with TermsEnum, 
> BytesRef etc. We should try to clean up this api step by step moving over to 
> BytesRef including the Lookup class and its interface...

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-tests-only-trunk - Build # 12493 - Still Failing

2012-02-22 Thread Apache Jenkins Server
Build: https://builds.apache.org/job/Lucene-Solr-tests-only-trunk/12493/

1 tests failed.
FAILED:  
org.apache.lucene.queries.function.TestOrdValues.testReverseOrdFieldRank

Error Message:
java.lang.AssertionError: 
org.apache.lucene.queries.function.TestOrdValues.testReverseOrdFieldRank: 
Insane FieldCache usage(s) found expected:<0> but was:<1>

Stack Trace:
java.lang.RuntimeException: java.lang.AssertionError: 
org.apache.lucene.queries.function.TestOrdValues.testReverseOrdFieldRank: 
Insane FieldCache usage(s) found expected:<0> but was:<1>
at 
org.apache.lucene.util.LuceneTestCase.tearDownInternal(LuceneTestCase.java:780)
at 
org.apache.lucene.util.LuceneTestCase.access$1000(LuceneTestCase.java:138)
at 
org.apache.lucene.util.LuceneTestCase$InternalSetupTeardownRule$1.evaluate(LuceneTestCase.java:607)
at 
org.apache.lucene.util.LuceneTestCase$TestResultInterceptorRule$1.evaluate(LuceneTestCase.java:511)
at 
org.apache.lucene.util.LuceneTestCase$RememberThreadRule$1.evaluate(LuceneTestCase.java:569)
at 
org.apache.lucene.util.LuceneTestCaseRunner.runChild(LuceneTestCaseRunner.java:165)
at 
org.apache.lucene.util.LuceneTestCaseRunner.runChild(LuceneTestCaseRunner.java:57)
at 
org.apache.lucene.util.LuceneTestCase.assertSaneFieldCaches(LuceneTestCase.java:907)
at 
org.apache.lucene.util.LuceneTestCase.tearDownInternal(LuceneTestCase.java:770)




Build Log (for compile errors):
[...truncated 5037 lines...]



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org

Re: [JENKINS] Lucene-trunk - Build # 1836 - Still Failing

2012-02-22 Thread Simon Willnauer
I committed a fix - test bug...

simon

On Wed, Feb 22, 2012 at 5:57 AM, Apache Jenkins Server <
jenk...@builds.apache.org> wrote:

> Build: https://builds.apache.org/job/Lucene-trunk/1836/
>
> 1 tests failed.
> FAILED:
>  org.apache.lucene.search.suggest.fst.FSTCompletionTest.testMultilingualInput
>
> Error Message:
> expected:<[74 68 72 65 65 62 79 74 65 f0 a4 ad a2 63 68 61 72]> but
> was:
>
> Stack Trace:
> junit.framework.AssertionFailedError: expected:<[74 68 72 65 65 62 79 74
> 65 f0 a4 ad a2 63 68 61 72]> but was:
>at
> org.apache.lucene.util.LuceneTestCaseRunner.runChild(LuceneTestCaseRunner.java:165)
>at
> org.apache.lucene.util.LuceneTestCaseRunner.runChild(LuceneTestCaseRunner.java:57)
>at
> org.apache.lucene.search.suggest.fst.FSTCompletionTest.testMultilingualInput(FSTCompletionTest.java:187)
>at
> org.apache.lucene.util.LuceneTestCase$SubclassSetupTeardownRule$1.evaluate(LuceneTestCase.java:707)
>at
> org.apache.lucene.util.LuceneTestCase$InternalSetupTeardownRule$1.evaluate(LuceneTestCase.java:606)
>at
> org.apache.lucene.util.LuceneTestCase$TestResultInterceptorRule$1.evaluate(LuceneTestCase.java:511)
>at
> org.apache.lucene.util.LuceneTestCase$RememberThreadRule$1.evaluate(LuceneTestCase.java:569)
>
>
>
>
> Build Log (for compile errors):
> [...truncated 16827 lines...]
>
>
>
>
> -
> To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
> For additional commands, e-mail: dev-h...@lucene.apache.org
>


[JENKINS] Lucene-Solr-tests-only-trunk-java7 - Build # 1787 - Failure

2012-02-22 Thread Apache Jenkins Server
Build: https://builds.apache.org/job/Lucene-Solr-tests-only-trunk-java7/1787/

1 tests failed.
FAILED:  junit.framework.TestSuite.org.apache.solr.cloud.RecoveryZkTest

Error Message:
ERROR: SolrIndexSearcher opens=15 closes=14

Stack Trace:
junit.framework.AssertionFailedError: ERROR: SolrIndexSearcher opens=15 
closes=14
at 
org.apache.solr.SolrTestCaseJ4.endTrackingSearchers(SolrTestCaseJ4.java:152)
at 
org.apache.solr.SolrTestCaseJ4.afterClassSolrTestCase(SolrTestCaseJ4.java:76)




Build Log (for compile errors):
[...truncated 10428 lines...]



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org

[jira] [Updated] (SOLR-3151) Replace zookeeper.jsp with a servlet

2012-02-22 Thread Stefan Matheis (steffkes) (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-3151?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Stefan Matheis (steffkes) updated SOLR-3151:


Attachment: SOLR-3151-invalid-json-for-solrconfig.json

Generated w/ SVN Revision 1245905, is not valid. I don't not all reasons why 
it's correct but one will be that it contains raw linebreaks, which doesn't fit 
the specs. more on http://jsonlint.com/

> Replace zookeeper.jsp with a servlet
> 
>
> Key: SOLR-3151
> URL: https://issues.apache.org/jira/browse/SOLR-3151
> Project: Solr
>  Issue Type: Improvement
>Reporter: Ryan McKinley
> Fix For: 4.0
>
> Attachments: SOLR-3151-invalid-json-for-solrconfig.json, 
> SOLR-3151-zookeeper-servlet.patch
>
>
> The zookeeper info is currently generated with a jsp file -- making it harder 
> to maintain then it should be, and harder to include in other applications

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-3151) Replace zookeeper.jsp with a servlet

2012-02-22 Thread Stefan Matheis (steffkes) (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-3151?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13213495#comment-13213495
 ] 

Stefan Matheis (steffkes) commented on SOLR-3151:
-

Just a quick note. Depending on which file you select, the current 
zookeeper.jsp renders invalid json. That will of course not work for the 
Javascript-Part, we can of course display a message in that case, but i guess 
it should be possible to ensure that we return valid json?

> Replace zookeeper.jsp with a servlet
> 
>
> Key: SOLR-3151
> URL: https://issues.apache.org/jira/browse/SOLR-3151
> Project: Solr
>  Issue Type: Improvement
>Reporter: Ryan McKinley
> Fix For: 4.0
>
> Attachments: SOLR-3151-zookeeper-servlet.patch
>
>
> The zookeeper info is currently generated with a jsp file -- making it harder 
> to maintain then it should be, and harder to include in other applications

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-3116) new Admin UI does not allow drill-down into ZooKeeper

2012-02-22 Thread Stefan Matheis (steffkes) (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-3116?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Stefan Matheis (steffkes) updated SOLR-3116:


Attachment: SOLR-3116-file-content.png

Erick, that's correct .. what about that one? Usable?

The current zookeeper.jsp responds sometimes w/ a invalid json-structure, that 
breaks the javascript .. will have a look on SOLR-3151 to ensure that this does 
not happen there.



> new Admin UI does not allow drill-down into ZooKeeper
> -
>
> Key: SOLR-3116
> URL: https://issues.apache.org/jira/browse/SOLR-3116
> Project: Solr
>  Issue Type: Bug
>Affects Versions: 4.0
> Environment: All
>Reporter: Erick Erickson
>Priority: Minor
> Attachments: SOLR-3116-file-content.png
>
>
> One thing missing from the old UI for the ZooKeeper view - you can no longer 
> see the data at each node (or at least I have not figured out how) - just the 
> node listing. (Mark Miller, broken out from SOLR-2667)

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Solr-trunk - Build # 1770 - Failure

2012-02-22 Thread Apache Jenkins Server
Build: https://builds.apache.org/job/Solr-trunk/1770/

1 tests failed.
FAILED:  junit.framework.TestSuite.org.apache.solr.cloud.RecoveryZkTest

Error Message:
ERROR: SolrIndexSearcher opens=15 closes=14

Stack Trace:
junit.framework.AssertionFailedError: ERROR: SolrIndexSearcher opens=15 
closes=14
at 
org.apache.solr.SolrTestCaseJ4.endTrackingSearchers(SolrTestCaseJ4.java:152)
at 
org.apache.solr.SolrTestCaseJ4.afterClassSolrTestCase(SolrTestCaseJ4.java:76)




Build Log (for compile errors):
[...truncated 10151 lines...]



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org

[jira] [Issue Comment Edited] (LUCENE-3807) Cleanup suggester API

2012-02-22 Thread Dawid Weiss (Issue Comment Edited) (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3807?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13213477#comment-13213477
 ] 

Dawid Weiss edited comment on LUCENE-3807 at 2/22/12 9:16 AM:
--

Sorry for being late, work. I like the patch. Comments:

{noformat}
+public Comparator getComparator() {
+  return null;
+}
{noformat}

This shows up in a number of places. I have mixed feelings about certain types 
having a comparator and others not having it, but it's minor.

BufferingTermFreqIteratorWrapper is a nuisance (buffers in memory). It would be 
nicer to have a sort on disk if something doesn't support sorted iteration 
order.

I also wonder float -> long = 4 -> 8 bytes... would this count as an 
incompatible API change (because what used to work for a given amount of RAM 
won't work anymore -- BufferingTermFreqIteratorWrapper again)?

{noformat}
+  if (l1 < l2) {
+aStop = l1;
+  } else {
+aStop = l2;
+  }
{noformat}

if I remember correctly Math.min/max are intrinsics, so you can afford to be 
explicit ;)

Why not a specific covariant here?

{noformat}
-  public Float get(String key) {
+  public Object get(CharSequence key) {
{noformat}

This doesn't seem necessary (lookup accepts a CharSequence?).
{noformat}
@@ -199,7 +199,7 @@ public class LookupBenchmarkTest extends LuceneTestCase {
 public Integer call() throws Exception {
   int v = 0;
   for (String term : input) {
-v += lookup.lookup(term, onlyMorePopular, num).size();
+v += lookup.lookup(new CharsRef(term), onlyMorePopular, 
num).size();
{noformat}

I like the rest, including the CharSequenceish evilness of bytesToCharSequence 
:)


  was (Author: dweiss):
Sorry for being late, work. I like the patch. Comments:

+public Comparator getComparator() {
+  return null;
+}

This shows up in a number of places. I have mixed feelings about certain types 
having a comparator and others not having it, but it's minor.

BufferingTermFreqIteratorWrapper is a nuisance (buffers in memory). It would be 
nicer to have a sort on disk if something doesn't support sorted iteration 
order.

I also wonder float -> long = 4 -> 8 bytes... would this count as an 
incompatible API change (because what used to work for a given amount of RAM 
won't work anymore -- BufferingTermFreqIteratorWrapper again)?

+  if (l1 < l2) {
+aStop = l1;
+  } else {
+aStop = l2;
+  }

if I remember correctly Math.min/max are intrinsics, so you can afford to be 
explicit ;)

Why not a specific covariant here?

-  public Float get(String key) {
+  public Object get(CharSequence key) {

@@ -199,7 +199,7 @@ public class LookupBenchmarkTest extends LuceneTestCase {
 public Integer call() throws Exception {
   int v = 0;
   for (String term : input) {
-v += lookup.lookup(term, onlyMorePopular, num).size();
+v += lookup.lookup(new CharsRef(term), onlyMorePopular, 
num).size();

This doesn't seem necessary (lookup accepts a CharSequence?).

I like the rest, including the CharSequenceish evilness of bytesToCharSequence 
:)


  
> Cleanup suggester API
> -
>
> Key: LUCENE-3807
> URL: https://issues.apache.org/jira/browse/LUCENE-3807
> Project: Lucene - Java
>  Issue Type: Improvement
>  Components: modules/other
>Affects Versions: 3.6, 4.0
>Reporter: Simon Willnauer
> Fix For: 4.0
>
> Attachments: LUCENE-3807.patch, LUCENE-3807.patch, LUCENE-3807.patch
>
>
> Currently the suggester api and especially TermFreqIterator don't play that 
> nice with BytesRef and other paradigms we use in lucene, further the java 
> iterator pattern isn't that useful when it gets to work with TermsEnum, 
> BytesRef etc. We should try to clean up this api step by step moving over to 
> BytesRef including the Lookup class and its interface...

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-3807) Cleanup suggester API

2012-02-22 Thread Dawid Weiss (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3807?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13213477#comment-13213477
 ] 

Dawid Weiss commented on LUCENE-3807:
-

Sorry for being late, work. I like the patch. Comments:

+public Comparator getComparator() {
+  return null;
+}

This shows up in a number of places. I have mixed feelings about certain types 
having a comparator and others not having it, but it's minor.

BufferingTermFreqIteratorWrapper is a nuisance (buffers in memory). It would be 
nicer to have a sort on disk if something doesn't support sorted iteration 
order.

I also wonder float -> long = 4 -> 8 bytes... would this count as an 
incompatible API change (because what used to work for a given amount of RAM 
won't work anymore -- BufferingTermFreqIteratorWrapper again)?

+  if (l1 < l2) {
+aStop = l1;
+  } else {
+aStop = l2;
+  }

if I remember correctly Math.min/max are intrinsics, so you can afford to be 
explicit ;)

Why not a specific covariant here?

-  public Float get(String key) {
+  public Object get(CharSequence key) {

@@ -199,7 +199,7 @@ public class LookupBenchmarkTest extends LuceneTestCase {
 public Integer call() throws Exception {
   int v = 0;
   for (String term : input) {
-v += lookup.lookup(term, onlyMorePopular, num).size();
+v += lookup.lookup(new CharsRef(term), onlyMorePopular, 
num).size();

This doesn't seem necessary (lookup accepts a CharSequence?).

I like the rest, including the CharSequenceish evilness of bytesToCharSequence 
:)



> Cleanup suggester API
> -
>
> Key: LUCENE-3807
> URL: https://issues.apache.org/jira/browse/LUCENE-3807
> Project: Lucene - Java
>  Issue Type: Improvement
>  Components: modules/other
>Affects Versions: 3.6, 4.0
>Reporter: Simon Willnauer
> Fix For: 4.0
>
> Attachments: LUCENE-3807.patch, LUCENE-3807.patch, LUCENE-3807.patch
>
>
> Currently the suggester api and especially TermFreqIterator don't play that 
> nice with BytesRef and other paradigms we use in lucene, further the java 
> iterator pattern isn't that useful when it gets to work with TermsEnum, 
> BytesRef etc. We should try to clean up this api step by step moving over to 
> BytesRef including the Lookup class and its interface...

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-tests-only-trunk - Build # 12492 - Failure

2012-02-22 Thread Apache Jenkins Server
Build: https://builds.apache.org/job/Lucene-Solr-tests-only-trunk/12492/

2 tests failed.
REGRESSION:  
org.apache.lucene.index.TestDeletionPolicy.testKeepLastNDeletionPolicyWithCreates

Error Message:
null

Stack Trace:
java.lang.NullPointerException
at 
org.apache.lucene.index.FilterAtomicReader$FilterFields.terms(FilterAtomicReader.java:53)
at org.apache.lucene.util.TermContext.build(TermContext.java:94)
at org.apache.lucene.search.TermQuery.createWeight(TermQuery.java:180)
at 
org.apache.lucene.search.IndexSearcher.createNormalizedWeight(IndexSearcher.java:584)
at 
org.apache.lucene.search.AssertingIndexSearcher.createNormalizedWeight(AssertingIndexSearcher.java:59)
at org.apache.lucene.search.IndexSearcher.search(IndexSearcher.java:258)
at 
org.apache.lucene.index.TestDeletionPolicy.testKeepLastNDeletionPolicyWithCreates(TestDeletionPolicy.java:701)
at 
org.apache.lucene.util.LuceneTestCase$SubclassSetupTeardownRule$1.evaluate(LuceneTestCase.java:707)
at 
org.apache.lucene.util.LuceneTestCase$InternalSetupTeardownRule$1.evaluate(LuceneTestCase.java:606)
at 
org.apache.lucene.util.LuceneTestCase$TestResultInterceptorRule$1.evaluate(LuceneTestCase.java:511)
at 
org.apache.lucene.util.LuceneTestCase$RememberThreadRule$1.evaluate(LuceneTestCase.java:569)
at 
org.apache.lucene.util.LuceneTestCaseRunner.runChild(LuceneTestCaseRunner.java:165)
at 
org.apache.lucene.util.LuceneTestCaseRunner.runChild(LuceneTestCaseRunner.java:57)


REGRESSION:  
org.apache.lucene.index.TestIndexWriterNRTIsCurrent.testIsCurrentWithThreads

Error Message:
java.lang.AssertionError: Some threads threw uncaught exceptions!

Stack Trace:
java.lang.RuntimeException: java.lang.AssertionError: Some threads threw 
uncaught exceptions!
at 
org.apache.lucene.util.LuceneTestCase.tearDownInternal(LuceneTestCase.java:780)
at 
org.apache.lucene.util.LuceneTestCase.access$1000(LuceneTestCase.java:138)
at 
org.apache.lucene.util.LuceneTestCase$InternalSetupTeardownRule$1.evaluate(LuceneTestCase.java:607)
at 
org.apache.lucene.util.LuceneTestCase$TestResultInterceptorRule$1.evaluate(LuceneTestCase.java:511)
at 
org.apache.lucene.util.LuceneTestCase$RememberThreadRule$1.evaluate(LuceneTestCase.java:569)
at 
org.apache.lucene.util.LuceneTestCaseRunner.runChild(LuceneTestCaseRunner.java:165)
at 
org.apache.lucene.util.LuceneTestCaseRunner.runChild(LuceneTestCaseRunner.java:57)
at 
org.apache.lucene.util.LuceneTestCase.checkUncaughtExceptionsAfter(LuceneTestCase.java:808)
at 
org.apache.lucene.util.LuceneTestCase.tearDownInternal(LuceneTestCase.java:752)




Build Log (for compile errors):
[...truncated 1547 lines...]



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org

[jira] [Commented] (LUCENE-3795) Replace spatial contrib module with LSP's spatial-lucene module

2012-02-22 Thread Chris Male (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3795?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13213445#comment-13213445
 ] 

Chris Male commented on LUCENE-3795:


bq. LSP Lucene spatial is in as a module, actually as two modules, spatial/base 
& spatial/strategy. That complicates the build but Chris & Ryan insist. The 
maven build works.

I don't recall insisting that here.  I think we should do what's best in this 
situation.  If having two sub modules is causing too much difficulty for no 
benefit, then +1 to reducing them to one.

> Replace spatial contrib module with LSP's spatial-lucene module
> ---
>
> Key: LUCENE-3795
> URL: https://issues.apache.org/jira/browse/LUCENE-3795
> Project: Lucene - Java
>  Issue Type: New Feature
>  Components: modules/spatial
>Reporter: David Smiley
>Assignee: David Smiley
> Fix For: 4.0
>
>
> I propose that Lucene's spatial contrib module be replaced with the 
> spatial-lucene module within Lucene Spatial Playground (LSP).  LSP has been 
> in development for approximately 1 year by David Smiley, Ryan McKinley, and 
> Chris Male and we feel it is ready.  LSP is here: 
> http://code.google.com/p/lucene-spatial-playground/  and the spatial-lucene 
> module is intuitively in svn/trunk/spatial-lucene/.
> I'll add more comments to prevent the issue description from being too long.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



  1   2   >