[jira] Commented: (SOLR-1425) Better exception messages for classloader mishaps

2009-09-11 Thread Hoss Man (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-1425?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12754483#action_12754483
 ] 

Hoss Man commented on SOLR-1425:


(FWIW: this is an offshoot of SOLR-1419)

Benson: can you post the *FULL* stacktrace of what you see in your VM when you 
encounter a problem like this?

The stack trace you posted in SOLR-1419 is only partial (note that it starts 
with "Caused by" ... there should have been more before that)  
SolrResourceLoader already catches ClassNotFound and logs appropriate info -- 
so that stack trace you posted should have started with something more useful.


> Better exception messages for classloader mishaps
> -
>
> Key: SOLR-1425
> URL: https://issues.apache.org/jira/browse/SOLR-1425
> Project: Solr
>  Issue Type: New Feature
>  Components: search
>Affects Versions: 1.3
>Reporter: Benson Margulies
>
> If an idiot such as myself tries to specify a filter or such that lives in a 
> parent classloader, such as the system classloader of a servlet container, 
> Solr will fail to load it. The JVM is prone to create an exception that 
> mentions some Solr interface as being missing instead of the filter itself. 
> It would be less confusing for the miscreant if Solr were to try/catch 
> ClassNotFound and NoClassDefError and throw its own exception with the name 
> of the thing specified in the schema included in the message.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (SOLR-1316) Create autosuggest component

2009-09-11 Thread Shalin Shekhar Mangar (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-1316?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12754482#action_12754482
 ] 

Shalin Shekhar Mangar commented on SOLR-1316:
-

This is a duplicate of SOLR-706.

Since we comments on both, which one should I close? :)

> Create autosuggest component
> 
>
> Key: SOLR-1316
> URL: https://issues.apache.org/jira/browse/SOLR-1316
> Project: Solr
>  Issue Type: New Feature
>  Components: search
>Affects Versions: 1.4
>Reporter: Jason Rutherglen
>Priority: Minor
> Fix For: 1.5
>
>   Original Estimate: 96h
>  Remaining Estimate: 96h
>
> Autosuggest is a common search function that can be integrated
> into Solr as a SearchComponent. Our first implementation will
> use the TernaryTree found in Lucene contrib. 
> * Enable creation of the dictionary from the index or via Solr's
> RPC mechanism
> * What types of parameters and settings are desirable?
> * Hopefully in the future we can include user click through
> rates to boost those terms/phrases higher

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (SOLR-1336) Add support for lucene's SmartChineseAnalyzer

2009-09-11 Thread Hoss Man (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-1336?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12754481#action_12754481
 ] 

Hoss Man commented on SOLR-1336:


bq. The downside is a 3MB jar in solr/lib and in the solr.war

contrib?

Chinese isn't something everybody needs, and 3MB would almost double the size 
of the solr.war. 

> Add support for lucene's SmartChineseAnalyzer
> -
>
> Key: SOLR-1336
> URL: https://issues.apache.org/jira/browse/SOLR-1336
> Project: Solr
>  Issue Type: New Feature
>  Components: Analysis
>Reporter: Robert Muir
> Attachments: SOLR-1336.patch, SOLR-1336.patch, SOLR-1336.patch
>
>
> SmartChineseAnalyzer was contributed to lucene, it indexes simplified chinese 
> text as words.
> if the factories for the tokenizer and word token filter are added to solr it 
> can be used, although there should be a sample config or wiki entry showing 
> how to apply the built-in stopwords list.
> this is because it doesn't contain actual stopwords, but must be used to 
> prevent indexing punctuation... 
> note: we did some refactoring/cleanup on this analyzer recently, so it would 
> be much easier to do this after the next lucene update.
> it has also been moved out of -analyzers.jar due to size, and now builds in 
> its own smartcn jar file, so that would need to be added if this feature is 
> desired.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (SOLR-788) MoreLikeThis should support distributed search

2009-09-11 Thread Shalin Shekhar Mangar (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-788?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Shalin Shekhar Mangar updated SOLR-788:
---

  Component/s: MoreLikeThis
Fix Version/s: 1.5

> MoreLikeThis should support distributed search
> --
>
> Key: SOLR-788
> URL: https://issues.apache.org/jira/browse/SOLR-788
> Project: Solr
>  Issue Type: Improvement
>  Components: MoreLikeThis
>Reporter: Grant Ingersoll
>Priority: Minor
> Fix For: 1.5
>
> Attachments: AlternateDistributedMLT.patch, 
> MoreLikeThisComponentTest.patch, SolrMoreLikeThisPatch.txt
>
>
> The MoreLikeThis component should support distributed processing.
> See SOLR-303.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Resolved: (SOLR-1424) ant generate-maven-artifacts fails on windows

2009-09-11 Thread Ryan McKinley (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-1424?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ryan McKinley resolved SOLR-1424.
-

   Resolution: Fixed
Fix Version/s: 1.4

> ant generate-maven-artifacts fails on windows
> -
>
> Key: SOLR-1424
> URL: https://issues.apache.org/jira/browse/SOLR-1424
> Project: Solr
>  Issue Type: Bug
>Reporter: Hoss Man
>Assignee: Ryan McKinley
> Fix For: 1.4
>
> Attachments: SOLR-1424.patch
>
>
> From solr-user...
> {noformat}
> generate-maven-artifacts:
> [mkdir] Created dir: c:\Downloads\solr_trunk\build\maven
> [mkdir] Created dir: c:\Downloads\solr_trunk\dist\maven
>  [copy] Copying 1 file to
> c:\Downloads\solr_trunk\build\maven\c:\Downloads\s
> olr_trunk\src\maven
> BUILD FAILED
> c:\Downloads\solr_trunk\build.xml:741: The following error occurred while
> execut
> ing this line:
> c:\Downloads\solr_trunk\common-build.xml:261: Failed to copy
> c:\Downloads\solr_t
> runk\src\maven\solr-parent-pom.xml.template to
> c:\Downloads\solr_trunk\build\mav
> en\c:\Downloads\solr_trunk\src\maven\solr-parent-pom.xml.template due to
> java.io
> .FileNotFoundException
> c:\Downloads\solr_trunk\build\maven\c:\Downloads\solr_tru
> nk\src\maven\solr-parent-pom.xml.template (The filename, directory name, or
> volu
> me label syntax is incorrect)
> {noformat}

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (SOLR-1424) ant generate-maven-artifacts fails on windows

2009-09-11 Thread Ryan McKinley (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-1424?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ryan McKinley updated SOLR-1424:


Attachment: SOLR-1424.patch

> ant generate-maven-artifacts fails on windows
> -
>
> Key: SOLR-1424
> URL: https://issues.apache.org/jira/browse/SOLR-1424
> Project: Solr
>  Issue Type: Bug
>Reporter: Hoss Man
>Assignee: Ryan McKinley
> Fix For: 1.4
>
> Attachments: SOLR-1424.patch
>
>
> From solr-user...
> {noformat}
> generate-maven-artifacts:
> [mkdir] Created dir: c:\Downloads\solr_trunk\build\maven
> [mkdir] Created dir: c:\Downloads\solr_trunk\dist\maven
>  [copy] Copying 1 file to
> c:\Downloads\solr_trunk\build\maven\c:\Downloads\s
> olr_trunk\src\maven
> BUILD FAILED
> c:\Downloads\solr_trunk\build.xml:741: The following error occurred while
> execut
> ing this line:
> c:\Downloads\solr_trunk\common-build.xml:261: Failed to copy
> c:\Downloads\solr_t
> runk\src\maven\solr-parent-pom.xml.template to
> c:\Downloads\solr_trunk\build\mav
> en\c:\Downloads\solr_trunk\src\maven\solr-parent-pom.xml.template due to
> java.io
> .FileNotFoundException
> c:\Downloads\solr_trunk\build\maven\c:\Downloads\solr_tru
> nk\src\maven\solr-parent-pom.xml.template (The filename, directory name, or
> volu
> me label syntax is incorrect)
> {noformat}

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Assigned: (SOLR-1424) ant generate-maven-artifacts fails on windows

2009-09-11 Thread Ryan McKinley (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-1424?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ryan McKinley reassigned SOLR-1424:
---

Assignee: Ryan McKinley

yup, that does it...  I will commit shortly

> ant generate-maven-artifacts fails on windows
> -
>
> Key: SOLR-1424
> URL: https://issues.apache.org/jira/browse/SOLR-1424
> Project: Solr
>  Issue Type: Bug
>Reporter: Hoss Man
>Assignee: Ryan McKinley
>
> From solr-user...
> {noformat}
> generate-maven-artifacts:
> [mkdir] Created dir: c:\Downloads\solr_trunk\build\maven
> [mkdir] Created dir: c:\Downloads\solr_trunk\dist\maven
>  [copy] Copying 1 file to
> c:\Downloads\solr_trunk\build\maven\c:\Downloads\s
> olr_trunk\src\maven
> BUILD FAILED
> c:\Downloads\solr_trunk\build.xml:741: The following error occurred while
> execut
> ing this line:
> c:\Downloads\solr_trunk\common-build.xml:261: Failed to copy
> c:\Downloads\solr_t
> runk\src\maven\solr-parent-pom.xml.template to
> c:\Downloads\solr_trunk\build\mav
> en\c:\Downloads\solr_trunk\src\maven\solr-parent-pom.xml.template due to
> java.io
> .FileNotFoundException
> c:\Downloads\solr_trunk\build\maven\c:\Downloads\solr_tru
> nk\src\maven\solr-parent-pom.xml.template (The filename, directory name, or
> volu
> me label syntax is incorrect)
> {noformat}

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



Re: Spec Version vs Implementation Version

2009-09-11 Thread Chris Hostetter

: What are the differences between specification version and implementation
: version

those are concepts from the Java specification for jars and wars (more 
info then you could ever possibly want in the URLs below)

: I downloaded the nightly build for September 05 2009 and it has a spec
: version of 1.3 and the implementation version states 1.4-dev

I think you are abreviating.  the spec version should be something like 
"1.3.0.2009.09.05.12.07.49" and the implementation version should be 
something like "1.4-dev 812759 - hudson - 2009-09-05 12:07:49"

In a nutshell: the spec version identifies the specification, in our case 
the Java APIs.  the Implementation version itentifies the implementation 
(in our case: the internals of those methods).  

a spec version must be purely numeric, and has rules about how to 
interpret when one version is newer then another.  For a relase, the spec 
version looks like "1.3.0" or "1.3.1" or "1.4.0" but for the nightlys we 
include the date to denote that the API is "newer" then it was in 1.3.0. 

An implementation version can be any string.  for releases it starts out 
the same as the spec version, but then includes other details about that 
particular build (the svn revision, who built it, and when it was build) 
... for dev versions it says all the same things, but the initial verion 
info tells you what development branch you're looking at - so 1.4-dev 
means it's working towards 1.4 (as opposed to a hypothetical 1.3.1-dev 
that might exist if someone created a maintence branch on 1.3 in 
anticipation of a 1.3.1 release)


http://java.sun.com/j2se/1.5.0/docs/guide/jar/jar.html#JAR%20Manifest
http://java.sun.com/j2se/1.5.0/docs/guide/versioning/spec/versioning2.html
http://java.sun.com/j2se/1.5.0/docs/api/java/lang/Package.html
http://java.sun.com/j2se/1.5.0/docs/api/java/util/jar/package-summary.html
http://java.sun.com/developer/Books/javaprogramming/JAR/basics/manifest.html

: 
: 
: -- 
: "Good Enough" is not good enough.
: To give anything less than your best is to sacrifice the gift.
: Quality First. Measure Twice. Cut Once.
: 



-Hoss



[jira] Commented: (SOLR-1424) ant generate-maven-artifacts fails on windows

2009-09-11 Thread Hoss Man (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-1424?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12754450#action_12754450
 ] 

Hoss Man commented on SOLR-1424:


Hmmm i think the way the macro uses the pom.xml argument is the problem ... 
in some cases it can be an absolute path (ie: when you're on windows) so 
concating with ${maven.build.dir} in the copy task is a bad idea.

we could change the copy task to use a target dir instead of a target file ... 
except the artifact:pom task also needs the final filename.

Try changing the  call something like this...

{noformat}
Index: build.xml
===
--- build.xml   (revision 813985)
+++ build.xml   (working copy)
@@ -747,7 +747,7 @@
 
   
 
-  
+  
 
   
 
{noformat}

...and if that works, just document the hell out of the m2-deploy macro that 
the pom.xml arg must be a relative path.


> ant generate-maven-artifacts fails on windows
> -
>
> Key: SOLR-1424
> URL: https://issues.apache.org/jira/browse/SOLR-1424
> Project: Solr
>  Issue Type: Bug
>Reporter: Hoss Man
>
> From solr-user...
> {noformat}
> generate-maven-artifacts:
> [mkdir] Created dir: c:\Downloads\solr_trunk\build\maven
> [mkdir] Created dir: c:\Downloads\solr_trunk\dist\maven
>  [copy] Copying 1 file to
> c:\Downloads\solr_trunk\build\maven\c:\Downloads\s
> olr_trunk\src\maven
> BUILD FAILED
> c:\Downloads\solr_trunk\build.xml:741: The following error occurred while
> execut
> ing this line:
> c:\Downloads\solr_trunk\common-build.xml:261: Failed to copy
> c:\Downloads\solr_t
> runk\src\maven\solr-parent-pom.xml.template to
> c:\Downloads\solr_trunk\build\mav
> en\c:\Downloads\solr_trunk\src\maven\solr-parent-pom.xml.template due to
> java.io
> .FileNotFoundException
> c:\Downloads\solr_trunk\build\maven\c:\Downloads\solr_tru
> nk\src\maven\solr-parent-pom.xml.template (The filename, directory name, or
> volu
> me label syntax is incorrect)
> {noformat}

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



Re: svn commit: r808988 - in /lucene/solr/trunk: CHANGES.txt src/java/org/apache/solr/request/PHPSerializedResponseWriter.java

2009-09-11 Thread Chris Hostetter

: > if the container can't correctly output
: > some characters, i see no reason to hide the bug
: 
: Another problem is that it won't reliably break.  The bug breaks our
: encapsulation (before the patch) and thus the client reads the wrong
: number of chars for the string, and who knows what happens after that.
:  The majority of the time will result in an exception, but it really
: depends.  This is the type of stuff (buffer underflows / overflows)
: that could be used to mess with security too... a carefully crafted
: request could inject / change fields in the response and have it look
: valid.

Isn't that an argument in favor of having an explicit option to control 
how we do the counting? otherwise we're still at risk of the scenerio i 
discribed (ie: jetty fixes the byte conversion code, but we're still 
counting the bytes "wrong" for them)

with an explicit option we put control in the hands of the solr admin to 
decide what's right for them (we can even give them a little shell script 
to test it with)


-Hoss



[jira] Commented: (SOLR-1316) Create autosuggest component

2009-09-11 Thread Jason Rutherglen (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-1316?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12754443#action_12754443
 ] 

Jason Rutherglen commented on SOLR-1316:


Basically we need an algorithm that does suffix compression as well?

> Create autosuggest component
> 
>
> Key: SOLR-1316
> URL: https://issues.apache.org/jira/browse/SOLR-1316
> Project: Solr
>  Issue Type: New Feature
>  Components: search
>Affects Versions: 1.4
>Reporter: Jason Rutherglen
>Priority: Minor
> Fix For: 1.5
>
>   Original Estimate: 96h
>  Remaining Estimate: 96h
>
> Autosuggest is a common search function that can be integrated
> into Solr as a SearchComponent. Our first implementation will
> use the TernaryTree found in Lucene contrib. 
> * Enable creation of the dictionary from the index or via Solr's
> RPC mechanism
> * What types of parameters and settings are desirable?
> * Hopefully in the future we can include user click through
> rates to boost those terms/phrases higher

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (SOLR-1316) Create autosuggest component

2009-09-11 Thread Jason Rutherglen (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-1316?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12754440#action_12754440
 ] 

Jason Rutherglen commented on SOLR-1316:


Andrzej,

There's the ternary tree which is supposed to be better?  There are other 
algorithms for compressed dictionaries that could be used (off hand I can't 
think of them).  

> Create autosuggest component
> 
>
> Key: SOLR-1316
> URL: https://issues.apache.org/jira/browse/SOLR-1316
> Project: Solr
>  Issue Type: New Feature
>  Components: search
>Affects Versions: 1.4
>Reporter: Jason Rutherglen
>Priority: Minor
> Fix For: 1.5
>
>   Original Estimate: 96h
>  Remaining Estimate: 96h
>
> Autosuggest is a common search function that can be integrated
> into Solr as a SearchComponent. Our first implementation will
> use the TernaryTree found in Lucene contrib. 
> * Enable creation of the dictionary from the index or via Solr's
> RPC mechanism
> * What types of parameters and settings are desirable?
> * Hopefully in the future we can include user click through
> rates to boost those terms/phrases higher

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (SOLR-1316) Create autosuggest component

2009-09-11 Thread Andrzej Bialecki (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-1316?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12754404#action_12754404
 ] 

Andrzej Bialecki  commented on SOLR-1316:
-

Jason, did you make any progress on this? I'm interested in this 
functionality.. I'm not sure tries are the best choice, unless heavily pruned 
they occupy a lot of RAM space. I had some moderate success using ngram based 
method (I reused the spellchecker, with slight modifications) - the method is 
fast and reuses the existing spellchecker index, but precision of lookups is 
not ideal.

> Create autosuggest component
> 
>
> Key: SOLR-1316
> URL: https://issues.apache.org/jira/browse/SOLR-1316
> Project: Solr
>  Issue Type: New Feature
>  Components: search
>Affects Versions: 1.4
>Reporter: Jason Rutherglen
>Priority: Minor
> Fix For: 1.5
>
>   Original Estimate: 96h
>  Remaining Estimate: 96h
>
> Autosuggest is a common search function that can be integrated
> into Solr as a SearchComponent. Our first implementation will
> use the TernaryTree found in Lucene contrib. 
> * Enable creation of the dictionary from the index or via Solr's
> RPC mechanism
> * What types of parameters and settings are desirable?
> * Hopefully in the future we can include user click through
> rates to boost those terms/phrases higher

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



Re: svn commit: r808988 - in /lucene/solr/trunk: CHANGES.txt src/java/org/apache/solr/request/PHPSerializedResponseWriter.java

2009-09-11 Thread Donovan Jimenez
you are correct, it was my misunderstanding of the problem - now that  
I've read more than I ever wanted to know about UCS-2, UTF-16 and  
modified UTF-8, I'm more upto speed.


Thanks for the patience.

On Sep 11, 2009, at 6:32 PM, Yonik Seeley wrote:


On Fri, Sep 11, 2009 at 6:21 PM, Donovan Jimenez
 wrote:
Is it possible (and would it even help)  to normalize all strings  
with

regards to surrogate pairs at indexing time instead?


Already done, in a way... there's only one way to represent a
character outside the BMP in UTF-16 (which is the in-memory encoding
used by Java String).  Unless I misunderstood what you meant by
normalization.

-Yonik
http://www.lucidimagination.com




Re: svn commit: r808988 - in /lucene/solr/trunk: CHANGES.txt src/java/org/apache/solr/request/PHPSerializedResponseWriter.java

2009-09-11 Thread Yonik Seeley
On Tue, Sep 8, 2009 at 7:46 PM, Chris Hostetter
 wrote:
> if the container can't correctly output
> some characters, i see no reason to hide the bug

Another problem is that it won't reliably break.  The bug breaks our
encapsulation (before the patch) and thus the client reads the wrong
number of chars for the string, and who knows what happens after that.
 The majority of the time will result in an exception, but it really
depends.  This is the type of stuff (buffer underflows / overflows)
that could be used to mess with security too... a carefully crafted
request could inject / change fields in the response and have it look
valid.

-Yonik
http://www.lucidimagination.com


Re: svn commit: r808988 - in /lucene/solr/trunk: CHANGES.txt src/java/org/apache/solr/request/PHPSerializedResponseWriter.java

2009-09-11 Thread Yonik Seeley
On Fri, Sep 11, 2009 at 6:21 PM, Donovan Jimenez
 wrote:
> Is it possible (and would it even help)  to normalize all strings with
> regards to surrogate pairs at indexing time instead?

Already done, in a way... there's only one way to represent a
character outside the BMP in UTF-16 (which is the in-memory encoding
used by Java String).  Unless I misunderstood what you meant by
normalization.

-Yonik
http://www.lucidimagination.com


[jira] Commented: (SOLR-1424) ant generate-maven-artifacts fails on windows

2009-09-11 Thread Ryan McKinley (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-1424?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12754386#action_12754386
 ] 

Ryan McKinley commented on SOLR-1424:
-

no dice.  with:

{code}
$ svn diff
Index: common-build.xml
===
--- common-build.xml(revision 814053)
+++ common-build.xml(working copy)
@@ -104,13 +104,12 @@
   

   
-  
-  
+  
+  
+  

-  
-
{code}

I still get:
{code}
generate-maven-artifacts:
 [copy] Copying 1 file to 
c:\workspace\apache\solr\build\maven\c:\workspace\apache\solr\src\maven

BUILD FAILED
c:\workspace\apache\solr\build.xml:750: The following error occurred while 
executing this line:
c:\workspace\apache\solr\common-build.xml:260: Failed to copy 
c:\workspace\apache\solr\src\maven\solr-parent-pom.xml.template to 
c:\workspace\apache\solr\build\maven\c:\workspace\apache\solr\src\maven
\solr-parent-pom.xml.template due to java.io.FileNotFoundException 
c:\workspace\apache\solr\build\maven\c:\workspace\apache\solr\src\maven\solr-parent-pom.xml.template
 (The filename, directory name, o
r volume label syntax is incorrect)

Total time: 4 minutes 7 seconds
{code}

looking around at some other options, but don't see anything obvious

> ant generate-maven-artifacts fails on windows
> -
>
> Key: SOLR-1424
> URL: https://issues.apache.org/jira/browse/SOLR-1424
> Project: Solr
>  Issue Type: Bug
>Reporter: Hoss Man
>
> From solr-user...
> {noformat}
> generate-maven-artifacts:
> [mkdir] Created dir: c:\Downloads\solr_trunk\build\maven
> [mkdir] Created dir: c:\Downloads\solr_trunk\dist\maven
>  [copy] Copying 1 file to
> c:\Downloads\solr_trunk\build\maven\c:\Downloads\s
> olr_trunk\src\maven
> BUILD FAILED
> c:\Downloads\solr_trunk\build.xml:741: The following error occurred while
> execut
> ing this line:
> c:\Downloads\solr_trunk\common-build.xml:261: Failed to copy
> c:\Downloads\solr_t
> runk\src\maven\solr-parent-pom.xml.template to
> c:\Downloads\solr_trunk\build\mav
> en\c:\Downloads\solr_trunk\src\maven\solr-parent-pom.xml.template due to
> java.io
> .FileNotFoundException
> c:\Downloads\solr_trunk\build\maven\c:\Downloads\solr_tru
> nk\src\maven\solr-parent-pom.xml.template (The filename, directory name, or
> volu
> me label syntax is incorrect)
> {noformat}

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



Re: svn commit: r808988 - in /lucene/solr/trunk: CHANGES.txt src/java/org/apache/solr/request/PHPSerializedResponseWriter.java

2009-09-11 Thread Donovan Jimenez
Is it possible (and would it even help)  to normalize all strings  
with regards to surrogate pairs at indexing time instead?  or will  
container still possibly differ in byte for byte output?


- Donovan

On Sep 11, 2009, at 5:34 PM, Chris Hostetter wrote:



: > why don't we just output the raw bytes ourselves?
:
: That would require writing TextResponseWriter and friends as binary
: writers, right?  Or did you have a different way in mind for  
injecting

: bytes into the output stream?


Grr you're right. i got so turned arround thinking about
counting the bytes and encapsulating it all in the PHPS markup i  
forgot
that we still want/need the *raw* bytes output as part of the  
character

stream ... not some sort of string representation of the bytes ... wow
that sounds even more comfusing when i look at it on paper.

character data sucks.

I still repeat my earlier vote: let's yank this patch and just  
document
that byte counts for strings included in the PHPS format are only  
accurate
for characters outside the BMP if the servlet container being used  
writes
them correctly -- that seems like a totally fair requirement to  
having in
place, and points the figner squarly where in belongs (without  
putting us

at risk for having broken code if/when jetty fixes this problem.

Alternately: have an option and/or system property to force/disable  
this

behavior even if jetty.home isn't/is set.

Even if we change nothing else: there needs to be a big fat freaking
disclaimer in the javadocs for the writer that it's looking at the
jetty.home variable, and explaining why.



-Hoss





[jira] Created: (SOLR-1428) make ValueSourceParser and QParserPlugin implement SolrInfoMBean so people can use registry.jsp to see which ones are loaded

2009-09-11 Thread Hoss Man (JIRA)
make ValueSourceParser and QParserPlugin implement SolrInfoMBean so people can 
use registry.jsp to see which ones are loaded


 Key: SOLR-1428
 URL: https://issues.apache.org/jira/browse/SOLR-1428
 Project: Solr
  Issue Type: Improvement
Reporter: Hoss Man


there are a lot of default QParserPlugins and ValueSourceParsers loaded by 
default in solr -- but there is no clear way to see which ones are loaded.  if 
we made the abstract base classes implement SolrInfoMBean (with sane defaults 
for all the methods, or at the very least have them return strings like "INFO 
NOT AVAILABLE") then people could use the infoRegistry to see what's available 
-- both by default, and when they load their own.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Created: (SOLR-1427) SearchComponents aren't listed on registry.jsp

2009-09-11 Thread Hoss Man (JIRA)
SearchComponents aren't listed on registry.jsp
--

 Key: SOLR-1427
 URL: https://issues.apache.org/jira/browse/SOLR-1427
 Project: Solr
  Issue Type: Bug
Reporter: Hoss Man
 Fix For: 1.4


SearchComponent implements SolrInfoMBean using getCategory() of "OTHER" but 
they aren't listed on the registry.jsp display of loaded plugins.

This may be a one-of-glitch because of the way SearchComponents get loaded, or 
it may indicate some other problem with the infoRegistry.


-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Resolved: (SOLR-1380) Extend StatsComponent to MultiValued Fields

2009-09-11 Thread Grant Ingersoll (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-1380?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Grant Ingersoll resolved SOLR-1380.
---

Resolution: Fixed

Committed revision 814042.

> Extend StatsComponent to MultiValued Fields
> ---
>
> Key: SOLR-1380
> URL: https://issues.apache.org/jira/browse/SOLR-1380
> Project: Solr
>  Issue Type: Improvement
>  Components: search
>Affects Versions: 1.4
>Reporter: Harish Agarwal
>Assignee: Grant Ingersoll
>Priority: Minor
> Fix For: 1.4
>
> Attachments: SOLR-1380.patch, SOLR-1380.patch, SOLR-1380.patch, 
> SOLR-1380.patch, SOLR-1380.patch
>
>   Original Estimate: 0h
>  Remaining Estimate: 0h
>
> The StatsComponent does not work on multivalued fields.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (SOLR-1426) Allow delta-import to run continously until aborted

2009-09-11 Thread Abdul Chaudhry (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-1426?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Abdul Chaudhry updated SOLR-1426:
-

Description: 
Modify the delta-import so that it takes a perpetual flag that makes it run 
continuously until its aborted.

http://localhost:8985/solr/tickets/select/?command=delta-import&clean=false&qt=/dataimport&commit=true&perpetual=true

perpetual means the delta import will keep running and pause for a few seconds 
when running queries.

The only way to stop delta import will be to explicitly issue an abort like so:-

http://localhost:8985/solr/tickets/select/?command=abort


  was:
Modify the delta-import so that it takes a perpetual flag that makes it run 
continuously until its aborted.

http://search-master.fansnap.com:8985/solr/tickets/select/?command=delta-import&clean=false&qt=/dataimport&commit=true&perpetual=true

perpetual means the delta import will keep running and pause for a few seconds 
when running queries.

The only way to stop delta import will be to explicitly issue an abort like so:-

http://search-master.fansnap.com:8985/solr/tickets/select/?command=abort



> Allow delta-import to run continously until aborted
> ---
>
> Key: SOLR-1426
> URL: https://issues.apache.org/jira/browse/SOLR-1426
> Project: Solr
>  Issue Type: Improvement
>  Components: contrib - DataImportHandler
>Affects Versions: 1.4
>Reporter: Abdul Chaudhry
> Fix For: 1.4
>
> Attachments: delta-import-perpetual.patch
>
>
> Modify the delta-import so that it takes a perpetual flag that makes it run 
> continuously until its aborted.
> http://localhost:8985/solr/tickets/select/?command=delta-import&clean=false&qt=/dataimport&commit=true&perpetual=true
> perpetual means the delta import will keep running and pause for a few 
> seconds when running queries.
> The only way to stop delta import will be to explicitly issue an abort like 
> so:-
> http://localhost:8985/solr/tickets/select/?command=abort

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (SOLR-1426) Allow delta-import to run continously until aborted

2009-09-11 Thread Abdul Chaudhry (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-1426?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Abdul Chaudhry updated SOLR-1426:
-

Description: 
Modify the delta-import so that it takes a perpetual flag that makes it run 
continuously until its aborted.

http://localhost:8985/solr/select/?command=delta-import&clean=false&qt=/dataimport&commit=true&perpetual=true

perpetual means the delta import will keep running and pause for a few seconds 
when running queries.

The only way to stop delta import will be to explicitly issue an abort like so:-

http://localhost:8985/solr/tickets/select/?command=abort


  was:
Modify the delta-import so that it takes a perpetual flag that makes it run 
continuously until its aborted.

http://localhost:8985/solr/tickets/select/?command=delta-import&clean=false&qt=/dataimport&commit=true&perpetual=true

perpetual means the delta import will keep running and pause for a few seconds 
when running queries.

The only way to stop delta import will be to explicitly issue an abort like so:-

http://localhost:8985/solr/tickets/select/?command=abort



> Allow delta-import to run continously until aborted
> ---
>
> Key: SOLR-1426
> URL: https://issues.apache.org/jira/browse/SOLR-1426
> Project: Solr
>  Issue Type: Improvement
>  Components: contrib - DataImportHandler
>Affects Versions: 1.4
>Reporter: Abdul Chaudhry
> Fix For: 1.4
>
> Attachments: delta-import-perpetual.patch
>
>
> Modify the delta-import so that it takes a perpetual flag that makes it run 
> continuously until its aborted.
> http://localhost:8985/solr/select/?command=delta-import&clean=false&qt=/dataimport&commit=true&perpetual=true
> perpetual means the delta import will keep running and pause for a few 
> seconds when running queries.
> The only way to stop delta import will be to explicitly issue an abort like 
> so:-
> http://localhost:8985/solr/tickets/select/?command=abort

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (SOLR-1426) Allow delta-import to run continously until aborted

2009-09-11 Thread Abdul Chaudhry (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-1426?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Abdul Chaudhry updated SOLR-1426:
-

Attachment: delta-import-perpetual.patch

Uploaded a patch that implements this feature.

Ran all unit tests on my tree and they pass.

The only thing I have hard-coded is the sleep interval which is :-
Thread.sleep(3000)

This should probably be configurable.

> Allow delta-import to run continously until aborted
> ---
>
> Key: SOLR-1426
> URL: https://issues.apache.org/jira/browse/SOLR-1426
> Project: Solr
>  Issue Type: Improvement
>  Components: contrib - DataImportHandler
>Affects Versions: 1.4
>Reporter: Abdul Chaudhry
> Fix For: 1.4
>
> Attachments: delta-import-perpetual.patch
>
>
> Modify the delta-import so that it takes a perpetual flag that makes it run 
> continuously until its aborted.
> http://search-master.fansnap.com:8985/solr/tickets/select/?command=delta-import&clean=false&qt=/dataimport&commit=true&perpetual=true
> perpetual means the delta import will keep running and pause for a few 
> seconds when running queries.
> The only way to stop delta import will be to explicitly issue an abort like 
> so:-
> http://search-master.fansnap.com:8985/solr/tickets/select/?command=abort

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Created: (SOLR-1426) Allow delta-import to run continously until aborted

2009-09-11 Thread Abdul Chaudhry (JIRA)
Allow delta-import to run continously until aborted
---

 Key: SOLR-1426
 URL: https://issues.apache.org/jira/browse/SOLR-1426
 Project: Solr
  Issue Type: Improvement
  Components: contrib - DataImportHandler
Affects Versions: 1.4
Reporter: Abdul Chaudhry
 Fix For: 1.4


Modify the delta-import so that it takes a perpetual flag that makes it run 
continuously until its aborted.

http://search-master.fansnap.com:8985/solr/tickets/select/?command=delta-import&clean=false&qt=/dataimport&commit=true&perpetual=true

perpetual means the delta import will keep running and pause for a few seconds 
when running queries.

The only way to stop delta import will be to explicitly issue an abort like so:-

http://search-master.fansnap.com:8985/solr/tickets/select/?command=abort


-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



Re: svn commit: r808988 - in /lucene/solr/trunk: CHANGES.txt src/java/org/apache/solr/request/PHPSerializedResponseWriter.java

2009-09-11 Thread Chris Hostetter

: > why don't we just output the raw bytes ourselves?
: 
: That would require writing TextResponseWriter and friends as binary
: writers, right?  Or did you have a different way in mind for injecting
: bytes into the output stream?


Grr you're right. i got so turned arround thinking about 
counting the bytes and encapsulating it all in the PHPS markup i forgot 
that we still want/need the *raw* bytes output as part of the character 
stream ... not some sort of string representation of the bytes ... wow 
that sounds even more comfusing when i look at it on paper.

character data sucks.

I still repeat my earlier vote: let's yank this patch and just document 
that byte counts for strings included in the PHPS format are only accurate 
for characters outside the BMP if the servlet container being used writes 
them correctly -- that seems like a totally fair requirement to having in 
place, and points the figner squarly where in belongs (without putting us 
at risk for having broken code if/when jetty fixes this problem.

Alternately: have an option and/or system property to force/disable this 
behavior even if jetty.home isn't/is set.

Even if we change nothing else: there needs to be a big fat freaking 
disclaimer in the javadocs for the writer that it's looking at the 
jetty.home variable, and explaining why.



-Hoss



Re: svn commit: r808988 - in /lucene/solr/trunk: CHANGES.txt src/java/org/apache/solr/request/PHPSerializedResponseWriter.java

2009-09-11 Thread Yonik Seeley
On Fri, Sep 11, 2009 at 5:06 PM, Chris Hostetter
 wrote:
> why don't we just output the raw bytes ourselves?

That would require writing TextResponseWriter and friends as binary
writers, right?  Or did you have a different way in mind for injecting
bytes into the output stream?

-Yonik
http://www.lucidimagination.com


Re: svn commit: r808988 - in /lucene/solr/trunk: CHANGES.txt src/java/org/apache/solr/request/PHPSerializedResponseWriter.java

2009-09-11 Thread Robert Muir
On Fri, Sep 11, 2009 at 5:06 PM, Chris Hostetter
 wrote:

> I must be missunderstanding something still ... based on your description,
> it sounds like it shouldn't matter if the encoder knows that it's one
> logical character or not, either way it should wind up outputing the same
> number of bytes

yes, the # of bytes is different: 6 bytes versus 4 bytes in UTF-8

-- 
Robert Muir
rcm...@gmail.com


Re: svn commit: r808988 - in /lucene/solr/trunk: CHANGES.txt src/java/org/apache/solr/request/PHPSerializedResponseWriter.java

2009-09-11 Thread Chris Hostetter

: A code point (unicode character) outside of the BMP (basic
: multilingual plane, fits in 16 bits) is represented as two java chars
: - a surrogate pair.  It's a single logical character - see
: String.codePointAt().  In correct UTF-8 it should be encoded as a
: single code point... but Jetty is ignoring the fact that it's a
: surrogate pair and encoding each Java char as it's own code point...
: this is often called modified-UTF8 or CESU-8.
: 
: So... say you have this incorrect CESU-8 that is masquerading as
: UTF-8: all is not lost.
...

I must be missunderstanding something still ... based on your description, 
it sounds like it shouldn't matter if the encoder knows that it's one 
logical character or not, either way it should wind up outputing the same 
number of bytes

Except that if that were the case, why would we have had this bug in the 
first place?  clearly i'm still missunderstanding something.

: Bottom line - if we correctly encapsulate whatever the servlet
: container is writing, it's certainly possible for clients to use the
: output correctly.

I still come back to not liking that this is a hardcoded hack just for 
jetty ... it seems like it would be easy for a future version of jetty to 
change this behavior in some way that makes solr start breaking -- jetty 
could fix this bug and break solr's byte counting ... that doesn't seem 
cool

why don't we just output the raw bytes ourselves?  the code to generate 
the byte[] was/is allready there, we're just ignoring it and only using 
the length.



-Hoss



[jira] Commented: (SOLR-756) Make DisjunctionMaxQueryParser generally useful by supporting all query types.

2009-09-11 Thread Hoss Man (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-756?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12754337#action_12754337
 ] 

Hoss Man commented on SOLR-756:
---

at this stage in the process for 1.4, i think most committers are focusing on 
bug fixes and avoiding adding new features ... especially new features that 
have been added to existing functionality, so they could trigger new bugs in 
*existing* use cases.

FWIW: 
* the patch doesn't seem to have any test cases demonstrating the new 
functionality, which makes me leary of committing
* i'm not certain just from skimming the patch, but it seems like this should 
break behavior for any existing users of DisjunctionMaxQueryParser who are 
expecting the field aliases to only affect getFieldQuery ... i could be wrong 
about that though.


> Make DisjunctionMaxQueryParser generally useful by supporting all query types.
> --
>
> Key: SOLR-756
> URL: https://issues.apache.org/jira/browse/SOLR-756
> Project: Solr
>  Issue Type: Improvement
>Affects Versions: 1.3
>Reporter: David Smiley
> Fix For: 1.5
>
> Attachments: SolrPluginUtilsDisMax.patch
>
>
> This is an enhancement to the DisjunctionMaxQueryParser to work on all the 
> query variants such as wildcard, prefix, and fuzzy queries, and to support 
> working in "AND" scenarios that are not processed by the min-should-match 
> DisMax QParser. This was not in Solr already because DisMax was only used for 
> a very limited syntax that didn't use those features. In my opinion, this 
> makes a more suitable base parser for general use because unlike the 
> Lucene/Solr parser, this one supports multiple default fields whereas other 
> ones (say Yonik's {!prefix} one for example, can't do dismax). The notion of 
> a single default field is antiquated and a technical under-the-hood detail of 
> Lucene that I think Solr should shield the user from by on-the-fly using a 
> DisMax when multiple fields are used. 
> (patch to be attached soon)

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Created: (SOLR-1425) Better exception messages for classloader mishaps

2009-09-11 Thread Benson Margulies (JIRA)
Better exception messages for classloader mishaps
-

 Key: SOLR-1425
 URL: https://issues.apache.org/jira/browse/SOLR-1425
 Project: Solr
  Issue Type: New Feature
  Components: search
Affects Versions: 1.3
Reporter: Benson Margulies


If an idiot such as myself tries to specify a filter or such that lives in a 
parent classloader, such as the system classloader of a servlet container, Solr 
will fail to load it. The JVM is prone to create an exception that mentions 
some Solr interface as being missing instead of the filter itself. It would be 
less confusing for the miscreant if Solr were to try/catch ClassNotFound and 
NoClassDefError and throw its own exception with the name of the thing 
specified in the schema included in the message.



-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (SOLR-1419) Solr won't load filters from parent class loader, and the resulting error stacktrace is very confusing

2009-09-11 Thread Benson Margulies (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-1419?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12754324#action_12754324
 ] 

Benson Margulies commented on SOLR-1419:


I'll comment here and then open a feature request. Solr knows that it just 
called into the class loader to locate a class by name. So it could certainly 
include the name of what it was looking for in its exception, even if the VM 
chooses to describe as the source of the problem.



> Solr won't load filters from parent class loader, and the resulting error 
> stacktrace is very confusing
> --
>
> Key: SOLR-1419
> URL: https://issues.apache.org/jira/browse/SOLR-1419
> Project: Solr
>  Issue Type: Bug
>  Components: search
>Affects Versions: 1.3
>Reporter: Benson Margulies
>
> I specified a token filter class in my schema, and provided that class in a 
> jar file in the system classpath of my jetty instance instead of in 
> WEB-INF/lib of my solr webapp.
> This did not work.
> To make matters odder, the logged error did not mention my filter, but rather 
> an internal solr interface:
> Caused by: java.lang.ClassNotFoundException: 
> org.apache.solr.util.plugin.ResourceLoaderAware
>   at java.net.URLClassLoader$1.run(URLClassLoader.java:200)
>   at java.security.AccessController.doPrivileged(Native Method)
>   at java.net.URLClassLoader.findClass(URLClassLoader.java:188)
>   at java.lang.ClassLoader.loadClass(ClassLoader.java:319)
>   at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:330)
>   at java.lang.ClassLoader.loadClass(ClassLoader.java:254)
>   at java.lang.ClassLoader.loadClassInternal(ClassLoader.java:399)
> I note in passing that the token filter in turn uses other classes which 
> stayed happily behind in the outer classloader after moved the immediate 
> filter class into the webapp.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (SOLR-1419) Solr won't load filters from parent class loader, and the resulting error stacktrace is very confusing

2009-09-11 Thread Hoss Man (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-1419?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12754294#action_12754294
 ] 

Hoss Man commented on SOLR-1419:


bq. The error message issue struck me as JIRA-worthy even if the underlying 
behavior is by design. 

the behavior is by design -- of the VM.  Solr doesn't have anything to do with 
it, so Solr can't really improve the message produced either.

there *MAY* be some tricks Solr can do to recognize when a class load error was 
caused by another class load error for a class Solr already know about -- and 
then solr could generate an additional err message explaining the problem -- 
but that would certainly be a new feature, not a bug.


> Solr won't load filters from parent class loader, and the resulting error 
> stacktrace is very confusing
> --
>
> Key: SOLR-1419
> URL: https://issues.apache.org/jira/browse/SOLR-1419
> Project: Solr
>  Issue Type: Bug
>  Components: search
>Affects Versions: 1.3
>Reporter: Benson Margulies
>
> I specified a token filter class in my schema, and provided that class in a 
> jar file in the system classpath of my jetty instance instead of in 
> WEB-INF/lib of my solr webapp.
> This did not work.
> To make matters odder, the logged error did not mention my filter, but rather 
> an internal solr interface:
> Caused by: java.lang.ClassNotFoundException: 
> org.apache.solr.util.plugin.ResourceLoaderAware
>   at java.net.URLClassLoader$1.run(URLClassLoader.java:200)
>   at java.security.AccessController.doPrivileged(Native Method)
>   at java.net.URLClassLoader.findClass(URLClassLoader.java:188)
>   at java.lang.ClassLoader.loadClass(ClassLoader.java:319)
>   at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:330)
>   at java.lang.ClassLoader.loadClass(ClassLoader.java:254)
>   at java.lang.ClassLoader.loadClassInternal(ClassLoader.java:399)
> I note in passing that the token filter in turn uses other classes which 
> stayed happily behind in the outer classloader after moved the immediate 
> filter class into the webapp.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (SOLR-1424) ant generate-maven-artifacts fails on windows

2009-09-11 Thread Hoss Man (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-1424?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12754287#action_12754287
 ] 

Hoss Man commented on SOLR-1424:


I believe the error is just because maven directory properties are defined like 
this...

{code}
  
  
  

  
{code}

...but i'm pretty sure they're suppose to be defined like this...

{code}
  
  
  

  
{code}

...so that ANt recognizes that they're paths.

but i don't have a windows box, so i can't test that that fixes the problem


> ant generate-maven-artifacts fails on windows
> -
>
> Key: SOLR-1424
> URL: https://issues.apache.org/jira/browse/SOLR-1424
> Project: Solr
>  Issue Type: Bug
>Reporter: Hoss Man
>
> From solr-user...
> {noformat}
> generate-maven-artifacts:
> [mkdir] Created dir: c:\Downloads\solr_trunk\build\maven
> [mkdir] Created dir: c:\Downloads\solr_trunk\dist\maven
>  [copy] Copying 1 file to
> c:\Downloads\solr_trunk\build\maven\c:\Downloads\s
> olr_trunk\src\maven
> BUILD FAILED
> c:\Downloads\solr_trunk\build.xml:741: The following error occurred while
> execut
> ing this line:
> c:\Downloads\solr_trunk\common-build.xml:261: Failed to copy
> c:\Downloads\solr_t
> runk\src\maven\solr-parent-pom.xml.template to
> c:\Downloads\solr_trunk\build\mav
> en\c:\Downloads\solr_trunk\src\maven\solr-parent-pom.xml.template due to
> java.io
> .FileNotFoundException
> c:\Downloads\solr_trunk\build\maven\c:\Downloads\solr_tru
> nk\src\maven\solr-parent-pom.xml.template (The filename, directory name, or
> volu
> me label syntax is incorrect)
> {noformat}

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (SOLR-1380) Extend StatsComponent to MultiValued Fields

2009-09-11 Thread Harish Agarwal (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-1380?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harish Agarwal updated SOLR-1380:
-

Attachment: SOLR-1380.patch

Fix to distributed search, null stats.

> Extend StatsComponent to MultiValued Fields
> ---
>
> Key: SOLR-1380
> URL: https://issues.apache.org/jira/browse/SOLR-1380
> Project: Solr
>  Issue Type: Improvement
>  Components: search
>Affects Versions: 1.4
>Reporter: Harish Agarwal
>Assignee: Grant Ingersoll
>Priority: Minor
> Fix For: 1.4
>
> Attachments: SOLR-1380.patch, SOLR-1380.patch, SOLR-1380.patch, 
> SOLR-1380.patch, SOLR-1380.patch
>
>   Original Estimate: 0h
>  Remaining Estimate: 0h
>
> The StatsComponent does not work on multivalued fields.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Reopened: (SOLR-1380) Extend StatsComponent to MultiValued Fields

2009-09-11 Thread Harish Agarwal (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-1380?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harish Agarwal reopened SOLR-1380:
--


Noticed that distributed search was crashing for me in the StatsComponent.  I 
believe this is related to the fuzziness around when the returned stats values 
should be null for a field.  I'm submitting a small patch which fixes the bug, 
and always returns null for stats values for which the count is equal to zero 
in both distributed and non-distributed search.  My rationale behind this is 
that when the count is equal to zero, the min/max/mean values reported are 
garbage and should not be relied on as valid statistics for the field.  
Hopefully this can still be incorporated into the trunk before the 1.4 release.

> Extend StatsComponent to MultiValued Fields
> ---
>
> Key: SOLR-1380
> URL: https://issues.apache.org/jira/browse/SOLR-1380
> Project: Solr
>  Issue Type: Improvement
>  Components: search
>Affects Versions: 1.4
>Reporter: Harish Agarwal
>Assignee: Grant Ingersoll
>Priority: Minor
> Fix For: 1.4
>
> Attachments: SOLR-1380.patch, SOLR-1380.patch, SOLR-1380.patch, 
> SOLR-1380.patch
>
>   Original Estimate: 0h
>  Remaining Estimate: 0h
>
> The StatsComponent does not work on multivalued fields.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Created: (SOLR-1424) ant generate-maven-artifacts fails on windows

2009-09-11 Thread Hoss Man (JIRA)
ant generate-maven-artifacts fails on windows
-

 Key: SOLR-1424
 URL: https://issues.apache.org/jira/browse/SOLR-1424
 Project: Solr
  Issue Type: Bug
Reporter: Hoss Man


>From solr-user...

{noformat}
generate-maven-artifacts:
[mkdir] Created dir: c:\Downloads\solr_trunk\build\maven
[mkdir] Created dir: c:\Downloads\solr_trunk\dist\maven
 [copy] Copying 1 file to
c:\Downloads\solr_trunk\build\maven\c:\Downloads\s
olr_trunk\src\maven

BUILD FAILED
c:\Downloads\solr_trunk\build.xml:741: The following error occurred while
execut
ing this line:
c:\Downloads\solr_trunk\common-build.xml:261: Failed to copy
c:\Downloads\solr_t
runk\src\maven\solr-parent-pom.xml.template to
c:\Downloads\solr_trunk\build\mav
en\c:\Downloads\solr_trunk\src\maven\solr-parent-pom.xml.template due to
java.io
.FileNotFoundException
c:\Downloads\solr_trunk\build\maven\c:\Downloads\solr_tru
nk\src\maven\solr-parent-pom.xml.template (The filename, directory name, or
volu
me label syntax is incorrect)
{noformat}

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (SOLR-1301) Solr + Hadoop

2009-09-11 Thread jv ning (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-1301?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12754276#action_12754276
 ] 

jv ning commented on SOLR-1301:
---

I have an updated version that uses a sequence number, to ensure uniqueness
in the multiple output format case.

Same concept, shorter path.






> Solr + Hadoop
> -
>
> Key: SOLR-1301
> URL: https://issues.apache.org/jira/browse/SOLR-1301
> Project: Solr
>  Issue Type: Improvement
>Affects Versions: 1.4
>Reporter: Andrzej Bialecki 
> Attachments: hadoop-0.19.1-core.jar, hadoop.patch, SOLR-1301.patch, 
> SOLR-1301.patch, SOLR-1301.patch, solr.zip
>
>
> This patch contains  a contrib module that provides distributed indexing 
> (using Hadoop) to Solr EmbeddedSolrServer. The idea behind this module is 
> twofold:
> * provide an API that is familiar to Hadoop developers, i.e. that of 
> OutputFormat
> * avoid unnecessary export and (de)serialization of data maintained on HDFS. 
> SolrOutputFormat consumes data produced by reduce tasks directly, without 
> storing it in intermediate files. Furthermore, by using an 
> EmbeddedSolrServer, the indexing task is split into as many parts as there 
> are reducers, and the data to be indexed is not sent over the network.
> Design
> --
> Key/value pairs produced by reduce tasks are passed to SolrOutputFormat, 
> which in turn uses SolrRecordWriter to write this data. SolrRecordWriter 
> instantiates an EmbeddedSolrServer, and it also instantiates an 
> implementation of SolrDocumentConverter, which is responsible for turning 
> Hadoop (key, value) into a SolrInputDocument. This data is then added to a 
> batch, which is periodically submitted to EmbeddedSolrServer. When reduce 
> task completes, and the OutputFormat is closed, SolrRecordWriter calls 
> commit() and optimize() on the EmbeddedSolrServer.
> The API provides facilities to specify an arbitrary existing solr.home 
> directory, from which the conf/ and lib/ files will be taken.
> This process results in the creation of as many partial Solr home directories 
> as there were reduce tasks. The output shards are placed in the output 
> directory on the default filesystem (e.g. HDFS). Such part-N directories 
> can be used to run N shard servers. Additionally, users can specify the 
> number of reduce tasks, in particular 1 reduce task, in which case the output 
> will consist of a single shard.
> An example application is provided that processes large CSV files and uses 
> this API. It uses a custom CSV processing to avoid (de)serialization overhead.
> This patch relies on hadoop-core-0.19.1.jar - I attached the jar to this 
> issue, you should put it in contrib/hadoop/lib.
> Note: the development of this patch was sponsored by an anonymous contributor 
> and approved for release under Apache License.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (SOLR-1301) Solr + Hadoop

2009-09-11 Thread Kris Jirapinyo (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-1301?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kris Jirapinyo updated SOLR-1301:
-

Attachment: SOLR-1301.patch

Because we are using MultipleOutputFormat, we can't have the data directory be 
the task_id, as when the indexes were building, the directories were 
conflicting.  This patch uses a random UUID instead as the data directory so 
that if there are more than one shards being created under a reducer, the 
directories will not conflict.

> Solr + Hadoop
> -
>
> Key: SOLR-1301
> URL: https://issues.apache.org/jira/browse/SOLR-1301
> Project: Solr
>  Issue Type: Improvement
>Affects Versions: 1.4
>Reporter: Andrzej Bialecki 
> Attachments: hadoop-0.19.1-core.jar, hadoop.patch, SOLR-1301.patch, 
> SOLR-1301.patch, SOLR-1301.patch, solr.zip
>
>
> This patch contains  a contrib module that provides distributed indexing 
> (using Hadoop) to Solr EmbeddedSolrServer. The idea behind this module is 
> twofold:
> * provide an API that is familiar to Hadoop developers, i.e. that of 
> OutputFormat
> * avoid unnecessary export and (de)serialization of data maintained on HDFS. 
> SolrOutputFormat consumes data produced by reduce tasks directly, without 
> storing it in intermediate files. Furthermore, by using an 
> EmbeddedSolrServer, the indexing task is split into as many parts as there 
> are reducers, and the data to be indexed is not sent over the network.
> Design
> --
> Key/value pairs produced by reduce tasks are passed to SolrOutputFormat, 
> which in turn uses SolrRecordWriter to write this data. SolrRecordWriter 
> instantiates an EmbeddedSolrServer, and it also instantiates an 
> implementation of SolrDocumentConverter, which is responsible for turning 
> Hadoop (key, value) into a SolrInputDocument. This data is then added to a 
> batch, which is periodically submitted to EmbeddedSolrServer. When reduce 
> task completes, and the OutputFormat is closed, SolrRecordWriter calls 
> commit() and optimize() on the EmbeddedSolrServer.
> The API provides facilities to specify an arbitrary existing solr.home 
> directory, from which the conf/ and lib/ files will be taken.
> This process results in the creation of as many partial Solr home directories 
> as there were reduce tasks. The output shards are placed in the output 
> directory on the default filesystem (e.g. HDFS). Such part-N directories 
> can be used to run N shard servers. Additionally, users can specify the 
> number of reduce tasks, in particular 1 reduce task, in which case the output 
> will consist of a single shard.
> An example application is provided that processes large CSV files and uses 
> this API. It uses a custom CSV processing to avoid (de)serialization overhead.
> This patch relies on hadoop-core-0.19.1.jar - I attached the jar to this 
> issue, you should put it in contrib/hadoop/lib.
> Note: the development of this patch was sponsored by an anonymous contributor 
> and approved for release under Apache License.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (SOLR-1395) Integrate Katta

2009-09-11 Thread Jason Rutherglen (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-1395?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jason Rutherglen updated SOLR-1395:
---

Attachment: hadoop-core-0.19.0.jar
log4j-1.2.13.jar
zookeeper-3.2.1.jar

These are the external libraries necessary to run the test

> Integrate Katta
> ---
>
> Key: SOLR-1395
> URL: https://issues.apache.org/jira/browse/SOLR-1395
> Project: Solr
>  Issue Type: New Feature
>Affects Versions: 1.4
>Reporter: Jason Rutherglen
>Priority: Minor
> Fix For: 1.5
>
> Attachments: hadoop-core-0.19.0.jar, KATTA-SOLR.patch, 
> log4j-1.2.13.jar, SOLR-1395.patch, zookeeper-3.2.1.jar
>
>   Original Estimate: 336h
>  Remaining Estimate: 336h
>
> We'll integrate Katta into Solr so that:
> * Distributed search uses Hadoop RPC
> * Shard/SolrCore distribution and management
> * Zookeeper based failover
> * Indexes may be built using Hadoop

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



Spec Version vs Implementation Version

2009-09-11 Thread Israel Ekpo
What are the differences between specification version and implementation
version

I downloaded the nightly build for September 05 2009 and it has a spec
version of 1.3 and the implementation version states 1.4-dev

What does that mean?


-- 
"Good Enough" is not good enough.
To give anything less than your best is to sacrifice the gift.
Quality First. Measure Twice. Cut Once.


[jira] Commented: (SOLR-1395) Integrate Katta

2009-09-11 Thread Jason Rutherglen (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-1395?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12754241#action_12754241
 ] 

Jason Rutherglen commented on SOLR-1395:


I added a wiki page at: http://wiki.apache.org/solr/KattaIntegration

> Integrate Katta
> ---
>
> Key: SOLR-1395
> URL: https://issues.apache.org/jira/browse/SOLR-1395
> Project: Solr
>  Issue Type: New Feature
>Affects Versions: 1.4
>Reporter: Jason Rutherglen
>Priority: Minor
> Fix For: 1.5
>
> Attachments: KATTA-SOLR.patch, SOLR-1395.patch
>
>   Original Estimate: 336h
>  Remaining Estimate: 336h
>
> We'll integrate Katta into Solr so that:
> * Distributed search uses Hadoop RPC
> * Shard/SolrCore distribution and management
> * Zookeeper based failover
> * Indexes may be built using Hadoop

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (SOLR-1394) HTML stripper is splitting tokens

2009-09-11 Thread Jason Rutherglen (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-1394?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12754215#action_12754215
 ] 

Jason Rutherglen commented on SOLR-1394:


bq. In which situation do you get this exception? 

We need to log the HTML.  I'll post it when we implement the logging.

> HTML stripper is splitting tokens
> -
>
> Key: SOLR-1394
> URL: https://issues.apache.org/jira/browse/SOLR-1394
> Project: Solr
>  Issue Type: Bug
>  Components: Analysis
>Affects Versions: 1.4
>Reporter: Anders Melchiorsen
> Attachments: SOLR-1394.patch
>
>
> I am having problems with the Solr HTML stripper.
> After some investigation, I have found the cause to be that the
> stripper is replacing the removed HTML with spaces. This obviously
> breaks when the HTML is in the middle of a word, like "Günther".
> So, without knowing what I was doing, I hacked together a fix that
> uses offset correction instead.
> That seemed to work, except that closing tags and attributes still
> broke the positioning. With even less of a clue, I replaced read()
> with next() in the two methods handling those.
> Finally, invalid HTML also gave wrong offsets, and I fixed that by
> restoring numRead when rolling back the input stream.
> At this point I stopped trying to break it, so there may still be more
> problems. Or I might have introduced some problem on my own. Anyway, I
> have put the three patches at the bottom of this mail, in case
> somebody wants to move along with this issue.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (SOLR-236) Field collapsing

2009-09-11 Thread Oleg Gnatovskiy (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-236?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12754212#action_12754212
 ] 

Oleg Gnatovskiy commented on SOLR-236:
--

Hey Martijn,
Have you made any progress on making field collapsing distributed?
Oleg

> Field collapsing
> 
>
> Key: SOLR-236
> URL: https://issues.apache.org/jira/browse/SOLR-236
> Project: Solr
>  Issue Type: New Feature
>  Components: search
>Affects Versions: 1.3
>Reporter: Emmanuel Keller
> Fix For: 1.5
>
> Attachments: collapsing-patch-to-1.3.0-dieter.patch, 
> collapsing-patch-to-1.3.0-ivan.patch, collapsing-patch-to-1.3.0-ivan_2.patch, 
> collapsing-patch-to-1.3.0-ivan_3.patch, field-collapse-3.patch, 
> field-collapse-4-with-solrj.patch, field-collapse-5.patch, 
> field-collapse-5.patch, field-collapse-solr-236-2.patch, 
> field-collapse-solr-236.patch, field-collapsing-extended-592129.patch, 
> field_collapsing_1.1.0.patch, field_collapsing_1.3.patch, 
> field_collapsing_dsteigerwald.diff, field_collapsing_dsteigerwald.diff, 
> field_collapsing_dsteigerwald.diff, SOLR-236-FieldCollapsing.patch, 
> SOLR-236-FieldCollapsing.patch, SOLR-236-FieldCollapsing.patch, 
> solr-236.patch, SOLR-236_collapsing.patch, SOLR-236_collapsing.patch
>
>
> This patch include a new feature called "Field collapsing".
> "Used in order to collapse a group of results with similar value for a given 
> field to a single entry in the result set. Site collapsing is a special case 
> of this, where all results for a given web site is collapsed into one or two 
> entries in the result set, typically with an associated "more documents from 
> this site" link. See also Duplicate detection."
> http://www.fastsearch.com/glossary.aspx?m=48&amid=299
> The implementation add 3 new query parameters (SolrParams):
> "collapse.field" to choose the field used to group results
> "collapse.type" normal (default value) or adjacent
> "collapse.max" to select how many continuous results are allowed before 
> collapsing
> TODO (in progress):
> - More documentation (on source code)
> - Test cases
> Two patches:
> - "field_collapsing.patch" for current development version
> - "field_collapsing_1.1.0.patch" for Solr-1.1.0
> P.S.: Feedback and misspelling correction are welcome ;-)

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (SOLR-1292) show lucene fieldcache entries and sizes

2009-09-11 Thread Grant Ingersoll (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-1292?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12754188#action_12754188
 ] 

Grant Ingersoll commented on SOLR-1292:
---

Looks like a reasonable first pass, Hoss.  I think we should commit and then we 
can iterate in 1.5.

> show lucene fieldcache entries and sizes
> 
>
> Key: SOLR-1292
> URL: https://issues.apache.org/jira/browse/SOLR-1292
> Project: Solr
>  Issue Type: Improvement
>Reporter: Yonik Seeley
>Assignee: Mark Miller
>Priority: Minor
> Fix For: 1.4
>
> Attachments: SOLR-1292.patch
>
>
> See LUCENE-1749, FieldCache introspection API

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (SOLR-236) Field collapsing

2009-09-11 Thread Paul Nelson (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-236?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12754182#action_12754182
 ] 

Paul Nelson commented on SOLR-236:
--

Thanks Martijn!

Also, while I was doing testing on collapse, I've noticed some threading issues 
as well. I think they are primarily centered around the collapseRequest field.

Specifically, when I run two collapse queries at the same time, I get the 
following exception:

{code}
java.lang.IllegalStateException: Invoke the collapse method before invoking 
getCollapseInfo method
at 
org.apache.solr.search.AbstractDocumentCollapser.getCollapseInfo(AbstractDocumentCollapser.java:183)
at 
org.apache.solr.handler.component.CollapseComponent.doProcess(CollapseComponent.java:115)
at 
org.apache.solr.handler.component.CollapseComponent.process(CollapseComponent.java:67)
at 
org.apache.solr.handler.component.SearchHandler.handleRequestBody(SearchHandler.java:195)
at 
org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:131)
at org.apache.solr.core.SolrCore.execute(SolrCore.java:1299)
at 
org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:338)
at 
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:241)
at 
org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:235)
at 
org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:206)
at 
org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:233)
at 
org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:191)
at 
org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:128)
at 
org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:102)
at 
org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:109)
at 
org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:293)
at 
org.apache.coyote.http11.Http11Processor.process(Http11Processor.java:849)
at 
org.apache.coyote.http11.Http11Protocol$Http11ConnectionHandler.process(Http11Protocol.java:583)
at 
org.apache.tomcat.util.net.JIoEndpoint$Worker.run(JIoEndpoint.java:454)
at java.lang.Thread.run(Thread.java:619)
{code}

And when I run a second (non-collapsing) query at the same time I run the 
collapse query I get this exception:

{code}
java.lang.NullPointerException
at 
org.apache.solr.handler.component.CollapseComponent.doProcess(CollapseComponent.java:109)
at 
org.apache.solr.handler.component.CollapseComponent.process(CollapseComponent.java:67)
at 
org.apache.solr.handler.component.SearchHandler.handleRequestBody(SearchHandler.java:195)
at 
org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:131)
at org.apache.solr.core.SolrCore.execute(SolrCore.java:1299)
at 
org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:338)
at 
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:241)
at 
org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:235)
at 
org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:206)
at 
org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:233)
at 
org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:191)
at 
org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:128)
at 
org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:102)
at 
org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:109)
at 
org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:293)
at 
org.apache.coyote.http11.Http11Processor.process(Http11Processor.java:849)
at 
org.apache.coyote.http11.Http11Protocol$Http11ConnectionHandler.process(Http11Protocol.java:583)
at 
org.apache.tomcat.util.net.JIoEndpoint$Worker.run(JIoEndpoint.java:454)
at java.lang.Thread.run(Thread.java:619)
{code}

These errors occurred with the 2009-08-24 patch, but (upon brief inspection) it 
looks like the same situation would occur with the latest patch.

If I get the chance, I'll try and debug further.

> Field collapsing
> 
>
> Key: SOLR-236
> URL: https://issues.apache.org/jira/browse/SOLR-236
> Project: Solr
>  Issue Type: New Feature
>  Components: search
>Affects Versions: 1.3
>Reporter: Emmanuel Keller
> Fix For: 1.5
>
> Attachments: collapsing-patch-to-1.3.0-dieter.patch, 
> c

[jira] Resolved: (SOLR-1380) Extend StatsComponent to MultiValued Fields

2009-09-11 Thread Grant Ingersoll (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-1380?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Grant Ingersoll resolved SOLR-1380.
---

Resolution: Fixed

Committed revision 813874.

Thanks Harish!

> Extend StatsComponent to MultiValued Fields
> ---
>
> Key: SOLR-1380
> URL: https://issues.apache.org/jira/browse/SOLR-1380
> Project: Solr
>  Issue Type: Improvement
>  Components: search
>Affects Versions: 1.4
>Reporter: Harish Agarwal
>Assignee: Grant Ingersoll
>Priority: Minor
> Fix For: 1.4
>
> Attachments: SOLR-1380.patch, SOLR-1380.patch, SOLR-1380.patch, 
> SOLR-1380.patch
>
>   Original Estimate: 0h
>  Remaining Estimate: 0h
>
> The StatsComponent does not work on multivalued fields.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



Re: solr 1.4 release schedule

2009-09-11 Thread Grant Ingersoll

https://issues.apache.org/jira/secure/BrowseVersion.jspa?id=12310230&versionId=12313351&showOpenIssuesOnly=true

We are down to 14 open.


On Sep 8, 2009, at 1:41 PM, Grant Ingersoll wrote:


I can be RM unless someone else is dying to do it.

On Sep 8, 2009, at 11:06 AM, Yonik Seeley wrote:


My day-job colleagues and I are all traveling this week (company
get-together) so that may slow things down a bit for some of us, and
perhaps cause the goal of releasing Solr 1 week after Lucene to  
slip a
little.  Still, if there are any issues that are assigned to you,  
and

that you can't get to this week (including the weekend) please
un-assign yourself as a signal that someone else should try and take
it up.


[moving release discussion to solr-dev]

OK... how about going into code freeze by the end of the week?
Again, if you have an issue that you won't get to, please un-assign  
yourself.


-Yonik
http://www.lucidimagination.com








[jira] Commented: (SOLR-1404) Random failures with highlighting

2009-09-11 Thread Uwe Schindler (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-1404?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12754162#action_12754162
 ] 

Uwe Schindler commented on SOLR-1404:
-

bq. Will LUCENE-1906 fix it (in an alternate way)?

It should fix it. Lucene Tokenizer now do not have separate methods for 
CharStream anymore. They are simply handled as Readers. The trap of overwriting 
the wrong method should be fixed now. The offset correction is now done 
conditionally if the Reader is a CharStream subclass.

> Random failures with highlighting
> -
>
> Key: SOLR-1404
> URL: https://issues.apache.org/jira/browse/SOLR-1404
> Project: Solr
>  Issue Type: Bug
>  Components: Analysis, highlighter
>Affects Versions: 1.4
>Reporter: Anders Melchiorsen
> Fix For: 1.4
>
> Attachments: SOLR-1404.patch
>
>
> With a recent Solr nightly, we started getting errors when highlighting.
> I have not been able to reduce our real setup to a minimal one that is 
> failing, but the same error seems to pop up with the configuration below. 
> Note that the QUERY will mostly fail, but it will work sometimes. Notably, 
> after running "java -jar start.jar", the QUERY will work the first time, but 
> then start failing for a while. Seems that something is not being reset 
> properly.
> The example uses the deprecated HTMLStripWhitespaceTokenizerFactory but the 
> problem apparently also exists with other tokenizers; I was just unable to 
> create a minimal example with other configurations.
> SCHEMA
> 
> 
>   
> 
> 
>   
> 
>   
> 
>  
>  
>
>
>  
>  id
> 
> INDEX
> URL=http://localhost:8983/solr/update
> curl $URL --data-binary '1 name="test">test' -H 'Content-type:text/xml; 
> charset=utf-8'
> curl $URL --data-binary '' -H 'Content-type:text/xml; charset=utf-8'
> QUERY
> curl 'http://localhost:8983/solr/select/?hl.fl=test&hl=true&q=id:1'
> ERROR
> org.apache.lucene.search.highlight.InvalidTokenOffsetsException: Token test 
> exceeds length of provided text sized 4
> org.apache.solr.common.SolrException: 
> org.apache.lucene.search.highlight.InvalidTokenOffsetsException: Token test 
> exceeds length of provided text sized 4
>   at 
> org.apache.solr.highlight.DefaultSolrHighlighter.doHighlighting(DefaultSolrHighlighter.java:328)
>   at 
> org.apache.solr.handler.component.HighlightComponent.process(HighlightComponent.java:89)
>   at 
> org.apache.solr.handler.component.SearchHandler.handleRequestBody(SearchHandler.java:195)
>   at 
> org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:131)
>   at org.apache.solr.core.SolrCore.execute(SolrCore.java:1299)
>   at 
> org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:338)
>   at 
> org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:241)
>   at 
> org.mortbay.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1089)
>   at 
> org.mortbay.jetty.servlet.ServletHandler.handle(ServletHandler.java:365)
>   at 
> org.mortbay.jetty.security.SecurityHandler.handle(SecurityHandler.java:216)
>   at 
> org.mortbay.jetty.servlet.SessionHandler.handle(SessionHandler.java:181)
>   at 
> org.mortbay.jetty.handler.ContextHandler.handle(ContextHandler.java:712)
>   at org.mortbay.jetty.webapp.WebAppContext.handle(WebAppContext.java:405)
>   at 
> org.mortbay.jetty.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:211)
>   at 
> org.mortbay.jetty.handler.HandlerCollection.handle(HandlerCollection.java:114)
>   at 
> org.mortbay.jetty.handler.HandlerWrapper.handle(HandlerWrapper.java:139)
>   at org.mortbay.jetty.Server.handle(Server.java:285)
>   at 
> org.mortbay.jetty.HttpConnection.handleRequest(HttpConnection.java:502)
>   at 
> org.mortbay.jetty.HttpConnection$RequestHandler.headerComplete(HttpConnection.java:821)
>   at org.mortbay.jetty.HttpParser.parseNext(HttpParser.java:513)
>   at org.mortbay.jetty.HttpParser.parseAvailable(HttpParser.java:208)
>   at org.mortbay.jetty.HttpConnection.handle(HttpConnection.java:378)
>   at 
> org.mortbay.jetty.bio.SocketConnector$Connection.run(SocketConnector.java:226)
>   at 
> org.mortbay.thread.BoundedThreadPool$PoolThread.run(BoundedThreadPool.java:442)
> Caused by: org.apache.lucene.search.highlight.InvalidTokenOffsetsException: 
> Token test exceeds length of provided text sized 4
>   at 
> org.apache.lucene.search.highlight.Highlighter.getBestTextFragments(Highlighter.java:254)
>   at 
> org.apache.solr.highlight.DefaultSolrHighlighter.doHighlighting(DefaultSolrHighlighter.java:321)
>   ... 23 more

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the i

[jira] Resolved: (SOLR-1411) SolrJ ContentStream Update Request

2009-09-11 Thread Grant Ingersoll (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-1411?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Grant Ingersoll resolved SOLR-1411.
---

Resolution: Fixed

Committed revision 813860.

> SolrJ ContentStream Update Request
> --
>
> Key: SOLR-1411
> URL: https://issues.apache.org/jira/browse/SOLR-1411
> Project: Solr
>  Issue Type: Improvement
>Reporter: Grant Ingersoll
>Assignee: Grant Ingersoll
>Priority: Minor
> Fix For: 1.4
>
> Attachments: SOLR-1411-build.xml, SOLR-1411.patch, SOLR-1411.patch, 
> SOLR-1411.patch
>
>
> Create a SolrRequest for SolrJ that can add content streams to Solr for 
> indexing.
> Patch shortly.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



Re: Solr nightly build failure

2009-09-11 Thread Robert Muir
thanks, this patch fixes the build issue

On Fri, Sep 11, 2009 at 10:14 AM, Grant Ingersoll  wrote:
> Please try https://issues.apache.org/jira/browse/SOLR-1411
>  SOLR-1411-build.patch

-- 
Robert Muir
rcm...@gmail.com


[jira] Updated: (SOLR-1423) Lucene 2.9 RC4 may need some changes in Solr Analyzers using CharStream & others

2009-09-11 Thread Koji Sekiguchi (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-1423?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Koji Sekiguchi updated SOLR-1423:
-

Affects Version/s: 1.4
Fix Version/s: 1.4
 Assignee: Koji Sekiguchi

I'd like to check it before 1.4 release. I'll look into it once RC4 is checked 
in Solr.

> Lucene 2.9 RC4 may need some changes in Solr Analyzers using CharStream & 
> others
> 
>
> Key: SOLR-1423
> URL: https://issues.apache.org/jira/browse/SOLR-1423
> Project: Solr
>  Issue Type: Task
>  Components: Analysis
>Affects Versions: 1.4
>Reporter: Uwe Schindler
>Assignee: Koji Sekiguchi
> Fix For: 1.4
>
>
> Because of some backwards compatibility problems (LUCENE-1906) we changed the 
> CharStream/CharFilter API a little bit. Tokenizer now only has a input field 
> of type java.io.Reader (as before the CharStream code). To correct offsets, 
> it is now needed to call the Tokenizer.correctOffset(int) method, which 
> delegates to the CharStream (if input is subclass of CharStream), else 
> returns an uncorrected offset. Normally it is enough to change all occurences 
> of input.correctOffset() to this.correctOffset() in Tokenizers. It should 
> also be checked, if custom Tokenizers in Solr do correct their offsets.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Resolved: (SOLR-1203) We should add an example of setting the update.processor for a given RequestHandler

2009-09-11 Thread Mark Miller (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-1203?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mark Miller resolved SOLR-1203.
---

Resolution: Fixed

thanks for the review - committed r813854

> We should add an example of setting the update.processor for a given 
> RequestHandler
> ---
>
> Key: SOLR-1203
> URL: https://issues.apache.org/jira/browse/SOLR-1203
> Project: Solr
>  Issue Type: Improvement
>Reporter: Mark Miller
>Assignee: Mark Miller
>Priority: Minor
> Fix For: 1.4
>
> Attachments: SOLR-1203.patch
>
>
> a commented out example that points to the commented out example update chain 
> or just as good: a comment above the current update chain example explaining 
> how to attach it to a handler.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



Re: Solr nightly build failure

2009-09-11 Thread Grant Ingersoll
Please try https://issues.apache.org/jira/browse/SOLR-1411  SOLR-1411- 
build.patch


On Sep 11, 2009, at 9:34 AM, Grant Ingersoll wrote:



On Sep 11, 2009, at 9:23 AM, Grant Ingersoll wrote:


I bet it is due to ant example not being run first.


I confirm that's the issue.  Argh.  Don't really want to make people  
run example first (because it is slow) but if the example is going  
to include /update/extract, then we need to make sure the libs are  
there.  I suppose we could run just the Solr Cell example target.








[jira] Reopened: (SOLR-1411) SolrJ ContentStream Update Request

2009-09-11 Thread Grant Ingersoll (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-1411?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Grant Ingersoll reopened SOLR-1411:
---


> SolrJ ContentStream Update Request
> --
>
> Key: SOLR-1411
> URL: https://issues.apache.org/jira/browse/SOLR-1411
> Project: Solr
>  Issue Type: Improvement
>Reporter: Grant Ingersoll
>Assignee: Grant Ingersoll
>Priority: Minor
> Fix For: 1.4
>
> Attachments: SOLR-1411-build.xml, SOLR-1411.patch, SOLR-1411.patch, 
> SOLR-1411.patch
>
>
> Create a SolrRequest for SolrJ that can add content streams to Solr for 
> indexing.
> Patch shortly.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (SOLR-1411) SolrJ ContentStream Update Request

2009-09-11 Thread Grant Ingersoll (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-1411?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Grant Ingersoll updated SOLR-1411:
--

Attachment: SOLR-1411-build.xml

Fixes the build issue

> SolrJ ContentStream Update Request
> --
>
> Key: SOLR-1411
> URL: https://issues.apache.org/jira/browse/SOLR-1411
> Project: Solr
>  Issue Type: Improvement
>Reporter: Grant Ingersoll
>Assignee: Grant Ingersoll
>Priority: Minor
> Fix For: 1.4
>
> Attachments: SOLR-1411-build.xml, SOLR-1411.patch, SOLR-1411.patch, 
> SOLR-1411.patch
>
>
> Create a SolrRequest for SolrJ that can add content streams to Solr for 
> indexing.
> Patch shortly.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Resolved: (SOLR-1321) Support for efficient leading wildcards search

2009-09-11 Thread Grant Ingersoll (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-1321?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Grant Ingersoll resolved SOLR-1321.
---

Resolution: Fixed

Committed revision 813830.

> Support for efficient leading wildcards search
> --
>
> Key: SOLR-1321
> URL: https://issues.apache.org/jira/browse/SOLR-1321
> Project: Solr
>  Issue Type: Improvement
>  Components: Analysis
>Affects Versions: 1.4
>Reporter: Andrzej Bialecki 
>Assignee: Grant Ingersoll
> Fix For: 1.4
>
> Attachments: SOLR-1321.patch, SOLR-1321.patch, SOLR-1321.patch, 
> wildcards-2.patch, wildcards-3.patch, wildcards.patch
>
>
> This patch is an implementation of the "reversed tokens" strategy for 
> efficient leading wildcards queries.
> ReversedWildcardsTokenFilter reverses tokens and returns both the original 
> token (optional) and the reversed token (with positionIncrement == 0). 
> Reversed tokens are prepended with a marker character to avoid collisions 
> between legitimate tokens and the reversed tokens - e.g. "DNA" would become 
> "and", thus colliding with the regular term "and", but with the marker 
> character it becomes "\u0001and".
> This TokenFilter can be added to the analyzer chain that it used during 
> indexing.
> SolrQueryParser has been modified to detect the presence of such fields in 
> the current schema, and treat them in a special way. First, SolrQueryParser 
> examines the schema and collects a map of fields where these reversed tokens 
> are indexed. If there is at least one such field, it also sets 
> QueryParser.setAllowLeadingWildcards(true). When building a wildcard query 
> (in getWildcardQuery) the term text may be optionally reversed to put 
> wildcards further along the term text. This happens when the field uses the 
> reversing filter during indexing (as detected above), AND if the wildcard 
> characters are either at 0-th or 1-st position in the term. Otherwise the 
> term text is processed as before, i.e. turned into a regular wildcard query.
> Unit tests are provided to test the TokenFilter and the query parsing.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (SOLR-1321) Support for efficient leading wildcards search

2009-09-11 Thread Grant Ingersoll (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-1321?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12754124#action_12754124
 ] 

Grant Ingersoll commented on SOLR-1321:
---

Only annoying thing about this solution is now that * check gets done twice, 
once in the SolrQueryParser method and once in the QueryParser method. 

Also, why not just && the two clauses (I realize it is a cut and paste from the 
parent).

I'll fix and commit.

> Support for efficient leading wildcards search
> --
>
> Key: SOLR-1321
> URL: https://issues.apache.org/jira/browse/SOLR-1321
> Project: Solr
>  Issue Type: Improvement
>  Components: Analysis
>Affects Versions: 1.4
>Reporter: Andrzej Bialecki 
>Assignee: Grant Ingersoll
> Fix For: 1.4
>
> Attachments: SOLR-1321.patch, SOLR-1321.patch, SOLR-1321.patch, 
> wildcards-2.patch, wildcards-3.patch, wildcards.patch
>
>
> This patch is an implementation of the "reversed tokens" strategy for 
> efficient leading wildcards queries.
> ReversedWildcardsTokenFilter reverses tokens and returns both the original 
> token (optional) and the reversed token (with positionIncrement == 0). 
> Reversed tokens are prepended with a marker character to avoid collisions 
> between legitimate tokens and the reversed tokens - e.g. "DNA" would become 
> "and", thus colliding with the regular term "and", but with the marker 
> character it becomes "\u0001and".
> This TokenFilter can be added to the analyzer chain that it used during 
> indexing.
> SolrQueryParser has been modified to detect the presence of such fields in 
> the current schema, and treat them in a special way. First, SolrQueryParser 
> examines the schema and collects a map of fields where these reversed tokens 
> are indexed. If there is at least one such field, it also sets 
> QueryParser.setAllowLeadingWildcards(true). When building a wildcard query 
> (in getWildcardQuery) the term text may be optionally reversed to put 
> wildcards further along the term text. This happens when the field uses the 
> reversing filter during indexing (as detected above), AND if the wildcard 
> characters are either at 0-th or 1-st position in the term. Otherwise the 
> term text is processed as before, i.e. turned into a regular wildcard query.
> Unit tests are provided to test the TokenFilter and the query parsing.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



Re: Solr nightly build failure

2009-09-11 Thread Grant Ingersoll


On Sep 11, 2009, at 9:23 AM, Grant Ingersoll wrote:


I bet it is due to ant example not being run first.


I confirm that's the issue.  Argh.  Don't really want to make people  
run example first (because it is slow) but if the example is going to  
include /update/extract, then we need to make sure the libs are  
there.  I suppose we could run just the Solr Cell example target.





Re: Solr nightly build failure

2009-09-11 Thread Grant Ingersoll

I bet it is due to ant example not being run first.

On Sep 11, 2009, at 9:02 AM, Robert Muir wrote:


i see this with a clean checkout and ant test

On Fri, Sep 11, 2009 at 8:37 AM, Grant Ingersoll  
 wrote:

Hmm, ant clean test passes for me.


On Sep 11, 2009, at 8:25 AM, Grant Ingersoll wrote:

Hmm, this looks like my commit.  I'll check into it.  Anyone else  
seeing

failures?

On Sep 11, 2009, at 4:38 AM, solr-dev@lucene.apache.org wrote:



init-forrest-entities:
 [mkdir] Created dir: /tmp/apache-solr-nightly/build
 [mkdir] Created dir: /tmp/apache-solr-nightly/build/web

compile-solrj:
 [mkdir] Created dir: /tmp/apache-solr-nightly/build/solrj
 [javac] Compiling 86 source files to
/tmp/apache-solr-nightly/build/solrj
 [javac] Note: Some input files use or override a deprecated API.
 [javac] Note: Recompile with -Xlint:deprecation for details.
 [javac] Note: Some input files use unchecked or unsafe operations.
 [javac] Note: Recompile with -Xlint:unchecked for details.

compile:
 [mkdir] Created dir: /tmp/apache-solr-nightly/build/solr
 [javac] Compiling 374 source files to
/tmp/apache-solr-nightly/build/solr
 [javac] Note: Some input files use or override a deprecated API.
 [javac] Note: Recompile with -Xlint:deprecation for details.
 [javac] Note: Some input files use unchecked or unsafe operations.
 [javac] Note: Recompile with -Xlint:unchecked for details.

compileTests:
 [mkdir] Created dir: /tmp/apache-solr-nightly/build/tests
 [javac] Compiling 168 source files to
/tmp/apache-solr-nightly/build/tests
 [javac] Note: Some input files use or override a deprecated API.
 [javac] Note: Recompile with -Xlint:deprecation for details.
 [javac] Note: Some input files use unchecked or unsafe operations.
 [javac] Note: Recompile with -Xlint:unchecked for details.

junit:
 [mkdir] Created dir: /tmp/apache-solr-nightly/build/test-results
 [junit] Running org.apache.solr.BasicFunctionalityTest
 [junit] Tests run: 20, Failures: 0, Errors: 0, Time elapsed:  
43.509 sec

 [junit] Running org.apache.solr.ConvertedLegacyTest
 [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed:  
24.636 sec

 [junit] Running org.apache.solr.DisMaxRequestHandlerTest
 [junit] Tests run: 3, Failures: 0, Errors: 0, Time elapsed:  
21.207 sec

 [junit] Running org.apache.solr.EchoParamsTest
 [junit] Tests run: 4, Failures: 0, Errors: 0, Time elapsed:  
6.797 sec

 [junit] Running org.apache.solr.MinimalSchemaTest
 [junit] Tests run: 3, Failures: 0, Errors: 0, Time elapsed:  
10.042 sec

 [junit] Running org.apache.solr.OutputWriterTest
 [junit] Tests run: 3, Failures: 0, Errors: 0, Time elapsed:  
7.022 sec

 [junit] Running org.apache.solr.SampleTest
 [junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed:  
4.666 sec

 [junit] Running org.apache.solr.SolrInfoMBeanTest
 [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed:  
2.302 sec

 [junit] Running org.apache.solr.TestDistributedSearch
 [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed:  
97.45 sec

 [junit] Running org.apache.solr.TestPluginEnable
 [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 2.54  
sec

 [junit] Running org.apache.solr.TestSolrCoreProperties
 [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed:  
9.094 sec

 [junit] Running org.apache.solr.TestTrie
 [junit] Tests run: 7, Failures: 0, Errors: 0, Time elapsed:  
14.825 sec

 [junit] Running
org.apache.solr.analysis.DoubleMetaphoneFilterFactoryTest
 [junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed:  
0.829 sec

 [junit] Running org.apache.solr.analysis.DoubleMetaphoneFilterTest
 [junit] Tests run: 6, Failures: 0, Errors: 0, Time elapsed:  
1.181 sec
 [junit] Running  
org.apache.solr.analysis.EnglishPorterFilterFactoryTest
 [junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed:  
2.589 sec

 [junit] Running org.apache.solr.analysis.HTMLStripCharFilterTest
 [junit] Tests run: 9, Failures: 0, Errors: 0, Time elapsed: 1.68  
sec

 [junit] Running org.apache.solr.analysis.LengthFilterTest
 [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed:  
4.704 sec
 [junit] Running  
org.apache.solr.analysis.SnowballPorterFilterFactoryTest
 [junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed:  
5.717 sec

 [junit] Running org.apache.solr.analysis.TestBufferedTokenStream
 [junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed:  
6.264 sec

 [junit] Running org.apache.solr.analysis.TestCapitalizationFilter
 [junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed:  
3.031 sec

 [junit] Running
org.apache.solr.analysis.TestDelimitedPayloadTokenFilterFactory
 [junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed:  
11.213 sec

 [junit] Running org.apache.solr.analysis.TestHyphenatedWordsFilter
 [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed:  
3.398 sec

 [junit] Running org.apache.solr.analysis.TestKeepFilterFactory
 [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed:  
6.732 sec

 [junit] Running org.apache.solr.analysis.T

[jira] Commented: (SOLR-1143) Return partial results when a connection to a shard is refused

2009-09-11 Thread Martijn van Groningen (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-1143?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12754118#action_12754118
 ] 

Martijn van Groningen commented on SOLR-1143:
-

Sure, that is a good idea. I think that also other types of exceptions should 
result in a partial result (currently just a connection timeout will result in 
a partial result). I think that this behaviour should be enabled with a 
parameter in the request. Something like _shards.requestTimeout=1500_ (time is 
in ms).

> Return partial results when a connection to a shard is refused
> --
>
> Key: SOLR-1143
> URL: https://issues.apache.org/jira/browse/SOLR-1143
> Project: Solr
>  Issue Type: Improvement
>  Components: search
>Reporter: Nicolas Dessaigne
>Assignee: Grant Ingersoll
> Fix For: 1.5
>
> Attachments: SOLR-1143-2.patch, SOLR-1143-3.patch, SOLR-1143.patch
>
>
> If any shard is down in a distributed search, a ConnectException it thrown.
> Here's a little patch that change this behaviour: if we can't connect to a 
> shard (ConnectException), we get partial results from the active shards. As 
> for TimeOut parameter (https://issues.apache.org/jira/browse/SOLR-502), we 
> set the parameter "partialResults" at true.
> This patch also adresses a problem expressed in the mailing list about a year 
> ago 
> (http://www.nabble.com/partialResults,-distributed-search---SOLR-502-td19002610.html)
> We have a use case that needs this behaviour and we would like to know your 
> thougths about such a behaviour? Should it be the default behaviour for 
> distributed search?

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



Re: Solr nightly build failure

2009-09-11 Thread Robert Muir
i see this with a clean checkout and ant test

On Fri, Sep 11, 2009 at 8:37 AM, Grant Ingersoll  wrote:
> Hmm, ant clean test passes for me.
>
>
> On Sep 11, 2009, at 8:25 AM, Grant Ingersoll wrote:
>
>> Hmm, this looks like my commit.  I'll check into it.  Anyone else seeing
>> failures?
>>
>> On Sep 11, 2009, at 4:38 AM, solr-dev@lucene.apache.org wrote:
>>
>>>
>>> init-forrest-entities:
>>>  [mkdir] Created dir: /tmp/apache-solr-nightly/build
>>>  [mkdir] Created dir: /tmp/apache-solr-nightly/build/web
>>>
>>> compile-solrj:
>>>  [mkdir] Created dir: /tmp/apache-solr-nightly/build/solrj
>>>  [javac] Compiling 86 source files to
>>> /tmp/apache-solr-nightly/build/solrj
>>>  [javac] Note: Some input files use or override a deprecated API.
>>>  [javac] Note: Recompile with -Xlint:deprecation for details.
>>>  [javac] Note: Some input files use unchecked or unsafe operations.
>>>  [javac] Note: Recompile with -Xlint:unchecked for details.
>>>
>>> compile:
>>>  [mkdir] Created dir: /tmp/apache-solr-nightly/build/solr
>>>  [javac] Compiling 374 source files to
>>> /tmp/apache-solr-nightly/build/solr
>>>  [javac] Note: Some input files use or override a deprecated API.
>>>  [javac] Note: Recompile with -Xlint:deprecation for details.
>>>  [javac] Note: Some input files use unchecked or unsafe operations.
>>>  [javac] Note: Recompile with -Xlint:unchecked for details.
>>>
>>> compileTests:
>>>  [mkdir] Created dir: /tmp/apache-solr-nightly/build/tests
>>>  [javac] Compiling 168 source files to
>>> /tmp/apache-solr-nightly/build/tests
>>>  [javac] Note: Some input files use or override a deprecated API.
>>>  [javac] Note: Recompile with -Xlint:deprecation for details.
>>>  [javac] Note: Some input files use unchecked or unsafe operations.
>>>  [javac] Note: Recompile with -Xlint:unchecked for details.
>>>
>>> junit:
>>>  [mkdir] Created dir: /tmp/apache-solr-nightly/build/test-results
>>>  [junit] Running org.apache.solr.BasicFunctionalityTest
>>>  [junit] Tests run: 20, Failures: 0, Errors: 0, Time elapsed: 43.509 sec
>>>  [junit] Running org.apache.solr.ConvertedLegacyTest
>>>  [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 24.636 sec
>>>  [junit] Running org.apache.solr.DisMaxRequestHandlerTest
>>>  [junit] Tests run: 3, Failures: 0, Errors: 0, Time elapsed: 21.207 sec
>>>  [junit] Running org.apache.solr.EchoParamsTest
>>>  [junit] Tests run: 4, Failures: 0, Errors: 0, Time elapsed: 6.797 sec
>>>  [junit] Running org.apache.solr.MinimalSchemaTest
>>>  [junit] Tests run: 3, Failures: 0, Errors: 0, Time elapsed: 10.042 sec
>>>  [junit] Running org.apache.solr.OutputWriterTest
>>>  [junit] Tests run: 3, Failures: 0, Errors: 0, Time elapsed: 7.022 sec
>>>  [junit] Running org.apache.solr.SampleTest
>>>  [junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed: 4.666 sec
>>>  [junit] Running org.apache.solr.SolrInfoMBeanTest
>>>  [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 2.302 sec
>>>  [junit] Running org.apache.solr.TestDistributedSearch
>>>  [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 97.45 sec
>>>  [junit] Running org.apache.solr.TestPluginEnable
>>>  [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 2.54 sec
>>>  [junit] Running org.apache.solr.TestSolrCoreProperties
>>>  [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 9.094 sec
>>>  [junit] Running org.apache.solr.TestTrie
>>>  [junit] Tests run: 7, Failures: 0, Errors: 0, Time elapsed: 14.825 sec
>>>  [junit] Running
>>> org.apache.solr.analysis.DoubleMetaphoneFilterFactoryTest
>>>  [junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed: 0.829 sec
>>>  [junit] Running org.apache.solr.analysis.DoubleMetaphoneFilterTest
>>>  [junit] Tests run: 6, Failures: 0, Errors: 0, Time elapsed: 1.181 sec
>>>  [junit] Running org.apache.solr.analysis.EnglishPorterFilterFactoryTest
>>>  [junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed: 2.589 sec
>>>  [junit] Running org.apache.solr.analysis.HTMLStripCharFilterTest
>>>  [junit] Tests run: 9, Failures: 0, Errors: 0, Time elapsed: 1.68 sec
>>>  [junit] Running org.apache.solr.analysis.LengthFilterTest
>>>  [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 4.704 sec
>>>  [junit] Running org.apache.solr.analysis.SnowballPorterFilterFactoryTest
>>>  [junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed: 5.717 sec
>>>  [junit] Running org.apache.solr.analysis.TestBufferedTokenStream
>>>  [junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed: 6.264 sec
>>>  [junit] Running org.apache.solr.analysis.TestCapitalizationFilter
>>>  [junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed: 3.031 sec
>>>  [junit] Running
>>> org.apache.solr.analysis.TestDelimitedPayloadTokenFilterFactory
>>>  [junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed: 11.213 sec
>>>  [junit] Running org.apache.solr.analysis.TestHyphenatedWordsFilter
>>>  [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 3.398 sec
>>>  [junit] Running org.

Re: Solr nightly build failure

2009-09-11 Thread Grant Ingersoll

Hmm, ant clean test passes for me.


On Sep 11, 2009, at 8:25 AM, Grant Ingersoll wrote:

Hmm, this looks like my commit.  I'll check into it.  Anyone else  
seeing failures?


On Sep 11, 2009, at 4:38 AM, solr-dev@lucene.apache.org wrote:



init-forrest-entities:
  [mkdir] Created dir: /tmp/apache-solr-nightly/build
  [mkdir] Created dir: /tmp/apache-solr-nightly/build/web

compile-solrj:
  [mkdir] Created dir: /tmp/apache-solr-nightly/build/solrj
  [javac] Compiling 86 source files to /tmp/apache-solr-nightly/ 
build/solrj

  [javac] Note: Some input files use or override a deprecated API.
  [javac] Note: Recompile with -Xlint:deprecation for details.
  [javac] Note: Some input files use unchecked or unsafe operations.
  [javac] Note: Recompile with -Xlint:unchecked for details.

compile:
  [mkdir] Created dir: /tmp/apache-solr-nightly/build/solr
  [javac] Compiling 374 source files to /tmp/apache-solr-nightly/ 
build/solr

  [javac] Note: Some input files use or override a deprecated API.
  [javac] Note: Recompile with -Xlint:deprecation for details.
  [javac] Note: Some input files use unchecked or unsafe operations.
  [javac] Note: Recompile with -Xlint:unchecked for details.

compileTests:
  [mkdir] Created dir: /tmp/apache-solr-nightly/build/tests
  [javac] Compiling 168 source files to /tmp/apache-solr-nightly/ 
build/tests

  [javac] Note: Some input files use or override a deprecated API.
  [javac] Note: Recompile with -Xlint:deprecation for details.
  [javac] Note: Some input files use unchecked or unsafe operations.
  [javac] Note: Recompile with -Xlint:unchecked for details.

junit:
  [mkdir] Created dir: /tmp/apache-solr-nightly/build/test-results
  [junit] Running org.apache.solr.BasicFunctionalityTest
  [junit] Tests run: 20, Failures: 0, Errors: 0, Time elapsed:  
43.509 sec

  [junit] Running org.apache.solr.ConvertedLegacyTest
  [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed:  
24.636 sec

  [junit] Running org.apache.solr.DisMaxRequestHandlerTest
  [junit] Tests run: 3, Failures: 0, Errors: 0, Time elapsed:  
21.207 sec

  [junit] Running org.apache.solr.EchoParamsTest
  [junit] Tests run: 4, Failures: 0, Errors: 0, Time elapsed: 6.797  
sec

  [junit] Running org.apache.solr.MinimalSchemaTest
  [junit] Tests run: 3, Failures: 0, Errors: 0, Time elapsed:  
10.042 sec

  [junit] Running org.apache.solr.OutputWriterTest
  [junit] Tests run: 3, Failures: 0, Errors: 0, Time elapsed: 7.022  
sec

  [junit] Running org.apache.solr.SampleTest
  [junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed: 4.666  
sec

  [junit] Running org.apache.solr.SolrInfoMBeanTest
  [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 2.302  
sec

  [junit] Running org.apache.solr.TestDistributedSearch
  [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 97.45  
sec

  [junit] Running org.apache.solr.TestPluginEnable
  [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 2.54  
sec

  [junit] Running org.apache.solr.TestSolrCoreProperties
  [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 9.094  
sec

  [junit] Running org.apache.solr.TestTrie
  [junit] Tests run: 7, Failures: 0, Errors: 0, Time elapsed:  
14.825 sec
  [junit] Running  
org.apache.solr.analysis.DoubleMetaphoneFilterFactoryTest
  [junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed: 0.829  
sec

  [junit] Running org.apache.solr.analysis.DoubleMetaphoneFilterTest
  [junit] Tests run: 6, Failures: 0, Errors: 0, Time elapsed: 1.181  
sec
  [junit] Running  
org.apache.solr.analysis.EnglishPorterFilterFactoryTest
  [junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed: 2.589  
sec

  [junit] Running org.apache.solr.analysis.HTMLStripCharFilterTest
  [junit] Tests run: 9, Failures: 0, Errors: 0, Time elapsed: 1.68  
sec

  [junit] Running org.apache.solr.analysis.LengthFilterTest
  [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 4.704  
sec
  [junit] Running  
org.apache.solr.analysis.SnowballPorterFilterFactoryTest
  [junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed: 5.717  
sec

  [junit] Running org.apache.solr.analysis.TestBufferedTokenStream
  [junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed: 6.264  
sec

  [junit] Running org.apache.solr.analysis.TestCapitalizationFilter
  [junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed: 3.031  
sec
  [junit] Running  
org.apache.solr.analysis.TestDelimitedPayloadTokenFilterFactory
  [junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed:  
11.213 sec

  [junit] Running org.apache.solr.analysis.TestHyphenatedWordsFilter
  [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 3.398  
sec

  [junit] Running org.apache.solr.analysis.TestKeepFilterFactory
  [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 6.732  
sec

  [junit] Running org.apache.solr.analysis.TestKeepWordFilter
  [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 8.458  
sec
  [junit] Running  
org.apache.solr.ana

Re: Solr nightly build failure

2009-09-11 Thread Robert Muir
i'm seeing this one too

On Fri, Sep 11, 2009 at 8:25 AM, Grant Ingersoll  wrote:
> Hmm, this looks like my commit.  I'll check into it.  Anyone else seeing
> failures?
>
> On Sep 11, 2009, at 4:38 AM, solr-dev@lucene.apache.org wrote:
>
>>
>> init-forrest-entities:
>>   [mkdir] Created dir: /tmp/apache-solr-nightly/build
>>   [mkdir] Created dir: /tmp/apache-solr-nightly/build/web
>>
>> compile-solrj:
>>   [mkdir] Created dir: /tmp/apache-solr-nightly/build/solrj
>>   [javac] Compiling 86 source files to
>> /tmp/apache-solr-nightly/build/solrj
>>   [javac] Note: Some input files use or override a deprecated API.
>>   [javac] Note: Recompile with -Xlint:deprecation for details.
>>   [javac] Note: Some input files use unchecked or unsafe operations.
>>   [javac] Note: Recompile with -Xlint:unchecked for details.
>>
>> compile:
>>   [mkdir] Created dir: /tmp/apache-solr-nightly/build/solr
>>   [javac] Compiling 374 source files to
>> /tmp/apache-solr-nightly/build/solr
>>   [javac] Note: Some input files use or override a deprecated API.
>>   [javac] Note: Recompile with -Xlint:deprecation for details.
>>   [javac] Note: Some input files use unchecked or unsafe operations.
>>   [javac] Note: Recompile with -Xlint:unchecked for details.
>>
>> compileTests:
>>   [mkdir] Created dir: /tmp/apache-solr-nightly/build/tests
>>   [javac] Compiling 168 source files to
>> /tmp/apache-solr-nightly/build/tests
>>   [javac] Note: Some input files use or override a deprecated API.
>>   [javac] Note: Recompile with -Xlint:deprecation for details.
>>   [javac] Note: Some input files use unchecked or unsafe operations.
>>   [javac] Note: Recompile with -Xlint:unchecked for details.
>>
>> junit:
>>   [mkdir] Created dir: /tmp/apache-solr-nightly/build/test-results
>>   [junit] Running org.apache.solr.BasicFunctionalityTest
>>   [junit] Tests run: 20, Failures: 0, Errors: 0, Time elapsed: 43.509 sec
>>   [junit] Running org.apache.solr.ConvertedLegacyTest
>>   [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 24.636 sec
>>   [junit] Running org.apache.solr.DisMaxRequestHandlerTest
>>   [junit] Tests run: 3, Failures: 0, Errors: 0, Time elapsed: 21.207 sec
>>   [junit] Running org.apache.solr.EchoParamsTest
>>   [junit] Tests run: 4, Failures: 0, Errors: 0, Time elapsed: 6.797 sec
>>   [junit] Running org.apache.solr.MinimalSchemaTest
>>   [junit] Tests run: 3, Failures: 0, Errors: 0, Time elapsed: 10.042 sec
>>   [junit] Running org.apache.solr.OutputWriterTest
>>   [junit] Tests run: 3, Failures: 0, Errors: 0, Time elapsed: 7.022 sec
>>   [junit] Running org.apache.solr.SampleTest
>>   [junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed: 4.666 sec
>>   [junit] Running org.apache.solr.SolrInfoMBeanTest
>>   [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 2.302 sec
>>   [junit] Running org.apache.solr.TestDistributedSearch
>>   [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 97.45 sec
>>   [junit] Running org.apache.solr.TestPluginEnable
>>   [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 2.54 sec
>>   [junit] Running org.apache.solr.TestSolrCoreProperties
>>   [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 9.094 sec
>>   [junit] Running org.apache.solr.TestTrie
>>   [junit] Tests run: 7, Failures: 0, Errors: 0, Time elapsed: 14.825 sec
>>   [junit] Running
>> org.apache.solr.analysis.DoubleMetaphoneFilterFactoryTest
>>   [junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed: 0.829 sec
>>   [junit] Running org.apache.solr.analysis.DoubleMetaphoneFilterTest
>>   [junit] Tests run: 6, Failures: 0, Errors: 0, Time elapsed: 1.181 sec
>>   [junit] Running org.apache.solr.analysis.EnglishPorterFilterFactoryTest
>>   [junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed: 2.589 sec
>>   [junit] Running org.apache.solr.analysis.HTMLStripCharFilterTest
>>   [junit] Tests run: 9, Failures: 0, Errors: 0, Time elapsed: 1.68 sec
>>   [junit] Running org.apache.solr.analysis.LengthFilterTest
>>   [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 4.704 sec
>>   [junit] Running org.apache.solr.analysis.SnowballPorterFilterFactoryTest
>>   [junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed: 5.717 sec
>>   [junit] Running org.apache.solr.analysis.TestBufferedTokenStream
>>   [junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed: 6.264 sec
>>   [junit] Running org.apache.solr.analysis.TestCapitalizationFilter
>>   [junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed: 3.031 sec
>>   [junit] Running
>> org.apache.solr.analysis.TestDelimitedPayloadTokenFilterFactory
>>   [junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed: 11.213 sec
>>   [junit] Running org.apache.solr.analysis.TestHyphenatedWordsFilter
>>   [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 3.398 sec
>>   [junit] Running org.apache.solr.analysis.TestKeepFilterFactory
>>   [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 6.732 sec
>>   [junit] Running 

Re: Solr nightly build failure

2009-09-11 Thread Grant Ingersoll
Hmm, this looks like my commit.  I'll check into it.  Anyone else  
seeing failures?


On Sep 11, 2009, at 4:38 AM, solr-dev@lucene.apache.org wrote:



init-forrest-entities:
   [mkdir] Created dir: /tmp/apache-solr-nightly/build
   [mkdir] Created dir: /tmp/apache-solr-nightly/build/web

compile-solrj:
   [mkdir] Created dir: /tmp/apache-solr-nightly/build/solrj
   [javac] Compiling 86 source files to /tmp/apache-solr-nightly/ 
build/solrj

   [javac] Note: Some input files use or override a deprecated API.
   [javac] Note: Recompile with -Xlint:deprecation for details.
   [javac] Note: Some input files use unchecked or unsafe operations.
   [javac] Note: Recompile with -Xlint:unchecked for details.

compile:
   [mkdir] Created dir: /tmp/apache-solr-nightly/build/solr
   [javac] Compiling 374 source files to /tmp/apache-solr-nightly/ 
build/solr

   [javac] Note: Some input files use or override a deprecated API.
   [javac] Note: Recompile with -Xlint:deprecation for details.
   [javac] Note: Some input files use unchecked or unsafe operations.
   [javac] Note: Recompile with -Xlint:unchecked for details.

compileTests:
   [mkdir] Created dir: /tmp/apache-solr-nightly/build/tests
   [javac] Compiling 168 source files to /tmp/apache-solr-nightly/ 
build/tests

   [javac] Note: Some input files use or override a deprecated API.
   [javac] Note: Recompile with -Xlint:deprecation for details.
   [javac] Note: Some input files use unchecked or unsafe operations.
   [javac] Note: Recompile with -Xlint:unchecked for details.

junit:
   [mkdir] Created dir: /tmp/apache-solr-nightly/build/test-results
   [junit] Running org.apache.solr.BasicFunctionalityTest
   [junit] Tests run: 20, Failures: 0, Errors: 0, Time elapsed:  
43.509 sec

   [junit] Running org.apache.solr.ConvertedLegacyTest
   [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed:  
24.636 sec

   [junit] Running org.apache.solr.DisMaxRequestHandlerTest
   [junit] Tests run: 3, Failures: 0, Errors: 0, Time elapsed:  
21.207 sec

   [junit] Running org.apache.solr.EchoParamsTest
   [junit] Tests run: 4, Failures: 0, Errors: 0, Time elapsed: 6.797  
sec

   [junit] Running org.apache.solr.MinimalSchemaTest
   [junit] Tests run: 3, Failures: 0, Errors: 0, Time elapsed:  
10.042 sec

   [junit] Running org.apache.solr.OutputWriterTest
   [junit] Tests run: 3, Failures: 0, Errors: 0, Time elapsed: 7.022  
sec

   [junit] Running org.apache.solr.SampleTest
   [junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed: 4.666  
sec

   [junit] Running org.apache.solr.SolrInfoMBeanTest
   [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 2.302  
sec

   [junit] Running org.apache.solr.TestDistributedSearch
   [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 97.45  
sec

   [junit] Running org.apache.solr.TestPluginEnable
   [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 2.54  
sec

   [junit] Running org.apache.solr.TestSolrCoreProperties
   [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 9.094  
sec

   [junit] Running org.apache.solr.TestTrie
   [junit] Tests run: 7, Failures: 0, Errors: 0, Time elapsed:  
14.825 sec
   [junit] Running  
org.apache.solr.analysis.DoubleMetaphoneFilterFactoryTest
   [junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed: 0.829  
sec

   [junit] Running org.apache.solr.analysis.DoubleMetaphoneFilterTest
   [junit] Tests run: 6, Failures: 0, Errors: 0, Time elapsed: 1.181  
sec
   [junit] Running  
org.apache.solr.analysis.EnglishPorterFilterFactoryTest
   [junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed: 2.589  
sec

   [junit] Running org.apache.solr.analysis.HTMLStripCharFilterTest
   [junit] Tests run: 9, Failures: 0, Errors: 0, Time elapsed: 1.68  
sec

   [junit] Running org.apache.solr.analysis.LengthFilterTest
   [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 4.704  
sec
   [junit] Running  
org.apache.solr.analysis.SnowballPorterFilterFactoryTest
   [junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed: 5.717  
sec

   [junit] Running org.apache.solr.analysis.TestBufferedTokenStream
   [junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed: 6.264  
sec

   [junit] Running org.apache.solr.analysis.TestCapitalizationFilter
   [junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed: 3.031  
sec
   [junit] Running  
org.apache.solr.analysis.TestDelimitedPayloadTokenFilterFactory
   [junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed:  
11.213 sec

   [junit] Running org.apache.solr.analysis.TestHyphenatedWordsFilter
   [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 3.398  
sec

   [junit] Running org.apache.solr.analysis.TestKeepFilterFactory
   [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 6.732  
sec

   [junit] Running org.apache.solr.analysis.TestKeepWordFilter
   [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 8.458  
sec
   [junit] Running  
org.apache.solr.analysis.TestMappingChar

[jira] Created: (SOLR-1423) Lucene 2.9 RC4 may need some changes in Solr Analyzers using CharStream & others

2009-09-11 Thread Uwe Schindler (JIRA)
Lucene 2.9 RC4 may need some changes in Solr Analyzers using CharStream & others


 Key: SOLR-1423
 URL: https://issues.apache.org/jira/browse/SOLR-1423
 Project: Solr
  Issue Type: Task
  Components: Analysis
Reporter: Uwe Schindler


Because of some backwards compatibility problems (LUCENE-1906) we changed the 
CharStream/CharFilter API a little bit. Tokenizer now only has a input field of 
type java.io.Reader (as before the CharStream code). To correct offsets, it is 
now needed to call the Tokenizer.correctOffset(int) method, which delegates to 
the CharStream (if input is subclass of CharStream), else returns an 
uncorrected offset. Normally it is enough to change all occurences of 
input.correctOffset() to this.correctOffset() in Tokenizers. It should also be 
checked, if custom Tokenizers in Solr do correct their offsets.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Created: (SOLR-1422) Add one QParser to Query Chinese Sentences

2009-09-11 Thread tom liu (JIRA)
Add one QParser to Query Chinese Sentences
--

 Key: SOLR-1422
 URL: https://issues.apache.org/jira/browse/SOLR-1422
 Project: Solr
  Issue Type: Improvement
  Components: search
 Environment: windows xp/ jdk1.6 / tomcat6
Reporter: tom liu


DisMaxQParser do not correctly analysis chinese sentence. So, i implement one 
QParser derived from DisMax.
Limis:
 in schema.xml, set defaultSearchField to chineseFieldType-Field
Result:
 if you input C1C2C3C4, then:
 in DisMaxQParser, we will find that qstr is "C1C2 C3C4"

1. SentenceDisMaxQParser Class::
package org.apache.solr.search;

import org.apache.lucene.queryParser.ParseException;
import org.apache.lucene.search.BooleanClause;
import org.apache.lucene.search.BooleanQuery;
import org.apache.lucene.search.Query;

import org.apache.lucene.analysis.Analyzer;
import org.apache.lucene.analysis.TokenStream;
import org.apache.lucene.analysis.Token;
import java.io.StringReader;

import org.apache.solr.common.SolrException;
import org.apache.solr.common.params.DefaultSolrParams;
import org.apache.solr.common.params.DisMaxParams;
import org.apache.solr.common.params.SolrParams;
import org.apache.solr.common.util.NamedList;
import org.apache.solr.request.SolrQueryRequest;
import org.apache.solr.util.SolrPluginUtils;

import java.util.ArrayList;
import java.util.List;
import java.util.Map;

import org.slf4j.Logger;
import org.slf4j.LoggerFactory;

public class SentenceDisMaxQParser extends DisMaxQParser {
  private static Logger log = 
LoggerFactory.getLogger(SentenceDisMaxQParser.class);
  public SentenceDisMaxQParser(String qstr, SolrParams localParams, SolrParams 
params, SolrQueryRequest req) {
super(qstr, localParams, params, req);

Analyzer analyzer = req.getSchema().getQueryAnalyzer();
if(null == analyzer)
return;

StringBuilder norm = new StringBuilder();
log.info("before analyzer, qstr="+this.qstr);
try{
TokenStream tokens = analyzer.reusableTokenStream( 
req.getSchema().getDefaultSearchFieldName(), new StringReader( this.qstr ) );
tokens.reset();
Token token = tokens.next();
while( token != null ) {
  norm.append( new String(token.termBuffer(), 0, 
token.termLength()) ).append(" ");
  token = tokens.next();
}
} catch(Exception ex){
log.info("Ex="+ex);
}
if(norm.length() > 0)
this.qstr = norm.toString();
log.info("after analyzer, qstr="+this.qstr);
  }
}

2. SentenceDisMaxQParserPlugin Class::
package org.apache.solr.search;

import org.apache.solr.common.params.SolrParams;
import org.apache.solr.common.util.NamedList;
import org.apache.solr.request.SolrQueryRequest;

public class SentenceDisMaxQParserPlugin extends QParserPlugin {
  public static String NAME = "sdismax";

  public void init(NamedList args) {
  }

  public QParser createParser(String qstr, SolrParams localParams, SolrParams 
params, SolrQueryRequest req) {
return new SentenceDisMaxQParser(qstr, localParams, params, req);
  }
}


-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



Build failed in Hudson: Solr-trunk #921

2009-09-11 Thread Apache Hudson Server
See http://hudson.zones.apache.org/hudson/job/Solr-trunk/921/changes

Changes:

[noble] SOLR-1421 FieldreaderDataSource uses a stale Variableresolver

[gsingers] SOLR-1381: Handle when term vecs are present, but not offsets

[gsingers] SOLR-868: removed sample.sgm

[gsingers] SOLR-1411: Add support for uploading ContentStreams via SolrJ

[noble] SOLR-1420 FieldreaderDataSource should throw an Exception if field has 
no data

--
[...truncated 2245 lines...]
[junit] Running org.apache.solr.analysis.TestSynonymMap
[junit] Tests run: 5, Failures: 0, Errors: 0, Time elapsed: 3.169 sec
[junit] Running org.apache.solr.analysis.TestTrimFilter
[junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 1.672 sec
[junit] Running org.apache.solr.analysis.TestWordDelimiterFilter
[junit] Tests run: 14, Failures: 0, Errors: 0, Time elapsed: 37.098 sec
[junit] Running org.apache.solr.client.solrj.SolrExceptionTest
[junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 1.274 sec
[junit] Running org.apache.solr.client.solrj.SolrQueryTest
[junit] Tests run: 5, Failures: 0, Errors: 0, Time elapsed: 0.532 sec
[junit] Running org.apache.solr.client.solrj.TestBatchUpdate
[junit] Tests run: 3, Failures: 0, Errors: 0, Time elapsed: 21.634 sec
[junit] Running org.apache.solr.client.solrj.TestLBHttpSolrServer
[junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed: 18.754 sec
[junit] Running org.apache.solr.client.solrj.beans.TestDocumentObjectBinder
[junit] Tests run: 4, Failures: 0, Errors: 0, Time elapsed: 0.935 sec
[junit] Running org.apache.solr.client.solrj.embedded.JettyWebappTest
[junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 14.625 sec
[junit] Running 
org.apache.solr.client.solrj.embedded.LargeVolumeBinaryJettyTest
[junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 13.22 sec
[junit] Running 
org.apache.solr.client.solrj.embedded.LargeVolumeEmbeddedTest
[junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 7.33 sec
[junit] Running org.apache.solr.client.solrj.embedded.LargeVolumeJettyTest
[junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 11.635 sec
[junit] Running 
org.apache.solr.client.solrj.embedded.MergeIndexesEmbeddedTest
[junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 7.514 sec
[junit] Running org.apache.solr.client.solrj.embedded.MultiCoreEmbeddedTest
[junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 6.35 sec
[junit] Running 
org.apache.solr.client.solrj.embedded.MultiCoreExampleJettyTest
[junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 7.54 sec
[junit] Running 
org.apache.solr.client.solrj.embedded.SolrExampleEmbeddedTest
[junit] Tests run: 9, Failures: 0, Errors: 1, Time elapsed: 21.947 sec
[junit] Test org.apache.solr.client.solrj.embedded.SolrExampleEmbeddedTest 
FAILED
[junit] Running org.apache.solr.client.solrj.embedded.SolrExampleJettyTest
[junit] 
[junit] 
lazy_loading_error__orgapachesolrcommonSolrException_lazy_loading_error__at_orgapachesolrcoreRequestHandlers$LazyRequestHandlerWrappergetWrappedHandlerRequestHandlersjava243__at_orgapachesolrcoreRequestHandlers$LazyRequestHandlerWrapperhandleRequestRequestHandlersjava225__at_orgapachesolrcoreSolrCoreexecuteSolrCorejava1301__at_orgapachesolrservletSolrDispatchFilterexecuteSolrDispatchFilterjava338__at_orgapachesolrservletSolrDispatchFilterdoFilterSolrDispatchFilterjava241__at_orgmortbayjettyservletServletHandler$CachedChaindoFilterServletHandlerjava1089__at_orgmortbayjettyservletServletHandlerhandleServletHandlerjava365__at_orgmortbayjettyservletSessionHandlerhandleSessionHandlerjava181__at_orgmortbayjettyhandlerContextHandlerhandleContextHandlerjava712__at_orgmortbayjettyhandlerHandlerWrapperhandleHandlerWrapperjava139__at_orgmortbayjettyServerhandleServerjava285__at_orgmortbayjettyHttpConnectionhandleRequestHttpConnectionjava502__at_orgmortbayjettyHttpConnection$RequestHandlercontentHttpConnectionjava835__at_orgmortbayjettyHttpParserparseNextHttpParserjava641__at_orgmortbayjettyHttpParserparseAvailableHttpParserjava202__at_orgmortbayjettyHttpConnectionhandleHttpConnectionjava378__at_orgmortbayjettybioSocketConnector$ConnectionrunSocketConnectorjava226__at_orgmortbaythreadBoundedThreadPool$PoolThreadrunBoundedThreadPooljava442_Caused_by_orgapachesolrcommonSolrException_Error_loading_class_orgapachesolrhandlerextractionExtractingRequestHandler__at_orgapachesolrcoreSolrResourceLoaderfindClassSolrResourceLoaderjava310__at_orgapachesolrcoreSolrCorecreateInstanceSolrCorejava411__at_orgapachesolrcoreSolrCorecreateRequestHandlerSolrCorejava436__at_orgapachesolrcoreRequestHandlers$LazyRequestHandlerWrappergetWrappedHandlerRequestHandlersjava234___17_more_Caused_by_javalangClassNotFoundException_orgapachesol
[junit] 
[junit] request: 
http://localhost:38026/example/update/extract?literal.id=mailing

Solr nightly build failure

2009-09-11 Thread solr-dev

init-forrest-entities:
[mkdir] Created dir: /tmp/apache-solr-nightly/build
[mkdir] Created dir: /tmp/apache-solr-nightly/build/web

compile-solrj:
[mkdir] Created dir: /tmp/apache-solr-nightly/build/solrj
[javac] Compiling 86 source files to /tmp/apache-solr-nightly/build/solrj
[javac] Note: Some input files use or override a deprecated API.
[javac] Note: Recompile with -Xlint:deprecation for details.
[javac] Note: Some input files use unchecked or unsafe operations.
[javac] Note: Recompile with -Xlint:unchecked for details.

compile:
[mkdir] Created dir: /tmp/apache-solr-nightly/build/solr
[javac] Compiling 374 source files to /tmp/apache-solr-nightly/build/solr
[javac] Note: Some input files use or override a deprecated API.
[javac] Note: Recompile with -Xlint:deprecation for details.
[javac] Note: Some input files use unchecked or unsafe operations.
[javac] Note: Recompile with -Xlint:unchecked for details.

compileTests:
[mkdir] Created dir: /tmp/apache-solr-nightly/build/tests
[javac] Compiling 168 source files to /tmp/apache-solr-nightly/build/tests
[javac] Note: Some input files use or override a deprecated API.
[javac] Note: Recompile with -Xlint:deprecation for details.
[javac] Note: Some input files use unchecked or unsafe operations.
[javac] Note: Recompile with -Xlint:unchecked for details.

junit:
[mkdir] Created dir: /tmp/apache-solr-nightly/build/test-results
[junit] Running org.apache.solr.BasicFunctionalityTest
[junit] Tests run: 20, Failures: 0, Errors: 0, Time elapsed: 43.509 sec
[junit] Running org.apache.solr.ConvertedLegacyTest
[junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 24.636 sec
[junit] Running org.apache.solr.DisMaxRequestHandlerTest
[junit] Tests run: 3, Failures: 0, Errors: 0, Time elapsed: 21.207 sec
[junit] Running org.apache.solr.EchoParamsTest
[junit] Tests run: 4, Failures: 0, Errors: 0, Time elapsed: 6.797 sec
[junit] Running org.apache.solr.MinimalSchemaTest
[junit] Tests run: 3, Failures: 0, Errors: 0, Time elapsed: 10.042 sec
[junit] Running org.apache.solr.OutputWriterTest
[junit] Tests run: 3, Failures: 0, Errors: 0, Time elapsed: 7.022 sec
[junit] Running org.apache.solr.SampleTest
[junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed: 4.666 sec
[junit] Running org.apache.solr.SolrInfoMBeanTest
[junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 2.302 sec
[junit] Running org.apache.solr.TestDistributedSearch
[junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 97.45 sec
[junit] Running org.apache.solr.TestPluginEnable
[junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 2.54 sec
[junit] Running org.apache.solr.TestSolrCoreProperties
[junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 9.094 sec
[junit] Running org.apache.solr.TestTrie
[junit] Tests run: 7, Failures: 0, Errors: 0, Time elapsed: 14.825 sec
[junit] Running org.apache.solr.analysis.DoubleMetaphoneFilterFactoryTest
[junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed: 0.829 sec
[junit] Running org.apache.solr.analysis.DoubleMetaphoneFilterTest
[junit] Tests run: 6, Failures: 0, Errors: 0, Time elapsed: 1.181 sec
[junit] Running org.apache.solr.analysis.EnglishPorterFilterFactoryTest
[junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed: 2.589 sec
[junit] Running org.apache.solr.analysis.HTMLStripCharFilterTest
[junit] Tests run: 9, Failures: 0, Errors: 0, Time elapsed: 1.68 sec
[junit] Running org.apache.solr.analysis.LengthFilterTest
[junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 4.704 sec
[junit] Running org.apache.solr.analysis.SnowballPorterFilterFactoryTest
[junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed: 5.717 sec
[junit] Running org.apache.solr.analysis.TestBufferedTokenStream
[junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed: 6.264 sec
[junit] Running org.apache.solr.analysis.TestCapitalizationFilter
[junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed: 3.031 sec
[junit] Running 
org.apache.solr.analysis.TestDelimitedPayloadTokenFilterFactory
[junit] Tests run: 2, Failures: 0, Errors: 0, Time elapsed: 11.213 sec
[junit] Running org.apache.solr.analysis.TestHyphenatedWordsFilter
[junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 3.398 sec
[junit] Running org.apache.solr.analysis.TestKeepFilterFactory
[junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 6.732 sec
[junit] Running org.apache.solr.analysis.TestKeepWordFilter
[junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 8.458 sec
[junit] Running org.apache.solr.analysis.TestMappingCharFilterFactory
[junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 0.764 sec
[junit] Running org.apache.solr.analysis.TestPatternReplaceFilter
[juni

[jira] Commented: (SOLR-1394) HTML stripper is splitting tokens

2009-09-11 Thread Anders Melchiorsen (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-1394?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12754032#action_12754032
 ] 

Anders Melchiorsen commented on SOLR-1394:
--

In which situation do you get this exception?

> HTML stripper is splitting tokens
> -
>
> Key: SOLR-1394
> URL: https://issues.apache.org/jira/browse/SOLR-1394
> Project: Solr
>  Issue Type: Bug
>  Components: Analysis
>Affects Versions: 1.4
>Reporter: Anders Melchiorsen
> Attachments: SOLR-1394.patch
>
>
> I am having problems with the Solr HTML stripper.
> After some investigation, I have found the cause to be that the
> stripper is replacing the removed HTML with spaces. This obviously
> breaks when the HTML is in the middle of a word, like "Günther".
> So, without knowing what I was doing, I hacked together a fix that
> uses offset correction instead.
> That seemed to work, except that closing tags and attributes still
> broke the positioning. With even less of a clue, I replaced read()
> with next() in the two methods handling those.
> Finally, invalid HTML also gave wrong offsets, and I fixed that by
> restoring numRead when rolling back the input stream.
> At this point I stopped trying to break it, so there may still be more
> problems. Or I might have introduced some problem on my own. Anyway, I
> have put the three patches at the bottom of this mail, in case
> somebody wants to move along with this issue.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.