[jira] [Updated] (SOLR-6892) Make update processors toplevel components

2014-12-28 Thread Noble Paul (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-6892?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Noble Paul updated SOLR-6892:
-
Description: 
The current update processor chain is rather cumbersome and we should be able 
to use the updateprocessors without a chain.

The scope of this ticket is 
* A new tag   becomes a toplevel tag and it will be equivalent 
to the {{}} tag inside {{}} . The only 
difference is that it should require a {{name}} attribute. The 
{{}} tag will continue to exist and it should be possible 
to define  inside as well . It should also be possible to reference 
a named URP in a chain.
* Any update request will be able  to pass a param {{processor=a,b,c}} , where 
a,b,c are names of update processors. A just in time chain will be created with 
those URPs
* Some in built update processors (wherever possible) will be predefined with 
standard names and can be directly used in requests 
* What happens when I say processor=a,b,c in a request? It will execute the 
default chain after the just-in-time chain {{a->b->c}} . 
* How to execute a different chain other than the default chain? the same old 
mechanism of update.chain=x means that the chain  {{x}} will be applied after 
{{a,b,c}}
* How to avoid the default processor chain from being executed ? There will be 
an implicit URP called {{STOP}} . send your request as processor=a,b,c,STOP. 

  was:
The current update processor chain is rather cumbersome and we should be able 
to use the updateprocessors without a chain.

The scope of this ticket is 
*  tag becomes a toplevel tag and it will be equivalent to the 
 tag inside  . The only difference is 
that it should require a {{name}} attribute
* Any update request will be able  to pass a param {{processor=a,b,c}} , where 
a,b,c are names of update processors. A just in time chain will be created with 
those update processors
* Some in built update processors (wherever possible) will be predefined with 
standard names and can be directly used in requests 


> Make update processors toplevel components 
> ---
>
> Key: SOLR-6892
> URL: https://issues.apache.org/jira/browse/SOLR-6892
> Project: Solr
>  Issue Type: Bug
>Reporter: Noble Paul
>Assignee: Noble Paul
>
> The current update processor chain is rather cumbersome and we should be able 
> to use the updateprocessors without a chain.
> The scope of this ticket is 
> * A new tag   becomes a toplevel tag and it will be 
> equivalent to the {{}} tag inside 
> {{}} . The only difference is that it should 
> require a {{name}} attribute. The {{}} tag will 
> continue to exist and it should be possible to define  inside as 
> well . It should also be possible to reference a named URP in a chain.
> * Any update request will be able  to pass a param {{processor=a,b,c}} , 
> where a,b,c are names of update processors. A just in time chain will be 
> created with those URPs
> * Some in built update processors (wherever possible) will be predefined with 
> standard names and can be directly used in requests 
> * What happens when I say processor=a,b,c in a request? It will execute the 
> default chain after the just-in-time chain {{a->b->c}} . 
> * How to execute a different chain other than the default chain? the same old 
> mechanism of update.chain=x means that the chain  {{x}} will be applied after 
> {{a,b,c}}
> * How to avoid the default processor chain from being executed ? There will 
> be an implicit URP called {{STOP}} . send your request as 
> processor=a,b,c,STOP. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-6892) Make it possible to define update request processors as toplevel components

2014-12-28 Thread Noble Paul (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-6892?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Noble Paul updated SOLR-6892:
-
Summary: Make it possible to define update request processors as toplevel 
components   (was: Make update processors toplevel components )

> Make it possible to define update request processors as toplevel components 
> 
>
> Key: SOLR-6892
> URL: https://issues.apache.org/jira/browse/SOLR-6892
> Project: Solr
>  Issue Type: Bug
>Reporter: Noble Paul
>Assignee: Noble Paul
>
> The current update processor chain is rather cumbersome and we should be able 
> to use the updateprocessors without a chain.
> The scope of this ticket is 
> * A new tag   becomes a toplevel tag and it will be 
> equivalent to the {{}} tag inside 
> {{}} . The only difference is that it should 
> require a {{name}} attribute. The {{}} tag will 
> continue to exist and it should be possible to define  inside as 
> well . It should also be possible to reference a named URP in a chain.
> * Any update request will be able  to pass a param {{processor=a,b,c}} , 
> where a,b,c are names of update processors. A just in time chain will be 
> created with those URPs
> * Some in built update processors (wherever possible) will be predefined with 
> standard names and can be directly used in requests 
> * What happens when I say processor=a,b,c in a request? It will execute the 
> default chain after the just-in-time chain {{a->b->c}} . 
> * How to execute a different chain other than the default chain? the same old 
> mechanism of update.chain=x means that the chain  {{x}} will be applied after 
> {{a,b,c}}
> * How to avoid the default processor chain from being executed ? There will 
> be an implicit URP called {{STOP}} . send your request as 
> processor=a,b,c,STOP. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-6892) Make update processors toplevel components

2014-12-28 Thread Noble Paul (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-6892?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14259600#comment-14259600
 ] 

Noble Paul commented on SOLR-6892:
--

I've updated the description with more meat this time. Please comment

> Make update processors toplevel components 
> ---
>
> Key: SOLR-6892
> URL: https://issues.apache.org/jira/browse/SOLR-6892
> Project: Solr
>  Issue Type: Bug
>Reporter: Noble Paul
>Assignee: Noble Paul
>
> The current update processor chain is rather cumbersome and we should be able 
> to use the updateprocessors without a chain.
> The scope of this ticket is 
> * A new tag   becomes a toplevel tag and it will be 
> equivalent to the {{}} tag inside 
> {{}} . The only difference is that it should 
> require a {{name}} attribute. The {{}} tag will 
> continue to exist and it should be possible to define  inside as 
> well . It should also be possible to reference a named URP in a chain.
> * Any update request will be able  to pass a param {{processor=a,b,c}} , 
> where a,b,c are names of update processors. A just in time chain will be 
> created with those URPs
> * Some in built update processors (wherever possible) will be predefined with 
> standard names and can be directly used in requests 
> * What happens when I say processor=a,b,c in a request? It will execute the 
> default chain after the just-in-time chain {{a->b->c}} . 
> * How to execute a different chain other than the default chain? the same old 
> mechanism of update.chain=x means that the chain  {{x}} will be applied after 
> {{a,b,c}}
> * How to avoid the default processor chain from being executed ? There will 
> be an implicit URP called {{STOP}} . send your request as 
> processor=a,b,c,STOP. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-6879) Have an option to disable autoAddReplicas temporarily for all collections

2014-12-28 Thread Varun Thacker (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-6879?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14259603#comment-14259603
 ] 

Varun Thacker commented on SOLR-6879:
-

Thanks Steve.

Couple of ref guide updates that we would need to make - 

1. 
https://cwiki.apache.org/confluence/display/solr/Collections+API#CollectionsAPI-api11
Instead of "As of Solr 4.7, only the property urlScheme is supported."
we could have something like - "The two supported properties are - 'urlScheme' 
and 'autoAddReplicas' "

2. Under "AutoAddReplica Settings" on 
https://cwiki.apache.org/confluence/display/solr/Running+Solr+on+HDFS we could 
add the following section- 


"Temporarily disable autoAddReplicas for the entire cluster" - When doing 
offline maintenance on the cluster and for various other use cases where an 
admin would like to disable autoAddReplicas temporarily, one could use the 
following APIs to disable and enable autoAddReplicas cluster wide.

API to disable autoAddReplicas:

{code}
admin/collections?action=CLUSTERPROP&name=autoAddReplicas&val=false
{code}

API to enable autoAddReplicas:

{code}
admin/collections?action=CLUSTERPROP&name=autoAddReplicas
{code}

> Have an option to disable autoAddReplicas temporarily for all collections
> -
>
> Key: SOLR-6879
> URL: https://issues.apache.org/jira/browse/SOLR-6879
> Project: Solr
>  Issue Type: Improvement
>Reporter: Varun Thacker
>Assignee: Steve Rowe
> Fix For: 5.0, Trunk
>
> Attachments: SOLR-6879.patch, SOLR-6879.patch
>
>
> When doing offline maintenance on my cluster I would like to disable 
> autoAddReplicas since I don't want replicas being created on nodes during 
> this time.
> It would be useful to have an option to enable/disable autoAddReplicas via an 
> API.
> This API would disable autoAddReplicas:
> {code}
> admin/collections?action=CLUSTERPROP&name=autoAddReplicas&val=false
> {code}
> Now when I want to enable it again the API call would look like this:
> {code}
> admin/collections?action=CLUSTERPROP&name=autoAddReplicas
> {code}
> This uses the CLUSTERPROP semantics of unsetting a property when the {{val}} 
> param is not provided.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-6135) re-number fields randomly in tests

2014-12-28 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-6135?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14259611#comment-14259611
 ] 

ASF subversion and git services commented on LUCENE-6135:
-

Commit 1648188 from [~rcmuir] in branch 'dev/trunk'
[ https://svn.apache.org/r1648188 ]

LUCENE-6135: renumber fields randomly in tests

> re-number fields randomly in tests
> --
>
> Key: LUCENE-6135
> URL: https://issues.apache.org/jira/browse/LUCENE-6135
> Project: Lucene - Core
>  Issue Type: Test
>Reporter: Robert Muir
> Attachments: LUCENE-6135.patch
>
>
> Currently some code (such as stored fields bulk merge) depends upon 
> consistent field number ordering. 
> In the case field numbers do not align, then optimizations are disabled, 
> because the would cause crazy corruption where values are mixed up across 
> different fields. 
> But this is hardly tested today. If i introduce an intentional bug into this 
> logic, then only one lone test fails: TestAddIndexes.testFieldNamesChanged, 
> and only about 10% of the time at best. In general tests pass.
> {code}
> --- 
> lucene/core/src/java/org/apache/lucene/codecs/compressing/MatchingReaders.java
> (revision 1647793)
> +++ 
> lucene/core/src/java/org/apache/lucene/codecs/compressing/MatchingReaders.java
> (working copy)
> @@ -52,6 +52,10 @@
>  for (int i = 0; i < numReaders; i++) {
>for (FieldInfo fi : mergeState.fieldInfos[i]) {
>  FieldInfo other = mergeState.mergeFieldInfos.fieldInfo(fi.number);
> +// nocommit:
> +if (true) {
> +  break;
> +}
>  if (other == null || !other.name.equals(fi.name)) {
>continue nextReader;
>  }
> {code}
> Don't get me wrong, its a great simple test, but it should not be the only 
> one doing this. And it would be great if it failed more often, i havent 
> looked as to why it only fails rarely if there is a bug in this stuff.
> But in general, we should simulate this more. My current idea is to shuffle 
> up field numbers in MockRandomMergePolicy. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-6135) re-number fields randomly in tests

2014-12-28 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-6135?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14259621#comment-14259621
 ] 

ASF subversion and git services commented on LUCENE-6135:
-

Commit 1648189 from [~rcmuir] in branch 'dev/branches/branch_5x'
[ https://svn.apache.org/r1648189 ]

LUCENE-6135: renumber fields randomly in tests

> re-number fields randomly in tests
> --
>
> Key: LUCENE-6135
> URL: https://issues.apache.org/jira/browse/LUCENE-6135
> Project: Lucene - Core
>  Issue Type: Test
>Reporter: Robert Muir
> Attachments: LUCENE-6135.patch
>
>
> Currently some code (such as stored fields bulk merge) depends upon 
> consistent field number ordering. 
> In the case field numbers do not align, then optimizations are disabled, 
> because the would cause crazy corruption where values are mixed up across 
> different fields. 
> But this is hardly tested today. If i introduce an intentional bug into this 
> logic, then only one lone test fails: TestAddIndexes.testFieldNamesChanged, 
> and only about 10% of the time at best. In general tests pass.
> {code}
> --- 
> lucene/core/src/java/org/apache/lucene/codecs/compressing/MatchingReaders.java
> (revision 1647793)
> +++ 
> lucene/core/src/java/org/apache/lucene/codecs/compressing/MatchingReaders.java
> (working copy)
> @@ -52,6 +52,10 @@
>  for (int i = 0; i < numReaders; i++) {
>for (FieldInfo fi : mergeState.fieldInfos[i]) {
>  FieldInfo other = mergeState.mergeFieldInfos.fieldInfo(fi.number);
> +// nocommit:
> +if (true) {
> +  break;
> +}
>  if (other == null || !other.name.equals(fi.name)) {
>continue nextReader;
>  }
> {code}
> Don't get me wrong, its a great simple test, but it should not be the only 
> one doing this. And it would be great if it failed more often, i havent 
> looked as to why it only fails rarely if there is a bug in this stuff.
> But in general, we should simulate this more. My current idea is to shuffle 
> up field numbers in MockRandomMergePolicy. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Resolved] (LUCENE-6135) re-number fields randomly in tests

2014-12-28 Thread Robert Muir (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-6135?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Robert Muir resolved LUCENE-6135.
-
   Resolution: Fixed
Fix Version/s: Trunk
   5.0

> re-number fields randomly in tests
> --
>
> Key: LUCENE-6135
> URL: https://issues.apache.org/jira/browse/LUCENE-6135
> Project: Lucene - Core
>  Issue Type: Test
>Reporter: Robert Muir
> Fix For: 5.0, Trunk
>
> Attachments: LUCENE-6135.patch
>
>
> Currently some code (such as stored fields bulk merge) depends upon 
> consistent field number ordering. 
> In the case field numbers do not align, then optimizations are disabled, 
> because the would cause crazy corruption where values are mixed up across 
> different fields. 
> But this is hardly tested today. If i introduce an intentional bug into this 
> logic, then only one lone test fails: TestAddIndexes.testFieldNamesChanged, 
> and only about 10% of the time at best. In general tests pass.
> {code}
> --- 
> lucene/core/src/java/org/apache/lucene/codecs/compressing/MatchingReaders.java
> (revision 1647793)
> +++ 
> lucene/core/src/java/org/apache/lucene/codecs/compressing/MatchingReaders.java
> (working copy)
> @@ -52,6 +52,10 @@
>  for (int i = 0; i < numReaders; i++) {
>for (FieldInfo fi : mergeState.fieldInfos[i]) {
>  FieldInfo other = mergeState.mergeFieldInfos.fieldInfo(fi.number);
> +// nocommit:
> +if (true) {
> +  break;
> +}
>  if (other == null || !other.name.equals(fi.name)) {
>continue nextReader;
>  }
> {code}
> Don't get me wrong, its a great simple test, but it should not be the only 
> one doing this. And it would be great if it failed more often, i havent 
> looked as to why it only fails rarely if there is a bug in this stuff.
> But in general, we should simulate this more. My current idea is to shuffle 
> up field numbers in MockRandomMergePolicy. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-6494) Query filters applied in a wrong order

2014-12-28 Thread Alexander S. (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-6494?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14259627#comment-14259627
 ] 

Alexander S. commented on SOLR-6494:


As I was told already, Solr does not apply filters incrementally, instead each 
filter runs through the entire data set, then Solr caches the results. In the 
case with filters that contain ranges cache is not effective, especially when 
we need NRT search and commits being triggered multiple times per minute. Then 
big caches make no sense and big autowarming numbers causing Solr to fail. My 
point is that cache is not always efficient and for such cases Solr need to use 
another strategy and apply filters incrementally (read as post filters).

So this:
{quote}
By design, fq clauses like this are calculated for the entire document set and 
the results cached, there is no "ordering" for that part. Otherwise, how could 
they be re-used for a different query?
{quote}
does not work in all cases.

Something like this:
{code}
fq={!cache=false cost=101}field:value # to run as a post filter
{code}
would definitely solve the problem, but this is not supported.

The frange parser has support for this, but it is not always suitable and fails 
with different errors, like "can not use FieldCache on multivalued field: 
type", etc.

Does that look like a missing feature? I mean for me it definitely does, but 
could this be considered as a wish and implemented some day? How can Solr 
community help with missing features?

> Query filters applied in a wrong order
> --
>
> Key: SOLR-6494
> URL: https://issues.apache.org/jira/browse/SOLR-6494
> Project: Solr
>  Issue Type: Bug
>Affects Versions: 4.8.1
>Reporter: Alexander S.
>
> This query:
> {code}
> {
>   fq: ["type:Award::Nomination"],
>   sort: "score desc",
>   start: 0,
>   rows: 20,
>   q: "*:*"
> }
> {code}
> takes just a few milliseconds, but this one:
> {code}
> {
>   fq: [
> "type:Award::Nomination",
> "created_at_d:[* TO 2014-09-08T23:59:59Z]"
>   ],
>   sort: "score desc",
>   start: 0,
>   rows: 20,
>   q: "*:*"
> }
> {code}
> takes almost 15 seconds.
> I have just ≈12k of documents with type "Award::Nomination", but around half 
> a billion with created_at_d field set. And it seems Solr applies the 
> created_at_d filter first going through all documents where this field is 
> set, which is not very smart.
> I think if it can't do anything better than applying filters in the alphabet 
> order it should apply them in the order they were received.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Comment Edited] (SOLR-6494) Query filters applied in a wrong order

2014-12-28 Thread Alexander S. (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-6494?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14259627#comment-14259627
 ] 

Alexander S. edited comment on SOLR-6494 at 12/28/14 12:50 PM:
---

As I was told already, Solr does not apply filters incrementally, instead each 
filter runs through the entire data set, then Solr caches the results. In the 
case with filters that contain ranges cache is not effective, especially when 
we need NRT search and commits being triggered multiple times per minute. Then 
big caches make no sense and big autowarming numbers causing Solr to fail. My 
point is that cache is not always efficient and for such cases Solr need to use 
another strategy and apply filters incrementally (read as post filters).

So this:
{quote}
By design, fq clauses like this are calculated for the entire document set and 
the results cached, there is no "ordering" for that part. Otherwise, how could 
they be re-used for a different query?
{quote}
does not work in all cases.

Something like this:
{code}
# cost > 100 to run as a post filter, but something like post=true would be 
better I think
fq={!cache=false cost=101}field:value
{code}
would definitely solve the problem, but this is not supported.

The frange parser has support for this, but it is not always suitable and fails 
with different errors, like "can not use FieldCache on multivalued field: 
type", etc.

Does that look like a missing feature? I mean for me it definitely does, but 
could this be considered as a wish and implemented some day? How can Solr 
community help with missing features?


was (Author: aheaven):
As I was told already, Solr does not apply filters incrementally, instead each 
filter runs through the entire data set, then Solr caches the results. In the 
case with filters that contain ranges cache is not effective, especially when 
we need NRT search and commits being triggered multiple times per minute. Then 
big caches make no sense and big autowarming numbers causing Solr to fail. My 
point is that cache is not always efficient and for such cases Solr need to use 
another strategy and apply filters incrementally (read as post filters).

So this:
{quote}
By design, fq clauses like this are calculated for the entire document set and 
the results cached, there is no "ordering" for that part. Otherwise, how could 
they be re-used for a different query?
{quote}
does not work in all cases.

Something like this:
{code}
fq={!cache=false cost=101}field:value # to run as a post filter
{code}
would definitely solve the problem, but this is not supported.

The frange parser has support for this, but it is not always suitable and fails 
with different errors, like "can not use FieldCache on multivalued field: 
type", etc.

Does that look like a missing feature? I mean for me it definitely does, but 
could this be considered as a wish and implemented some day? How can Solr 
community help with missing features?

> Query filters applied in a wrong order
> --
>
> Key: SOLR-6494
> URL: https://issues.apache.org/jira/browse/SOLR-6494
> Project: Solr
>  Issue Type: Bug
>Affects Versions: 4.8.1
>Reporter: Alexander S.
>
> This query:
> {code}
> {
>   fq: ["type:Award::Nomination"],
>   sort: "score desc",
>   start: 0,
>   rows: 20,
>   q: "*:*"
> }
> {code}
> takes just a few milliseconds, but this one:
> {code}
> {
>   fq: [
> "type:Award::Nomination",
> "created_at_d:[* TO 2014-09-08T23:59:59Z]"
>   ],
>   sort: "score desc",
>   start: 0,
>   rows: 20,
>   q: "*:*"
> }
> {code}
> takes almost 15 seconds.
> I have just ≈12k of documents with type "Award::Nomination", but around half 
> a billion with created_at_d field set. And it seems Solr applies the 
> created_at_d filter first going through all documents where this field is 
> set, which is not very smart.
> I think if it can't do anything better than applying filters in the alphabet 
> order it should apply them in the order they were received.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-5507) Admin UI - Refactoring using AngularJS

2014-12-28 Thread Upayavira (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-5507?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14259635#comment-14259635
 ] 

Upayavira commented on SOLR-5507:
-

[~erickerickson] please feel free to make a dependent issue for your SolrCloud 
enhancement, it is a good idea that shouldn't be lost.

> Admin UI - Refactoring using AngularJS
> --
>
> Key: SOLR-5507
> URL: https://issues.apache.org/jira/browse/SOLR-5507
> Project: Solr
>  Issue Type: Improvement
>  Components: web gui
>Reporter: Stefan Matheis (steffkes)
>Assignee: Stefan Matheis (steffkes)
>Priority: Minor
> Attachments: SOLR-5507.patch
>
>
> On the LSR in Dublin, i've talked again to [~upayavira] and this time we 
> talked about Refactoring the existing UI - using AngularJS: providing (more, 
> internal) structure and what not ;>
> He already started working on the Refactoring, so this is more a 'tracking' 
> issue about the progress he/we do there.
> Will extend this issue with a bit more context & additional information, w/ 
> thoughts about the possible integration in the existing UI and more (:



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-6138) ItalianLightStemmer doesn't apply on words shorter then 6 chars in length

2014-12-28 Thread Massimo Pasquini (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-6138?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14259655#comment-14259655
 ] 

Massimo Pasquini commented on LUCENE-6138:
--

The issue you pointed out is related to a different stemmer for Russian 
language. I see no connection to the Italian light stemmer. According to the 
rules of the Italian grammar, I think the bug I found can be fixed (it possibly 
cannot be done in the Russian stemmer according to what I've read on the other 
post).

So I suppose the ItalianLightStemmer can evolve to a better implementation: it 
is possible to find some simple rules on words suffixes in order to make a 
decision about applying the stemming on short words (shorter then 6 characters).

Notice my thoughts are not based on a deep and accurate study of the problem 
though. But I think it could be worth to investigate about it. I may suggest to 
submit this issue to the author of the code and see if he got a better solution 
(I saw he's in the field of computational linguistics). According to the notes 
in the source, the algorithm was written in 2005 as the result of some 
experiments. We don't know if they've put further efforts in investigating the 
problem and they possibly wrote a best algorithm they agree to publish 
according to Lucene's license.

I don't expect the stemmer to be 100% successful, but the issue I pointed out 
affects an important range on words.

> ItalianLightStemmer doesn't apply on words shorter then 6 chars in length
> -
>
> Key: LUCENE-6138
> URL: https://issues.apache.org/jira/browse/LUCENE-6138
> Project: Lucene - Core
>  Issue Type: Bug
>  Components: modules/analysis
>Affects Versions: 4.10.2
>Reporter: Massimo Pasquini
>Priority: Minor
>
> I expect a stemmer to transform nouns in their singular and plural forms into 
> a shorter common form. The implementation of the ItalianLightStemmer doesn't 
> apply any stemming to words shorter then 6 characters in length. This leads 
> to some annoying results:
> singular form | plural form
> 4|5 chars in length (no stemming)
> alga -> alga | alghe -> alghe
> fuga -> fuga | fughe -> fughe
> lega -> lega | leghe -> leghe
> 5|6 chars in length (stemming only on plural form)
> vanga -> vanga | vanghe -> vang
> verga -> verga | verghe -> verg
> I suppose that such limitation on words length is to avoid other side effects 
> on shorter words not in the set above, but I think something must be reviewed 
> in the code for better results.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-5507) Admin UI - Refactoring using AngularJS

2014-12-28 Thread Jack Krupansky (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-5507?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14259657#comment-14259657
 ] 

Jack Krupansky commented on SOLR-5507:
--

bq.  All I ask, though, is that you forgive the occasional burst of ebullient 
enthusiasm!

No need for it to be forgiven... all ebullient enthusiasm is always welcome and 
encouraged.


> Admin UI - Refactoring using AngularJS
> --
>
> Key: SOLR-5507
> URL: https://issues.apache.org/jira/browse/SOLR-5507
> Project: Solr
>  Issue Type: Improvement
>  Components: web gui
>Reporter: Stefan Matheis (steffkes)
>Assignee: Stefan Matheis (steffkes)
>Priority: Minor
> Attachments: SOLR-5507.patch
>
>
> On the LSR in Dublin, i've talked again to [~upayavira] and this time we 
> talked about Refactoring the existing UI - using AngularJS: providing (more, 
> internal) structure and what not ;>
> He already started working on the Refactoring, so this is more a 'tracking' 
> issue about the progress he/we do there.
> Will extend this issue with a bit more context & additional information, w/ 
> thoughts about the possible integration in the existing UI and more (:



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-6141) simplify stored fields bulk merge logic

2014-12-28 Thread Robert Muir (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-6141?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14259660#comment-14259660
 ] 

Robert Muir commented on LUCENE-6141:
-

We already have a test: testMixedCompressions. I will improve it with the 
commit though.

> simplify stored fields bulk merge logic
> ---
>
> Key: LUCENE-6141
> URL: https://issues.apache.org/jira/browse/LUCENE-6141
> Project: Lucene - Core
>  Issue Type: Task
>Reporter: Robert Muir
> Attachments: LUCENE-6141.patch
>
>
> The current logic checks that the same chunksize and compression algorithm 
> were used, but this is obselete as it no longer iterates chunks.
> We only need to check that the format version is the same.
> This also allows for bulk merging across compression algorithms. A good use 
> case is BEST_SPEED -> BEST_COMPRESSION for archiving.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-6141) simplify stored fields bulk merge logic

2014-12-28 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-6141?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14259664#comment-14259664
 ] 

ASF subversion and git services commented on LUCENE-6141:
-

Commit 1648221 from [~rcmuir] in branch 'dev/trunk'
[ https://svn.apache.org/r1648221 ]

LUCENE-6141: simplify stored fields bulk merge logic

> simplify stored fields bulk merge logic
> ---
>
> Key: LUCENE-6141
> URL: https://issues.apache.org/jira/browse/LUCENE-6141
> Project: Lucene - Core
>  Issue Type: Task
>Reporter: Robert Muir
> Attachments: LUCENE-6141.patch
>
>
> The current logic checks that the same chunksize and compression algorithm 
> were used, but this is obselete as it no longer iterates chunks.
> We only need to check that the format version is the same.
> This also allows for bulk merging across compression algorithms. A good use 
> case is BEST_SPEED -> BEST_COMPRESSION for archiving.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-6892) Make it possible to define update request processors as toplevel components

2014-12-28 Thread Jack Krupansky (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-6892?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14259665#comment-14259665
 ] 

Jack Krupansky commented on SOLR-6892:
--

Thanks for the description updates. Comments...

1. We need to be explicit about how and when the hard-wired processors are 
invoked. In particular the "run update" processor. The "log update" processor 
is somewhat special in that it is not mandatory, but a lot of people are not 
explicitly aware of it, so if they leave it out, they will be wondering why 
they don't get logging of updates.

2. I suggest three parameters: "pre.processors" to specify processors before 
the default chain, "post.processors" to specify processors after the default 
chain (before or after "run update" and "log update"??), and "processors" to 
specify a processor list to completely replace the default chain.

3. Make "log update" be automatically added at the end unless a "nolog" 
processor is specified.

4. Make "run update" be automatically added at the end unless a "norun" 
processor is specified.

5. Discuss "processor" vs. "processors" - I prefer the latter since it is 
explicit, but maybe allow both since the singular/plural can be confusing.

6. Consider supporting both a single parameter with a csv list as well as 
multiple parameters each with a single value. I prefer having the choice. 
Having a separate parameter for each processor can be more explicit sometimes.

7. Consider a single-processor parameter with the option to specify the 
parameters for that processor. That would make it possible to invoke the 
various field mutating update processors, which would be especially cool and 
convenient.


> Make it possible to define update request processors as toplevel components 
> 
>
> Key: SOLR-6892
> URL: https://issues.apache.org/jira/browse/SOLR-6892
> Project: Solr
>  Issue Type: Bug
>Reporter: Noble Paul
>Assignee: Noble Paul
>
> The current update processor chain is rather cumbersome and we should be able 
> to use the updateprocessors without a chain.
> The scope of this ticket is 
> * A new tag   becomes a toplevel tag and it will be 
> equivalent to the {{}} tag inside 
> {{}} . The only difference is that it should 
> require a {{name}} attribute. The {{}} tag will 
> continue to exist and it should be possible to define  inside as 
> well . It should also be possible to reference a named URP in a chain.
> * Any update request will be able  to pass a param {{processor=a,b,c}} , 
> where a,b,c are names of update processors. A just in time chain will be 
> created with those URPs
> * Some in built update processors (wherever possible) will be predefined with 
> standard names and can be directly used in requests 
> * What happens when I say processor=a,b,c in a request? It will execute the 
> default chain after the just-in-time chain {{a->b->c}} . 
> * How to execute a different chain other than the default chain? the same old 
> mechanism of update.chain=x means that the chain  {{x}} will be applied after 
> {{a,b,c}}
> * How to avoid the default processor chain from being executed ? There will 
> be an implicit URP called {{STOP}} . send your request as 
> processor=a,b,c,STOP. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-6141) simplify stored fields bulk merge logic

2014-12-28 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-6141?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14259668#comment-14259668
 ] 

ASF subversion and git services commented on LUCENE-6141:
-

Commit 1648222 from [~rcmuir] in branch 'dev/branches/branch_5x'
[ https://svn.apache.org/r1648222 ]

LUCENE-6141: simplify stored fields bulk merge logic

> simplify stored fields bulk merge logic
> ---
>
> Key: LUCENE-6141
> URL: https://issues.apache.org/jira/browse/LUCENE-6141
> Project: Lucene - Core
>  Issue Type: Task
>Reporter: Robert Muir
> Fix For: 5.0, Trunk
>
> Attachments: LUCENE-6141.patch
>
>
> The current logic checks that the same chunksize and compression algorithm 
> were used, but this is obselete as it no longer iterates chunks.
> We only need to check that the format version is the same.
> This also allows for bulk merging across compression algorithms. A good use 
> case is BEST_SPEED -> BEST_COMPRESSION for archiving.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Resolved] (LUCENE-6141) simplify stored fields bulk merge logic

2014-12-28 Thread Robert Muir (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-6141?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Robert Muir resolved LUCENE-6141.
-
   Resolution: Fixed
Fix Version/s: Trunk
   5.0

> simplify stored fields bulk merge logic
> ---
>
> Key: LUCENE-6141
> URL: https://issues.apache.org/jira/browse/LUCENE-6141
> Project: Lucene - Core
>  Issue Type: Task
>Reporter: Robert Muir
> Fix For: 5.0, Trunk
>
> Attachments: LUCENE-6141.patch
>
>
> The current logic checks that the same chunksize and compression algorithm 
> were used, but this is obselete as it no longer iterates chunks.
> We only need to check that the format version is the same.
> This also allows for bulk merging across compression algorithms. A good use 
> case is BEST_SPEED -> BEST_COMPRESSION for archiving.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-6892) Make it possible to define update request processors as toplevel components

2014-12-28 Thread Noble Paul (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-6892?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14259670#comment-14259670
 ] 

Noble Paul commented on SOLR-6892:
--

[~jkrupan] 
A lot of your suggestions are not realy in the scope of this ticket and should 
be addressed separately

bq. Consider supporting both a single parameter with a csv list as well as 
multiple parameters each with a single value. I prefer having the choice.

I think having choices is not really good if it doesn't offer anything better.

bq.Consider a single-processor parameter with the option to specify the 
parameters for that processor. 

This is just too complex to handle with the flat http parameter structure. 
Again beyond the scope

URPs are really powerful but they are a bit hard to configure and use. I wish 
to make it easier and more widely used. That is the  objective of this ticket 

> Make it possible to define update request processors as toplevel components 
> 
>
> Key: SOLR-6892
> URL: https://issues.apache.org/jira/browse/SOLR-6892
> Project: Solr
>  Issue Type: Bug
>Reporter: Noble Paul
>Assignee: Noble Paul
>
> The current update processor chain is rather cumbersome and we should be able 
> to use the updateprocessors without a chain.
> The scope of this ticket is 
> * A new tag   becomes a toplevel tag and it will be 
> equivalent to the {{}} tag inside 
> {{}} . The only difference is that it should 
> require a {{name}} attribute. The {{}} tag will 
> continue to exist and it should be possible to define  inside as 
> well . It should also be possible to reference a named URP in a chain.
> * Any update request will be able  to pass a param {{processor=a,b,c}} , 
> where a,b,c are names of update processors. A just in time chain will be 
> created with those URPs
> * Some in built update processors (wherever possible) will be predefined with 
> standard names and can be directly used in requests 
> * What happens when I say processor=a,b,c in a request? It will execute the 
> default chain after the just-in-time chain {{a->b->c}} . 
> * How to execute a different chain other than the default chain? the same old 
> mechanism of update.chain=x means that the chain  {{x}} will be applied after 
> {{a,b,c}}
> * How to avoid the default processor chain from being executed ? There will 
> be an implicit URP called {{STOP}} . send your request as 
> processor=a,b,c,STOP. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-6140) simplify inflater usage in deflate CompressionMode

2014-12-28 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-6140?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14259671#comment-14259671
 ] 

ASF subversion and git services commented on LUCENE-6140:
-

Commit 1648224 from [~rcmuir] in branch 'dev/trunk'
[ https://svn.apache.org/r1648224 ]

LUCENE-6140: simplify inflater usage in CompressionMode

> simplify inflater usage in deflate CompressionMode
> --
>
> Key: LUCENE-6140
> URL: https://issues.apache.org/jira/browse/LUCENE-6140
> Project: Lucene - Core
>  Issue Type: Task
>Reporter: Robert Muir
> Attachments: LUCENE-6140.patch
>
>
> This currently loops-n-grows the output byte[]. But we always decompress the 
> whole block (we dont emit flushes or anything to allow otherwise) and ignore 
> offset/length to the end, and we know how big the uncompressed size is 
> up-front... we can just call inflate one time.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Resolved] (LUCENE-6140) simplify inflater usage in deflate CompressionMode

2014-12-28 Thread Robert Muir (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-6140?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Robert Muir resolved LUCENE-6140.
-
   Resolution: Fixed
Fix Version/s: Trunk
   5.0

> simplify inflater usage in deflate CompressionMode
> --
>
> Key: LUCENE-6140
> URL: https://issues.apache.org/jira/browse/LUCENE-6140
> Project: Lucene - Core
>  Issue Type: Task
>Reporter: Robert Muir
> Fix For: 5.0, Trunk
>
> Attachments: LUCENE-6140.patch
>
>
> This currently loops-n-grows the output byte[]. But we always decompress the 
> whole block (we dont emit flushes or anything to allow otherwise) and ignore 
> offset/length to the end, and we know how big the uncompressed size is 
> up-front... we can just call inflate one time.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-6140) simplify inflater usage in deflate CompressionMode

2014-12-28 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-6140?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14259673#comment-14259673
 ] 

ASF subversion and git services commented on LUCENE-6140:
-

Commit 1648226 from [~rcmuir] in branch 'dev/branches/branch_5x'
[ https://svn.apache.org/r1648226 ]

LUCENE-6140: simplify inflater usage in CompressionMode

> simplify inflater usage in deflate CompressionMode
> --
>
> Key: LUCENE-6140
> URL: https://issues.apache.org/jira/browse/LUCENE-6140
> Project: Lucene - Core
>  Issue Type: Task
>Reporter: Robert Muir
> Fix For: 5.0, Trunk
>
> Attachments: LUCENE-6140.patch
>
>
> This currently loops-n-grows the output byte[]. But we always decompress the 
> whole block (we dont emit flushes or anything to allow otherwise) and ignore 
> offset/length to the end, and we know how big the uncompressed size is 
> up-front... we can just call inflate one time.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-6892) Make it possible to define update request processors as toplevel components

2014-12-28 Thread Noble Paul (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-6892?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Noble Paul updated SOLR-6892:
-
Issue Type: Improvement  (was: Bug)

> Make it possible to define update request processors as toplevel components 
> 
>
> Key: SOLR-6892
> URL: https://issues.apache.org/jira/browse/SOLR-6892
> Project: Solr
>  Issue Type: Improvement
>Reporter: Noble Paul
>Assignee: Noble Paul
>
> The current update processor chain is rather cumbersome and we should be able 
> to use the updateprocessors without a chain.
> The scope of this ticket is 
> * A new tag   becomes a toplevel tag and it will be 
> equivalent to the {{}} tag inside 
> {{}} . The only difference is that it should 
> require a {{name}} attribute. The {{}} tag will 
> continue to exist and it should be possible to define  inside as 
> well . It should also be possible to reference a named URP in a chain.
> * Any update request will be able  to pass a param {{processor=a,b,c}} , 
> where a,b,c are names of update processors. A just in time chain will be 
> created with those URPs
> * Some in built update processors (wherever possible) will be predefined with 
> standard names and can be directly used in requests 
> * What happens when I say processor=a,b,c in a request? It will execute the 
> default chain after the just-in-time chain {{a->b->c}} . 
> * How to execute a different chain other than the default chain? the same old 
> mechanism of update.chain=x means that the chain  {{x}} will be applied after 
> {{a,b,c}}
> * How to avoid the default processor chain from being executed ? There will 
> be an implicit URP called {{STOP}} . send your request as 
> processor=a,b,c,STOP. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (SOLR-6893) DIH doen't work using URL and wget

2014-12-28 Thread Dani (JIRA)
Dani created SOLR-6893:
--

 Summary: DIH doen't work using URL and wget
 Key: SOLR-6893
 URL: https://issues.apache.org/jira/browse/SOLR-6893
 Project: Solr
  Issue Type: Bug
Affects Versions: 4.10.2, 4.8.1, 4.6.1
 Environment: Linux (Ubuntu, CentOS)
Reporter: Dani
Priority: Minor


[ related to #SOLR-2305 ]

I put this URL on browser and import correctly starts:

http://localhost:8983/solr/#/sintesicontratti/dataimport//importazione?command=full-import&entity=fileImport&clean=true&commit=true

But if I use wget it doens't work:

wget 
"http://localhost:8983/solr/#/sintesicontratti/dataimport//importazione?command=full-import&entity=fileImport&clean=true&commit=true";

nor also using escape:

wget 
"http://localhost:8983/solr/#/sintesicontratti/dataimport//importazione?command=full-import\&entity=fileImport\&clean=true\&commit=true";



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-6138) ItalianLightStemmer doesn't apply on words shorter then 6 chars in length

2014-12-28 Thread Erick Erickson (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-6138?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14259684#comment-14259684
 ] 

Erick Erickson commented on LUCENE-6138:


Right, the important part of the discussion (I should have pointed it out) was 
that the stemmers are not part of the Solr code base, they're another project 
and that project would be the place to raise possible bugs or submit patches, 

bq: Can you propose your changes to 
http://members.unine.ch/jacques.savoy/clef/index.html?

Sorry for the confusion



> ItalianLightStemmer doesn't apply on words shorter then 6 chars in length
> -
>
> Key: LUCENE-6138
> URL: https://issues.apache.org/jira/browse/LUCENE-6138
> Project: Lucene - Core
>  Issue Type: Bug
>  Components: modules/analysis
>Affects Versions: 4.10.2
>Reporter: Massimo Pasquini
>Priority: Minor
>
> I expect a stemmer to transform nouns in their singular and plural forms into 
> a shorter common form. The implementation of the ItalianLightStemmer doesn't 
> apply any stemming to words shorter then 6 characters in length. This leads 
> to some annoying results:
> singular form | plural form
> 4|5 chars in length (no stemming)
> alga -> alga | alghe -> alghe
> fuga -> fuga | fughe -> fughe
> lega -> lega | leghe -> leghe
> 5|6 chars in length (stemming only on plural form)
> vanga -> vanga | vanghe -> vang
> verga -> verga | verghe -> verg
> I suppose that such limitation on words length is to avoid other side effects 
> on shorter words not in the set above, but I think something must be reviewed 
> in the code for better results.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-6893) DIH doen't work using URL and wget

2014-12-28 Thread Dani (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-6893?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dani updated SOLR-6893:
---
Description: 
I put this URL on browser and import correctly starts:

http://localhost:8983/solr/#/sintesicontratti/dataimport//importazione?command=full-import&entity=fileImport&clean=true&commit=true

But if I use wget it doens't work:

wget 
"http://localhost:8983/solr/#/sintesicontratti/dataimport//importazione?command=full-import&entity=fileImport&clean=true&commit=true";

nor also using escape:

wget 
"http://localhost:8983/solr/#/sintesicontratti/dataimport//importazione?command=full-import\&entity=fileImport\&clean=true\&commit=true";

  was:
[ related to #SOLR-2305 ]

I put this URL on browser and import correctly starts:

http://localhost:8983/solr/#/sintesicontratti/dataimport//importazione?command=full-import&entity=fileImport&clean=true&commit=true

But if I use wget it doens't work:

wget 
"http://localhost:8983/solr/#/sintesicontratti/dataimport//importazione?command=full-import&entity=fileImport&clean=true&commit=true";

nor also using escape:

wget 
"http://localhost:8983/solr/#/sintesicontratti/dataimport//importazione?command=full-import\&entity=fileImport\&clean=true\&commit=true";


> DIH doen't work using URL and wget
> --
>
> Key: SOLR-6893
> URL: https://issues.apache.org/jira/browse/SOLR-6893
> Project: Solr
>  Issue Type: Bug
>Affects Versions: 4.6.1, 4.8.1, 4.10.2
> Environment: Linux (Ubuntu, CentOS)
>Reporter: Dani
>Priority: Minor
>
> I put this URL on browser and import correctly starts:
> http://localhost:8983/solr/#/sintesicontratti/dataimport//importazione?command=full-import&entity=fileImport&clean=true&commit=true
> But if I use wget it doens't work:
> wget 
> "http://localhost:8983/solr/#/sintesicontratti/dataimport//importazione?command=full-import&entity=fileImport&clean=true&commit=true";
> nor also using escape:
> wget 
> "http://localhost:8983/solr/#/sintesicontratti/dataimport//importazione?command=full-import\&entity=fileImport\&clean=true\&commit=true";



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-5507) Admin UI - Refactoring using AngularJS

2014-12-28 Thread Erick Erickson (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-5507?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14259692#comment-14259692
 ] 

Erick Erickson commented on SOLR-5507:
--

My desire here is for the really annoying error messages to go away if the 
files are misplaced...

I have no real feel for whether they are useful enough to try to keep around 
anyway, up to the people doing the work of course.

> Admin UI - Refactoring using AngularJS
> --
>
> Key: SOLR-5507
> URL: https://issues.apache.org/jira/browse/SOLR-5507
> Project: Solr
>  Issue Type: Improvement
>  Components: web gui
>Reporter: Stefan Matheis (steffkes)
>Assignee: Stefan Matheis (steffkes)
>Priority: Minor
> Attachments: SOLR-5507.patch
>
>
> On the LSR in Dublin, i've talked again to [~upayavira] and this time we 
> talked about Refactoring the existing UI - using AngularJS: providing (more, 
> internal) structure and what not ;>
> He already started working on the Refactoring, so this is more a 'tracking' 
> issue about the progress he/we do there.
> Will extend this issue with a bit more context & additional information, w/ 
> thoughts about the possible integration in the existing UI and more (:



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-6082) Umbrella JIRA for Admin UI and SolrCloud.

2014-12-28 Thread Erick Erickson (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-6082?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14259693#comment-14259693
 ] 

Erick Erickson commented on SOLR-6082:
--

[~shalinmangar] Don't know whether this is getting much of your attention, but 
see SOLR-5507. Upayavira is working on moving the admin UI to AngularJS, 
coordinating the two efforts seems like A Good Thing

An open question is how to prioritize this and the work on SOLR-5507. How/when 
should priority be given to reproducing the current admin UI which looks at 
things largely on the basis of individual nodes with SolrCloud bolted on as 
opposed to adding SolrCloud-centric features to the admin UI?

Recording here for posterity, as always the people doing the work are the ones 
who get to choose.

> Umbrella JIRA for Admin UI and SolrCloud.
> -
>
> Key: SOLR-6082
> URL: https://issues.apache.org/jira/browse/SOLR-6082
> Project: Solr
>  Issue Type: Improvement
>  Components: web gui
>Affects Versions: 4.9, Trunk
>Reporter: Erick Erickson
>Assignee: Shalin Shekhar Mangar
>
> It would be very helpful if the admin UI were more "cloud friendly". This is 
> an umbrella JIRA so we can collect sub-tasks as necessary. I think there 
> might be scattered JIRAs about this, let's link them in as we find them.
> [~steffkes] - I've taken the liberty of assigning it to you since you 
> expressed some interest. Feel free to assign it back if you want...
> Let's imagine that a user has a cluster with _no_ collections assigned and 
> start from there.
> Here's a simple way to set this up. Basically you follow the reference guide 
> tutorial but _don't_ define a collection.
> 1> completely delete the "collection1" directory from example
> 2> cp -r example example2
> 3> in example, execute "java -DzkRun -jar start.jar"
> 4> in example2, execute "java -Djetty.port=7574 -DzkHost=localhost:9983 -jar 
> start.jar"
> Now the "cloud link" appears. If you expand the tree view, you see the two 
> live nodes. But, there's nothing in the graph view, no cores are selectable, 
> etc.
> First problem (need to solve before any sub-jiras, so including it here): You 
> have to push a configuration directory to ZK.
> [~thetapi] The _last_ time Stefan and I started allowing files to be written 
> to Solr from the UI it was...unfortunate. I'm assuming that there's something 
> similar here. That is, we shouldn't allow pushing the Solr config _to_ 
> ZooKeeper through the Admin UI, where they'd be distributed to all the solr 
> nodes. Is that true? If this is a security issue, we can keep pushing the 
> config dirs to ZK a manual step for now...
> Once we determine how to get configurations up, we can work on the various 
> sub-jiras.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-6876) Remove unused legacy scripts.conf

2014-12-28 Thread Erick Erickson (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-6876?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Erick Erickson updated SOLR-6876:
-
Attachment: SOLR-6876.patch

Patch with entry in CHANGES.txt

> Remove unused legacy scripts.conf
> -
>
> Key: SOLR-6876
> URL: https://issues.apache.org/jira/browse/SOLR-6876
> Project: Solr
>  Issue Type: Bug
>Affects Versions: 4.10.2, 5.0, Trunk
>Reporter: Alexandre Rafalovitch
>Assignee: Erick Erickson
>Priority: Minor
> Attachments: SOLR-6876.patch
>
>
> Some of the example collections include *scripts.conf* in the *conf* 
> directory. It is not used by anything in the distribution and is somehow left 
> over from the Solr 1.x legacy days.
> It should be possible to safe delete it to avoid confusing users trying to 
> understand what different files actually do.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-6876) Remove unused legacy scripts.conf

2014-12-28 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-6876?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14259709#comment-14259709
 ] 

ASF subversion and git services commented on SOLR-6876:
---

Commit 1648246 from [~erickoerickson] in branch 'dev/trunk'
[ https://svn.apache.org/r1648246 ]

SOLR-6876: Remove unused legacy scripts.conf

> Remove unused legacy scripts.conf
> -
>
> Key: SOLR-6876
> URL: https://issues.apache.org/jira/browse/SOLR-6876
> Project: Solr
>  Issue Type: Bug
>Affects Versions: 4.10.2, 5.0, Trunk
>Reporter: Alexandre Rafalovitch
>Assignee: Erick Erickson
>Priority: Minor
> Attachments: SOLR-6876.patch
>
>
> Some of the example collections include *scripts.conf* in the *conf* 
> directory. It is not used by anything in the distribution and is somehow left 
> over from the Solr 1.x legacy days.
> It should be possible to safe delete it to avoid confusing users trying to 
> understand what different files actually do.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-6893) DIH doen't work using URL and wget

2014-12-28 Thread Tomoko Uchida (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-6893?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14259713#comment-14259713
 ] 

Tomoko Uchida commented on SOLR-6893:
-

Hi,
At my environment (Mac OS), DIH works well with wget command.
Maybe you should remove '/#/' from request url.

I have tested two patterns with /#/ and without /#/
It works.
{code}
$ wget -O response 
"http://localhost:8983/solr/wikipedia/dataimport?command=full-import&indent=true";
$ wget -O response 
"http://localhost:8983/solr/wikipedia/dataimport?command=status&indent=true";
$ cat response 




  0
  0


  
db-data-config.xml
  

status
busy
A command is still running...

  0:6:59.792
  1
  121772
  60885
  0
  2014-12-28 19:38:24


{code}

It does not work.
{code}
$ wget -O response 
"http://localhost:8983/solr/#/wikipedia/dataimport?command=full-import&indent=true";
{code}

Please try again after removing '#' character (this character has special 
meaning in URI specification.)

> DIH doen't work using URL and wget
> --
>
> Key: SOLR-6893
> URL: https://issues.apache.org/jira/browse/SOLR-6893
> Project: Solr
>  Issue Type: Bug
>Affects Versions: 4.6.1, 4.8.1, 4.10.2
> Environment: Linux (Ubuntu, CentOS)
>Reporter: Dani
>Priority: Minor
>
> I put this URL on browser and import correctly starts:
> http://localhost:8983/solr/#/sintesicontratti/dataimport//importazione?command=full-import&entity=fileImport&clean=true&commit=true
> But if I use wget it doens't work:
> wget 
> "http://localhost:8983/solr/#/sintesicontratti/dataimport//importazione?command=full-import&entity=fileImport&clean=true&commit=true";
> nor also using escape:
> wget 
> "http://localhost:8983/solr/#/sintesicontratti/dataimport//importazione?command=full-import\&entity=fileImport\&clean=true\&commit=true";



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-6876) Remove unused legacy scripts.conf

2014-12-28 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-6876?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14259726#comment-14259726
 ] 

ASF subversion and git services commented on SOLR-6876:
---

Commit 1648252 from [~erickoerickson] in branch 'dev/branches/branch_5x'
[ https://svn.apache.org/r1648252 ]

SOLR-6876: Remove unused legacy scripts.conf

> Remove unused legacy scripts.conf
> -
>
> Key: SOLR-6876
> URL: https://issues.apache.org/jira/browse/SOLR-6876
> Project: Solr
>  Issue Type: Bug
>Affects Versions: 4.10.2, 5.0, Trunk
>Reporter: Alexandre Rafalovitch
>Assignee: Erick Erickson
>Priority: Minor
> Fix For: 5.0, Trunk
>
> Attachments: SOLR-6876.patch
>
>
> Some of the example collections include *scripts.conf* in the *conf* 
> directory. It is not used by anything in the distribution and is somehow left 
> over from the Solr 1.x legacy days.
> It should be possible to safe delete it to avoid confusing users trying to 
> understand what different files actually do.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Resolved] (SOLR-6876) Remove unused legacy scripts.conf

2014-12-28 Thread Erick Erickson (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-6876?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Erick Erickson resolved SOLR-6876.
--
   Resolution: Fixed
Fix Version/s: Trunk
   5.0

Thanks Alexandre!

> Remove unused legacy scripts.conf
> -
>
> Key: SOLR-6876
> URL: https://issues.apache.org/jira/browse/SOLR-6876
> Project: Solr
>  Issue Type: Bug
>Affects Versions: 4.10.2, 5.0, Trunk
>Reporter: Alexandre Rafalovitch
>Assignee: Erick Erickson
>Priority: Minor
> Fix For: 5.0, Trunk
>
> Attachments: SOLR-6876.patch
>
>
> Some of the example collections include *scripts.conf* in the *conf* 
> directory. It is not used by anything in the distribution and is somehow left 
> over from the Solr 1.x legacy days.
> It should be possible to safe delete it to avoid confusing users trying to 
> understand what different files actually do.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-5507) Admin UI - Refactoring using AngularJS

2014-12-28 Thread Erick Erickson (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-5507?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14259729#comment-14259729
 ] 

Erick Erickson commented on SOLR-5507:
--

[~upayavira] Added a comment to SOLR-6082 about prioritization, I think that's 
enough.



> Admin UI - Refactoring using AngularJS
> --
>
> Key: SOLR-5507
> URL: https://issues.apache.org/jira/browse/SOLR-5507
> Project: Solr
>  Issue Type: Improvement
>  Components: web gui
>Reporter: Stefan Matheis (steffkes)
>Assignee: Stefan Matheis (steffkes)
>Priority: Minor
> Attachments: SOLR-5507.patch
>
>
> On the LSR in Dublin, i've talked again to [~upayavira] and this time we 
> talked about Refactoring the existing UI - using AngularJS: providing (more, 
> internal) structure and what not ;>
> He already started working on the Refactoring, so this is more a 'tracking' 
> issue about the progress he/we do there.
> Will extend this issue with a bit more context & additional information, w/ 
> thoughts about the possible integration in the existing UI and more (:



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-5.x-Windows (32bit/jdk1.8.0_40-ea-b09) - Build # 4412 - Failure!

2014-12-28 Thread Policeman Jenkins Server
Build: http://jenkins.thetaphi.de/job/Lucene-Solr-5.x-Windows/4412/
Java: 32bit/jdk1.8.0_40-ea-b09 -client -XX:+UseParallelGC (asserts: false)

No tests ran.

Build Log:
[...truncated 10932 lines...]
FATAL: java.io.IOException: Unexpected termination of the channel
hudson.remoting.RequestAbortedException: java.io.IOException: Unexpected 
termination of the channel
at hudson.remoting.Request.abort(Request.java:295)
at hudson.remoting.Channel.terminate(Channel.java:814)
at 
hudson.remoting.SynchronousCommandTransport$ReaderThread.run(SynchronousCommandTransport.java:69)
at ..remote call to Windows VBOX(Native Method)
at hudson.remoting.Channel.attachCallSiteStackTrace(Channel.java:1356)
at hudson.remoting.Request.call(Request.java:171)
at hudson.remoting.Channel.call(Channel.java:751)
at 
hudson.remoting.RemoteInvocationHandler.invoke(RemoteInvocationHandler.java:179)
at com.sun.proxy.$Proxy75.join(Unknown Source)
at hudson.Launcher$RemoteLauncher$ProcImpl.join(Launcher.java:979)
at hudson.Launcher$ProcStarter.join(Launcher.java:388)
at hudson.tasks.Ant.perform(Ant.java:217)
at hudson.tasks.BuildStepMonitor$1.perform(BuildStepMonitor.java:20)
at 
hudson.model.AbstractBuild$AbstractBuildExecution.perform(AbstractBuild.java:770)
at hudson.model.Build$BuildExecution.build(Build.java:199)
at hudson.model.Build$BuildExecution.doRun(Build.java:160)
at 
hudson.model.AbstractBuild$AbstractBuildExecution.run(AbstractBuild.java:533)
at hudson.model.Run.execute(Run.java:1759)
at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:43)
at hudson.model.ResourceController.execute(ResourceController.java:89)
at hudson.model.Executor.run(Executor.java:240)
Caused by: java.io.IOException: Unexpected termination of the channel
at 
hudson.remoting.SynchronousCommandTransport$ReaderThread.run(SynchronousCommandTransport.java:50)
Caused by: java.io.EOFException
at 
java.io.ObjectInputStream$PeekInputStream.readFully(ObjectInputStream.java:2325)
at 
java.io.ObjectInputStream$BlockDataInputStream.readShort(ObjectInputStream.java:2794)
at 
java.io.ObjectInputStream.readStreamHeader(ObjectInputStream.java:801)
at java.io.ObjectInputStream.(ObjectInputStream.java:299)
at 
hudson.remoting.ObjectInputStreamEx.(ObjectInputStreamEx.java:40)
at 
hudson.remoting.AbstractSynchronousByteArrayCommandTransport.read(AbstractSynchronousByteArrayCommandTransport.java:34)
at 
hudson.remoting.SynchronousCommandTransport$ReaderThread.run(SynchronousCommandTransport.java:48)



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org

[JENKINS-MAVEN] Lucene-Solr-Maven-5.x #801: POMs out of sync

2014-12-28 Thread Apache Jenkins Server
Build: https://builds.apache.org/job/Lucene-Solr-Maven-5.x/801/

3 tests failed.
FAILED:  
org.apache.solr.hadoop.MorphlineGoLiveMiniMRTest.org.apache.solr.hadoop.MorphlineGoLiveMiniMRTest

Error Message:
null

Stack Trace:
java.lang.AssertionError: null
at __randomizedtesting.SeedInfo.seed([42DA430FA16CA04]:0)
at 
org.apache.lucene.util.TestRuleTemporaryFilesCleanup.before(TestRuleTemporaryFilesCleanup.java:105)
at 
com.carrotsearch.randomizedtesting.rules.TestRuleAdapter$1.before(TestRuleAdapter.java:26)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:35)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:54)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65)
at 
org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:55)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365)
at java.lang.Thread.run(Thread.java:745)


FAILED:  org.apache.solr.hadoop.MorphlineBasicMiniMRTest.testPathParts

Error Message:
Test abandoned because suite timeout was reached.

Stack Trace:
java.lang.Exception: Test abandoned because suite timeout was reached.
at __randomizedtesting.SeedInfo.seed([B98A46E3988AA203]:0)


FAILED:  
org.apache.solr.hadoop.MorphlineBasicMiniMRTest.org.apache.solr.hadoop.MorphlineBasicMiniMRTest

Error Message:
Suite timeout exceeded (>= 720 msec).

Stack Trace:
java.lang.Exception: Suite timeout exceeded (>= 720 msec).
at __randomizedtesting.SeedInfo.seed([B98A46E3988AA203]:0)




Build Log:
[...truncated 53961 lines...]
BUILD FAILED
/usr/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Maven-5.x/build.xml:552: 
The following error occurred while executing this line:
/usr/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Maven-5.x/build.xml:204: 
The following error occurred while executing this line:
: Java returned: 1

Total time: 380 minutes 16 seconds
Build step 'Invoke Ant' marked build as failure
Recording test results
Email was triggered for: Failure
Sending email for trigger: Failure



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org

[JENKINS] Lucene-Solr-trunk-Linux (64bit/jdk1.9.0-ea-b34) - Build # 11824 - Failure!

2014-12-28 Thread Policeman Jenkins Server
Build: http://jenkins.thetaphi.de/job/Lucene-Solr-trunk-Linux/11824/
Java: 64bit/jdk1.9.0-ea-b34 -XX:-UseCompressedOops -XX:+UseG1GC (asserts: true)

2 tests failed.
FAILED:  
junit.framework.TestSuite.org.apache.solr.cloud.ChaosMonkeySafeLeaderTest

Error Message:
Suite timeout exceeded (>= 720 msec).

Stack Trace:
java.lang.Exception: Suite timeout exceeded (>= 720 msec).
at __randomizedtesting.SeedInfo.seed([78D62EC33B461E33]:0)


FAILED:  org.apache.solr.cloud.ChaosMonkeySafeLeaderTest.testDistribSearch

Error Message:
Test abandoned because suite timeout was reached.

Stack Trace:
java.lang.Exception: Test abandoned because suite timeout was reached.
at __randomizedtesting.SeedInfo.seed([78D62EC33B461E33]:0)




Build Log:
[...truncated 10414 lines...]
   [junit4] Suite: org.apache.solr.cloud.ChaosMonkeySafeLeaderTest
   [junit4]   2> Creating dataDir: 
/mnt/ssd/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.ChaosMonkeySafeLeaderTest-78D62EC33B461E33-001/init-core-data-001
   [junit4]   2> 1691221 T5088 oas.SolrTestCaseJ4.buildSSLConfig Randomized ssl 
(false) and clientAuth (false)
   [junit4]   2> 1691221 T5088 
oas.BaseDistributedSearchTestCase.initHostContext Setting hostContext system 
property: /
   [junit4]   2> 1691223 T5088 oas.SolrTestCaseJ4.setUp ###Starting 
testDistribSearch
   [junit4]   2> 1691224 T5088 oasc.ZkTestServer.run STARTING ZK TEST SERVER
   [junit4]   1> client port:0.0.0.0/0.0.0.0:0
   [junit4]   2> 1691224 T5089 oasc.ZkTestServer$ZKServerMain.runFromConfig 
Starting server
   [junit4]   2> 1691324 T5088 oasc.ZkTestServer.run start zk server on 
port:45567
   [junit4]   2> 1691325 T5088 
oascc.SolrZkClient.createZkCredentialsToAddAutomatically Using default 
ZkCredentialsProvider
   [junit4]   2> 1691325 T5088 oascc.ConnectionManager.waitForConnected Waiting 
for client to connect to ZooKeeper
   [junit4]   2> 1691328 T5096 oascc.ConnectionManager.process Watcher 
org.apache.solr.common.cloud.ConnectionManager@733c650b 
name:ZooKeeperConnection Watcher:127.0.0.1:45567 got event WatchedEvent 
state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 1691328 T5088 oascc.ConnectionManager.waitForConnected Client 
is connected to ZooKeeper
   [junit4]   2> 1691329 T5088 oascc.SolrZkClient.createZkACLProvider Using 
default ZkACLProvider
   [junit4]   2> 1691329 T5088 oascc.SolrZkClient.makePath makePath: /solr
   [junit4]   2> 1691332 T5088 
oascc.SolrZkClient.createZkCredentialsToAddAutomatically Using default 
ZkCredentialsProvider
   [junit4]   2> 1691333 T5088 oascc.ConnectionManager.waitForConnected Waiting 
for client to connect to ZooKeeper
   [junit4]   2> 1691333 T5099 oascc.ConnectionManager.process Watcher 
org.apache.solr.common.cloud.ConnectionManager@7099daea 
name:ZooKeeperConnection Watcher:127.0.0.1:45567/solr got event WatchedEvent 
state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 1691334 T5088 oascc.ConnectionManager.waitForConnected Client 
is connected to ZooKeeper
   [junit4]   2> 1691334 T5088 oascc.SolrZkClient.createZkACLProvider Using 
default ZkACLProvider
   [junit4]   2> 1691334 T5088 oascc.SolrZkClient.makePath makePath: 
/collections/collection1
   [junit4]   2> 1691336 T5088 oascc.SolrZkClient.makePath makePath: 
/collections/collection1/shards
   [junit4]   2> 1691338 T5088 oascc.SolrZkClient.makePath makePath: 
/collections/control_collection
   [junit4]   2> 1691339 T5088 oascc.SolrZkClient.makePath makePath: 
/collections/control_collection/shards
   [junit4]   2> 1691340 T5088 oasc.AbstractZkTestCase.putConfig put 
/mnt/ssd/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/core/src/test-files/solr/collection1/conf/solrconfig-tlog.xml
 to /configs/conf1/solrconfig.xml
   [junit4]   2> 1691340 T5088 oascc.SolrZkClient.makePath makePath: 
/configs/conf1/solrconfig.xml
   [junit4]   2> 1691342 T5088 oasc.AbstractZkTestCase.putConfig put 
/mnt/ssd/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/core/src/test-files/solr/collection1/conf/schema15.xml
 to /configs/conf1/schema.xml
   [junit4]   2> 1691342 T5088 oascc.SolrZkClient.makePath makePath: 
/configs/conf1/schema.xml
   [junit4]   2> 1691343 T5088 oasc.AbstractZkTestCase.putConfig put 
/mnt/ssd/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/core/src/test-files/solr/collection1/conf/solrconfig.snippet.randomindexconfig.xml
 to /configs/conf1/solrconfig.snippet.randomindexconfig.xml
   [junit4]   2> 1691344 T5088 oascc.SolrZkClient.makePath makePath: 
/configs/conf1/solrconfig.snippet.randomindexconfig.xml
   [junit4]   2> 1691345 T5088 oasc.AbstractZkTestCase.putConfig put 
/mnt/ssd/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/core/src/test-files/solr/collection1/conf/stopwords.txt
 to /configs/conf1/stopwords.txt
   [junit4]   2> 1691345 T5088 oascc.SolrZkClient.makePath makePath: 
/configs/conf1/stopwords.txt
   [junit4]   2> 1691346 T5088 oasc.AbstractZkTestCase.putConfig put 
/mnt/s

[jira] [Commented] (SOLR-6888) Decompressing documents on first-pass distributed queries to get docId is inefficient, use indexed values instead?

2014-12-28 Thread David Smiley (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-6888?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14259787#comment-14259787
 ] 

David Smiley commented on SOLR-6888:


Erick,
I enjoyed reading what you have here.  I think this issue duplicates SOLR-5478, 
for which I have a patch in fact. I encourage you to review it and kick the 
tires on it!

> Decompressing documents on first-pass distributed queries to get docId is 
> inefficient, use indexed values instead?
> --
>
> Key: SOLR-6888
> URL: https://issues.apache.org/jira/browse/SOLR-6888
> Project: Solr
>  Issue Type: Improvement
>Affects Versions: 5.0, Trunk
>Reporter: Erick Erickson
>Assignee: Erick Erickson
> Attachments: SOLR-6888-hacktiming.patch
>
>
> Assigning this to myself to just not lose track of it, but I won't be working 
> on this in the near term; anyone feeling ambitious should feel free to grab 
> it.
> Note, docId used here is whatever is defined for ...
> Since Solr 4.1, the compression/decompression process is based on 16K blocks 
> and is automatic, and not configurable. So, to get a single stored value one 
> must decompress an entire 16K block. At least.
> For SolrCloud (and distributed processing in general), we make two trips, one 
> to get the doc id and score (or other sort criteria) and one to return the 
> actual data.
> The first pass here requires that we return the top N docIDs and sort 
> criteria, which means that each and every sub-request has to unpack at least 
> one 16K block (and sometimes more) to get just the doc ID. So if we have 20 
> shards and only want 20 rows, 95% of the decompression cycles will be wasted. 
> Not to mention all the disk reads.
> It seems like we should be able to do better than that. Can we argue that doc 
> ids are 'special' and should be cached somehow? Let's discuss what this would 
> look like. I can think of a couple of approaches:
> 1> Since doc IDs are "special", can we say that for this purpose returning 
> the indexed version is OK? We'd need to return the actual stored value when 
> the full doc was requested, but for the sub-request only what about returning 
> the indexed value instead of the stored one? On the surface I don't see a 
> problem here, but what do I know? Storing these as DocValues seems useful in 
> this case.
> 1a> A variant is treating numeric docIds specially since the indexed value 
> and the stored value should be identical. And DocValues here would be useful 
> it seems. But this seems an unnecessary specialization if <1> is implemented 
> well.
> 2> We could cache individual doc IDs, although I'm not sure what use that 
> really is. Would maintaining the cache overwhelm the savings of not 
> decompressing? I really don't like this idea, but am throwing it out there. 
> Doing this from stored data up front would essentially mean decompressing 
> every doc so that seems untenable to try up-front.
> 3> We could maintain an array[maxDoc] that held document IDs, perhaps lazily 
> initializing it. I'm not particularly a fan of this either, doesn't seem like 
> a Good Thing. I can see lazy loading being almost, but not quite totally, 
> useless, i.e. a hit ratio near 0, especially since it'd be thrown out on 
> every openSearcher.
> Really, the only one of these that seems viable is <1>/<1a>. The others would 
> all involve decompressing the docs anyway to get the ID, and I suspect that 
> caching would be of very limited usefulness. I guess <1>'s viability hinges 
> on whether, for internal use, the indexed form of DocId is interchangeable 
> with the stored value.
> Or are there other ways to approach this? Or isn't it something to really 
> worry about?



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-NightlyTests-5.x - Build # 717 - Failure

2014-12-28 Thread Apache Jenkins Server
Build: https://builds.apache.org/job/Lucene-Solr-NightlyTests-5.x/717/

1 tests failed.
REGRESSION:  
org.apache.solr.cloud.CollectionsAPIDistributedZkTest.testDistribSearch

Error Message:
Error from server at http://127.0.0.1:54363: Error CREATEing SolrCore 
'halfcollection_shard1_replica1': Unable to create core 
[halfcollection_shard1_replica1] Caused by: Could not get shard id for core: 
halfcollection_shard1_replica1

Stack Trace:
org.apache.solr.client.solrj.impl.HttpSolrServer$RemoteSolrException: Error 
from server at http://127.0.0.1:54363: Error CREATEing SolrCore 
'halfcollection_shard1_replica1': Unable to create core 
[halfcollection_shard1_replica1] Caused by: Could not get shard id for core: 
halfcollection_shard1_replica1
at 
org.apache.solr.client.solrj.impl.HttpSolrServer.executeMethod(HttpSolrServer.java:566)
at 
org.apache.solr.client.solrj.impl.HttpSolrServer.request(HttpSolrServer.java:213)
at 
org.apache.solr.client.solrj.impl.HttpSolrServer.request(HttpSolrServer.java:209)
at 
org.apache.solr.cloud.CollectionsAPIDistributedZkTest.testErrorHandling(CollectionsAPIDistributedZkTest.java:452)
at 
org.apache.solr.cloud.CollectionsAPIDistributedZkTest.doTest(CollectionsAPIDistributedZkTest.java:201)
at 
org.apache.solr.BaseDistributedSearchTestCase.testDistribSearch(BaseDistributedSearchTestCase.java:869)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1618)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:827)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:863)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:877)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:53)
at 
org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:50)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55)
at 
org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:49)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:798)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:458)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:836)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$3.evaluate(RandomizedRunner.java:738)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:772)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:783)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:53)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
at 
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
org.apache.lucene.util.TestRuleAssertionsRequired$1

[JENKINS] Lucene-Solr-Tests-5.x-Java7 - Build # 2394 - Failure

2014-12-28 Thread Apache Jenkins Server
Build: https://builds.apache.org/job/Lucene-Solr-Tests-5.x-Java7/2394/

2 tests failed.
REGRESSION:  org.apache.solr.cloud.ChaosMonkeySafeLeaderTest.testDistribSearch

Error Message:
Test abandoned because suite timeout was reached.

Stack Trace:
java.lang.Exception: Test abandoned because suite timeout was reached.
at __randomizedtesting.SeedInfo.seed([2FE41655C33758DD]:0)


FAILED:  
junit.framework.TestSuite.org.apache.solr.cloud.ChaosMonkeySafeLeaderTest

Error Message:
Suite timeout exceeded (>= 720 msec).

Stack Trace:
java.lang.Exception: Suite timeout exceeded (>= 720 msec).
at __randomizedtesting.SeedInfo.seed([2FE41655C33758DD]:0)




Build Log:
[...truncated 10475 lines...]
   [junit4] Suite: org.apache.solr.cloud.ChaosMonkeySafeLeaderTest
   [junit4]   2> Creating dataDir: 
/usr/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-5.x-Java7/solr/build/solr-core/test/J2/temp/solr.cloud.ChaosMonkeySafeLeaderTest-2FE41655C33758DD-001/init-core-data-001
   [junit4]   2> 1667952 T3808 oas.SolrTestCaseJ4.buildSSLConfig Randomized ssl 
(false) and clientAuth (false)
   [junit4]   2> 1667952 T3808 
oas.BaseDistributedSearchTestCase.initHostContext Setting hostContext system 
property: /dt_/lw
   [junit4]   2> 1667959 T3808 oas.SolrTestCaseJ4.setUp ###Starting 
testDistribSearch
   [junit4]   2> 1667960 T3808 oasc.ZkTestServer.run STARTING ZK TEST SERVER
   [junit4]   1> client port:0.0.0.0/0.0.0.0:0
   [junit4]   2> 1667961 T3809 oasc.ZkTestServer$ZKServerMain.runFromConfig 
Starting server
   [junit4]   2> 1668060 T3808 oasc.ZkTestServer.run start zk server on 
port:11132
   [junit4]   2> 1668061 T3808 
oascc.SolrZkClient.createZkCredentialsToAddAutomatically Using default 
ZkCredentialsProvider
   [junit4]   2> 1668062 T3808 oascc.ConnectionManager.waitForConnected Waiting 
for client to connect to ZooKeeper
   [junit4]   2> 1668065 T3816 oascc.ConnectionManager.process Watcher 
org.apache.solr.common.cloud.ConnectionManager@418cfca8 
name:ZooKeeperConnection Watcher:127.0.0.1:11132 got event WatchedEvent 
state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 1668066 T3808 oascc.ConnectionManager.waitForConnected Client 
is connected to ZooKeeper
   [junit4]   2> 1668066 T3808 oascc.SolrZkClient.createZkACLProvider Using 
default ZkACLProvider
   [junit4]   2> 1668066 T3808 oascc.SolrZkClient.makePath makePath: /solr
   [junit4]   2> 1668069 T3808 
oascc.SolrZkClient.createZkCredentialsToAddAutomatically Using default 
ZkCredentialsProvider
   [junit4]   2> 1668070 T3808 oascc.ConnectionManager.waitForConnected Waiting 
for client to connect to ZooKeeper
   [junit4]   2> 1668072 T3819 oascc.ConnectionManager.process Watcher 
org.apache.solr.common.cloud.ConnectionManager@4b266679 
name:ZooKeeperConnection Watcher:127.0.0.1:11132/solr got event WatchedEvent 
state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 1668072 T3808 oascc.ConnectionManager.waitForConnected Client 
is connected to ZooKeeper
   [junit4]   2> 1668072 T3808 oascc.SolrZkClient.createZkACLProvider Using 
default ZkACLProvider
   [junit4]   2> 1668073 T3808 oascc.SolrZkClient.makePath makePath: 
/collections/collection1
   [junit4]   2> 1668075 T3808 oascc.SolrZkClient.makePath makePath: 
/collections/collection1/shards
   [junit4]   2> 1668076 T3808 oascc.SolrZkClient.makePath makePath: 
/collections/control_collection
   [junit4]   2> 1668078 T3808 oascc.SolrZkClient.makePath makePath: 
/collections/control_collection/shards
   [junit4]   2> 1668080 T3808 oasc.AbstractZkTestCase.putConfig put 
/usr/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-5.x-Java7/solr/core/src/test-files/solr/collection1/conf/solrconfig-tlog.xml
 to /configs/conf1/solrconfig.xml
   [junit4]   2> 1668080 T3808 oascc.SolrZkClient.makePath makePath: 
/configs/conf1/solrconfig.xml
   [junit4]   2> 1668083 T3808 oasc.AbstractZkTestCase.putConfig put 
/usr/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-5.x-Java7/solr/core/src/test-files/solr/collection1/conf/schema15.xml
 to /configs/conf1/schema.xml
   [junit4]   2> 1668083 T3808 oascc.SolrZkClient.makePath makePath: 
/configs/conf1/schema.xml
   [junit4]   2> 1668185 T3808 oasc.AbstractZkTestCase.putConfig put 
/usr/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-5.x-Java7/solr/core/src/test-files/solr/collection1/conf/solrconfig.snippet.randomindexconfig.xml
 to /configs/conf1/solrconfig.snippet.randomindexconfig.xml
   [junit4]   2> 1668186 T3808 oascc.SolrZkClient.makePath makePath: 
/configs/conf1/solrconfig.snippet.randomindexconfig.xml
   [junit4]   2> 1668188 T3808 oasc.AbstractZkTestCase.putConfig put 
/usr/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-5.x-Java7/solr/core/src/test-files/solr/collection1/conf/stopwords.txt
 to /configs/conf1/stopwords.txt
   [junit4]   2> 1668189 T3808 oascc.SolrZkClient.makePath makePath: 
/configs/conf1/stopwords.txt
   [junit4]   2> 1668190 T3808 oasc.AbstractZkTes

Multiple CFS files are generated in lucene 4.10.2

2014-12-28 Thread jaga_india
Hi

I am trying index around 612 record using lucene 4.10.2. It is creating lot
of CF files in index directory.
around 626 CFS file are created. It is taking more time to index.

ENV : java 8, window 7

 

Directory dir = FSDirectory.open(file);
IndexWriterConfig config = new
IndexWriterConfig(Version.LUCENE_4_10_2, new ClassicAnalyzer());
if(bufferSizeMB != 0 && bufferSizeMB != -1){
config.setRAMBufferSizeMB(bufferSizeMB);
}  else {
config.setRAMBufferSizeMB(DEFAULT_RAM_BUFFER_SIZE_MB);
}  
config.setMaxBufferedDocs(1000);
config.setMaxBufferedDeleteTerms(1000);
config.setMergePolicy(new LogDocMergePolicy());
IndexWriter iwriter = new IndexWriter(dir, config);
iwriter.getConfig().setMaxBufferedDeleteTerms(1000);
iwriter.getConfig().setMaxBufferedDocs(1000);
iwriter.getConfig().setRAMBufferSizeMB(bufferSizeMB)


Thanks and Regards
jaga india



--
View this message in context: 
http://lucene.472066.n3.nabble.com/Multiple-CFS-files-are-generated-in-lucene-4-10-2-tp4176336.html
Sent from the Lucene - Java Developer mailing list archive at Nabble.com.

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org