Re: DIH timezone offset

2014-05-19 Thread rulinma
good.



--
View this message in context: 
http://lucene.472066.n3.nabble.com/DIH-timezone-offset-tp504958p4137077.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: Question on 3-level composite-id routing

2014-05-19 Thread shamik
Awesome, thanks a lot Anshum, makes total sense now. Appreciate your help.



--
View this message in context: 
http://lucene.472066.n3.nabble.com/Question-on-3-level-composite-id-routing-tp4137044p4137071.html
Sent from the Solr - User mailing list archive at Nabble.com.


ReversedWildcardFilter to TestRandomChains

2014-05-19 Thread Ahmet Arslan
Hi,

How can we add ReversedWildcardFilterFactory to TestRandomChains?

I believe there is an offset bug in RWF : SOLR-3193 SOLR-3901 SOLR-5426

If a filter (like RWF) does not set offset attribute, how it is calculated? How 
to debug/prevent offset bugs?

Thanks,
Ahmet


Re: How to delete docs only in a spec shard within a collection

2014-05-19 Thread YouPeng Yang
Hi

  It seems that ther _route_ parameter make work as [1]. In  my last
posts,It said that  the shards should be same with the  _route_ and
shard.keys during update process to  make sure consistent.

 I will  should file a JIRA to improve that.

[1]:http://10.1.22.1:8082/solr/tv_201402/update?stream.body=BEGINTIME:["2014-03-01
00:00:00" TO "2014-03-01 00:00:10"]&_route_
=tv_201402&commit=true





2014-05-19 19:23 GMT+08:00 Jack Krupansky :

> Normally a delete-by-query needs to be distributed to all shards since it
> is not known which document keys it will affect.
>
> It sounds like SolrCloud does not support the feature you would really
> like: support for distrb=false.
>
> You can file a Jira request for a feature "improvement."
>
> -- Jack Krupansky
>
> -Original Message- From: YouPeng Yang
> Sent: Monday, May 19, 2014 3:46 AM
> To: solr-user@lucene.apache.org
> Subject: Re: How to delete docs only in a spec shard within a collection
>
>
> Hi
>  Anyone gives some suggestions.
>
>
> Regards
>
>
> 2014-05-19 11:31 GMT+08:00 YouPeng Yang :
>
>  Hi
>>   I have a colloection with 3 shards.
>>   I want to delete some docs in one shard with the command:
>>
>>
>> http://10.1.22.1:8082/solr/tv_201402/update?&stream.body=<
>> delete>BEGINTIME:["2014-03-01
>> 00:00:00" TO *]&shards=tv_201402&commit=true
>>  As the red font exepression, It is supposed that docs only in the shard
>> tv_201402  whose BEGINTIME field were larger than "2014-03-01 00:00:00"
>> would be deleted.However ,as a result, the docs on the other shards such
>> astv_201403 ,tv_201404 whose BEGINTIME were also larger than 2014-03-01
>> 00:00:00  were deleted too.
>>   This is not the effientlty right thing I want.
>>
>>   How to delete docs only in a spec shard  within a collection which
>> contains several shards?
>>
>> Regards.
>>
>>
>


Re: Question on 3-level composite-id routing

2014-05-19 Thread Anshum Gupta
Hi Shamik,

Your assumptions on that are correct.

As far as explicit '/8' at query time is concerned, that's the only
way the router would get to know that it's a 3-level id and not a
2-level one i.e. e.g.

shard.keys='myapp!'
Hash range to be fetched: 
to 
Intended Hash Range: <8 bits from the hash of myapp>00 to <8 bits
from hash of myapp>ff (larger range).

If you don't specify a '/8', it will operate as a regular composite id
and try to use 16 bits from the 1st component instead of the intended
8 (in case of tri-level).

Does that make sense?

On Mon, May 19, 2014 at 5:06 PM, Shamik Bandopadhyay  wrote:
> Hi,
>
>   Need some clarification on multilevel composite-id routing in SolrCloud
> .I'm currently composite id routing using the following pattern *topic!url* .
> This is aimed at run-time de-duplication based on topic field. As I'm
> adding support for language search, I felt the need to include language
> parameter for better multi-tenancy. Here's the new key structure I'm
> thinking of --> *language!topic!url*.
> An example would be : english!12345!www.testurl.com
>
> Now, during query time, I'll always have language parameter at my disposal.
> I was thinking,of leveraging the shard.key parameter to specify
> *shard.keys=language!
> *, which will route the request to the right shard and bring back english
> content. Is this a valid assumption ?
>
> Also, as per my understanding, the three fields will default to 8,8 and 16
> bits of the routing hash. What'll be a valid scenario for providing
> a custom allocation of bits for these fields. I was referring to the
> following article http://searchhub.org/2014/01/06/10590/ , but was not
> entirely sure on this section.
>
>
> *At query time:*
>
> *To query all records for myapp: shard.keys=myapp/8!*
>
> *Note the explicit mention of 8 bits in case of querying by component 1
> only i.e. app level. This is required because the usage of the router as 2
> or 3 level isn’t implicit. Specifying ’8′ bits for the component highlights
> the use of ’3′ level router*.
> Any feedback will be much appreciated.
>
> Thanks,
> Shamik



-- 

Anshum Gupta
http://www.anshumgupta.net


Re: Is there a way to change transientCacheSize dynamically without restarting Solr

2014-05-19 Thread Erick Erickson
Nope. I never thought of the need when we put together transient cores

Best,
Erick

On Mon, May 19, 2014 at 1:13 AM, Elran Dvir  wrote:
> Thank you, Shalin.
> -Original Message-
> From: Shalin Shekhar Mangar [mailto:shalinman...@gmail.com]
> Sent: Monday, May 19, 2014 10:18 AM
> To: solr-user@lucene.apache.org
> Subject: Re: Is there a way to change transientCacheSize dynamically without 
> restarting Solr
>
> Hi Elran,
>
> No, I don't think there is way to do that right now.
>
>
> On Mon, May 19, 2014 at 12:18 PM, Elran Dvir  wrote:
>
>> Does anyone have an answer for this?
>> Thanks.
>>
>> From: Elran Dvir
>> Sent: Thursday, May 15, 2014 12:32 PM
>> To: solr-user@lucene.apache.org
>> Subject: Is there a way to change transientCacheSize dynamically
>> without restarting Solr
>>
>> Hi All,
>>
>> Is there an API in Solr to change transientCacheSize dynamically
>> without the need to restart Solr?
>> Are there other Solr configuration parameters that can be changed
>> dynamically?
>>
>> Thanks.
>>
>
>
>
> --
> Regards,
> Shalin Shekhar Mangar.
>
>
> Email secured by Check Point


Re: date range queries efficiency

2014-05-19 Thread Erick Erickson
This might be useful:
http://searchhub.org/2012/02/23/date-math-now-and-filter-queries/

Best,
Erick

On Mon, May 19, 2014 at 12:09 AM, Dmitry Kan  wrote:
> Thanks, Jack, Alex and Shawn.
>
> This makes proper sense. One win of rounding down on indexing side is
> saving index space, according to hoss (reply over IRC):
>
> "with the TrieDateFields, rounding dates at indexing time won't have any
> effect on the cachability of the rounded queries, and even for non cached
> queries it shouldn't affect the performance much -- but yes, it would help
> reduce index size"
>
> I haven't tried it myself, just thought to ask if somebody tried it already.
>
> Dmitry
>
>
> On Sat, May 17, 2014 at 8:37 AM, Shawn Heisey  wrote:
>
>> On 5/15/2014 1:34 AM, Alexandre Rafalovitch wrote:
>> > I thought the date math rounding was for _caching_ the repeated
>> > queries, not so much the speed of the query itself.
>>
>> Absolutely correct.  When NOW is used without rounding, caching is
>> completely ineffective.  This is because if the same query using NOW is
>> sent multiple times several seconds apart, every one of those queries
>> will be different after they are parsed and NOW is converted to an
>> actual timestamp.
>>
>> > Also, if you are using TrieDateField, precisionStep value is how
>> > optimization is done. There is bucketing at different level of
>> > precision, so the range search works at the least granular level
>> > first, etc.
>>
>> Some nitty-gritty details of how range queries are accelerated with the
>> Trie data types and precisionStep are described in the Javadoc for
>> NumericRangeQuery:
>>
>>
>> http://lucene.apache.org/core/4_8_0/core/org/apache/lucene/search/NumericRangeQuery.html
>>
>> Thanks,
>> Shawn
>>
>>
>
>
> --
> Dmitry Kan
> Blog: http://dmitrykan.blogspot.com
> Twitter: http://twitter.com/dmitrykan


Re: Replica as a "leader"

2014-05-19 Thread Erick Erickson
bq: Is there a way that solr can recover without losing docs in this scenario?

Not that I know of currently. SolrCloud is designed to _not_ lose
documents as long
as all leaders are present. And when a leader goes down, assuming
there's a replica
handy docs shouldn't be lost either. But taking down the leader then
starting an out-of-date
replica up and hoping that Solr has somehow magically cached all the
intervening updates
is not a supported scenario. Perhaps SOLR-5468 will help here, I'm not
entirely sure. This
scenario seems out-of-band though.

Best,
Erick

On Sun, May 18, 2014 at 3:12 AM, Anshum Gupta  wrote:
> SOLR-5468  might be useful
> for you.
>
>
> On Sun, May 18, 2014 at 1:54 AM, adfel70  wrote:
>
>> *one of the most impotent requirements in my system is not to lose docs and
>> not to retrieve part of the data at query time.*
>>
>> I expect the replica to wait until the real leader will start or
>> at least to sync the real leader with the docs indexed in the replica after
>> starting and syncing the replica with the docs that were indexed to the
>> leader.
>>
>> Is there a way that solr can recover without losing docs in this scenario?
>>
>> Thanks.
>>
>>
>>
>> --
>> View this message in context:
>> http://lucene.472066.n3.nabble.com/Replica-as-a-leader-tp4135614p4136729.html
>> Sent from the Solr - User mailing list archive at Nabble.com.
>>
>
>
>
> --
>
> Anshum Gupta
> http://www.anshumgupta.net


Re: Solr Commiter

2014-05-19 Thread Erick Erickson
The short form is "join the community". Shawn did so as he explained.

A _great_ place to contribute is to write unit tests. That's also a
bit "safe" in that it doesn't change Solr but does improve it!

Poke around the build site and look at the coverage reports, pick a
part of Solr that isn't tested and create some.

Improve the documentation. Join the effort to revamp the website. Pick
the part of Solr that was least clear to you and improve it. And on
and on

And I flat guarantee that if you could get to the bottom of the
distributed test failures that seem to occur fairly regularly, you'd
have the undying gratitude of multiple committers :).. Warning though,
that code is fairly hairy!

Best,
Erick


On Fri, May 16, 2014 at 10:05 PM, Shawn Heisey  wrote:
> On 5/15/2014 6:10 AM, Mukundaraman valakumaresan wrote:
>> How to become a solr committer? Any suggestions?
>
> For me, this question has personal relevance.
>
> In 2010, I began to integrate Solr into our environment.  I joined the
> mailing list, asked questions, stumbled around quite a lot.  Eventually
> I got my install working very well, and I discovered that when others
> would ask questions, I sometimes knew the answer, so I started answering
> a lot more questions than I asked.
>
> Eventually, I also joined the dev list, began to learn Java, and started
> contributing patches, mostly to issues that I would file myself, but
> sometimes for other issues.  My name ended up in the CHANGES.txt more
> than once.  A little over a year ago, the Lucene PMC asked me to become
> a committer.  I was not pursuing this as a goal, so it was completely
> unexpected.  I accepted the offer.
>
> My advice would be to put some serious time and effort into making Solr
> better.  As the following wiki page says, this involves a lot more than
> writing code.
>
> http://wiki.apache.org/solr/HowToContribute
>
> Thanks,
> Shawn
>


Re: core.properties setup help

2014-05-19 Thread Erick Erickson
You can actually just remove those entries from solr.xml (and all the
 stuff).

Then create a core.properties with
name=cat
and
name=mcat

respectively.

Actually, you don't even need to put that in. Just put an empty
core.properties file in the respective dirs. The name of the core
defaults to the directory it's in.

Best,
Erick

On Thu, May 15, 2014 at 11:47 PM, Aman Tandon  wrote:
> Any help here.??
>
> With Regards
> Aman Tandon
>
>
> On Thu, May 15, 2014 at 7:33 PM, Aman Tandon wrote:
>
>> Hi,
>>
>> In my solr-4.2 we were using the two cores as described below:
>>
>> 
>>   > hostContext="${hostContext:}" zkClientTimeout="${zkClientTimeout:15000}">
>> 
>> 
>> 
>>
>>
>> How should i setup the solr 4.7 for the core.properties of mcat and cat
>> cores for using the solrcloud?
>>
>> With Regards
>> Aman Tandon
>>


Re: Cloudera Manager install

2014-05-19 Thread Erick Erickson
Michael:

You really have to ask Cloudera that. The Solr project has nothing to
say about what versions of Solr work with Cloudera Manager and what
ones don't.

Best,
Erick

On Thu, May 15, 2014 at 8:17 AM, Michael Della Bitta
 wrote:
> Hi everyone,
>
> I'm investigating migrating over to an HDFS-based Solr Cloud install.
>
> We use Cloudera Manager here to maintain a few other clusters, so
> maintaining our Solr cluster with it as well is attractive. However, just
> from reading the documentation, it's not totally clear to me what
> version(s) of Solr I can install and manage with Cloudera Manager. I saw in
> one place in the documentation an indication that Cloudera Search uses 4.4,
> but then elsewhere I see the opportunity to use custom versions, and
> finally, one indication that Cloudera Manager uses the "latest version."
>
> I'm wondering if anybody has experience with installing a fairly new
> version of Solr, say 4.7 or 4.8, through Cloudera Manager.
>
>
> Michael Della Bitta
>
> Applications Developer
>
> o: +1 646 532 3062
>
> appinions inc.
>
> “The Science of Influence Marketing”
>
> 18 East 41st Street
>
> New York, NY 10017
>
> t: @appinions  | g+:
> plus.google.com/appinions
> w: appinions.com 


Question on 3-level composite-id routing

2014-05-19 Thread Shamik Bandopadhyay
Hi,

  Need some clarification on multilevel composite-id routing in SolrCloud
.I'm currently composite id routing using the following pattern *topic!url* .
This is aimed at run-time de-duplication based on topic field. As I'm
adding support for language search, I felt the need to include language
parameter for better multi-tenancy. Here's the new key structure I'm
thinking of --> *language!topic!url*.
An example would be : english!12345!www.testurl.com

Now, during query time, I'll always have language parameter at my disposal.
I was thinking,of leveraging the shard.key parameter to specify
*shard.keys=language!
*, which will route the request to the right shard and bring back english
content. Is this a valid assumption ?

Also, as per my understanding, the three fields will default to 8,8 and 16
bits of the routing hash. What'll be a valid scenario for providing
a custom allocation of bits for these fields. I was referring to the
following article http://searchhub.org/2014/01/06/10590/ , but was not
entirely sure on this section.


*At query time:*

*To query all records for myapp: shard.keys=myapp/8!*

*Note the explicit mention of 8 bits in case of querying by component 1
only i.e. app level. This is required because the usage of the router as 2
or 3 level isn’t implicit. Specifying ’8′ bits for the component highlights
the use of ’3′ level router*.
Any feedback will be much appreciated.

Thanks,
Shamik


Re: Error when creating collection

2014-05-19 Thread Mark Olsen
Looking on the logs of the other server, I see this:

SEVERE: org.apache.solr.common.SolrException: I was asked to wait on state 
recovering for 192.168.1.155:8983_solr but I still do not see the requested 
state. I see state: active live:true
   at 
org.apache.solr.handler.admin.CoreAdminHandler.handleWaitForStateAction(CoreAdminHandler.java:890)
   at 
org.apache.solr.handler.admin.CoreAdminHandler.handleRequestBody(CoreAdminHandler.java:186)
   at 
org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:135)
   at 
org.apache.solr.servlet.SolrDispatchFilter.handleAdminRequest(SolrDispatchFilter.java:591)
   at 
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:192)
   at 
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:141)


- Original Message -
From: "Mark Olsen" 
To: solr-user@lucene.apache.org
Sent: Monday, May 19, 2014 3:54:30 PM
Subject: Re: Error when creating collection

Shawn,

I set a shard ID however am still getting the same issue. The logs on the solr 
server are reporting the following:

The request is being performed on 192.168.1.155 and it is timing out when 
contacting another server in the zookeeper group (192.168.1.152). There are no 
firewall rules in place and I am able to lynx from 192.168.1.155 to 
192.168.1.152 on that port.

SEVERE: Error while trying to recover. 
core=collection1:org.apache.solr.client.solrj.SolrServerException: Timeout 
occured while waiting response from server at: http://192.168.1.152:8983/solr
at 
org.apache.solr.client.solrj.impl.HttpSolrServer.request(HttpSolrServer.java:412)
at 
org.apache.solr.client.solrj.impl.HttpSolrServer.request(HttpSolrServer.java:181)
at 
org.apache.solr.cloud.RecoveryStrategy.sendPrepRecoveryCmd(RecoveryStrategy.java:202)
at 
org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:346)
at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:223)
Caused by: java.net.SocketTimeoutException: Read timed out

Thanks,
Mark


- Original Message -
From: "Shawn Heisey" 
To: solr-user@lucene.apache.org
Sent: Wednesday, May 14, 2014 12:43:00 PM
Subject: Re: Error when creating collection

On 5/13/2014 4:39 PM, Mark Olsen wrote:
> I'm creating a collection via Java using this function call: 
>
> String collection = "profile-2"; 
> CoreAdminRequest.Create createRequest = new CoreAdminRequest.Create(); 
> createRequest.setCoreName(collection); 
> createRequest.setCollection(collection); 
> createRequest.setInstanceDir(collection); 
> createRequest.setNumShards(1); 
> createRequest.process(server); 
>
> It is timing out with this exception (from the solr.out logs): 
>
> SEVERE: org.apache.solr.common.SolrException: Error CREATEing SolrCore 
> 'profile-2': Could not get shard_id for core: profile-2 
> coreNodeName:192.168.1.152:8983_solr_profile-2 
> at 
> org.apache.solr.handler.admin.CoreAdminHandler.handleCreateAction(CoreAdminHandler.java:483)
>  
> at 
> org.apache.solr.handler.admin.CoreAdminHandler.handleRequestBody(CoreAdminHandler.java:140)
>  
> at 
> org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:135)
>  
> at 
> org.apache.solr.servlet.SolrDispatchFilter.handleAdminRequest(SolrDispatchFilter.java:591)
>  
> at 
> org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:192)
>  
> at 
> org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:141)
>  
> ... 
> Caused by: org.apache.solr.common.SolrException: Could not get shard_id for 
> core: profile-2 coreNodeName:192.168.1.152:8983_solr_profile-2 
> at 
> org.apache.solr.cloud.ZkController.doGetShardIdProcess(ZkController.java:1221)
>  
> at org.apache.solr.cloud.ZkController.preRegister(ZkController.java:1290) 
> at org.apache.solr.core.CoreContainer.registerCore(CoreContainer.java:861) 
>
> In a "development" environment the zookeeper/solr instances are running with 
> elevated permissions and this function worked without error. 
> In a "test" environment (which matches the "production" environment) the 
> permissions are more restricted. I made sure the group/owner of the 
> /usr/local/solr directory are set up to be the correct user. 

This is happening because you never set the shard ID.  See the "Caused
by" message above.  There is a setShardID method on the class that you
are using.  I believe this would typically get something like "shard1"
as a value.

The user that runs Solr must typically have write permissions to the
solr home and all of its descendants.

Note that with the CoreAdminRequest class, you are not creating a
collection.  You are creating a core.  If you want to create an entire
collection (which will typically create at least two cores on different
Solr instances), you need to use CollectionAdminRequest instead.

https://lucene.apache.org/solr/4_8_0/solr-solrj/org/apache/solr/client/solrj/request/CollectionAdminRequest.Create.

Re: Error when creating collection

2014-05-19 Thread Mark Olsen
Shawn,

I set a shard ID however am still getting the same issue. The logs on the solr 
server are reporting the following:

The request is being performed on 192.168.1.155 and it is timing out when 
contacting another server in the zookeeper group (192.168.1.152). There are no 
firewall rules in place and I am able to lynx from 192.168.1.155 to 
192.168.1.152 on that port.

SEVERE: Error while trying to recover. 
core=collection1:org.apache.solr.client.solrj.SolrServerException: Timeout 
occured while waiting response from server at: http://192.168.1.152:8983/solr
at 
org.apache.solr.client.solrj.impl.HttpSolrServer.request(HttpSolrServer.java:412)
at 
org.apache.solr.client.solrj.impl.HttpSolrServer.request(HttpSolrServer.java:181)
at 
org.apache.solr.cloud.RecoveryStrategy.sendPrepRecoveryCmd(RecoveryStrategy.java:202)
at 
org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:346)
at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:223)
Caused by: java.net.SocketTimeoutException: Read timed out

Thanks,
Mark


- Original Message -
From: "Shawn Heisey" 
To: solr-user@lucene.apache.org
Sent: Wednesday, May 14, 2014 12:43:00 PM
Subject: Re: Error when creating collection

On 5/13/2014 4:39 PM, Mark Olsen wrote:
> I'm creating a collection via Java using this function call: 
>
> String collection = "profile-2"; 
> CoreAdminRequest.Create createRequest = new CoreAdminRequest.Create(); 
> createRequest.setCoreName(collection); 
> createRequest.setCollection(collection); 
> createRequest.setInstanceDir(collection); 
> createRequest.setNumShards(1); 
> createRequest.process(server); 
>
> It is timing out with this exception (from the solr.out logs): 
>
> SEVERE: org.apache.solr.common.SolrException: Error CREATEing SolrCore 
> 'profile-2': Could not get shard_id for core: profile-2 
> coreNodeName:192.168.1.152:8983_solr_profile-2 
> at 
> org.apache.solr.handler.admin.CoreAdminHandler.handleCreateAction(CoreAdminHandler.java:483)
>  
> at 
> org.apache.solr.handler.admin.CoreAdminHandler.handleRequestBody(CoreAdminHandler.java:140)
>  
> at 
> org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:135)
>  
> at 
> org.apache.solr.servlet.SolrDispatchFilter.handleAdminRequest(SolrDispatchFilter.java:591)
>  
> at 
> org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:192)
>  
> at 
> org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:141)
>  
> ... 
> Caused by: org.apache.solr.common.SolrException: Could not get shard_id for 
> core: profile-2 coreNodeName:192.168.1.152:8983_solr_profile-2 
> at 
> org.apache.solr.cloud.ZkController.doGetShardIdProcess(ZkController.java:1221)
>  
> at org.apache.solr.cloud.ZkController.preRegister(ZkController.java:1290) 
> at org.apache.solr.core.CoreContainer.registerCore(CoreContainer.java:861) 
>
> In a "development" environment the zookeeper/solr instances are running with 
> elevated permissions and this function worked without error. 
> In a "test" environment (which matches the "production" environment) the 
> permissions are more restricted. I made sure the group/owner of the 
> /usr/local/solr directory are set up to be the correct user. 

This is happening because you never set the shard ID.  See the "Caused
by" message above.  There is a setShardID method on the class that you
are using.  I believe this would typically get something like "shard1"
as a value.

The user that runs Solr must typically have write permissions to the
solr home and all of its descendants.

Note that with the CoreAdminRequest class, you are not creating a
collection.  You are creating a core.  If you want to create an entire
collection (which will typically create at least two cores on different
Solr instances), you need to use CollectionAdminRequest instead.

https://lucene.apache.org/solr/4_8_0/solr-solrj/org/apache/solr/client/solrj/request/CollectionAdminRequest.Create.html

http://wiki.apache.org/solr/SolrTerminology

Thanks,
Shawn



Re: Index / Query IP Address as number.

2014-05-19 Thread Jack Krupansky
Consider an update processor - either raw Java or a snippet of JavaScript 
with the stateless script update processor. The update processor could be 
hard-coded or take parameters as to which source value to examine and what 
field to output. It could use a simple regex to extract only IP addresses. 
And then you could output to multiple fields - one for the raw string for 
wildcard matches, say, and one as an integer for proximity or range checks.


-- Jack Krupansky

-Original Message- 
From: SolrUser1543

Sent: Monday, May 19, 2014 3:04 PM
To: solr-user@lucene.apache.org
Subject: Re: Index / Query IP Address as number.

I have a text field containing a large piece of mixed text , like :

test test 12/12/2001 12345 192.168.1.1 1234324


I need to  create a copy field which will capture only all IPs from the text
( may be more than one IP ) .

What will be the best way to do ?

I dont see any option to make WordDelimiter to not break down the IP , so as
alternative I will use a copy field .




--
View this message in context: 
http://lucene.472066.n3.nabble.com/Index-Query-IP-Address-as-number-tp4136760p4136974.html
Sent from the Solr - User mailing list archive at Nabble.com. 



Re: Issue paging when sorting on a Date field

2014-05-19 Thread Chris Hostetter

: Using Solr 4.6.1 and in my schema I have a date field storing the time a
: document was added to Solr.

what *exactly* does your schema look like?  are you using "solr.DateField" 
or "solr.TrieDateField" ? what field options do you have specified?

: I have a utility program which:
: - queries for all of the documents in the previous day sorted by create date
: - pages through the results keeping track of the unique document ids
: - compare the total number of unique doc ids to the numFound to see if it
: they match

what *exactly* do your queries look like?  show us some examples please 
(URL & results).  Are you using distributed searching across multiple 
nodes, or a single node?  do you have concurrent updates going on during 
your test?

: It is not consistent between tests, the number of occurrences changes and
: the locations of the occurrences can change as well. The larger the result
: set, and smaller the page size, the more frequent the occurrences are.

if you bring up a test instance of Solr using your current configs, can 
you reproduce (even occasionally) with some synthetic data you can share 
with us?  If so please provide your full configs & sample data (ie: create 
a Jira & attach all the neccessary files i na ZIP)


-Hoss
http://www.lucidworks.com/


Issue paging when sorting on a Date field

2014-05-19 Thread Bryan Bende
Using Solr 4.6.1 and in my schema I have a date field storing the time a
document was added to Solr.

I have a utility program which:
- queries for all of the documents in the previous day sorted by create date
- pages through the results keeping track of the unique document ids
- compare the total number of unique doc ids to the numFound to see if it
they match

I've noticed that if I use a page size larger than the number of documents
for the given day (aka get everything in one query), then everything works
as expected (results sorted correctly, unique doc ids size == numFound).

However, when I use a smaller page say, say 10 rows per page, I randomly
see cases where the last document of a page will be duplicated as the first
document of the next page, even though the "start" and "rows" parameters
increased correctly. So I might see something like numFound=100 but unique
doc ids is 97, and then I see three occurrences where the last doc id on a
page was also the first on the next page.

It is not consistent between tests, the number of occurrences changes and
the locations of the occurrences can change as well. The larger the result
set, and smaller the page size, the more frequent the occurrences are.

The only thing I have noticed is that if I change the sorting of the
initial query to use a non-date field, then this doesn't happen anymore.

Are there any know issues/limitations sorting/paging on a date field ?

The only mention I can find is this thread:
http://mail-archives.apache.org/mod_mbox/lucene-solr-user/200909.mbox/%3c57912a0644b6ab4381816de07cb1c38d02a00...@s2na1excluster.na1.ad.group%3E


Re: Index / Query IP Address as number.

2014-05-19 Thread SolrUser1543
I have a text field containing a large piece of mixed text , like : 

test test 12/12/2001 12345 192.168.1.1 1234324


I need to  create a copy field which will capture only all IPs from the text
( may be more than one IP ) . 

What will be the best way to do ? 

I dont see any option to make WordDelimiter to not break down the IP , so as
alternative I will use a copy field . 




--
View this message in context: 
http://lucene.472066.n3.nabble.com/Index-Query-IP-Address-as-number-tp4136760p4136974.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: Index / Query IP Address as number.

2014-05-19 Thread SolrUser1543
I dont have autogeneratephrasequeries set to true .  I tried both false /
true for it  , but nothing changed

Capture.JPG   

the same chain defined for both query / index : 


  
  
" />
   

  



--
View this message in context: 
http://lucene.472066.n3.nabble.com/Index-Query-IP-Address-as-number-tp4136760p4136971.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: trigger delete on nested documents

2014-05-19 Thread Mikhail Khludnev
Thomas,

Vanilla way to override a blocks is to send it with the same unique-key (I
guess it's "id" for your case, btw don't you have unique-key defined in the
schema?), but it must have at least one child. It seems like analysis issue
to me https://issues.apache.org/jira/browse/SOLR-5211

While block is indexed the special field _root_ equal to the 
is added across the whole block (caveat, it's not stored by default). At
least you can issue

_root_:PK_VAL

to wipe the whole block.

19.05.2014 10:37 пользователь "Thomas Scheffler" <
thomas.scheff...@uni-jena.de> написал:

> Hi,
>
> I plan to use nested documents to group some of my fields
>
> 
> art0001
> My first article
>   
> art0001-foo
> Smith, John
> author
>   
>   
> art0001-bar
> Power, Max
> reviewer
>   
> 
>
> This way can ask for any documents that are reviewed by Max Power. However
> to simplify update and deletes I want to ensure that nested documents are
> deleted automatically on update and delete of the parent document.
> Does anyone had to deal with this problem and found a solution?
>
> regards,
>
> Thomas
>


Re: Slow file-import

2014-05-19 Thread Ahmet Arslan
Hi,

Try http://wiki.apache.org/solr/UpdateCSV , it should be faster. 
See 'Tab-delimited importing' at the end of the wiki page.

Cheers,
Ahmet

On Monday, May 19, 2014 1:31 PM, Hal Arres  wrote:



Hallo there,

I am working on an import-configuration for my solr-index and I got
some issues with that.

In the first step I configured an import-handler to import data from a
database into the solr-index and it worked just fine, but it is very
slow (7K documents per second). So I wanted to change that towards a
data-import-handler using a FileDataSource. (i am running solr 4.6.1)

I have to import nearly 150_000_000 lines each night and each line has
the following characteristics:
- fields are seperated by tabulator
- 70 fields each line
- one line is nearly 600 characters long
- each line contains multiple data-types (date, int, string...)

In the moment the files are imported into the database, from which
they are imported by solr (database import-handler).
To improve the import performance I wanted to import the files directly.


This is the first approach I tested:
---
        
            
                
            
-
If I import less than 10 fields this works just fine. But as soon as I
extend the import to 30 fields, the time to import one line increases
to more than 10sec!


So I checked another way, in which I moved the transformation to a script:


-
But this was just slightly faster than the database import.


Has someone of you an idea, how I can improve my import performance?

Thank you very, very much,
Sebastian


Re: Custom filter not working with solr 4.7.1

2014-05-19 Thread Shawn Heisey
On 5/19/2014 1:10 AM, Kamal Kishore Aggarwal wrote:
> I have created a custom filter factory for solr 4.2. It is working good.
> But when I am trying to upgarde solr-4.2 to 4.7 version, it is reporting
> errors.
>
> Caused by: org.apache.solr.common.SolrException: Plugin init failure for
> [schema.xml] analyzer/filter: Error instantiating class:
> 'org.apache.lucene.analysis.ExtendedNameFilterFactory'
>
> Here's the java code :http://pastebin.com/REu6cJxR,
> http://pastebin.com/380YZaAM
>
> It was working good with solr 4.2. Can anybody tell me the changes that I
> need to make for running it in solr 4.7.1?

I probably would have used something like com.company.lucene.filter as
the package name, but what you name the package is up to you, and will
have no effect on your problem.

I dropped your code into a trunk checkout (the code that will eventually
become Lucene/Solr 5.0).  The Eclipse editor looked like this for your
factory class:

https://www.dropbox.com/s/tobjrk0riq7vqqt/extendednamefilterfactory.png

The errors shown for the three red marks are:

* Implicit super constructor TokenFilterFactory() is undefined for
default constructor. Must define an explicit constructor.
* The method init(Map) is undefined for the type
TokenFilterFactory.
* The method getInt(Map, String, int) in the type
AbstractAnalysisFactory is not applicable for the arguments (String, int)

I don't have easy access to a 4.x checkout right this minute, but the
errors there are probably similar.

Your actual filter class was all good except for three eclipse warnings
that are fixed by adding a type argument of  to the code.

Thanks,
Shawn



Re: Custom filter not working with solr 4.7.1

2014-05-19 Thread Kamal Kishore Aggarwal
I am new to solr. I am not getting your answer. Can you please explain in
more detail.


On Mon, May 19, 2014 at 12:40 PM, Kamal Kishore Aggarwal <
kkroyal@gmail.com> wrote:

> Dear Team,
>
> I have created a custom filter factory for solr 4.2. It is working good.
> But when I am trying to upgarde solr-4.2 to 4.7 version, it is reporting
> errors.
>
> Caused by: org.apache.solr.common.SolrException: Plugin init failure for
> [schema.xml] analyzer/filter: Error instantiating class:
> 'org.apache.lucene.analysis.ExtendedNameFilterFactory'
>
> Here's the java code :http://pastebin.com/REu6cJxR,
> http://pastebin.com/380YZaAM
>
> It was working good with solr 4.2. Can anybody tell me the changes that I
> need to make for running it in solr 4.7.1?
>
>
> With Regards & Thanks
>
> Kamal Kishore
>
>
>


Re: Custom filter not working with solr 4.7.1

2014-05-19 Thread Kamal Kishore Aggarwal
I am new to solr. I am not getting your answer. Can you please explain in
more detail.


On Mon, May 19, 2014 at 4:02 PM, Kamal Kishore Aggarwal <
kkroyal@gmail.com> wrote:

> I am new to solr. I am not getting your answer. Can you please explain in
> more detail.
>
>
> On Mon, May 19, 2014 at 12:40 PM, Kamal Kishore Aggarwal <
> kkroyal@gmail.com> wrote:
>
>> Dear Team,
>>
>> I have created a custom filter factory for solr 4.2. It is working good.
>> But when I am trying to upgarde solr-4.2 to 4.7 version, it is reporting
>> errors.
>>
>> Caused by: org.apache.solr.common.SolrException: Plugin init failure for
>> [schema.xml] analyzer/filter: Error instantiating class:
>> 'org.apache.lucene.analysis.ExtendedNameFilterFactory'
>>
>> Here's the java code :http://pastebin.com/REu6cJxR,
>> http://pastebin.com/380YZaAM
>>
>> It was working good with solr 4.2. Can anybody tell me the changes that I
>> need to make for running it in solr 4.7.1?
>>
>>
>> With Regards & Thanks
>>
>> Kamal Kishore
>>
>>
>>
>


Re: score retrieval performance

2014-05-19 Thread Gora Mohanty
On 19 May 2014 19:52, Michael Ryan  wrote:
> Is there any significant difference in query speed when retrieving the score 
> pseudo-field? E.g., does...
>
> q=foo&sort=date+desc&fl=*,score
>
> ...take longer to run than...
>
> q=foo&sort=date+desc&fl=*
>
> I know there's different code paths in Solr depending on whether the score is 
> needed or not, but not sure how much slower the score path is...

If you are using the q parameter, the score is being calculated in any
case, and while we
have never measured this, I would imagine that the performance hit
from retrieving the score
would be miniscule.

Are you perchance thinking of using filter queries using the fq
parameter? As filter queries are
cached, using these can indeed make a difference.

Regards,
Gora


score retrieval performance

2014-05-19 Thread Michael Ryan
Is there any significant difference in query speed when retrieving the score 
pseudo-field? E.g., does...

q=foo&sort=date+desc&fl=*,score

...take longer to run than...

q=foo&sort=date+desc&fl=*

I know there's different code paths in Solr depending on whether the score is 
needed or not, but not sure how much slower the score path is...

-Michael


Re: Solr 4.7.2 Core Creation Issue on SC - ZK

2014-05-19 Thread Shawn Heisey
On 5/19/2014 2:31 AM, deniz wrote:
> I am using SC with version 4.7.2. There is already one collection/core
> running on the cloud, and I am trying to add a new core according to
> https://wiki.apache.org/solr/CoreAdmin#CREATE 
> 
> When I add the core, I can see that it is added to collections in Cloud/file
> menu, but the config part still shows the existing core, so the new one that
> I add shares the same config. So the result looks like this:
> 
>   /collections
>   newcore
>   oldcore
>   /configs
>  oldcore
> 
> Are there any other settings that I need to do to see my newcore's configs
> on the cloud? In this way newcore is only seeing oldcore settings, which is
> not something wanted...

If SC means SolrCloud (which it probably does, because you then mention
cloud), then you should not be using the CoreAdmin API, you should be
using the CollectionsAdmin API.

https://cwiki.apache.org/confluence/display/solr/Collections+API

The basic operation for a long-term sustainable SolrCloud is:

* Start a zookeeper ensemble, minimum 3 hosts.
* Start the cloud members with an appropriate -DzkHost parameter
referencing all zookeeper hosts.
* Upload one or more configuration sets to zookeeper with zkCli.
* Create one or more collections with the Collections API.

After you already have a collection, it may make sense to use the
CoreAdmin API to add additional replicas to it, but creating an entire
collection from scratch with the CoreAdmin API is not something you
should try until you see how SolrCloud does it with the Collections API
... and most people will *never* need to do it that way.

Some other pages that will probably be helpful:

https://wiki.apache.org/solr/SolrTerminology
http://wiki.apache.org/solr/SolrCloud#Managing_collections_via_the_Collections_API
https://cwiki.apache.org/confluence/display/solr/Command+Line+Utilities

Thanks,
Shawn



Re: Subscribe for Solr Mails

2014-05-19 Thread Jack Krupansky

Read this web page for information about subscribing:
https://lucene.apache.org/solr/discussion.html

-- Jack Krupansky

-Original Message- 
From: Kamal Kishore Aggarwal 
Sent: Monday, May 19, 2014 6:12 AM 
To: solr-user@lucene.apache.org 
Subject: Subscribe for Solr Mails 


Dear Team,


I am new to solr. Please subscribe me for solr mails.


With Regards & Thanks
Kamal Kishore


How does query on AND work

2014-05-19 Thread Per Steffensen

Hi

Lets say I have a Solr collection (running across several servers) 
containing 5 billion documents. A.o. each document have a value for 
field "no_dlng_doc_ind_sto" (a long) and field 
"timestamp_dlng_doc_ind_sto" (also a long). Both "no_dlng_doc_ind_sto" 
and "timestamp_dlng_doc_ind_sto" are doc-value, indexed and stored. Like 
this in schema.xml
stored="true" required="true" docValues="true"/>
positionIncrementGap="0" docValuesFormat="Disk"/>


I make queries like this: no_dlng_doc_ind_sto:() AND 
timestamp_dlng_doc_ind_sto:([ TO ])
* The "no_dlng_doc_ind_sto:()"-part of a typical query will hit 
between 500 and 1000 documents out of the total 5 billion
* The "timestamp_dlng_doc_ind_sto:([ TO ])"-part 
of a typical query will hit between 3-4 billion documents out of the 
total 5 billion


Question is how Solr/Lucene deals with such requests?
I am thinking that using the indices on both "no_dlng_doc_ind_sto" and 
"timestamp_dlng_doc_ind_sto" to get two sets of doc-ids and then make an 
intersection of those might not be the most efficient. You are making an 
intersection of two doc-id-sets of size 500-1000 and 3-4 billion. It 
might be faster to just use the index for "no_dlng_doc_ind_sto" to get 
the doc-ids for the 500-1000 documents, then for each of those fetch 
their "timestamp_dlng_doc_ind_sto"-value (using doc-value) to filter out 
the ones among the 500-1000 that does not match the timestamp-part of 
the query.
But what does Solr/Lucene actually do? Is it Solr- or Lucene-code that 
make the decision on what to do? Can you somehow "hint" the 
search-engine that you want one or the other method used?


Solr 4.4 (and corresponding Lucene), BTW, if that makes a difference

Regards, Per Steffensen


Re: How to delete docs only in a spec shard within a collection

2014-05-19 Thread Jack Krupansky
Normally a delete-by-query needs to be distributed to all shards since it is 
not known which document keys it will affect.


It sounds like SolrCloud does not support the feature you would really like: 
support for distrb=false.


You can file a Jira request for a feature "improvement."

-- Jack Krupansky

-Original Message- 
From: YouPeng Yang

Sent: Monday, May 19, 2014 3:46 AM
To: solr-user@lucene.apache.org
Subject: Re: How to delete docs only in a spec shard within a collection

Hi
 Anyone gives some suggestions.


Regards


2014-05-19 11:31 GMT+08:00 YouPeng Yang :


Hi
  I have a colloection with 3 shards.
  I want to delete some docs in one shard with the command:


http://10.1.22.1:8082/solr/tv_201402/update?&stream.body=BEGINTIME:["2014-03-01
00:00:00" TO *]&shards=tv_201402&commit=true
 As the red font exepression, It is supposed that docs only in the shard
tv_201402  whose BEGINTIME field were larger than "2014-03-01 00:00:00"
would be deleted.However ,as a result, the docs on the other shards such
astv_201403 ,tv_201404 whose BEGINTIME were also larger than 2014-03-01
00:00:00  were deleted too.
  This is not the effientlty right thing I want.

  How to delete docs only in a spec shard  within a collection which
contains several shards?

Regards.





Slow file-import

2014-05-19 Thread Hal Arres
Hallo there,

I am working on an import-configuration for my solr-index and I got
some issues with that.

In the first step I configured an import-handler to import data from a
database into the solr-index and it worked just fine, but it is very
slow (7K documents per second). So I wanted to change that towards a
data-import-handler using a FileDataSource. (i am running solr 4.6.1)

I have to import nearly 150_000_000 lines each night and each line has
the following characteristics:
- fields are seperated by tabulator
- 70 fields each line
- one line is nearly 600 characters long
- each line contains multiple data-types (date, int, string...)

In the moment the files are imported into the database, from which
they are imported by solr (database import-handler).
To improve the import performance I wanted to import the files directly.


This is the first approach I tested:
---




-
If I import less than 10 fields this works just fine. But as soon as I
extend the import to 30 fields, the time to import one line increases
to more than 10sec!


So I checked another way, in which I moved the transformation to a script:


-
But this was just slightly faster than the database import.


Has someone of you an idea, how I can improve my import performance?

Thank you very, very much,
Sebastian


Subscribe for Solr Mails

2014-05-19 Thread Kamal Kishore Aggarwal
Dear Team,


I am new to solr. Please subscribe me for solr mails.


With Regards & Thanks
Kamal Kishore


Re: How to delete docs only in a spec shard within a collection

2014-05-19 Thread YouPeng Yang
Hi
  I have checked the src in
DistributedUpdateProcessor.doDeleteByQuery(DeleteUpdateCommand cmd) as[1]

  According to the source code,it seems that  ShardParams._ROUTE_ (_route_)
and the ShardParams.SHARD_KEYS(shard.keys)
would be used to find the right shard.
  Compared to my last two posts,the shards is not the right properity to
identify the right shards that really need to delete the docs.


  Am I right?
  Does it need to add the processeing  about the shards to unify the
process?


[1]---
..omitted..
  String route = params.get(ShardParams._ROUTE_);
  if(route == null) route = params.get(ShardParams.SHARD_KEYS);//
deprecated . kept for backcompat
  Collection slices = coll.getRouter().getSearchSlices(route,
params, coll);
  List leaders =  new ArrayList<>(slices.size());
  for (Slice slice : slices) {
 ..omitted..
 cmdDistrib.distribDelete(cmd, leaders, outParams);
 ..omitted..
   }

..omitted..

-


Regards.


2014-05-19 15:46 GMT+08:00 YouPeng Yang :

> Hi
>   Anyone gives some suggestions.
>
>
> Regards
>
>
> 2014-05-19 11:31 GMT+08:00 YouPeng Yang :
>
> Hi
>>   I have a colloection with 3 shards.
>>   I want to delete some docs in one shard with the command:
>>
>>   
>> http://10.1.22.1:8082/solr/tv_201402/update?&stream.body=BEGINTIME:["2014-03-01
>> 00:00:00" TO *]&shards=tv_201402&commit=true
>>  As the red font exepression, It is supposed that docs only in the shard
>> tv_201402  whose BEGINTIME field were larger than "2014-03-01 00:00:00"
>> would be deleted.However ,as a result, the docs on the other shards such
>> astv_201403 ,tv_201404 whose BEGINTIME were also larger than 2014-03-01
>> 00:00:00  were deleted too.
>>   This is not the effientlty right thing I want.
>>
>>   How to delete docs only in a spec shard  within a collection which
>> contains several shards?
>>
>> Regards.
>>
>
>


Solr 4.7.2 Core Creation Issue on SC - ZK

2014-05-19 Thread deniz
Hello,

I am using SC with version 4.7.2. There is already one collection/core
running on the cloud, and I am trying to add a new core according to
https://wiki.apache.org/solr/CoreAdmin#CREATE 

When I add the core, I can see that it is added to collections in Cloud/file
menu, but the config part still shows the existing core, so the new one that
I add shares the same config. So the result looks like this:

  /collections
  newcore
  oldcore
  /configs
 oldcore

Are there any other settings that I need to do to see my newcore's configs
on the cloud? In this way newcore is only seeing oldcore settings, which is
not something wanted...



-
Zeki ama calismiyor... Calissa yapar...
--
View this message in context: 
http://lucene.472066.n3.nabble.com/Solr-4-7-2-Core-Creation-Issue-on-SC-ZK-tp4136833.html
Sent from the Solr - User mailing list archive at Nabble.com.


RE: Is there a way to change transientCacheSize dynamically without restarting Solr

2014-05-19 Thread Elran Dvir
Thank you, Shalin.
-Original Message-
From: Shalin Shekhar Mangar [mailto:shalinman...@gmail.com] 
Sent: Monday, May 19, 2014 10:18 AM
To: solr-user@lucene.apache.org
Subject: Re: Is there a way to change transientCacheSize dynamically without 
restarting Solr

Hi Elran,

No, I don't think there is way to do that right now.


On Mon, May 19, 2014 at 12:18 PM, Elran Dvir  wrote:

> Does anyone have an answer for this?
> Thanks.
>
> From: Elran Dvir
> Sent: Thursday, May 15, 2014 12:32 PM
> To: solr-user@lucene.apache.org
> Subject: Is there a way to change transientCacheSize dynamically 
> without restarting Solr
>
> Hi All,
>
> Is there an API in Solr to change transientCacheSize dynamically 
> without the need to restart Solr?
> Are there other Solr configuration parameters that can be changed 
> dynamically?
>
> Thanks.
>



--
Regards,
Shalin Shekhar Mangar.


Email secured by Check Point


RE: Reloading core with param vs unloading and creating with params

2014-05-19 Thread Elran Dvir
Hi,

Did anyone have a chance to take a look at this issue?

Thanks.

From: Elran Dvir
Sent: Thursday, May 15, 2014 12:22 PM
To: solr-user@lucene.apache.org
Subject: Reloading core with param vs unloading and creating with params

Hi all,

I created a new patch https://issues.apache.org/jira/browse/SOLR-6063 , 
enabling changes in core properties without the need to unload and create it.
Considering the change in patch,
is reloading a core with transient=true and loadOnStartup=false equivalent in 
memory footprint to unloading the core and creating it with the same parameters?
Thanks.


Re: How to delete docs only in a spec shard within a collection

2014-05-19 Thread YouPeng Yang
Hi
  Anyone gives some suggestions.


Regards


2014-05-19 11:31 GMT+08:00 YouPeng Yang :

> Hi
>   I have a colloection with 3 shards.
>   I want to delete some docs in one shard with the command:
>
>   
> http://10.1.22.1:8082/solr/tv_201402/update?&stream.body=BEGINTIME:["2014-03-01
> 00:00:00" TO *]&shards=tv_201402&commit=true
>  As the red font exepression, It is supposed that docs only in the shard
> tv_201402  whose BEGINTIME field were larger than "2014-03-01 00:00:00"
> would be deleted.However ,as a result, the docs on the other shards such
> astv_201403 ,tv_201404 whose BEGINTIME were also larger than 2014-03-01
> 00:00:00  were deleted too.
>   This is not the effientlty right thing I want.
>
>   How to delete docs only in a spec shard  within a collection which
> contains several shards?
>
> Regards.
>


Re: deep paging without sorting / keep IRs open

2014-05-19 Thread Tommaso Teofili
thanks Yonik, that looks promising, I'll have a look at it.

Tommaso


2014-05-17 17:57 GMT+02:00 Yonik Seeley :

> On Sat, May 17, 2014 at 10:30 AM, Yonik Seeley 
> wrote:
> > I think searcher leases would fit the bill here?
> > https://issues.apache.org/jira/browse/SOLR-2809
> >
> > Not yet implemented though...
>
> FYI, I just put up a simple LeaseManager implementation on that issue.
>
> -Yonik
> http://heliosearch.org - facet functions, subfacets, off-heap
> filters&fieldcache
>


Re: Replica active during warming

2014-05-19 Thread lboutros
Hi Erick,

I do not pass the LBHttpSolrServer to the c'tor of CloudSolrServer.

thx,

Ludovic.



-
Jouve
France.
--
View this message in context: 
http://lucene.472066.n3.nabble.com/Replica-active-during-warming-tp4135274p4136828.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: Custom filter not working with solr 4.7.1

2014-05-19 Thread Alexandre Rafalovitch
I think the init method signatures changed. Just compare the same
factory across two Solr versions and you will see.

Regards,
   Alex.
Personal website: http://www.outerthoughts.com/
Current project: http://www.solr-start.com/ - Accelerating your Solr proficiency


On Mon, May 19, 2014 at 2:10 PM, Kamal Kishore Aggarwal
 wrote:
> Dear Team,
>
> I have created a custom filter factory for solr 4.2. It is working good.
> But when I am trying to upgarde solr-4.2 to 4.7 version, it is reporting
> errors.
>
> Caused by: org.apache.solr.common.SolrException: Plugin init failure for
> [schema.xml] analyzer/filter: Error instantiating class:
> 'org.apache.lucene.analysis.ExtendedNameFilterFactory'
>
> Here's the java code :http://pastebin.com/REu6cJxR,
> http://pastebin.com/380YZaAM
>
> It was working good with solr 4.2. Can anybody tell me the changes that I
> need to make for running it in solr 4.7.1?
>
>
> With Regards & Thanks
>
> Kamal Kishore


Re: Is there a way to change transientCacheSize dynamically without restarting Solr

2014-05-19 Thread Shalin Shekhar Mangar
Hi Elran,

No, I don't think there is way to do that right now.


On Mon, May 19, 2014 at 12:18 PM, Elran Dvir  wrote:

> Does anyone have an answer for this?
> Thanks.
>
> From: Elran Dvir
> Sent: Thursday, May 15, 2014 12:32 PM
> To: solr-user@lucene.apache.org
> Subject: Is there a way to change transientCacheSize dynamically without
> restarting Solr
>
> Hi All,
>
> Is there an API in Solr to change transientCacheSize dynamically without
> the need to restart Solr?
> Are there other Solr configuration parameters that can be changed
> dynamically?
>
> Thanks.
>



-- 
Regards,
Shalin Shekhar Mangar.


Custom filter not working with solr 4.7.1

2014-05-19 Thread Kamal Kishore Aggarwal
Dear Team,

I have created a custom filter factory for solr 4.2. It is working good.
But when I am trying to upgarde solr-4.2 to 4.7 version, it is reporting
errors.

Caused by: org.apache.solr.common.SolrException: Plugin init failure for
[schema.xml] analyzer/filter: Error instantiating class:
'org.apache.lucene.analysis.ExtendedNameFilterFactory'

Here's the java code :http://pastebin.com/REu6cJxR,
http://pastebin.com/380YZaAM

It was working good with solr 4.2. Can anybody tell me the changes that I
need to make for running it in solr 4.7.1?


With Regards & Thanks

Kamal Kishore


Re: date range queries efficiency

2014-05-19 Thread Dmitry Kan
Thanks, Jack, Alex and Shawn.

This makes proper sense. One win of rounding down on indexing side is
saving index space, according to hoss (reply over IRC):

"with the TrieDateFields, rounding dates at indexing time won't have any
effect on the cachability of the rounded queries, and even for non cached
queries it shouldn't affect the performance much -- but yes, it would help
reduce index size"

I haven't tried it myself, just thought to ask if somebody tried it already.

Dmitry


On Sat, May 17, 2014 at 8:37 AM, Shawn Heisey  wrote:

> On 5/15/2014 1:34 AM, Alexandre Rafalovitch wrote:
> > I thought the date math rounding was for _caching_ the repeated
> > queries, not so much the speed of the query itself.
>
> Absolutely correct.  When NOW is used without rounding, caching is
> completely ineffective.  This is because if the same query using NOW is
> sent multiple times several seconds apart, every one of those queries
> will be different after they are parsed and NOW is converted to an
> actual timestamp.
>
> > Also, if you are using TrieDateField, precisionStep value is how
> > optimization is done. There is bucketing at different level of
> > precision, so the range search works at the least granular level
> > first, etc.
>
> Some nitty-gritty details of how range queries are accelerated with the
> Trie data types and precisionStep are described in the Javadoc for
> NumericRangeQuery:
>
>
> http://lucene.apache.org/core/4_8_0/core/org/apache/lucene/search/NumericRangeQuery.html
>
> Thanks,
> Shawn
>
>


-- 
Dmitry Kan
Blog: http://dmitrykan.blogspot.com
Twitter: http://twitter.com/dmitrykan