Hi,
I think stock example from official ref guide will do the trick.
Please see :
https://cwiki.apache.org/confluence/display/solr/Tokenizers#Tokenizers-RegularExpressionPatternTokenizer
Ahmet
On Tuesday, May 17, 2016 8:31 AM, SRINI SOLR wrote:
Hi Ahmet /
Hi Ahmet / Team -
Thanks for your quick response...
Can you please help me out on this PatternTokenizer configuration...
Here we are using configuration as below ...
And also - I have made changes to the field value so that it is separated
by space instead of commas and
bq: One thing that still feels a bit odd though is that the health
check query was referencing a collection that no longer existed in the
cluster. So it seems like it was downloading the state for ALL
non-hosted collections, not a requested one.
This is a bit odd, I don't know whether there's
Ah, I tracked this down to an haproxy that was set up on a load server during
development and still running. It was configured with a health check every 10
seconds, so that’s pretty clearly the cause. Thanks for the pointer.
One thing that still feels a bit odd though is that the health check
With the per-collection state.json, if "something" goes to a node that doesn't
host a replica for a node, it downloads the state for the "other"
collection then
throws it away.
In this case, "something" is apparently asking the nodes hosting collectionA to
do "something" with collections B and/or
Not enough info. What is the field type for that string? 10-to-1
it's some kind of numeric type.
Best,
Erick
On Mon, May 16, 2016 at 2:14 PM, Boman wrote:
> Something's not right.
> org.apache.solr.client.solrj.impl.CloudSolrClient$RouteException: Error from
> server at
Also, depending on your use cases, whether stemming
would do what you want.
On Mon, May 16, 2016 at 2:27 PM, Ahmet Arslan wrote:
> Hi Gates,
>
> There are two approaches:
>
> 1) Use a wildcard query with star operator q=consult*
>
> 2) Create an index with
OK, this is very strange. There's no _good_ reason that
restarting the servers should make a difference. The fact
that it took 1/2 hour leads me to believe, though, that your
shards are somehow "incomplete", especially that you
are indexing to the system and don't have, say,
your autocommit
On 5/16/2016 6:29 PM, Li Ding wrote:
> This happened when the second time I'm performing restart. But after that,
> every time this collection is stuck at here. If I restart the leader node
> as well, the core can get out of the recovering state
>
> On Mon, May 16, 2016 at 5:00 PM, Li Ding
OK! Thanks for reminding. I'll stick to the convention.
scott.chu,scott@udngroup.com
2016/5/17 (週二)
- Original Message -
From: Chris Hostetter
To: solr-user ; scott(自己)
CC:
Date: 2016/5/17 (週二) 02:43
Subject: Re: [scottchu] Cab I migrate solrcloud by just copying whole
This happened when the second time I'm performing restart. But after that,
every time this collection is stuck at here. If I restart the leader node
as well, the core can get out of the recovering state
On Mon, May 16, 2016 at 5:00 PM, Li Ding wrote:
> Hi Anshum,
>
>
Hi Anshum,
This is for restart solr with 1000 collections. I created an environment
with 1023 collections today All collections are empty. During repeated
restart test, one of the cores are marked as "recovering" and stuck there
for ever. The solr is 4.6.1 and we have 3 zk hosts and 8 solr
Just one more note - while experimenting, I found that if I stopped all nodes
(full cluster shutdown), and then startup all nodes, they do in fact seem to
repair themselves then. We have a script to monitor the differences between
replicas (just looking at numDocs) and before the full shutdown
Hi Solr mailling list,
This is a non-technical question but please let me know if I should post it
somewhere else.
I really don't think any place other than the Solr mailling list will be a
better place to post this type of question.
The Arrow Big Data team desperately is seeking for a talented
Each node has one JVM with 16GB of RAM. Are you suggesting we would put each
shard into a separate JVM (something like 32 nodes)?
We aren't encountering any OOMs. We are testing this in a separate cloud which
no one is even using, the only activity is this very small amount of indexing
and
Hi Gates,
There are two approaches:
1) Use a wildcard query with star operator q=consult*
2) Create an index with EdgeNGramFilterFactory and issue a regular search
q=consult
(2) will be faster at the cost of bigger index size
You don't need to change anything for (1) if the execution time is
Thank you very much Erick and Chris.
On 12 May 2016 at 03:57, Erick Erickson wrote:
> I just added you to the contributors group, you should be able to post now.
>
> On Wed, May 11, 2016 at 4:22 PM, Chris Hostetter
> wrote:
> >
> > If you
Something's not right.
org.apache.solr.client.solrj.impl.CloudSolrClient$RouteException: Error from
server at http://10.204.10.19:8983/solr/00D61007927_shard1_replica2:
java.lang.NumberFormatException: For input string: "A1"
at
Hi
Wondering if someone can guide me to how I can search words by part.
ie: how to return the word ‘consultation' by say entering a query with just the
word ‘consult’.
How does one do this in Solr ?
Thanks,
Mark
Hi,
I'm hoping someone has seen/encountered a similar problem. We have
solr instances with all Jetty threads in BLOCKED state. The
application does not respond to any http requests.
It is SOLR 4.9 running inside docker on Amazon EC2. Jetty is 8.1 and
there is an nginx proxy in front of it (with
I think you are approaching the problem all wrong. This seems, what is
described as an x-y problem (https://people.apache.org/~hossman/#xyproblem).
Can you tell us more about :
* What's your setup like? SolrCloud - Version, number of shards, is there
any custom code, etc.
* Did you start seeing
Hi all,
We have an unique scenario where we don't need leaders in every collection
to recover from failures. The indexing never changes. But we have faced
problems where either zk marked a core as down while the core is fine in
non-distributed query or during restart, the core never comes up.
Hi, I am having trouble performing a search across multiple cores on a single
server running 5.4.0
I have erased, rebuilt & optimized the indexes, and and according to the
schema browser for both cores, every document has a unique key (id). However I
am still getting the same error. I would
: Message-Id: <7fd5fd02628b55831193271f9b39a...@udngroup.com>
: Subject: [scottchu] Cab I migrate solrcloud by just copying whole package
: folder?
: References:
: ,
: <856da447-e7b7-49a4-af87-f161b0fe5...@elyograg.org>
: In-Reply-To:
: Is there a way to add a flag (or name value pair to Json resonse) in result
: being returned from Solr based on some security handling which happens in
: custom post filter code?
:
: How about if in custom post filter code I added the information to request
: context and then in a custom
Is there a way to add a flag (or name value pair to Json resonse) in result
being returned from Solr based on some security handling which happens in
custom post filter code?
How about if in custom post filter code I added the information to request
context and then in a custom response builder,
Hmmm... is baseDir empty before you call new MiniSolrCloudCluster ?
My best guess is you are getting bit by this...
https://issues.apache.org/jira/browse/SOLR-8999
: I am only setting up a MiniSolrCloudCluster with 2 servers like this:
:
: JettyConfig jettyConfig =
The scenario you described should be done with the suggester component.
Nothing prevent you to configure multiple dictionaries for the suggester as
well.
The you pass the dictionary to the suggester at query time as a request
parameter for your suggester request handler.
Cheers
On 16 May 2016
I have a solr 5.4 cluster with three collections, A, B, C.
Nodes either host replicas for collection A, or B and C. Collections B and C
are not currently used - no inserts or queries. Collection A is getting
significant query traffic, but no insert traffic, and queries are only directed
to
Thanks Renaud.
Here is my setup:
1- I have created 2 sites: Main (source) and DR (traget).
2- Both sites are the same before configuring CDCR
3- The collections (source and target) are created before configuring CDCR
4- collections are created using interactive mode: accepting most defaults
The following message addressed to you was quarantined because it likely
contains a virus:
Subject: RE: Need Help with Solr 6.0 Cross Data Center Replication
From: Satvinder Singh
However, if you know the sender and are expecting an attachment, please reply
to this
Thank you.
To summarize this is what I have, all VMS running on Centos7 :
Source Side
|___ 1 VM running 3 Zookeeper instances on port 2181, 2182 and 2183
(ZOOKEEPER 3.4.8)(Java 1.8.0_91)
|___ 1 VM running 2 solr 6.0 instances on port 8501, 8502 (Solr 6.0)
(Java 1.8.0_91)
8 nodes, 4 shards apiece? All in the same JVM? People have gotten by
the GC pain by running in separate JVMs with less Java memory each on
big beefy machines That's not a recommendation as much as an
observation.
That aside, unless you have some very strange stuff going on this is
totally
Thanks Satvinder,
Tomorrow, I'll try to reproduce the issue with your steps and will let
you know.
Regards
--
Renaud Delbru
On 16/05/16 16:53, Satvinder Singh wrote:
Hi,
So the way I am doing it is, for both for the Target and Source side, I took a
copy of the sample_techproducts_config
I know the sender, he often posts to this list, and I don't download
attachments until I've vetted them anyway.
-Original Message-
From: postmas...@ssww.com [mailto:postmas...@ssww.com] On Behalf Of
h...@ssww.com
Sent: Monday, May 16, 2016 11:54 AM
To: solr-user@lucene.apache.org
I also am using a zk ensemble with 3 nodes on each side.
Thanks
Satvinder Singh
Security Systems Engineer
satvinder.si...@nc4.com
703.682.6000 x276 direct
703.989.8030 cell
www.NC4.com
-Original Message-
From: Satvinder Singh [mailto:satvinder.si...@nc4.com]
Sent:
Thanks to Binoy and Erick. I'll go use external zk tomorrow and do what you
suggest.
Scott Chu,scott@udngroup.com
2016/5/16 (週一)
- Original Message -
From: Erick Erickson
To: solr-user ; scott.chu
CC:
Date: 2016/5/16 (週一) 23:41
Subject: Re: [scottchu] Cab I migrate solrcloud by
Thanks Renaud, fixed.
On Mon, May 16, 2016 at 8:45 AM, Renaud Delbru wrote:
> Hi,
>
> I have tried to reproduce the problem, but was unable to.
> I have downloaded the Solr 6.0 distribution, added to the solr config the
> cdcr request handler and modified the update
Hi,
I have tried to reproduce the problem, but was unable to.
I have downloaded the Solr 6.0 distribution, added to the solr config
the cdcr request handler and modified the update handler to register the
CdcrUpdateLog, then start Solr in cloud mode and created a new
collection using my solr
bq: How can I run that embedded zk node independently?
You don't. You run ZK independently. It _seems_ like
you should be able to copy the zk_data directory over "to
the right place" and have it found, I suspect somehow
you're not. Take a look at where it is on your source
machine and see that
This depends on where zookeeper is storing the data. I'm not sure exactly
how the embedded zookeeper stores data locally, but given your situation we
can be sure that it's not stored in the solr folder.
Try to find out where the embedded zookeeper stores it's data and then copy
over that
But I'm using embeded zk nodes provided by solr start command. I thought they
are all under d:\solr5. How can I run that embedded zk node independently?
Scott Chu,scott@udngroup.com
2016/5/16 (週一)
- Original Message -
From: Binoy Dalal
To: scott.chu ; solr-user
CC:
Date:
What you copied is just the index.
Your configurations are stored on zookeeper.
You need to upload these to your zookeeper on your other machine and link
it to your collection.
Then it'll work.
On Mon, 16 May 2016, 20:55 Scott Chu, wrote:
> On my office pc, I install
Clever and real cool.
--Abdel
On Sun, May 15, 2016 at 10:42 AM, Lasitha Wattaladeniya
wrote:
> Hello all,
>
> I found a way of doing this and thought of sharing this info with you. I
> found a way to dynamically change the field which gives the suggestions.
> It's using the
bq: ...my primary customer for the deleted docs functionality wants to
see it immediately...
I don't quite know how docs get deleted, but presumably you have a
uniquekey. The
fastest would be to have a delete trigger on your database table that
puts that uniqueKey
and timestamp in a "delete_docs"
On my office pc, I install Solr 5 on d:\solr5 and create myconfigsets and
mynodes under it. Then run a solrcloud with 2 nodes and embedded zk nodes by
executing these commands:
cd /d d:\solr5
bin\solr start -c -s mynode\node1
bin\solr start -c -s mynode\node2 -p 7973 -z
We likely have the same laptop :-)
There must be something weird with my schema or usage but even if I had 10x
the throughput I have now, throwing around that many docs for a single join
isn't conducive to desired latency, concurrent requests, network bandwidth,
etc. I feel like I'm not using
So, with that setup you're getting around 150,000 docs per second
throughput. On my laptop with a similar query I was able to stream around
650,000 docs per second. I have an SSD and 16 Gigs of RAM. Also I did lots
of experimenting with different numbers of workers and tested after warming
the
Hi everyone,
I'm running into a problem with SolrCloud replicas and thought I would ask the
list to see if anyone else has seen this / gotten past it.
Right now, we are running with only one replica per shard. This is obviously a
problem because if one node goes down anywhere, the whole
Thanks Matteo, looks like you found a bug.
I can reproduce this with simpler queries too:
_query_:"ABC" name_t:"white cat"~3
is parsed to
text:abc name_t:"white cat"
Can you open a JIRA for this?
-Yonik
On Mon, May 16, 2016 at 10:23 AM, Matteo Grolla wrote:
> Hi
Thanks for all this info, Joel. I found if I artificially limit the
triples stream to 3M and use the /export handler with only 2 workers, I can
get results in @ 20 seconds and Solr doesn't tip over. That seems to be
the best config for this local/single instance.
It's also clear I'm not using
Hi everyone,
I have a problem with nested queries
If the order is:
1) query
2) nested query (embedded in _query_:"...")
everything works fine
if it is the opposite, like this
Hi,
Its all about how you tokenize the category field.
It looks like you are using a string type, which does not tokenize at all (e.g.
verbatim)
Please use a PatterTokenizer and configure it so that it splits on comma.
Ahmet
On Monday, May 16, 2016 2:11 PM, SRINI SOLR
Hi Team -
Can you please help me out on the following ...
I have a following field in the solr document which has the comma seperated
values like below ..
1,456,768,345 doc1
456 doc2
1,456 doc3
So - Here I need to filter the search docs which contains category is
Dear Solr users/developers,
Hi,
I have tried to implement the Page and Post relation in single Solr Schema.
In my use case each page has multiple posts. Page and Post fields are as
follows:
Post:{post_content, owner_page_id, document_type}
Page:{page_id, document_type}
Suppose I want to query
Thanks Eric! That worked.
On Mon, May 16, 2016 at 3:11 PM, Erik Hatcher
wrote:
> Quick reply: Use a different tag/ex value for each field.
>
> > On May 16, 2016, at 04:42, Manohar Sripada wrote:
> >
> > We have a similar requirement as that of
Quick reply: Use a different tag/ex value for each field.
> On May 16, 2016, at 04:42, Manohar Sripada wrote:
>
> We have a similar requirement as that of mentioned in solr wiki -
> http://wiki.apache.org/solr/SimpleFacetParameters#Multi-Select_Faceting_and_LocalParams
>
We have a similar requirement as that of mentioned in solr wiki -
http://wiki.apache.org/solr/SimpleFacetParameters#Multi-Select_Faceting_and_LocalParams
Here, the example given works when single facet is selected, where I can
get the counts of other facet values.
*q=*:*=0=10=true=5=1&*
>
Mikhail
It was caused by an endless loop in the page's codes that is triggered
only under certain conditions.
On 5/11/2016 4:07 PM, Mikhail Khludnev wrote:
On Wed, May 11, 2016 at 10:16 AM, Derek Poh wrote:
Hi Erick
Yes we have identified and fixed the page slow
59 matches
Mail list logo