Hi,
We migrated from solr 3.6 to solr 4.3, when we fire a backup command
(/replication?command=backuplocation=/disk4/backups ) on the salve the slave
stops replicating OR starts full replication from the master. This was not the
behavior in the earlier version or solr, I have check the logs
Hello!
Just to be perfectly clear here - Solr has advantage over
ElasticSearch because it can split the live index. However, this
sentence from Jack's mail is not true:
In other words, even though you can easily “add shards” on ES,
those are really just replicas of existing primary shards and
(Given that hardware is sufficient) The upper limit of documents in Norch is
determined by the capacity of levelDB, the underlying data store. I have heard
tell of a slight performance drop off in LevelDB after 200 000 000 million
entries. If you say that one Norch document generates roughly
Norch is based on the node module search-index which is like a simplified
lucene, built with Google's levelDB library
(https://github.com/fergiemcdowall/search-index)
(so posting here is a bit cheeky- but I figured the solr/lucene readership
might be interested :-)
F
On Jul 6, 2013, at
If you are asking about how to use Solrj with SolrCloud, see
http://wiki.apache.org/solr/Solrj#Using_with_SolrCloud
If you are not using SolrCloud then add
shards=solr_server1:port/solr,solr_server2:port/solr as a parameter
to a query request. Indexing doesn't need anything special -- you just
Hi,
I am working on solr 3.3. i am getting total 120 records with below query,
in response xml numFound is showing 540 records.
http://localhost:8080/test/select?q=*:*rows=*120*
response
lst name=responseHeader
int name=status0/int
int name=QTime1/int
lst name=params
str
Well, you've just restated the problem. I'm asking
what use-case this is supporting? You've said:
he/she wants to know the occurrence of a
specific term in the result set of that 'X1' search criteria
OK, _why_? Idle curiosity? Ranking the docs?
Choosing the most relevant? I don't think you can
bq: Reason being If a write request is sent to the replica it
relays it to the leader and then the leader relays it to all the replicas.
This will help me in saving some network traffic as my application performs
continuous writes
How are you indexing? SolrJ already is leader aware and sends
Not saying it's always one way or the other, just
that one shouldn't automatically _assume_
putting the most recent data on a single node
is automatically good. It may well be, but
not in all cases.
On Wed, Jul 3, 2013 at 12:21 PM, Otis Gospodnetic
otis.gospodne...@gmail.com wrote:
Exactly.
Neither ORing or ANDing the terms together will help you
regardless of whether you do it or the query parser
you use does it behind the scenes. That's
because you have _string_ fields. So
Google Cloud Storage
is a single token in your index, not three.
You probably want to spend some time on the
kill -9 is evil, you aren't certain what the state is afterwards,
so the presence of the lock file is not surprising..
solrconfig.xml has a commented-out entry
unlockOnStartupfalse/unlockOnStartup
I haven't personally used it, but it looks like it might
help if you insist on kill -9.
The process looks like this:
each shard returns the top 100K documents
(actually the doc ID and whatever your
sort criteria is, often just the score).
_from every shard_ and the node that
distributes that request then takes those
900K items and merges the list to get the
100K that satisfy the
Not a problem. index time boosts are boosts made
_when you're indexing_, not when you're querying
so omitting norms should stil have your query boosting
work.
Also, try adding debug=all and examining the results,
it'll show you exactly how scores were calculated. It does
take a bit to work though
That's exactly what turned out to be the problem. We thought we had
already tried that permutation but apparently hadn't. I know it's obvious
in retrospect. Thanks for the suggestion.
Thanks,
Andy Pickler
On Wed, Jul 3, 2013 at 2:38 PM, Alexandre Rafalovitch arafa...@gmail.comwrote:
On Tue,
Ok. If you can create a test setting to repro the bug, I would suggest
filing a jira. Sounds like a regression if your config remained otherwise
same.
On 5 Jul 2013 16:57, adityab aditya_ba...@yahoo.com wrote:
well our observation leads us that this happens only during spell check.
If we turn
What does this have to to with removing the
int32 limit? It's still the same problem if you have
a high start parameter, it's the deep paging
issue that's part of Solr.
I know there has been work on this (you'll have to
search the JIRAs), the basic idea is that you pass
enough information that
Oh, P.S. Solr is a great search engine, but it's
certainly not the perfect answer to all problems.
Mayhap you've hit on a case where it isn't the
best solution
On Sat, Jul 6, 2013 at 8:22 AM, Erick Erickson erickerick...@gmail.comwrote:
What does this have to to with removing the
int32
uniqueKey is used to enforce there
being only a single copy of a doc. Say
a doc changes and you re-index it. If
there is a doc in the index already _with
the same uniqueKey_ it'll be deleted
and the new one will be the only one
visible.
Which implies that if you do implement
the suggestions, be
Solr only returns the number of documents specified by the rows
parameter.
You can page through your results by specifying, say,
start=20rows=20
then
start=40rows=20
etc.
Or you can bump rows, but you really don't want to return
huge result sets so you'll probably be paging sometime.
But it is
I know this stuff can be confusing, but I think my statement is still true -
it uses the phrase NEW primary shard - the promotion of a shard can only
be to replace an existing primary shard, not to create a new and separate
partitioning of key values.
You left out the next sentence from my
On Fri, Jul 5, 2013 at 10:57 AM, adityab aditya_ba...@yahoo.com wrote:
well our observation leads us that this happens only during spell check.
If we turn off the spell check we don't see this issue occurring at all from
our 24hrs test run.
Are you using any custom components / plugins, or is
Hello,
I have 3 indices that form a hierarchy. Basically these were constructed from 3
tables: parent, child1 and child2 and between parent and children there is a
one to many relationship.
parent (id,name)
child1(id,join_key,field1)
child2(id,join_key,field2)
join_key is the foreign key
If you flatten this to make a single table with rows for each combination of
parent, child1, and child2, the query will be simple and probably very fast.
You've said it makes more sense to have three tables, and that is true for a
relational database. For a search engine, like Solr, it makes
On Sat, Jul 6, 2013 at 2:22 PM, mihaela olteanu mihaela...@yahoo.com wrote:
Hello,
I have 3 indices that form a hierarchy. Basically these were constructed from
3 tables: parent, child1 and child2 and between parent and children there is
a one to many relationship.
parent (id,name)
Does anyone have any idea how I can access the schema version info using SolrJ?
Thanks.
On Jul 3, 2013, at 4:16 PM, Steven Glass wrote:
I'm using a Solr 4.3 server and accessing it from both a Java based desktop
application using SolrJ and an Android based mobile application using my
Steven,
Some information can be gleaned from the system admin request handler:
http://localhost:8983/solr/admin/system
I am specifically looking at this:
lst name=corestr name=schemaexample/str
Mind you, that is a manually-set value in the schema file. But just in case
you want to get crazy
Thanks for your response.
But it seems like there should be a way to issue the equivalent of
http://localhost:8983/solr/schema/version
which returns
{
responseHeader:{
status:0,
QTime:4},
version:1.5}
from the server. I know how to do it using HTTPGet in
You _should_ be able to use an HttpSolrServer, set
the base URL and then go ahead and make the
request. Haven't done it myself, but the SolrJ
support for various Solr features often consists of
just convenience methods over the underlying HTTP,
saving you the necessity of, say, parsing the
Nope no custom plugin. Just use the DirectSpellCheck component.
We have raised a ticket with LucidWorks i will followup with that and once
have a JIRA will update this post.
--
View this message in context:
Can it do Geo Spatial searching? (i.e. Find documents within 10 miles of a
lat,long?)
On Fri, Jul 5, 2013 at 12:53 PM, Fergus McDowall
fergusmcdow...@gmail.comwrote:
Here is some news that might be of interest to users and implementers of
Solr
I submitted a JIRA ticket a while ago, since I thought that having a way to
use the Lucene facets in SOLR could speed up our faceting. However, no one
seems to have picked up the development.
https://issues.apache.org/jira/browse/SOLR-4774
What is involved with hooking it into SOLR ? Similar to
Why is LUCENE-474 not committed?
On Thu, Jul 4, 2013 at 4:21 PM, Koji Sekiguchi k...@r.email.ne.jp wrote:
Hi Dotan,
(13/07/04 23:51), Dotan Cohen wrote:
Thank you Jack and Koji. I will take a look at MLT and also at the
.zip files from LUCENE-474. Koji, did you have to modify the code
thanks Erick.
My tomcat setup is running on Ubuntu and has nothing else deployed other than
the Solr war. My suspicion is it takes a long time to de-allocate the memory it
has reserved for itself, but will get dump to find better.
regards,
Ayush
Date: Sat, 6 Jul 2013 07:55:57 -0400
Subject:
Hi Erik,
I just wanted to clarify if u got my concern right. If i send some documents
to the replica core wont it first have to send the documents to the leader
core which in turn would be sending it back to the replica cores. If yes
then this will lead to additional network traffic which can be
There are three concepts to grasp:
1. You can send Solr update requests to ANY node of the cluster. Period.
2. Any extra network traffic (within the cluster) is likely to be negligible
and absolutely not worrying about unless you have definitive evidence to the
contrary.
3. Leader nodes in
35 matches
Mail list logo