Hi All,
WE have setup a 4.2 Solr cloud with 4 nodes and while the add/update/delete
operations are working we are not able to perform a rollback. Is there
something different for this operation vs the 3.x sole master/slave config?
Thanks,
Dipti
phone: 408.678.1595 | cell: 408.806.1970 |
Hi all,
I have recently migrated from solr 3.6 to solr 4.0. The documents in my core
are getting constantly updated and so I fire a code commit after every 10
thousand docs . However moving from 3.6 to 4.0 I have noticed that for the
same core size it takes about twice the time to commit in
Can you explain more about your document size, shard and replica sizes, and
auto/soft commit time parameters?
2013/5/2 vicky desai vicky.de...@germinait.com
Hi all,
I have recently migrated from solr 3.6 to solr 4.0. The documents in my
core
are getting constantly updated and so I fire a
I am wondering if it was possible to achieve SolrJ/Solr Two Phase Commit.
Any examples? Any best practices?
What I know:
* Lucene offers Two Phase Commitvia it's index writer (prepareCommit()
followed by either commit() or rollback()).
What version of Solr are you using? 4.2.0 or 4.2.1?
The following might be of interest to you:
* https://issues.apache.org/jira/browse/SOLR-4605
https://issues.apache.org/jira/browse/SOLR-4605
* https://issues.apache.org/jira/browse/SOLR-4733
https://issues.apache.org/jira/browse/SOLR-4733
Hi,
I am using 1 shard and two replicas. Document size is around 6 lakhs
My solrconfig.xml is as follows
?xml version=1.0 encoding=UTF-8 ?
config
luceneMatchVersionLUCENE_40/luceneMatchVersion
indexConfig
maxFieldLength2147483647/maxFieldLength
The Integer.MAX_VALUE-1 limit is set by Lucene. As hardware capacity and
performance continues to advance, I think it's only a matter of time before
Lucene (and then Solr) relaxes the limit, but I don't imagine it will
happened real soon. Maybe in Lucene/Solr 6.0?
-- Jack Krupansky
Hi,
We're planning on upgrading our solr cluster from 4.0 to 4.2.1
Is 4.3 coming any soon?
thanks.
--
View this message in context:
http://lucene.472066.n3.nabble.com/Any-estimation-for-solr-4-3-tp4060408.html
Sent from the Solr - User mailing list archive at Nabble.com.
RC4 of 4.3 is available now. The final release of 4.3 is likely to be within
days.
-- Jack Krupansky
-Original Message-
From: adfel70
Sent: Thursday, May 02, 2013 1:32 AM
To: solr-user@lucene.apache.org
Subject: Any estimation for solr 4.3?
Hi,
We're planning on upgrading our solr
I use the same folder naming convention of Solr example for my Solr 4.2.1
cloud. I have a collection1 folder and under it I have a conf folder. When
I starting up my first node, I indicate that:
-Dsolr.solr.home=./solr -Dsolr.data.dir=./solr/data -DnumShards=5
Does Near Real Time get not supported at SolrCloud?
I mean when a soft commit occurs at a leader I think that it doesn't
distribute it to replicas(because it is not at storage, does indexes at RAM
distributes to replicas too?) and a search query comes what happens?
Hi guys!
We have solr router and shards. I see this in jetty log on the router:
May 02, 2013 1:30:22 PM org.apache.commons.httpclient.HttpMethodDirector
executeWithRetry
INFO: I/O exception (java.net.SocketException) caught when processing
request: Connection reset by peer: socket write error
Hi,
In previous versions of solr, indexing documents with large fields caused
performance degradation.
Is this still the case in solr 4.2?
If so, and I'll need to chunk the document and index many document parts,
can anyony give a general idea of what field/document size solr CAN handle?
The only issue I ran into was returning the content field. Once I modified
my query to avoid that, I got good performance.
Admittedly, I only have about 15-20k documents in my index ATM, but most of
them are in the multiMB range with a current max of 250MB.
On Thu, May 2, 2013 at 7:05 AM,
Well, returning the content field for highlighting is within my requirements.
Did you solve this in some other way? or you just didn't have to?
Bai Shen wrote
The only issue I ran into was returning the content field. Once I
modified
my query to avoid that, I got good performance.
Here is a part from wiki:
1) Just forward credentials from the super-request which caused the
inter-solr-node sub-requests
2) Use internal credentials provided to the solr-node by the
administrator at startup
what do you use and is there any code example for it?
The leader would not be behind replica because the old leader would not
come back and take over the leader role. It would ne just a replica and it
would replicate the index from whichever node is the leader.
Otis
Solr ElasticSearch Support
http://sematext.com/
On Apr 29, 2013 5:31 PM, Furkan
After some searching around, I see this:
http://search-lucene.com/m/ErEZUl7P5f2/%2522socket+write+error%2522subj=Long+list+of+shards+breaks+solrj+query
Seems like this has happened in the past with large amount of shards.
To make it clear: the distributed search works with 20 shards.
On Thu,
DF uses maxDoc which is updated when segments merge so DF is almost never
accurate in a dynamic index.
-Original message-
From:Furkan KAMACI furkankam...@gmail.com
Sent: Thu 02-May-2013 14:05
To: solr-user@lucene.apache.org
Subject: DF is not updated when a document is marked for
Thanks for the answer. This is what I try to say:
time = t
Node A (Leader): version is 100
Node B (Replica): version is 90
time = t+1
Node A (Killing): version is 100 and killed
Node B (Replica): version is 90
time = t+2
Node A (Killed): version is 100 and killed
Node B (Become Leader):
If you're using zookeeper, this should not be allowed to happen (I think).
On Thu, May 2, 2013 at 2:12 PM, Furkan KAMACI furkankam...@gmail.comwrote:
Thanks for the answer. This is what I try to say:
time = t
Node A (Leader): version is 100
Node B (Replica): version is 90
time = t+1
On May 2, 2013, at 3:36 AM, Jack Krupansky j...@basetechnology.com wrote:
RC4 of 4.3 is available now. The final release of 4.3 is likely to be within
days.
How can I see the Changelog of what will be in it?
Thanks,
xoa
--
Andy Lester = a...@petdance.com = www.petdance.com = AIM:petdance
The road map has this release note, but I think that most of it will be move to
4.3.1 or 4.4
https://issues.apache.org/jira/secure/ReleaseNote.jspa?projectId=12310230version=12324128
Regards
--
Yago Riveiro
Sent with Sparrow (http://www.sparrowmailapp.com/?sig)
On Thursday, May 2, 2013 at
On May 2, 2013, at 9:03 AM, Yago Riveiro yago.rive...@gmail.com wrote:
The road map has this release note, but I think that most of it will be move
to 4.3.1 or 4.4
https://issues.apache.org/jira/secure/ReleaseNote.jspa?projectId=12310230version=12324128
So, is there a way I can see
In attachment the change log of solr 4.3 RC3
Regards
--
Yago Riveiro
Sent with Sparrow (http://www.sparrowmailapp.com/?sig)
On Thursday, May 2, 2013 at 3:06 PM, Andy Lester wrote:
On May 2, 2013, at 9:03 AM, Yago Riveiro yago.rive...@gmail.com
(mailto:yago.rive...@gmail.com) wrote:
On May 2, 2013, at 9:11 AM, Yago Riveiro yago.rive...@gmail.com wrote:
In attachment the change log of solr 4.3 RC3
And where would I find that? I don't see anything at
http://lucene.apache.org/solr/downloads.html to download? Do I need to check
out Subversion repo? Is there a page
Hopefully, this is not a secret, but the RCs are built and available
for download and announced on the dev mailing list.
So, the changes for RC4 (not RC3 anymore) are here:
http://people.apache.org/~simonw/staging_area/lucene-solr-4.3.0-RC4-rev1477023/solr/changes/Changes.html
Regards,
Alex.
I get the RC3 zip file from the the mailing list.
The 4.3 has not yet been released. Therefore you can't download it from the
regular channels
Regards
--
Yago Riveiro
Sent with Sparrow (http://www.sparrowmailapp.com/?sig)
On Thursday, May 2, 2013 at 3:14 PM, Andy Lester wrote:
On May
On May 2, 2013, at 9:20 AM, Alexandre Rafalovitch arafa...@gmail.com wrote:
Hopefully, this is not a secret, but the RCs are built and available
for download and announced on the dev mailing list.
Thanks for the link.
I don't think it's a secret, but I sure don't see anything that says This
One thing I do know is that commits in Solr are global, so there's no way
to do this with concurrency.
That being said, Solr doesn't tend to accept updates that would generate
errors once committed in my experience.
Michael Della Bitta
Appinions
Hi everyone,
I am working on an internal project in my company that requires solr, but I
could not manage to link it to Tika. I bought the apache solr 4 cookbook
yet I couldn't figure out the solution.
1) I copied the required jar files into a lib directory 2) I added the lib
directory in
On Thu, May 2, 2013 at 10:23 AM, Andy Lester a...@petdance.com wrote:
I don't think it's a secret, but I sure don't see anything that says This is
how the dev process works.
I suspect this is somewhere in the Apache operating charter/standard
operating procedures for all projects and we (Solr
First, I would upgrade to 4.2.1 and remember to change luceneMatchVersion to
LUCENE_42.
There were a LOT of fixes between 4.0 and 4.2.1.
wunder
On May 2, 2013, at 12:16 AM, vicky desai wrote:
Hi,
I am using 1 shard and two replicas. Document size is around 6 lakhs
My solrconfig.xml
$100 for anyone who gets me a working Long.MAX_VALUE branch! ;-)
I know that for many of the SOLR with faceting use cases, things will
not scale to Long documents, but there are a number of more
straightforward use cases, where SOLR/Lucene will scale to Long. Like
simple searches, small numbers
you might want to added openSearcher=false for hard commit, so hard commit
also act like soft commit
autoCommit
maxDocs5/maxDocs
maxTime30/maxTime
openSearcherfalse/openSearcher
/autoCommit
What happens exactly when you don't open searcher at commit?
2013/5/2 Gopal Patwa gopalpa...@gmail.com
you might want to added openSearcher=false for hard commit, so hard commit
also act like soft commit
autoCommit
maxDocs5/maxDocs
If you don't re-open the searcher, you will not see new changes. So,
if you only have hard commit, you never see those changes (until
restart). But if you also have soft commit enabled, that will re-open
your searcher for you.
Regards,
Alex.
Personal blog: http://blog.outerthoughts.com/
Hi Vicky,
I faced this issue as well and after some playing around I found the
autowarm count in cache sizes to be a problem.
I changed that from a fixed count (3072) to percentage (10%) and all commit
times were stable then onwards.
filterCache class=solr.FastLRUCache size=8192 initialSize=3072
Hi,
I want to set up a master / slave configuration for solr 3.6
Is there a best practice for the Raid config and the Linux partitions for the
master server?
Cheers,
Torsten
On Wed, May 1, 2013 at 7:01 PM, 李威 li...@antvision.cn wrote:
For facet seach, solr would create cache which is based on the whole docs.
If I import a new doc into index, the cache would out of time and need to
create again.
For real time seach, the docs would be import to index anytime. In
Hi Jan
my question is when I tweak pf and qf parameter and the results change
slightly and I do not think for exact match you need to implement the
solution that you mentioned in your reply. you can always have string field
and in your pf parameter you can boost that field to get the exact match
Hi Jan. Thank's again for your reply.
You're right. It is almost impossible to an user exclude 200.000 documents.
I'll do some tests with NOT IN query.
Thank you again.
*
--
*
*E conhecereis a verdade,
Hi,
First, which version of Solr are you using?
I also has 60 shards+ on Solr 4.2.1 and it doesn't seems to be a problem
for me.
- Make sure you use POST to send a query to Solr.
- 'connection reset by peer' from client can indicate that there is
something wrong with server e.g. server closes
My first guess would be that your tomcat container timeouts need to be
lengthened, but that's mostly a guess based on the socket timeout
error message. Not sure where in Tomcat that needs to be configured
though...
Best
Erick
On Tue, Apr 30, 2013 at 12:37 PM, Arun Rangarajan
because some of the underlying classes in SolrJ try to communicate
with Zookeeper to intelligently route requests to leaders.
It looks like you don't have your classpath pointed at the
dist/solrj-lib, at least that would be my first guess...
Best
Erick
On Wed, May 1, 2013 at 7:51 AM, Peri
None that I know of. But if you hit issues I know where you can get help!
:)
Otis
Solr ElasticSearch Support
http://sematext.com/
On May 2, 2013 12:41 PM, Torsten Albrecht tors...@soahc.eu wrote:
Hi,
I want to set up a master / slave configuration for solr 3.6
Is there a best practice for
Hi,
We have one master and 2 slaves with solr3.6. The below messages are logged
in solr log .
ERROR: Master at: http://server:port/solr/pe/replication is not available.
Index fetch failed. Exception: Connection reset
ERROR: Master at: http://server:port/solr/pe/replication is not available.
Val,
Haven't seen this mentioned in a while...
I'm curious...what sort of index, queries, hardware, and latency
requirements do you have?
Otis
Solr ElasticSearch Support
http://sematext.com/
On May 1, 2013 4:36 PM, Valery Giner valgi...@research.att.com wrote:
Dear Solr Developers,
I've
Actually, I found it very hard to figure out the exact Jar
requirements for SolrJ. I ended up basically pointing at expanded
webapp's lib directory, which is a total overkill.
Would be nice to have some specific guidance on this issue.
Regards,
Alex.
Personal blog:
Hi,
Can you actually make this happen?
Otis
Solr ElasticSearch Support
http://sematext.com/
On May 2, 2013 8:12 AM, Furkan KAMACI furkankam...@gmail.com wrote:
Thanks for the answer. This is what I try to say:
time = t
Node A (Leader): version is 100
Node B (Replica): version is 90
Looks like a network issue, especially if this is not happening
consistently.
Otis
Solr ElasticSearch Support
http://sematext.com/
On May 2, 2013 3:42 PM, gpssolr2020 psgoms...@gmail.com wrote:
Hi,
We have one master and 2 slaves with solr3.6. The below messages are logged
in solr log .
Nah, not until and IF you see issues. Most users are not even aware of this.
Otis
Solr ElasticSearch Support
http://sematext.com/
On May 2, 2013 8:05 AM, Furkan KAMACI furkankam...@gmail.com wrote:
When I look at here: http://localhost:8983/solr/admin/luke
I see that Note: Document Frequency
Otis,
The documents themselves are relatively small, tens of fields, only a
few of them could be up to a hundred bytes.
Lunix Servers with relatively large RAM (256),
Minutes on the searches are fine for our purposes, adding a few tens of
millions of records in tens of minutes are also fine.
Hi Otis;
I see that at my admin page:
Replication (Slave) Version GenSize
Master: 1367307652512 82 778.04 MB
Slave: 1367307658862 82 781.05 MB
and I started to figure about it so that's why I asked this question.
2013/5/2 Otis
Hi,
I've been very interested in the transient core feature of solr to manage a
large number of cores. I'm especially interested in this use case, that the
wiki lists at http://wiki.apache.org/solr/LotsOfCores (looks to be down
now):
loadOnStartup=false transient=true: This is really the
This feature is not yet part of Solr, but a feature under development in
SOLR-4470. We encourage you to try it out and report back what worked best for
you.
--
Jan Høydahl, search solution architect
Cominvent AS - www.cominvent.com
Solr Training - www.solrtraining.com
2. mai 2013 kl. 13:58
I use Solr 4.2.1 as SolrCloud. I crawl huge data with Nutch and index them
with SolrCloud. I wonder about Solr's deduplication mechanism. What exactly
it does and does it results with a slow indexing or is it beneficial for my
situation?
Distributed deduplication does not work right now:
https://issues.apache.org/jira/browse/SOLR-3473
We've chosen not do use update processors for deduplication anymore and rely on
several custom mapreduce jobs in Nutch and some custom collectors in Solr to do
some on-demand online deduplication.
I know that I can use boosting at query for a field, for a searching term,
at solrconfig.xml and query elevator so I can arrange the results of a
search. However after I get top documents how can I change the order of a
results? Does Lucene's postfilter stands for that?
On 5/2/2013 1:43 PM, Alexandre Rafalovitch wrote:
Actually, I found it very hard to figure out the exact Jar
requirements for SolrJ. I ended up basically pointing at expanded
webapp's lib directory, which is a total overkill.
Would be nice to have some specific guidance on this issue.
I
Hi folks;
I want to use Solr to index any other language except for English. I will
use Turkish documents to index with Solr. I will implement some algorithms
that is more suitable to Turkish rather than English. Is there any wiki
page that explains to steps for it? I mean what are the main parts
Hello,
I updated my schema to use a copyField and have triggered a reindex, 80% of
the reindexing is complete. Although when I query the data, I don't see
myNewCopyFieldName being returned with the documents.
Is there something wrong with my schema or I need to wait for the indexing
to complete
I read that at wiki:
Sometimes a smaller cache size will help avoid full garbage collections at
the cost of more evictions. Load testing should be used to help determine
proper cache sizes throughout the searching/indexing lifecycle.
Could anybody give me an example scenario of how can I make a
We are using Solr 4.2.1, which claims to have fixed this issue. I have
some logging indicating that the rollback is not broadcasted to other
nodes in solr. So only one node in the cluster gets the rollback but not
the others.
Thanks,
Dipti
phone: 408.678.1595 | cell: 408.806.1970 |
On 5/2/2013 3:13 PM, Utkarsh Sengar wrote:
Hello,
I updated my schema to use a copyField and have triggered a reindex, 80% of
the reindexing is complete. Although when I query the data, I don't see
myNewCopyFieldName being returned with the documents.
Is there something wrong with my
Have you looked at the main example that comes with Solr? It contains
a specific configuration for Turkish. Perhaps you could try that and
narrow the question to more precise issues?
I don't remember any Turkish-specific discussions, but perhaps
something can be learned from searching for
On 5/2/2013 2:19 PM, Furkan KAMACI wrote:
I see that at my admin page:
Replication (Slave) Version GenSize
Master: 1367307652512 82 778.04 MB
Slave: 1367307658862 82 781.05 MB
and I started to figure about it so that's why
Thanks Shawn. Find my answers below.
On Thu, May 2, 2013 at 2:34 PM, Shawn Heisey s...@elyograg.org wrote:
On 5/2/2013 3:13 PM, Utkarsh Sengar wrote:
Hello,
I updated my schema to use a copyField and have triggered a reindex, 80%
of
the reindexing is complete. Although when I query
Hi,
You can try to increase the pf boost for your string field, I don't think
you'll have success in having it boosted with pf since it's a string? Check
explain output with debugQuery=true and see whether you get a phrase boost.
--
Jan Høydahl, search solution architect
Cominvent AS -
On 5/2/2013 4:24 AM, Annette Newton wrote:
Hi Shawn,
Thanks so much for your response. We basically are very write intensive
and write throughput is pretty essential to our product. Reads are
sporadic and actually is functioning really well.
We write on average (at the moment) 8-12
On 5/2/2013 7:56 AM, Andy Lester wrote:
On May 2, 2013, at 3:36 AM, Jack Krupansky j...@basetechnology.com wrote:
RC4 of 4.3 is available now. The final release of 4.3 is likely to be within
days.
How can I see the Changelog of what will be in it?
Here's the latest CHANGES.txt file
NRT works with SolrCloud.
Otis
Solr ElasticSearch Support
http://sematext.com/
On May 2, 2013 5:34 AM, Furkan KAMACI furkankam...@gmail.com wrote:
Does Near Real Time get not supported at SolrCloud?
I mean when a soft commit occurs at a leader I think that it doesn't
distribute it to
You simply need to monitor and adjust. Both during testing and in
production because search patterns change over time. Hook up alerting to it
to get notified of high evictions and low cache hit rate so you don't have
to actively look at stats all day.
Here is the graph of Query Cache metrics for
Hi,
You should use search more often :)
http://search-lucene.com/?q=scriptable+collectorsort=newestOnTopfc_project=Solrfc_type=issue
Coincidentally, what you see there happens to be a good example of a
Solr component that does something behind the scenes to deliver those
search results even
By saying commits in Solr are global, do you mean per Solr deployment, per
HttpSolrServer instance, per thread, or something else?
--
View this message in context:
http://lucene.472066.n3.nabble.com/SolrJ-Solr-Two-Phase-Commit-tp4060399p4060584.html
Sent from the Solr - User mailing list
I actually have a maven project with a declared solrj dependency (4.2.1); Do I
need anything extra to get rid of the Zookeeper exception? I didn't see jars
specific to zookeeper in the list below that I would need. Any more ideas
please?
Thank you,
Peri Subrahmanya
On May 2, 2013, at 4:48
Peer core or collection, depending on whether we're talking about Cloud or
not.
Basically, commits in Solr are about controlling visibility more than
anything, although now with Cloud, they have resource consumption and
lifecycle ramifications as well.
On May 2, 2013 10:01 PM, mark12345
Question: Just to clarify. Are you saying that if I have multiple threads
using multiple instances of HttpSolrServer each making calls to add
SolrInputDocuments (For example, httpSolrServer.add(SolrInputDocument
doc). ), and one server calls httpSolrServer.commit(), all documents
added are now
One thing I noticed is that while the HttpSolrServer add(SolrInputDocument
doc) method is atomic (Either a bean is added or an exception is thrown),
the HttpSolrServer add(CollectionSolrInputDocument docs) method is not
atomic.
Question: Is there a way to commit multiple documents/beans in a
Sorry I don't know enough of the system to help you directly. I was only
going to suggest upgrading to 4.2.1 if you where using 4.2.0.
It might be worthwhile to create a JIRA issue for what you are experiencing.
https://issues.apache.org/jira/browse/SOLR
Dipti Srivastava wrote
We are using
On 5/2/2013 8:07 PM, Peri Subrahmanya wrote:
I actually have a maven project with a declared solrj dependency (4.2.1); Do
I need anything extra to get rid of the Zookeeper exception? I didn't see
jars specific to zookeeper in the list below that I would need. Any more
ideas please?
SolrJ
Hi, thanks.
Solr 3.4.
There is POST request everywhere, between client and router, router and
shards.
Do you do faceting across all shards? How many documents approx you have?
On 2 May 2013 22:02, Patanachai Tangchaisin
patanachai.tangchai...@wizecommerce.com wrote:
Hi,
First, which version
Hi,
Hm, I *think* you can't do it in one go with Solr's Suggester, but I'm
not expert there. I can only point you to something like our
AutoComplete - http://sematext.com/products/autocomplete/index.html -
which, as you can see on that screenshot, has the grouping you seem to
be after. Maybe
Yes, that is correct. --wunder
On May 2, 2013, at 7:46 PM, mark12345 wrote:
Question: Just to clarify. Are you saying that if I have multiple threads
using multiple instances of HttpSolrServer each making calls to add
SolrInputDocuments (For example, httpSolrServer.add(SolrInputDocument
Hallo,
Peter, try sorting them only using one sort parameter, separating the
fields by comma.
sort=zip+asc,street+asc
This was it, thank you.
Ciao
Peter Schütt
85 matches
Mail list logo