I've successfully made extensive use of load balancers in sharded,
replicated slave setups - see [1].
My question is how that might work with a master. You can have a load
balancer, but you'd need to configure it into a 'fail over but please
don't fail back' configuration. I'm not sure if that is
you can put table names in a different table and use like this
entity name=tab query=select table_name from sometable
rootEntity=false
entity query=select * from ${tab.table_name}
-
-
-
-
Can anyone tell me hoe the dismax score is computed? Is it the maximum score
for any of the component fields that are searched? Thank You.
If you wish to view the St. James's Place email disclaimer, please use the link
below
http://www.sjp.co.uk/portal/internet/SJPemaildisclaimer
Can anyone tell me hoe the dismax score is computed? Is it
the maximum score for any of the component fields that are
searched? Thank You.
http://www.lucidimagination.com/blog/2010/05/23/whats-a-dismax/
hi. i'm looking for a technology who could have high performance in
searching a high amount of data (nearly 10 milion lines in a convencional
database like sql server) and i think PHP running under apache solr is a
good choice. i have only a doubt about its possibilities.
i need to show in
Hi Erick,
Thanks for the explanation.
At which point does the index get too big where sharding is appropriate where
it affects performance?
Tri
--- On Sun, 12/19/10, Erick Erickson erickerick...@gmail.com wrote:
From: Erick Erickson erickerick...@gmail.com
Subject: Re: shard versus core
Thought about it some more and after some reading. I suppose the answer
depends on what kind of response time is expected to be good enough.
I can do some stress testing and see if disk i/o is the bottleneck as the index
grows. I can also look into optimizing/configuring solr parameters to
Dennis,
If you need to search a key/value pair, you'll have to put them both in the
same field, somehow. One way is to re-index them using the key in the
fieldname. For instance, suppose you have:
contributor: dyer, james
contributor: smith, sam
role: author
role: editor
...but you want
This is the response I get...Does it matter that the configuration file is
called something other than data-config.xml? After I get this I still have
to restart the service. I wonder...do I need to commit the change?
?xml version=1.0 encoding=UTF-8 ?
You probably do not want this ranking, because any query with a common word,
like the, will match most of the corpus in step two.
Instead, use Solr to weight better quality matches more heavily, maybe 4X for
exact matches, 2X for stemmed matches, and 1X for phonetic matches.
wunder
On Dec 20,
thank you for your help... this search will be published in Portuguese, and
in this language we can clean up the sentence from words shorter than 3
characters.
Paula C. Laun : Dataprisma
pa...@dataprisma.com.br
(47) 3035.1868
www.dataprisma.com.br
- Original Message -
From: Walter
The problem happens with any common word, not just short words. What happens
with Brasil?
If this was a good way to do search, Solr would already implement it. It is not
that hard to build. But it is not a good way to do search. I have been working
on search for almost 15 years, and I hear
brasil will return companies with this word in any part of its name. this
search (randomic in 4 different levels) is only for promoted records (1
records to be searched at all). free records (10 milion) are the fifth level
and will respect the common search mode.
Um Abraço,
Paula C. Laun
Hi,
In SOLR XML the blank spaces are displayed with just str/ tags
Is there a way I can make SOLR XML to display the blank values as
str/str
instead of just
str/
Also has anyone parsed the blank value tags using SOLRNET before?
If anyone can help me with my question or provide pointers
There's another problem, i'm not sure i was clear: i need these records
randomic, each level randomic alone. (one level cannot random with another
level)
Is it possible for the same request?
Um Abraço,
Paula C. Laun : Dataprisma
pa...@dataprisma.com.br
(47) 3035.1868
www.dataprisma.com.br
Oops! That seems to be the problem, since I am using 1.4.
Thanks!
Juan M.
On Tue, Dec 14, 2010 at 8:40 PM, Alexey Serba ase...@gmail.com wrote:
What Solr version do you use?
It seems that sync flag has been added to 3.1 and 4.0 (trunk) branches
and not to 1.4
No. But why is it a problem? A standard XML parser won't feel the difference.
Hi,
In SOLR XML the blank spaces are displayed with just str/ tags
Is there a way I can make SOLR XML to display the blank values as
str/str
instead of just
str/
Also has anyone parsed the blank value
Thanks James.
So being accurate with fields with fields(mulitvalues) is probably not possible
using all the currently made analyzers.
- Original Message
From: Dyer, James james.d...@ingrambook.com
To: solr-user@lucene.apache.org solr-user@lucene.apache.org
Sent: Mon, December 20,
All,
I have a couple websites that I need to crawl and the following command line
used to work I think. Solr is up and running and everything is fine there
and I can go through and index the site but I really need the results added
to Solr after the crawl. Does anyone have any idea on how to make
why are using solrindex in the argument.? It is used when we need to index
the crawled data in Solr
For more read http://wiki.apache.org/nutch/NutchTutorial .
Also for nutch-solr integration this is very useful blog
http://www.lucidimagination.com/blog/2009/03/09/nutch-solr/
I integrated nutch
Hello,
When using a writer type of json, SOLR (1.4.1) sets the content type
header of the response as text/plain although it should be
application/json. This is not a very big problem, but it writes many
warnings in Chrome logs: Resource interpreted as script but
transferred with MIME type
Well, right now, I'm using SOLR in a LOT of my projects.
I'm VERY fond of it, proud of it and VERY happy that such a team exists to
make it work.
Of course the pagination issue is a bit frustrating on the field
collapsing... But... heck... I'm currently de-normalizing my postgresql
database
Here is a thread on this subject that I did not find earlier. Sometimes
discussion, thought, and 'mulling' in the subconcious gets me better Google
searches.
http://lucene.472066.n3.nabble.com/multi-valued-associated-fields-td811883.html
Dennis Gearon
Signature Warning
It
Ah, thanks for pointing that out.
Each indexer needs its own marker for where is new data in this
stream? This way, when either the primary or secondary starts, it can
restart indexing from where it left off. The most reliable way to do
this is to search the indexer Solr for its last update.
2x the index size is required for optimizing.
Things that increase with index size: indexing time, query time and
disk index size. My 500GB index at a previous job worked. Indexing was
a little slow, queries were much slower. What finally made us split it
up was that one binary blob of 500GB was
bin/nutch crawl urls -dir crawl -threads 10 -depth 100 -topN 50 -solrindex
http://localhost:8983/solr
I've run that command before and it worked...that's why I asked.
grab nutch from trunk and run bin/nutch and see that it is in fact an
option. It looks like Hadoop is the culprit now and I am at
On Wed, Dec 15, 2010 at 9:57 AM, Stephen Green eelstretch...@gmail.com wrote:
Otis pointed out that the patch can't be applied against the current
source, so I need to go back and make it work with the current source
(new job = no time). I'll see if I can find the time this weekend to
do
Based on more searches and manual consolidation, I've put together some of
the ideas for this already suggested in a summary below. The last item in the
summary
seems to be interesting, low technical cost way of doing it.
Basically, it treats the index like a 'BigTable', a la No SQL.
Erick
Then what will be when we filter out only some result and want to group ,how
your index time group count will help.
-
Grijesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/solr-4-0-pagination-tp1812384p2124747.html
Sent from the Solr - User mailing list archive at
29 matches
Mail list logo