Hi,
answering my own question for the records: the experiments show that the
described functionality is achievable with the TokenFilter class
implementation. The only caveat though, is that Highlighter component stops
working properly, if the match position goes beyond the length of the text
Hi,
Have a look onto TokenFilter. Extending it will give you access to a
TokenStream.
Regards,
Dmitry Kan
On Fri, Dec 21, 2012 at 9:05 AM, Xi Shen davidshe...@gmail.com wrote:
Hi,
I am looking for a token filter that can combine 2 terms into 1? E.g.
the input has been tokenized by white
Hi,
I have a text field with value O O Jaane Jaane. When i search with *q=Jaane
Jaane* it is giving the results. But if i give *q=O O Jaane Jaane* it is not
working? What could be the reason?
Thanks,
Sangeetha
--
View this message in context:
Which Analyzer is being used in the field that was indexed ?
May be you can use solradmin to analyze and see how is your index
thanks
On Thu, Dec 27, 2012 at 2:30 PM, Sangeetha sangeetha...@gmail.com wrote:
Hi,
I have a text field with value O O Jaane Jaane. When i search with *q=Jaane
Hi,
I am having trouble with getting solr + jetty to work. I am following all
instructions to the letter from - http://wiki.apache.org/solr/SolrJetty. I
also created a work folder - /opt/solr/work. I am also setting tmpdir to a
new path in /etc/default/jetty . I am confirming the tmpdir is set to
Thanks Gora,
I can definitely trigger the full re-indexing using CURL for multiple cores
although if i try to index multiple cores (more than 4-5 cores)
simultaneously then the re-indexing fails due to DB connection pool
problems( Connection not available ). Thus I need to schedule indexing once
Unfortunately to track the
status of
indexing for a core one need to keeping pinging the server
to check
completion status. Is there a way to get a response from
SOLR once the
indexing is complete ?
Yes it is possible :
http://wiki.apache.org/solr/DataImportHandler#EventListeners
Added https://issues.apache.org/jira/browse/SOLR-4237
Otis
--
Performance Monitoring - http://sematext.com/spm/index.html
Search Analytics - http://sematext.com/search-analytics/index.html
On Tue, Dec 25, 2012 at 9:13 PM, Mark Miller markrmil...@gmail.com wrote:
I've been thinking about
Hi Guys,
I also worked on a CombiningTokenFilter, see:
https://issues.apache.org/jira/browse/LUCENE-3413
Patch has been up and available for a while.
HTH!
Cheers,
Chris
On 12/27/12 12:26 AM, Dmitry Kan solrexp...@gmail.com wrote:
Hi,
Have a look onto TokenFilter. Extending it will give
Hi Lance,
Thanks for the response.
I didn't quite understand how to issue the queries from DirectSpellChecker
with the fq params applied like you were suggesting - could you point me to
the API that can be used for this?
Also, we haven't benchmarked the DirectSpellChecker against the
I think the answer is yes, that there's a better way to doing all of this. But
I'm not yet sure what this all entails in your situation. What are you
overriding with the Lucene searches? I imagine Solr has the flexibility to
handle what you're trying to do without overriding anything core
Hi Eric,
Sorry, I think I wasn't very clear in explaining what we need to do.
We don't really need to do any complicated overriding, just want to change
the DirectSpellChecker to issue a query for every correction it finds *with
fq params from the original query taken into account* so that we
Apologies for misunderstanding.
Does what you're trying to do already work this way using the
http://wiki.apache.org/solr/SpellCheckComponent#spellcheck.maxCollationTries
maxCollationTries feature of the spellcheck component?
It looks like it passes through the fq's even, so that the hit
Nalini,
You could take the code from SpellCheckCollator#collate and have it issue a
test query for each word individually instead of for each collation. This
would do exactly what you want. See
Hi James,
Yup, that was what I tried to do initially but it seems like calling
through to those Solr methods from DirectSpellChecker was not a good idea -
am I wrong? And like you mentioned, this seemed like it wasn't low-level
enough.
Eric: Unfortunately the collate functionality does not work
Nalini,
Assuming that you're using Solr, the hook into the collate functionality is in
SpellCheckComponent#addCollationsToResponse . To do what you want, you would
have to modify the call to SpellCheckCollator to issue test queries against the
individual words instead of the collations.
See
That's debugQuery=true or debug=query.
-- Jack Krupansky
-Original Message-
From: Otis Gospodnetic
Sent: Thursday, December 27, 2012 10:56 AM
To: solr-user@lucene.apache.org
Subject: Re: search with spaces
Hi,
Add debugQuery=query to your search requests. That will point you in the
Hello,
I am seeing frequent OOMs for the past 2 days on a SolrCloud Cluster
(Solr4.0 with a patch from Solr-2592) setup (3 shards, each shard with 2
instances. Each instance is running CentOS with 30GB memory, 500GB disk
space), with a separate Zoo Keeper ensemble of 3.
Here is the stacktrace:
Hi,
I'm using master/slave replication on Solr 4.0.
Replication is successfully run.
But old index not cleaned up.
Is that bug or not?
My slave index directory is below...
$ ls -l solr_kr/krg01/data/index/
total 23472512
-rw-r--r--. 1 tomcat tomcat563722625 Dec 24 21:48 _15.fdt
-rw-r--r--.
Do you see any errors coming in on the console, stderr?
I start solr this way and redirect the stdout and stderr to log files, when
I have a problem stderr generally has the answer:
java \
-server \
-Djetty.port=8080 \
-Dsolr.solr.home=/opt/solr \
I'm doing a query like this for MoreLikeThis, sending it a document ID. But
the only result I ever get back is the document ID I sent it. The debug
response is below.
If I read it correctly, it's taking id:1004401713626 as the term (not the
document ID) and only finding it once. But I want it to
Hi David,
From what I see in the log and threaddump it seems that getSearcher method
in SolrCore is not able to acquire required lock and because of that its
blocking startup of the server. Here is threaddump -
http://pastebin.com/GPnAzF1q .
On Fri, Dec 28, 2012 at 8:01 AM, David Parks
Sounds like it is simply dispatching to the normal search request handler.
Although you specified qt=mlt, make sure you enable the legacy select
handler dispatching in solrconfig.xml.
Change:
requestDispatcher handleSelect=false
to
requestDispatcher handleSelect=true
Or, simply
Ok, that worked, I had the /mlt request handler misconfigured (forgot a
'/'). It's working now. Thanks!
-Original Message-
From: Jack Krupansky [mailto:j...@basetechnology.com]
Sent: Friday, December 28, 2012 11:38 AM
To: solr-user@lucene.apache.org
Subject: Re: MoreLikeThis only returns
I'm somewhat new to Solr (it's running, I've been through the books, but I'm
no master). What I hear you say is that MLT *can* accept, say 5, documents
and provide results, but the results would essentially be the same as
running the query 5 times for each document?
If that's the case, I might
Here is latest threaddump taken after setting up latest nightly build
version - apache-solr-4.1-2012-12-27_04-32-37 - http://pastebin.com/eum7CxX4
Kind of stuck with this from few days now, so can use little help.
Here is more details on the issue -
1. Setting up jetty + solr using instructions
Hi Dave,
Think of search components as a chain of Java classes that get executed
during each search request. If you open solrconfig.xml you will see how
they are defined and used.
HTH
Otis
Solr ElasticSearch Support
http://sematext.com/
On Dec 28, 2012 12:06 AM, David Parks
If I comment out the /browse requesthandler from solrconfig.xml, problem
goes away. So issue is definitely with the way I am configuring
solrconfig.xml. I will debug into on my side.
On Fri, Dec 28, 2012 at 11:55 AM, Sushrut Bidwai
bidwai.sush...@gmail.comwrote:
Here is latest threaddump taken
So the Search Components are executed in series an _every_ request. I
presume then that they look at the request parameters and decide what and
whether to take action.
So in the case of the MLT component this was said:
The MLT search component returns similar documents for each of the
29 matches
Mail list logo