Hi All,
I have a small test index of 400 documents , it happens to have an entry
for "wrangler", When I search for "wranglr", I correctly get the collation
suggestion as "wrangler", however when I search for "wrangle" , I do not
get a suggestion for "wrangler".
The Levenstien distance between wr
James,
Thanks , there is no error in the logs, it just that I do not get the start
up message in the log.
I do not see any warm up related configuration for any spell checker in my
solrconfig.xml , I have also pasted the auto warm related configuration
data below .
1024
I added the primaryKey as the and still same result.
On May 30, 2014, at 8:38 PM, Ahmet Arslan wrote:
> Hi,
>
> Sure, have a look at definition in example schema.xml
>
> http://wiki.apache.org/solr/UniqueKey
>
>
>
>
> On Saturday, May 31, 2014 3:35 AM, PeriS wrote:
> No. Is there a wa
Hi,
Sure, have a look at definition in example schema.xml
http://wiki.apache.org/solr/UniqueKey
On Saturday, May 31, 2014 3:35 AM, PeriS wrote:
No. Is there a way to have the primary key of my entity be the unique key?
On May 30, 2014, at 7:00 PM, Ahmet Arslan wrote:
> Hi,
>
> Do you
No. Is there a way to have the primary key of my entity be the unique key?
On May 30, 2014, at 7:00 PM, Ahmet Arslan wrote:
> Hi,
>
> Do you have uniqueKey defined in schema.xml ?
>
>
>
> On Saturday, May 31, 2014 1:23 AM, PeriS wrote:
> Hi,
>
> I have followed the documentation to set up
Hi,
Do you have uniqueKey defined in schema.xml ?
On Saturday, May 31, 2014 1:23 AM, PeriS wrote:
Hi,
I have followed the documentation to set up my delta query, but when I call the
delta-import, the index is happening again for the same record and ends up
being indexed twice. Any clues pl
Hi,
I have followed the documentation to set up my delta query, but when I call the
delta-import, the index is happening again for the same record and ends up
being indexed twice. Any clues please?
Thanks
-Peri
*** DISCLAIMER *** This is a PRIVATE message. If you are not the intended
rec
Gregg,
I don’t have an answer to your question but I’m very curious what use case you
have that permits such arbitrary partial-results. Is it just an edge case or
do you want to permit a common occurrence?
Jason
On May 30, 2014, at 3:05 PM, Gregg Donovan wrote:
> I'd like a to add a hard ti
I'd like a to add a hard timeout on some of my sharded requests. E.g.: for
about 30% of the requests, I want to wait no longer than 120ms before a
response comes back, but aggregating results from as many shards as
possible in that 120ms.
My first attempt was to use timeAllowed=120&shards.tolerant
bq: it's not getting replicated
This is definitely not what's expected. Are you by chance _configuring_
replication while at the same time using SolrCloud? Posting your
solrconfig.xml would help answer that.
This should be all that's in your solrconfig.xml file when running under
SolrCloud:
No
We've switched to CMS GC to see if there is any improvement.
Looking at this use case, G1GC might have been a better option, but we are
running JDK 1.6
http://blog.sematext.com/2013/06/24/g1-cms-java-garbage-collector/
Jeremy D. Branham
Tel: **DOTNET
-Original Message-
From: Branham,
Erick,
Thanks a bunch. Good to know the internals.
Best,
Robin
--
View this message in context:
http://lucene.472066.n3.nabble.com/DataImportHandler-while-Replication-tp4138763p4138984.html
Sent from the Solr - User mailing list archive at Nabble.com.
As of now I'm trying reindexing everything. Basically I have now an empty
collection and when I add a document it's not getting replicated. Not
trying anymore to load my old index anymore.
Marc
2014-05-30 16:44 GMT-04:00 Jason Hellman :
> Marc,
>
> Fundamentally it’s a good solution design to a
I just realized I failed my own reading comprehension :)
You have maxDocs, not maxTime for hard commit. Please disregard.
On May 30, 2014, at 1:46 PM, Jason Hellman
wrote:
> I’m also not sure I understand the practical purpose of your hard/soft auto
> commit settings. You are stating the fo
Hi,
Custom similarity returning 1 in coord factor?
http://lucene.apache.org/core/4_6_0/core/org/apache/lucene/search/similarities/TFIDFSimilarity.html#coord(int,%20int)
On Friday, May 30, 2014 11:41 PM, Diego Fernandez wrote:
Hi! I have a question which I posted on
http://stackoverflow.com/q
I’m also not sure I understand the practical purpose of your hard/soft auto
commit settings. You are stating the following:
Every 10 seconds I want data written to disk, but not be searchable.
Every 15 seconds I want data to be written into memory and searchable.
I would consider whether your s
Hi all,
What's the cleanest way to solve this problem: localize the 'display
names' for facet values without storing the localized names in solr.
Example:
-store 'country code' field in solr document
-facet on country code
-translate the country code based on the user's locale
For the facets
Marc,
Fundamentally it’s a good solution design to always be capable of reposting
(reindexing) your data to Solr. You are demonstrating a classic use case of
this, which is upgrade. Is there a critical reason why you are avoiding this
step?
Jason
On May 30, 2014, at 10:38 AM, Marc Campeau
Hi! I have a question which I posted on
http://stackoverflow.com/questions/23959727/sum-of-nested-queries-in-solr about
taking the sum of OR'd nested queries. I'll repeat it here, but if you want
some SO points and have an answer, feel free to answer there.
[quote]
We have a search that takes
No, I'm not indexing. My index has been static for most than 2 weeks.
I'm running the exact same query over and over. The CloudSolrServer is
constructed through spring, build my query using the SolrQuery class, and
then execute a the query a few times in a loop.
Query: atp:20
Start: 0
Rows: 20
are you indexing at the same time?
And when you say "exact same queries", is this the same _form_ or the same
query over and over? If the latter, then the responses are being served
from the queryResultCache probably. Which doesn't explain why 2 and 3 take
more time.
It is _vaguely_ possible that
bq: because despite the issue title being about non-SolrCloud mode, later
comments include a lot of discussions about zookeeper.
Right. SOLR-4478 is, indeed, about allowing named configuration sets in
non-cloud mode. The ZooKeeper comments were mostly about getting the two
different modes to play
You don't need to do anything. Essentially, when replication starts the
current state of the master's index is "frozen" in the sense that the
_closed_ segments that make up the index at the time replication starts are
the only ones that are replicated to the slave.
All indexing happens into an _op
I don't think that's the case. Ignoring the results from the admin screen,
I'm still left with inconsistencies within the SolrJ calls.
I ran the exact same query (a simple integer equality returning 9 documents)
5 times every 10 seconds through solrJ and I get the following results (3
separate
Hi Shawn,
Thanks very much for the feedback.
Have tested using the routing mechanism/composite-id on a larger scale.
Unfortunately the same behaviour.
Regards
Ronald
-Original Message-
From: Shawn Heisey [mailto:s...@elyograg.org]
Sent: 29 May 2014 20:16
To: solr-user@lucene.apache.or
On 5/30/2014 12:25 PM, drmangrum wrote:
> -Running a 2 shard (each with a replica) setup with 100k records on JBoss7.
> -Using SolrJ and CloudSolrServer to interface with SOLR.
>
> Executing queries through the admin screen query tool is amazingly fast,
> usually less than 40ms. However, when I ex
On 5/30/2014 12:00 PM, Jim.Musil wrote:
> I’m attempting to define a core using the new core discovery method described
> here:
>
> http://wiki.apache.org/solr/Core%20Discovery%20(4.4%20and%20beyond)
>
> At the bottom of the page is a parameter named configName that should allow
> me to specify a
-Running a 2 shard (each with a replica) setup with 100k records on JBoss7.
-Using SolrJ and CloudSolrServer to interface with SOLR.
Executing queries through the admin screen query tool is amazingly fast,
usually less than 40ms. However, when I execute the queries through my java
application I g
On 5/29/2014 9:55 PM, M, Arjun (NSN - IN/Bangalore) wrote:
> Thanks a lot for your nice explanation.. Now I understood the
> difference between autoCommit and autoSoftCommit.. Now my config looks like
> below.
>
>
>1
>false
>
>
>
>15000
>
>
>
Hi,
I’m attempting to define a core using the new core discovery method described
here:
http://wiki.apache.org/solr/Core%20Discovery%20(4.4%20and%20beyond)
At the bottom of the page is a parameter named configName that should allow me
to specify a configuration name to use for a collection. Th
2014-05-30 12:24 GMT-04:00 Erick Erickson :
> Let's back up a bit here. Why are you copying your indexes around?
> SolrCloud does all this for you. I suspect you've somehow made a mis-step.
>
I started by copying the index around because my 4.5.1 instance is not
setup as Cloud and I wanted to avo
Hello Shalin,
Appreciate your reply. I've not implemented DIH in production and now we are
considering. hence the question, if we configure DIH on the master from
which the Slave is replicating the index, we will need to control this
externally? or is there any setting in DIH that will allow us to
Sorry didn't format it correctly.. here is the output without facet.threads:
query: q=*:*&fq={!tag=ha15
df=attributes}"15=14"&fl=id&facet=on&facet.sort=index&facet.mincount=1&facet.field={!key=attrib}attributes&facet.field={!ex=ha15
key=ha15 facet.prefix=15=}attributes&rows=0
0
Sure. I have a multi-valued string field called "attributes" with values
stored in the format "id=value", e.g. "15=14". When one of the values is
selected, I use ex to get the counts for the other values of the currently
selected field (by using facet.prefix=15=).
q=*:*&fq={!tag=ha15
df=attri
We saw the file descriptors peak out and full GCs running causing DOS on our
SOLR servers this morning.
* Does this stack trace give enough information for some ideas?
* solr-spec
4.5.1-SNAPSHOT
* solr-impl
4.5.1-SNAPSHOT ${svnversion} - kx101435 - 2013-11-04 17:39:36
*
Sure, boosts < 1 "de-boost" matches on those terms.
Best
Erick
P.S. Your use-case still isn't very clear, some examples would help.
On Fri, May 30, 2014 at 2:57 AM, Christian Loock wrote:
> An alternative might be to drop down its relevancy for certain words.
> Would that be possible?
>
>
>
Let's back up a bit here. Why are you copying your indexes around?
SolrCloud does all this for you. I suspect you've somehow made a mis-step.
So here's what I'd do by preference; Just set up a new collection and
re-index. Make sure all of the nodes are up and then just go ahead and
index to any of
Make sure you checkout from the "tagged" branch, not trunk or the actual
release branch which has updates since the release.
See:
http://svn.apache.org/viewvc/lucene/dev/tags/
-- Jack Krupansky
-Original Message-
From: eShard
Sent: Friday, May 30, 2014 11:55 AM
To: solr-user@lucene.a
OK, I see what you're trying to do. Unfortunately grouping is just not
built to support multivalued fields. The nearest thing I can think of is
"pivot faceting", but that only returns counts, not the actual documents.
Sorry I can't be more help
Erick
On Fri, May 30, 2014 at 3:44 AM, Bhoomit Vas
Better to implement this first at your application layer, and when you are
finally happy with the results then consider pushing it down into Solr, or
even better, leave it at the application level.
You might want to take a look at the "Unsupervised Feedback" feature of the
LucidWorks Search So
Ok, I think I figured it out.
Somehow my Solr4.0Final project was accidentally updated to 5.0.
The solr/build.xml was fine.
the build.xml file at the top level was pointed at 5.0-snapshot.
I need to pull down the 4.0 and start from scratch.
--
View this message in context:
http://lucene.47206
I'd spend some time with the admin/analysis page to understand the exact
tokenization going on here. For instance, sequencing the
shinglefilterfactory before worddelimiterfilterfactory may produce
"interesting" resutls. And then throwing the Snowball factory at it and
putting synonyms in front
Please review: http://wiki.apache.org/solr/UsingMailingLists
There's not much info to go on here. At a guess, you haven't issued a
"commit".
Best,
Erick
On Fri, May 30, 2014 at 3:26 AM, Sohan Kalsariya
wrote:
> Hi all:
> I'm using solr 4.7.1
> While indexing my data on solr, I am not able to
BTW, memory usage is _much_ better in Solr 4.x, if it's at all an option
you might consider upgrading. See:
http://searchhub.org/2012/04/06/memory-comparisons-between-solr-3x-and-trunk/
Best,
Erick
On Thu, May 29, 2014 at 11:40 AM, Toke Eskildsen
wrote:
> bihan.chandu [bihan.cha...@gmail.com]
Good morning,
My company uses Solr4.0Final and I need to add some code to it and
recompile.
However, when I rebuild, all of the jars and the war file say Solr 5.0!
I'm using the old build.xml file from 4.0 so I don't know why it's
automatically upgrading.
How do I force it to build the older versi
Not unless you disable the update handler completely. Assuming you have
sensitive data, you really need to secure your Solr server behind an app
though.
Best,
Erick
On Thu, May 29, 2014 at 9:12 AM, Joshi, Shital wrote:
> Hi,
>
> What are ways to prevent someone executing random delete commands
You can copy to the shards and use the mergindexes command, the
MapReduceIndexerTool follows that approach.
But really, what is the higher-level use-case you're trying to support?
This feels a little like an XY problem. You could do things like
1> index to a different collection then use collectio
Consider using the edismax query parser, which allows you to specify q.alt
which is an alternative query such as "*:". Or use an application layer to
handle the situation before it gets to Solr.
-- Jack Krupansky
-Original Message-
From: vit
Sent: Friday, May 30, 2014 9:01 AM
To: sol
I see errors in both cases. Do you
1> have schemaless configured
or
2> have a dynamic field pattern that matches your "non_exist_field"?
Best,
Erick
On Tue, May 27, 2014 at 8:54 PM, 小川修 wrote:
> Hi.
>
> I use Solr4.7.2.
>
> When I search by wrong query
> for example
> [non_exists_field:value].
Do these forbidden documents have 'certain search words' in them? May be
KeepWordFilterFactory could help.
On Friday, May 30, 2014 12:57 PM, Christian Loock wrote:
Hi,
well we have a product search which often will return products one might
not expect because they contain somesort of referen
Hi,
Where is the fix for SOLR-5426 ?
On Friday, May 30, 2014 10:23 AM, Mistminds wrote:
These are the two issues we fixed for SOLR API
https://issues.apache.org/jira/browse/SOLR-5426
https://issues.apache.org/jira/browse/SOLR-5423
--
View this message in context:
http://lucene.472066.
I am not sure why changing spellcheck parameters would prevent your server from
restarting. One thing to check is to see if you have warming queries running
that involve spellcheck. I think I remember from long ago there was (maybe
still is) an obscure bug where sometimes it will lock up in ra
Hi, forgot to mention that I'm migrating the index from Solr 4.5.1 to 4.8.1.
Thanks,
Marc Campeau
2014-05-30 9:54 GMT-04:00 Marc Campeau :
> Hi,
>
> I currently have a standalone SOLR 4.5.1deployment on an EC2 instance with
> a single collection and core containing an index that's roughly 10G.
Hi,
I currently have a standalone SOLR 4.5.1deployment on an EC2 instance with
a single collection and core containing an index that's roughly 10G. I've
used this as a proof of concept, prototype and staging during development
phases and I'm about to release to production.
For this release, I've
In my code sometimes I get an emty query. Namely ... q=& which causes
"bad request" exception.
How to gracefully check query correctness and bypass execution if it is not
correct.
--
View this message in context:
http://lucene.472066.n3.nabble.com/How-to-check-and-handle-empty-Solr-query-t
Thank you very much. I will take a look at that.
On Fri, May 30, 2014 at 4:24 PM, Ahmet Arslan wrote:
> Hi Ali,
>
> I did a similar user filtering by indexing user table once per hour, and
> filtering results by solr query time join query parser.
>
> Assuming there is no easy way to transfer US
Hi Ali,
I did a similar user filtering by indexing user table once per hour, and
filtering results by solr query time join query parser.
Assuming there is no easy way to transfer USER table to solr, Solr post
filtering is the way to :
http://searchhub.org/2012/02/22/custom-security-filtering-
Is the issue SOLR-5478 what you were looking for?
Hi every body,
I was wondering is there any way for using cross doc join on integraion of
one solr core and a relational database.
Suppose I have a table in relational database (my sql) name USER. I want to
keep track of news that each user can have access. Assume news are stored
inside solr and th
Hi Erick,
Thanks for the response.
What am I trying to do? :
I have a multivalued field called tags for e.g.
- {
- id: "1",
- *tags*:["solr","lucene","facet"],
*type*:"post"
},
- {
- id: "2",
- *tags*:["solr","multi-valued","grouping"],
*type*:"qu
Hi Folks,
Any updates ??
On Wed, May 28, 2014 at 12:13 PM, sunshine glass <
sunshineglassof2...@gmail.com> wrote:
> Dear Team,
>
> How can I handle compound word searches in solr ?.
> How can i search "hand bag" if I have "handbag" in my index. While using
> shingle in query analyzer, the query
Hi all:
I'm using solr 4.7.1
While indexing my data on solr, I am not able to get query results on solr.
So how can I get query results while indexing the data on the solr?
Sounds like a bug. Can you paste an example request/response that can help
us reproduce the problem?
On Fri, May 30, 2014 at 11:55 AM, Kenneth Yung wrote:
> Hi all:
> I'm using solr 4.7 and my application uses local param syntax to specify
> different facet.prefix on the same field. It works f
They are completely separate components in Solr. Are you seeing performance
problems in replication due to indexing or vice versa?
On Fri, May 30, 2014 at 10:10 AM, Robin Woods wrote:
> Hi,
>
> What would happen to DataImportHandler that is setup on the master when the
> slave is in the process
An alternative might be to drop down its relevancy for certain words.
Would that be possible?
On 30.05.2014 11:55, Christian Loock wrote:
Hi,
well we have a product search which often will return products one
might not expect because they contain somesort of reference to other
products.
I
Hi,
well we have a product search which often will return products one might
not expect because they contain somesort of reference to other products.
I want it to appear in the index, so it can be found in general, but I
dont want it to be found for certain search words.
On 30.05.2014 11:5
Explain your use case a little more, but you can define terms as stop words
with a stop filter, which means they won't appear in the index.
-- Jack Krupansky
-Original Message-
From: Christian Loock
Sent: Friday, May 30, 2014 5:38 AM
To: solr-user@lucene.apache.org
Subject: Block docu
Hi,
is there a way to block a document for being found when you search for a
certain way?
Cheers,
Christian
--
Christian Loock
Web Developer
Renzel Agentur
www.renzel-agentur.de
I am using solr3.6.2 and I found the default highlight does not work like
solr1.4.1 or FastVectorHighlighter.
Is it a bug?
I use 2gram tokenizer.
example:
text:testabctest123456testabc
index:te es st ta ab bc ct te es st t1 12 23 ・・・
query:test
parameters:
hl=true
hl.fragsize=200
hl.simple.pre={
solrconfig.xml:
solr.FileBasedSpellChecker
file
spellings.txt
UTF-8
./spellcheckerFile
file
5
spellcheck
spellings.txt:
apple
apples
apple's
applet
applets
appliers
1) http://192.168.1.99:8983/so
These are the two issues we fixed for SOLR API
https://issues.apache.org/jira/browse/SOLR-5426
https://issues.apache.org/jira/browse/SOLR-5423
--
View this message in context:
http://lucene.472066.n3.nabble.com/We-fixed-two-issues-for-SOLR-API-tp4138628.html
Sent from the Solr - User mailing
71 matches
Mail list logo