Very interesting! Thanks for sharing, I'll ponder on it.
--
View this message in context:
http://lucene.472066.n3.nabble.com/Wildcard-query-vs-facet-prefix-for-autocomplete-tp3995199p3995899.html
Sent from the Solr - User mailing list archive at Nabble.com.
Hello,
I have DIH reading an XML file and getting fields with empty values.
My definition is:
/text here is actual node name, not text() (e.g. )
Right now, I get the field (of type string) with empty value
indexed/stored/returned. Plus, all the copy fields get the empties as
well.
Can I get DI
I had tried with splitBy for numeric field, but that also did not worked
for me. However I got rid of group_concat and it was all good to go.
Thanks a lot!! I really had a difficult time understanding this behavior.
*Pranav Prakash*
"temet nosce"
On Thu, Jul 19, 2012 at 1:34 AM, Dyer, James
Greetings,
I've been digging in to this for two days now and have come up short -
hopefully there is some simple answer I am just not seeing:
I have a solr 1.4.1 instance and a solr 3.6.0 instance, both configured as
identically as possible (given deprecations) and indexing the same document.
Fo
Yury,
Thank you so much! That was it. Man, I spent a good long while trouble
shooting this. Probably would have spent quite a bit more time. I
appreciate your help!!
-Briggs
On Wed, Jul 18, 2012 at 9:35 PM, Yury Kats wrote:
> On 7/18/2012 7:11 PM, Briggs Thompson wrote:
> > I have realized thi
Put my question wrong.. Excuse me for spamming.. its been a tiring couple
of days and I am almost sleep typing.. Please read the snippet again.
This might be a dumb question. But I would like to confirm.
>
> Will the following snippet cause a index searcher leak and end up in an
> out of memory e
On 7/18/2012 7:11 PM, Briggs Thompson wrote:
> I have realized this is not specific to SolrJ but to my instance of Solr.
> Using curl to delete by query is not working either.
Can be this: https://issues.apache.org/jira/browse/SOLR-3432
Increasing the polling interval does help. But the requirement is to get a
document indexed and searchable instantly ( sounds like RTS), 30 sec is
acceptable.I need to look at Solr NRT and cloud.
I created a new core to accept daily updates and replicate every 10 sec.
Two other cores with 234 Mill
Yury and Shashi,
Thanks very much for helps! I am studying the options pointed
out by you (Solr multiple cores and Elasticsearch).
Best regards, Lisheng
-Original Message-
From: Yury Kats [mailto:yuryk...@yahoo.com]
Sent: Tuesday, July 17, 2012 7:19 PM
To: solr-user@lucene.apache.org
Sub
I am implementing a search engine with Nutch as web crawler and Solr for
searching. Now,since Nutch has no search-user-interface any more, I came to
know about Ajax-Solr as search-user-interface.
I implemented Ajax-Solr with no hindrance, but during its search operation
its only search under reute
When I setup a 2 shard cluster using the example and run it through its
paces, I find two features that do not work as I expect. Any suggestions on
adjusting my configuration or expectations would be appreciated.
/terms does not return any terms when issued as follows:
http://hostname:8983/solr/t
Hi Briggs,
I'm not sure about Solr 4.0, but do you need to commit?
> curl http://localhost:8983/solr/coupon/update?commit=true -H "Content-Type:
> text/xml" --data-binary '*:*'
Brendan
www.kuripai.com
On Jul 18, 2012, at 7:11 PM, Briggs Thompson wrote:
> I have realized this is not specifi
Replication will indeed be incremental. But if you commit too often (and
committing too often a common mistake) then the merging will
eventually merge everything into new segments and the whole thing will
be replicated.
Additionally, optimizing (or forceMerge in 4.x) will make a single segment
and
Hi,
I am trying out the Solr Alpha release against some custom and Junit codes
I have written. I am seeing my custom JUnit tests failing once in a while.
The tests are based on Solr Junit test code where they are extending
SolrTestCaseJ4. My guess is the Randomized Testing coming across some issu
I have realized this is not specific to SolrJ but to my instance of Solr.
Using curl to delete by query is not working either.
Running
curl http://localhost:8983/solr/coupon/update -H "Content-Type: text/xml"
--data-binary '*:*'
Yields this in the logs:
INFO: [coupon] webapp=/solr path=/update
pa
On 07/18/2012 02:39 PM, Richard Frovarp wrote:
On 07/18/2012 11:20 AM, Erick Erickson wrote:
the ~2 is the mm parameter I'm pretty sure. So I'd guess your
configuration has
a mm parameter set on the core that isn't doing what you want..
I'm not setting the mm parameter or the q.op paramet
Thank you! I haven't really understood the LuceneTestCase.classRules
before this.
roman
On Wed, Jul 18, 2012 at 3:11 PM, Chris Hostetter
wrote:
>
> : I am porting 3x unittests to the solr/lucene trunk. My unittests are
> : OK and pass, but in the end fail because the new rule checks for
> : modi
Don't you want to specify "splitBy" for the integer field too?
Actually though, you shouldn't need to use GROUP_CONCAT and RegexTransformer at
all. DIH is designed to handle "1>many" relations between parent and child
entities by populating all the child fields as multi-valued automatically. I
I'd guess the getSearcher call you are making is incrementing the ref count and
you are not decrementing it?
On Jul 18, 2012, at 12:17 PM, Karthick Duraisamy Soundararaj wrote:
> Hi All,
>The SolrCore seems to have a reference counted searcher with it. I
> had to write a customSear
That approach would work for core dependent parameters. In my case, the
params are environment dependent. I think a simpler approach would be to
pass the url param as JVM options, and these XMLs get it from there.
I haven't tried it yet.
*Pranav Prakash*
"temet nosce"
On Tue, Jul 17, 2012 at
On 07/18/2012 11:20 AM, Erick Erickson wrote:
the ~2 is the mm parameter I'm pretty sure. So I'd guess your configuration has
a mm parameter set on the core that isn't doing what you want..
I'm not setting the mm parameter or the q.op parameter. All three cores
have a defaultOperator of O
I have a multivalued integer field and a multivalued string field defined
in my schema as
The DIH entity and field defn for the same goes as
The value for field community_tags comes correctly as an array of strings.
However the value of field community_tag_ids
: I am porting 3x unittests to the solr/lucene trunk. My unittests are
: OK and pass, but in the end fail because the new rule checks for
: modifier properties. I know what the problem is, I am creating new
: system properties in the @beforeClass, but I think I need to do it
: there, because the p
Could anyone suggest the options available to handle the following situation:
1. Say we have 1,000 authors
2. 65% of these authors have 10-100 titles they authored; the others have
not authored any titles but provide only their biography and writing
capability.
3. We want to search for authors,
I have a "keyword" field type that I made:
Hi Eric,
I totally agree. That's what I also figured ultimately. One thing I am not
clear. The replication is supposed to be incremental ? But looks like it
is trying to replicate the whole index. May be I am changing the index so
frequently, it is triggering auto merge and a full replication ?
See:
http://wiki.apache.org/solr/SolrReplication#enable.2BAC8-disable_master.2BAC8-slave_in_a_node
I'll admit that I haven't tried this personally, but I think it'll work.
Although I'm pretty sure that if you just disable the master,
disabling the polling on the slave isn't necessary.
Best
Eric
But I did run across an idea a while ago... Either with a custom
update processor
or on the client side, you permute the title so you index something like:
Shadows of the Damned
of the Damned&Shadows
the Damned&Shadows of
Damned&Shadows of the
Index these with KeywordTokenizer and LowercaseFilter.
bq: This index is only used for searching and being replicated every 7 sec from
the master.
This is a red-flag. 7 second replication times are likely forcing your
app to spend
all its time opening new searchers. Your cached filter queries are
likely rarely being re-used
because they're being throw
the ~2 is the mm parameter I'm pretty sure. So I'd guess your configuration has
a mm parameter set on the core that isn't doing what you want..
Best
Erick
On Tue, Jul 17, 2012 at 3:05 PM, Richard Frovarp wrote:
> On 07/14/2012 05:32 PM, Erick Erickson wrote:
>>
>> Really hard to say. Try exe
Have you tried the analysis window to debug.
I believe you are doing something wrong in the fieldType.
On Wed, Jul 18, 2012 at 8:07 PM, Husain, Yavar wrote:
> Thanks Sahi. I have replaced my EdgeNGramFilterFactory to
> NGramFilterFactory as I need substrings not just in front or back but
> anyw
Thanks Sahi. I have replaced my EdgeNGramFilterFactory to NGramFilterFactory as
I need substrings not just in front or back but anywhere.
You are right I put the same NGramFilterFactory in both Query and Index however
now it does not return any results not even the basic one.
-Original Messa
You are creating grams only while indexing and not querying hence 'ludlwo'
would not match. Your analyzer will create the following grams while
indexing for 'ludlow': lu lud ludl ludlo ludlow and hence would not match
to 'ludlwo'.
Either you need to create gram while querying also or use Edit Dist
Hi Guys,
Steps to reproduce:
1) Download apache-solr-4.0.0-ALPHA
2) cd example; java -jar start.jar
3) cd exampledocs; ./post.sh *.xml
4) Use statsComponent to get the stats info for field 'popularity' based on
facet 'cat'. And the 'count' for 'electronics' is 3
http://localhost:8983/solr/coll
I have configured NGram Indexing for some fields.
Say I search for the city Ludlow, I get the results (normal search)
If I search for Ludlo (with w ommitted) I get the results
If I search for Ludl (with ow ommitted) I still get the results
I know that they are all partial strings of the main
Well silly me... you're right.
On Wed, Jul 18, 2012 at 6:44 PM, Erick Erickson [via Lucene] <
ml-node+s472066n399570...@n3.nabble.com> wrote:
> Well, option 2 won't do you any good, so speed doesn't really matter.
> Your response would have a facet count for "dam", all by itself, something
> like
Well, option 2 won't do you any good, so speed doesn't really matter.
Your response would have a facet count for "dam", all by itself, something like
2
1
etc.
which does not contain anything that lets you reconstruct the title
for autosuggest.
Best
Erick
On Tue, Jul 17, 2012 at 3:18 AM, santam
Hi folks,
i have this case:
i want to update my solr 4.0 from trunk to solr 4.0 alpha. the index
structure has changed, i can't replicate.
10 cores are in use, each with 30Mio docs. We assume that all fields
are stored and indexed.
What is the best way to export the docs from all cores on one mach
Nick,
to solve out of memory issue, i think you can make below change:
1) in solrsconfig.xml, reduce ramBufferSizeMB (there are two, change both)
2) in solrsconfig.xml, reduce documentCache value
to solve call commit slow down index issue, i think you can change new
search default queyr:
in solrs
Dear developers,
while upgrading from 3.6.x to 4.x I have to rewrite some of my code and
search for the new methods and/or classes. In 3.6.x and older versions
the API Javadoc interface had an "Index" which made it easy to find the
appropriate methods. The button to call the "Index" was located in
Hi,
Well, if I understand correctly, only the search term is important for
elevation, not the query.
Anyway, we ended up modifying QueryElevationComponent class, extracting the
search term from the query using regex.
After that, it turned out that elevation doesn't work with grouped results,
so w
Hi,
It's possible to start the solr master and slave with the following
configuration
- replication on master disabled when we start solr --> the replication
feature must be available
- polling on slave disabled --> the replication feature must be available
-- Best Regards
-- Jamel
--
View t
I think what makes the most sense is to limit the number of
connections to another host. A host only has so many CPU resources,
and beyond a certain point throughput would start to suffer anyway
(and then only make the problem worse). It also makes sense in that a
client could generate documents
Hi,
I am using solr for indexing. Index size is small and it is around
50GB. I need to use solr for high query throughput system. I am using
twitter api and I need to search incoming tweet in solr. So I want to
know how should I design such system ? Does solr supports HDFS
natively ? How can I ind
44 matches
Mail list logo