Eh, this replies through Nabble are really problematic. I don't recall what
the original error was any more. java-u...@lucene is the best place to ask
Lucene questions.
Otis
--
Sematext -- http://sematext.com/ -- Lucene - Solr - Nutch
From: Qingdi
To: sol
Mauricio - are you aware of SolrSharp - a Solr client for .NET? Would it be
better to contribute to SolrSharp instead of creating another .NET client, or
is your client going to be built very differently?
Otis
--
Sematext -- http://sematext.com/ -- Lucene - Solr - Nutch
___
this link is not working example.solrstuff.org. who is maintaining it?
On Thu, Feb 12, 2009 at 7:57 AM, Mauricio Scheffer
wrote:
> Hi everyone. I'm developing a Solr client for .NET (
> http://code.google.com/p/solrnet/) and I was wondering if I could use the
> Solr instance at example.solrstuff.
This was a false alarm, sorry. I misinterpreted some results.
wojtekpia wrote:
>
> Has there been a recent change (since Dec 2/08) in the paging algorithm?
> I'm seeing much worse performance (75% drop in throughput) when I request
> 20 records starting at record 180 (page 10 in my application
Hi everyone. I'm developing a Solr client for .NET (
http://code.google.com/p/solrnet/) and I was wondering if I could use the
Solr instance at example.solrstuff.org (the one used by solrjs) to build an
online demo of my library... Of course, this would be just read-only access,
no updates. I would
Hi Otis,
Thanks for your quick response. We are on solr 1.3.
We cannot upgrade to solr 1.4-dev at this moment. Do you know where we can
find more details on how lucene optimization process work? We want to check
if there is any solr config parameter we could adjust to avoid this problem.
Thanks
+1. Of course it is doable, but that doesn't mean you should, which
is what I was trying to say before, (but was typing on my iPod so it
wasn't fast) and which Walter has done so. It is entirely conceivable
to me that someone could search for a very common word such that the
score of all
Hi,
I'm trying to get some information how boost is used in the ranking
calculation via the debugQuery parameter for the following query:
(bodytext:iphone OR bodytext:firmware)^2.0 OR dateCreatedYear:2009^5.0
For one of the matching documents I can see:
4.7144237 = (MATCH) sum of:
2.2903786
Don't bother doing this. It doesn't work.
This seems like a good idea, something that would be useful for
almost every Lucene installation, but it isn't in Lucene because
it does not work in the real world.
A few problems:
* Some users want every match and don't care how many pages of
results th
Hi,
I'm trying to understand the internal Sturcture of the lucene indexer.
Well according to "Lucene in action" book , the Document are first converted
into lucene Document Format, then analysed with the standardAnalyser.
I don't understand how the analysed Documents added to the inverted index,
ar
I'm trying to exercise the termOffset functions in the nightly build
(2009-02-11) but it doesn't seem to do anything. I have an item in my
schema like so:
And I attempt this query:
qt=tvrh&
tv=true&
tv.offsets=true&
indent=true&
wt=json&
facet.mincount=1&
facet=true&
hl=on&
hl.fl=document&
hl.
Yes, I commit roughly every 15 minutes (via a data update). This update is
consistent between my tests, and only causes a performance drop when I'm
sorting on fields with many unique values. I've examined my GC logs, and
they are also consistent between my tests.
Otis Gospodnetic wrote:
>
> Hi
I'll run a profiler on new and old code and let you know what I find.
I have changed my schema between tests: I used to have termVectors turned on
for several fields, and now they are always off. My underlying data has not
changed.
--
View this message in context:
http://www.nabble.com/Recent-P
If a naive user enters a string that contains typical stopwords like "and"
and "the", these seem to be included in the word count for the must
match criteria of the the dismax query.
So, if for example the mm parameter is the default " 2>-1 5>-2
6>90%" and the user enters something like "Jason
Thanks,
Here is a ruby translation for those that want it:
solr_query = ""
doi_part.each_char do |c|
if (c == '\\' || c == '+' || c == '-' || c == '!' || c == '(' || c
== ')' || c == ':' || c == '^' || c == '[' || c == ']' || c == '\"' || c ==
'{' || c == '}' || c == '~' || c == '*'
Just did some research. It seems that it's doable with additional code added to
Solr but not out of box. Thank you, Grant.
- Original Message
From: Grant Ingersoll
To: "solr-user@lucene.apache.org"
Sent: Wednesday, February 11, 2009 8:14:01 AM
Subject: Re: score filter
At what point
At what point do you draw the line? 0.01 is too low, but what about
0.5 or 0.3? In fact, there may be queries where 0.01 is relevant.
Relevance is a tricky thing and putting in arbitrary cutoffs is
usually not a good thing. An alternative might be to instead look at
the difference between
Hi Grant,
In my case, for example searching a book. Some of the returned documents are
with high relevance (score > 3), but some of document with low score (<0.01)
are useless.
Without a "score filter", I have to go through each document to find out the
number of documents I'm interested (sco
What's the motivation for wanting to do this? The reason I ask, is
score is a relative thing determined by Lucene based on your index
statistics. It is only meaningful for comparing the results of a
specific query with a specific instance of the index. In other words,
it isn't useful to
It looks like you are running out of memory. What is your heap size?
On Feb 11, 2009, at 4:09 AM, sunnyfr wrote:
Hi
Have you an idea why after a night with solr running, but just
commit every
five minute??
It looks like process never shutdown ???
root 29428 0.0 0.0 53988 2648 ?
Has anything else changed index-wise? For instance, do you have
larger stored fields or are you retrieving more fields?
On Feb 10, 2009, at 8:26 PM, wojtekpia wrote:
Has there been a recent change (since Dec 2/08) in the paging
algorithm? I'm
seeing much worse performance (75% drop in thr
Grant Ingersoll schrieb:
What's your "textSpell FieldType look like?
Spelling is definitely something that needs tuning, so you might have
to play with some of the knobs like accuracy, etc.
As for JaroWinkler, and I suppose the default, your field is "spell",
but based on your configuration,
Thanks,
That fixed it.
>On Wed, Feb 11, 2009 at 4:19 PM, Fergus McMenemie wrote:
>
>
>> java.lang.NoSuchFieldError: docCount
>>at
>> org.apache.solr.handler.dataimport.SolrWriter.getDocCount(SolrWriter.java:231)
>>at
>> org.apache.solr.handler.dataimport.DataImportHandlerExcepti
Hi,
I have been looking at the nightly downloads, trying to work
backwards through the nightly's till my code starts working
again!
I have downloaded all the available nightly's and they all fail
to "ant dist" as follows:-
>root: ant dist
>Buildfile: build.xml
>
>init-forrest-entities:
>
>comp
What's your "textSpell FieldType look like?
Spelling is definitely something that needs tuning, so you might have
to play with some of the knobs like accuracy, etc.
As for JaroWinkler, and I suppose the default, your field is "spell",
but based on your configuration, I gather you really wan
On Wed, Feb 11, 2009 at 4:19 PM, Fergus McMenemie wrote:
> java.lang.NoSuchFieldError: docCount
>at
> org.apache.solr.handler.dataimport.SolrWriter.getDocCount(SolrWriter.java:231)
>at
> org.apache.solr.handler.dataimport.DataImportHandlerException.(DataImportHandlerException.jav
Hello,
I had a nice working version of SOLR building from trunk, I think
it was from about 2-4th Feb, On the 7th I performed a "svn update"
and it now fails as follows when performing
get 'http://localhost:8080/apache-solr-1.4-dev/dataimport?command=full-import'
I have performed a "svn update"
Hi,
My SOLRCONFIG.XML
spellcheck
textSpell
default
spell
./spellchecker1
true
jarowinkler
spell
name="distanceMeasure
Hi
Have you an idea why after a night with solr running, but just commit every
five minute??
It looks like process never shutdown ???
root 29428 0.0 0.0 53988 2648 ?S01:05 0:00 curl
http://localhost:8180/solr/book/update -s -H Content-type:text/xml;
charset=utf-8 -d
roo
29 matches
Mail list logo