Thanks for the replies. I did look at caching but our commit time time is 90
seconds. It's definitely possible for someone to make a search, change the
page, and have wonky results. How about getting it to autowarm the x most
recent searches in the queryResultCache and that can hopefully reduce
If I'm searching for users based on last login time, and I search once, then
go to the second page with a new offset, I could potentially see the same
users on page 2 if the index has changed. What is the best way to anchor it
so I avoid this?
--
View this message in context:
Stupid me. The output file was named something else. I really need to make
a proper servlet mapping. Works now :D
--
View this message in context:
http://lucene.472066.n3.nabble.com/Problems-generating-war-distribution-using-ant-tp3260070p3260843.html
Sent from the Solr - User mailing list
So the way I generate war files now is by running an 'ant dist' in the solr
folder. It generates the war fine and I get a build success, and then I
deploy it to tomcat and once again the logs show it was successful (from the
looks of it). However, when I go to 'myip:8080/solr/admin' I get an
Interesting. I can use this as an option and create a custom 'war' target if
need be but I'd like to avoid this. I'd rather do a full build from the
source code I have checked out from the SVN. Any reason why 'ant dist'
doesn't produce a good war file?
--
View this message in context:
Interesting. I can use this as an option and create a custom 'war' target if
need be but I'd like to avoid this. I'd rather do a full build from the
source code I have checked out from the SVN. Any reason why 'ant dist'
doesn't produce a good war file?
--
View this message in context:
I downloaded the official build (4.0) and I've been customizing it for my
needs. I'm not really sure how to use these scripts. Is there somewhere in
Hudson where I can apply these scripts or something?
--
View this message in context:
Thanks for the advice paul, but post processing is a must for me given the
nature of my application. I haven't had problems yet though.
--
View this message in context:
http://lucene.472066.n3.nabble.com/Cache-replication-tp3240708p3244202.html
Sent from the Solr - User mailing list archive
Whenever I try to build this on our hudson server it says it can't find
org.apache.lucene:lucene-xercesImpl:jar:4.0-SNAPSHOT. Is the Apache repo
lacking this artifact?
--
View this message in context:
http://lucene.472066.n3.nabble.com/Hudson-build-issues-tp3244563p3244563.html
Sent from the
I'm wondering if the caches on all the slaves are replicated across (such as
queryResultCache). That is to say, if I hit one of my slaves and cache a
result, and I make a search later and that search happens to hit a different
slave, will that first cached result be available for use?
This is
Thanks for the informative response. I'll consider using the 'sticky'
addressing as you suggested. The reason cache is so important for me is
because I'm actually doing more processing after the query component to come
up with my query result and I want to avoid that processing as much as
Lets say my sort is something like:
sort=sum(indexedField, constant). If I have a component that runs right
after the QueryComponent, is it possible to know what this value was for
each of the documents IF the field is not stored, and only indexed? I
scoured through the code and it didn't look
Sorry for being vague. Okay so these scores exist on an external server and
they change often enough. The score for each returned user is actually
dependent on the user doing the searching (if I'm making the request, and
you make the same request, the scores are different). So what I'm doing is
Sorry for my ignorance, but do you have any lead in the code on where to look
for this? Also, I'd still need a way of finding out how long its been in
the cache because I don't want it to regenerate every time. I'd want it to
regenerate only if its been in the cache for less then 6 hours (or
It depends on how many queries you'd be making per second. I know for us, I
have a gradient of index sizes. The first machine, which gets hit most
often is about 2.5 gigs. Most of the queries would only ever need to hit
this index but then I have a bigger indices of about 5-10 gigs each which
I guess I'll have to use something other then SolrCache to get what I want
then. Or I could use SolrCache and just change the code (I've already done
so much of this anwyways...). Anyways thanks for the reply.
--
View this message in context:
I know the queryResultCache and stuff live only so long as a commit happens
but I'm wondering if the custom caches are like this as well? I'd actually
rather have a custom cache which is not cleared at all. I want to give the
elements of this Cache a 6 hour TTL (or some time frame) but I never
So it seems the things in the queryResultCache have no TTL, I'm just curious
how it works if I reindex something with new info? I am going to be
reindexing things often (I'd sort by last login and this changes fast).
I've been stepping through the code and of course if the same queries come
in
Thanks for the quick reply! I see theres no way to access the result cache,
I actually want to access the result the cache in a new component I have
which runs after the query but it seems this is impossible. I guess I'm
just going to rebuild the code to make it public or something as I need the
unfortunately the userIdsToScore updates very often. I'd get more Ids almost
every single query (hence why I made the new component). But I see the
problem of not being able to score the whole resultSet. I'd actually need
to do this now that I think about it. I want to get a whole whack of
Sorry for the double post but in this case, is it possible for me to access
the queryResultCache in my component and play with it? Ideally what I want
is this:
1) I have 1 (just a random large number) total results.
2) In my component I access all of these results, score them, and take the
bump
--
View this message in context:
http://lucene.472066.n3.nabble.com/After-the-query-component-has-the-results-can-I-do-more-filtering-on-them-tp3114775p3123502.html
Sent from the Solr - User mailing list archive at Nabble.com.
So I made a custom search component which runs right after the query
component and this custom component will update the score of each based on
some things (and no, I definitely can't use existing components). I didn't
see any easy way to just update the score so what I currently do is
something
Thanks, this is exactly what I'm looking for!
--
View this message in context:
http://lucene.472066.n3.nabble.com/Caching-queries-tp3078271p3087497.html
Sent from the Solr - User mailing list archive at Nabble.com.
I'm wondering if something like this is possible. Lets say I want to query
5000 objects all pertaining to a specific search and I want to return the
top 100 or something and cache the rest on my solr server. The next time I
get the same query or something with a new offset (lets say start from
Basically I don't want the hits and the scores at the same time. I want to
get a list of hits but I want to score them myself externally (there is a
dedicated server that will do the scoring given a list of id's). Thanks!
--
View this message in context:
To clarify. I want to do this all underneath solr. I don't want to get a
bunch of hits from solr in my app and then go to my server and score them
again. I'd like to score them myself underneath solr before I return the
results to my app.
--
View this message in context:
Actually I was thinking I wanted to do something before the sharding (like in
the layer where faceting happens for example). I wanna hack a plugin in the
middle to go to my server after I have a bunch of hits. Just not sure where
to do this...
Though I've decided I can do scoring from solr
Hmm, looks like I can inherit the Similarity Class and do my own thing there.
--
View this message in context:
http://lucene.472066.n3.nabble.com/Is-there-a-way-to-get-all-the-hits-and-score-them-later-tp3016424p3018001.html
Sent from the Solr - User mailing list archive at Nabble.com.
bump
--
View this message in context:
http://lucene.472066.n3.nabble.com/Custom-Scoring-relying-on-another-server-tp2994546p3006873.html
Sent from the Solr - User mailing list archive at Nabble.com.
I know this question has been asked before but I think my situation is a
little different. Basically I need to do custom scores that the traditional
function queries simply won't allow me to do. I actually need to hit
another server from Java (passing in a bunch of things mostly relying on how
bump
--
View this message in context:
http://lucene.472066.n3.nabble.com/Field-collapsing-on-multiple-fields-and-or-ranges-tp2929793p2958029.html
Sent from the Solr - User mailing list archive at Nabble.com.
Thanks for the reply! How exactly do I open an issue?
--
View this message in context:
http://lucene.472066.n3.nabble.com/Field-collapsing-on-multiple-fields-and-or-ranges-tp2929793p2958277.html
Sent from the Solr - User mailing list archive at Nabble.com.
https://issues.apache.org/jira/browse/SOLR-2526
modules/grouping was not a valid component so I just put it in search.
Thanks!
--
View this message in context:
http://lucene.472066.n3.nabble.com/Field-collapsing-on-multiple-fields-and-or-ranges-tp2929793p2958408.html
Sent from the Solr - User
Ah, my mistake. Thanks alot, this would be a really cool feature :)
For now I'm resorting to like making more then one query and cross
referencing the two separate queries.
--
View this message in context:
I'm wondering if there is a way to get the field collapsing to collapse on
multiple things? For example, is there a way to get it to collapse on a
field (lets say 'domain') but ALSO something else (maybe time or something)?
To visualize maybe something like this:
Group1 has common field
I'm actually using php but I get what you're saying. I think I understand
what I need to do. Thanks a lot man!
--
View this message in context:
http://lucene.472066.n3.nabble.com/SolrQuery-API-for-adding-group-filter-tp2921539p2923701.html
Sent from the Solr - User mailing list archive at
I actually have another question unrelated to this (but related to grouping).
I'm wondering if I can do a more complex grouping, such as grouping by a
field and also making sure it matches some other criteria (such as date).
For example, currently it might group 5 items from some field, but the
There doesn't seem to be API to add a group (like group.field or group=true).
I'm very new to this so I'm wondering how I'd go about adding a group query
much like how I use 'addFilterQuery' to add an fq. Thanks.
--
View this message in context:
39 matches
Mail list logo