Re: bf calculation
Hi Georg, So obvious I totally forgot about that option. Thanks. Jan. On 2 May 2016 at 11:15, Georg Sorst wrote: > Hi Jan, > > have you tried Solr's debug output? ie. add > "...&debug=true&debug.explain.structured=true" to your query. This should > answer your question. > > Best, > Georg > > Jan Verweij - Reeleez schrieb am Mo., 2. Mai 2016 um > 09:47 Uhr: > > > Hi, > > I'm trying to understand the exact calculation that takes place when > using > > edismax and the bf parameter. > > When searching I get a product returned with a score of 0.625 > > Now, I have a field called productranking with a value of 0.5 for this > > specific > > product. If I add &bf=field(productranking) to the request the score > > becomes 0.7954515 > > How is this calculated? > > Cheers, > > Jan Verweij > > -- > *Georg M. Sorst I CTO* > FINDOLOGIC GmbH > > > > Jakob-Haringer-Str. 5a | 5020 Salzburg I T.: +43 662 456708 > E.: g.so...@findologic.com > www.findologic.com Folgen Sie uns auf: XING > <https://www.xing.com/profile/Georg_Sorst>facebook > <https://www.facebook.com/Findologic> Twitter > <https://twitter.com/findologic> > > Wir sehen uns auf dem *Shopware Community Day in Ahaus am 20.05.2016!* Hier > Termin > vereinbaren! > Wir sehen uns auf der* dmexco in Köln am 14.09. und 15.09.2016!* Hier > Termin > vereinbaren! >
bf calculation
Hi, I'm trying to understand the exact calculation that takes place when using edismax and the bf parameter. When searching I get a product returned with a score of 0.625 Now, I have a field called productranking with a value of 0.5 for this specific product. If I add &bf=field(productranking) to the request the score becomes 0.7954515 How is this calculated? Cheers, Jan Verweij
Re: solr | backup and restoration
Hi Prateek, To me it feels like the backup/restore is still an open item should be higher on the agenda. Yes, there are work-arounds like copying data from/into the index folder but this doesn' t seem very stable. I'm using the following approach in solrcloud since I ran into an issue with restoring the same backup multiple times: rm $SOLRDATA_LOCATION/$INDEXNAME/data/index.properties rm -rf $SOLRDATA_LOCATION/$INDEXNAME/data/restore.snapshot. curl " http://localhost:8983/solr/$INDEXNAME/replication?command=restore&location=$BACKUP_LOCATION&name=$INDEXNAME.$TIMESTAMP " Though this is working, I still think it should work with a simple restore command and no need to tweak the current core/index directory/files. Met vriendelijke groet / Kind regards, *Jan Verweij* +31 (0)6 460 010 86 j...@reeleez.nl www.reeleez.nl *Disclaimer*: The information contained in this message is for the intended addressee only and may contain confidential and/or privileged information. If you are not the intended addressee, please delete this message and notify the sender; do not copy or distribute this message or disclose its contents to anyone. On 27 April 2016 at 15:45, Prateek Jain J wrote: > > Manually copying files under index directory fixed the issue. > > > Regards, > Prateek Jain > > -Original Message- > From: Prateek Jain J [mailto:prateek.j.j...@ericsson.com] > Sent: 27 April 2016 02:08 PM > To: solr-user@lucene.apache.org > Subject: solr | backup and restoration > > > Hi, > > We are using solr 4.8.1 in production and want to create backups at > runtime. As per the reference guide, we can create backup using something > like this: > > > http://localhost:8983/solr/myCore/replication?command=backup&location=/tmp/myBackup&numberToKeep=1 > > and we verified that some file are getting created in /tmp/myBackup > directory. The issue that we are facing is, how to restore everything using > this backup. > Admin guide does talk about "Merging Indexes" using two methods: > > > a. indexDir for example, > > > http://localhost:8983/solr/admin/cores?action=mergeindexes&core=core0&indexDir=/home/solr/core1/data/index&; > > indexDir=/home/solr/core2/data/index > > > > b. srcCore for example, > > > http://localhost:8983/solr/admin/cores?action=mergeindexes&core=core0&srcCore=core1&srcCore=core2 > > > > these are not working in our case as, we want entire data should also be > back there for example, if we want to re-create core from a snapshot. I do > see there is such functionality available in later versions as, described > here > > > https://cwiki.apache.org/confluence/display/solr/Making+and+Restoring+Backups+of+SolrCores > > > Regards, > Prateek Jain > >
Re: large number of fields
First of all thank you for the repsonses/ideas. The plain idea is to boost products that are sold more for a specific customer. I'll discuss this with my customer to see if we can start with limited functionality as suggested by Walter. Cheers, Jan. On 5 February 2016 at 17:33, Walter Underwood wrote: > I would add a multiValued field for buying_customers. Add the customer ID > for each relevant customer to that field. Then use a boost query “bq”, to > boost those. > > Try that first before using the hit rate. Always try on/off control before > going proportional. The simple approach will probably give you 80% of the > benefit. Then you can declare victory and go on to the next idea. > > If you do need hit rate, try quantizing that into high/medium/low, or > deciles, or something. Then you have one multiValued field for each level > and one bq for each level. The bq will include a weight: > bq=customer_hi:1234^8. Logarithmic levels are probably your friend here. > > You can get some unwanted idf scoring with bq. Customers that only buy a > few things get a higher bq weight that customers that buy a lot of things. > You can fix that with function queries, but I’d get it working with a boost > query first. > > wunder > Walter Underwood > wun...@wunderwood.org > http://observer.wunderwood.org/ (my blog) > > > > On Feb 5, 2016, at 8:13 AM, Jack Krupansky > wrote: > > > > This doesn't sound like a great use case for Solr - or any other search > > engine for that matter. I'm not sure what you are really trying to > > accomplish, but you are trying to put way too many balls in the air to > > juggle efficiently. You really need to re-conceptualize your problem so > > that it has far fewer moving parts. Sure, Solr can handle many millions > or > > even billions of documents, but the focus for scaling Solr is on more > > documents and more nodes, not incredibly complex or large documents. The > > key to effective and efficient use of Solr is that queries are "quite > > short", definitely not "quite long." > > > > That said, the starting point for any data modeling effort is to look at > > the full range of desired queries and that should drive the data model. > So, > > give us more info on queries, in terms of plain English descriptions of > > what the user is trying to achieve. > > > > > > -- Jack Krupansky > > > > On Fri, Feb 5, 2016 at 8:20 AM, Jan Verweij - Experts in search < > > j...@searchxperts.nl> wrote: > > > >> Hi, > >> We store 50K products stored in Solr. We have 10K customers and each > >> customer buys up to 10K of these products. Now we want to influence the > >> results by adding a field for every customer. > >> So we end up with 10K fields to influence the results on the buying > >> behavior of > >> each customer (personal results). Don't think this is the way to go so > I'm > >> looking for suggestions how to solve > >> this. > >> One other option would be to: 1. create one multivaluefield > >> 'company_hitrate' > >> 2. store for each company their [companyID]_[hitrate] > >> > >> During search use boostfields [companyID]_50 …. [companyID]_100 So in > this > >> case the query can become quit long (51 options) but the number of > >> fields is limited to 1. What kind of effect would this have on the > search > >> performance > >> Any other suggestions? > >> Jan. > >
statuscode list
Hi, If I'm correct you will get a statuscode="0" in the response if you use XML messages for updating the solr index. Is there a list of possible other statuscodes you can receive in case anything fails and what these errorcodes mean? THNX, Jan.
spellcheck if docsfound below threshold
Hi, Is there a setting to only include spellcheck if the number of documents found is below a certain threshold? Or would we need to rerun the request with the spellcheck parameters based on the docs found? Kind regards, Jan Verweij
get sub-facets based on main-facet selections
Dear, I'm implementing a productcatalog and have 5 main facets and 60+ possible subfacets. If I select a specific value from one of my main facets, let's say, productgroupX, I want to show the facets related to this productgroup, say length and height. But if productgroupY is selected I have to show weight and color. To make it even more complex if I run a query and one of my main facets returns with just a single value it's the same as selecting this single value and should already come back with the additional subfacets. I know how todo this with two requests to solr but perhaps there are more dynamic ways within solr I haven't thought about. Cheers, Jan Verweij