Re: UI for solr core admin?
Well sure, Solr is no webapp framework. But you can do some things with the ShowFileRequestHandler, like have static textual content (like CSS and JavaScript, or just an HTML page) served up. And you can simply have a Velocity template with static text in it, and map it as a request handler, so /solr/view.html can get served up from a dummy request handler with wt=velocity and v.template=view set in solrconfig.xml, like this: velocity view One issue with ShowFileRequestHandler currently is that it cannot serve up binary content, though. So images are an issue unfortunately. I think this should be fixed, as there could be binary data files (an optimized Chinese dictionary, maybe) that you'd want accessible externally, not just for this fancy view stuff. So, in summary: HTML handled. Images still an issue. Erik On Dec 11, 2009, at 4:49 AM, Lance Norskog wrote: There is a general problem here in that Solr has no configuration for content management. There is nowhere to throw in images and html pages as part of your velocity app either. On Wed, Dec 9, 2009 at 11:09 PM, Erik Hatcher wrote: After I sent that, though, I realized that the core admin is special in that it isn't within the context of a single core. I'll have to research this and see, but I suspect there may be an issue with using VwR for this particular handler, as it wouldn't have a solr-home/conf/velocity directory to pull templates from. I'll look into it. Erik On Dec 10, 2009, at 7:51 AM, Mattmann, Chris A (388J) wrote: Nice, Erik! Cheers, Chris On 12/9/09 10:39 PM, "Erik Hatcher" wrote: Just a note about the hidden gem that I haven't taken as far as I'd like... With the VelocityResponseWriter, it's as easy as creating a Velocity template (and wiring in VwR in solrconfig, which I'll set up by default in 1.5). It will even default to the template named after the handler name, so all you have to do is &wt=velocity. Erik On Dec 10, 2009, at 7:33 AM, Mattmann, Chris A (388J) wrote: Hi Jason, Patches welcome, though! :) Cheers, Chris On 12/9/09 10:31 PM, "Shalin Shekhar Mangar" wrote: On Thu, Dec 10, 2009 at 11:52 AM, Jason Rutherglen < jason.rutherg...@gmail.com> wrote: I assume there isn't one? Anything in the works? Nope. -- Regards, Shalin Shekhar Mangar. ++ Chris Mattmann, Ph.D. Senior Computer Scientist NASA Jet Propulsion Laboratory Pasadena, CA 91109 USA Office: 171-266B, Mailstop: 171-246 Email: chris.mattm...@jpl.nasa.gov WWW: http://sunset.usc.edu/~mattmann/ ++ Adjunct Assistant Professor, Computer Science Department University of Southern California, Los Angeles, CA 90089 USA ++ ++ Chris Mattmann, Ph.D. Senior Computer Scientist NASA Jet Propulsion Laboratory Pasadena, CA 91109 USA Office: 171-266B, Mailstop: 171-246 Email: chris.mattm...@jpl.nasa.gov WWW: http://sunset.usc.edu/~mattmann/ ++ Adjunct Assistant Professor, Computer Science Department University of Southern California, Los Angeles, CA 90089 USA ++ -- Lance Norskog goks...@gmail.com
Re: Results after using Field Collapsing are not matching the results without using Field Collapsing
Here is the field type configuration of ctype: In solrconfig.xml, this is how I am enabling field collapsing: Apart from this, I made no changes in solrconfig.xml for field collapse. I am currently not using the field collapse cache. I have applied the patch on the Solr 1.4 build. I am not using the latest solr nightly build. Can that cause any problem? -- Thanks Varun Gupta On Fri, Dec 11, 2009 at 3:44 AM, Martijn v Groningen < martijn.is.h...@gmail.com> wrote: > I tried to reproduce a similar situation here, but I got the expected > and correct results. Those three documents that you saw in your first > search result should be the first in your second search result (unless > the index changes or the sort changes ) when fq on that specific > category. I'm not sure what is causing this problem. Can you give me > some more information like the field type configuration for the ctype > field and how have configured field collapsing? > > I did find another problem to do with field collapse caching. The > collapse.threshold or collapse.maxdocs parameters are not taken into > account when caching, which is off course wrong because they do matter > when collapsing. Based on the information you have given me this > caching problem is not the cause of the situation you have. I will > update the patch that fixes this problem shortly. > > Martijn > > 2009/12/10 Varun Gupta : > > Hi Martijn, > > > > I am not sending the collapse parameters for the second query. Here are > the > > queries I am using: > > > > *When using field collapsing (searching over all categories):* > > > spellcheck=true&collapse.info.doc=true&facet=true&collapse.threshold=3&facet.mincount=1&spellcheck.q=weight+loss&collapse.facet=before&wt=xml&f.content.hl.snippets=2&hl=true&version=2.2&rows=20&collapse.field=ctype&fl=id,sid,title,image,ctype,score&start=0&q=weight+loss&collapse.info.count=false&facet.field=ctype&qt=contentsearch > > > > categories is represented as the field "ctype" above. > > > > *Without using field collapsing:* > > > spellcheck=true&facet=true&facet.mincount=1&spellcheck.q=weight+loss&wt=xml&hl=true&rows=10&version=2.2&fl=id,sid,title,image,ctype,score&start=0&q=weight+loss&facet.field=ctype&qt=contentsearch > > > > I append "&fq=ctype:1" to the above queries when trying to get results > for a > > particular category. > > > > -- > > Thanks > > Varun Gupta > > > > > > On Thu, Dec 10, 2009 at 5:58 PM, Martijn v Groningen < > > martijn.is.h...@gmail.com> wrote: > > > >> Hi Varun, > >> > >> Can you send the whole requests (with params), that you send to Solr > >> for both queries? > >> In your situation the collapse parameters only have to be used for the > >> first query and not the second query. > >> > >> Martijn > >> > >> 2009/12/10 Varun Gupta : > >> > Hi, > >> > > >> > I have documents under 6 different categories. While searching, I want > to > >> > show 3 documents from each category along with a link to see all the > >> > documents under a single category. I decided to use field collapsing > so > >> that > >> > I don't have to make 6 queries (one for each category). Currently I am > >> using > >> > the field collapsing patch uploaded on 29th Nov. > >> > > >> > Now, the results that are coming after using field collapsing are not > >> > matching the results for a single category. For example, for category > C1, > >> I > >> > am getting results R1, R2 and R3 using field collapsing, but after I > see > >> > results only from the category C1 (without using field collapsing) > these > >> > results are nowhere in the first 10 results. > >> > > >> > Am I doing something wrong or using the field collapsing for the wrong > >> > feature? > >> > > >> > I am using the following field collapsing parameters while querying: > >> > collapse.field=category > >> > collapse.facet=before > >> > collapse.threshold=3 > >> > > >> > -- > >> > Thanks > >> > Varun Gupta > >> > > >> > >> > >> > >> -- > >> Met vriendelijke groet, > >> > >> Martijn van Groningen > >> > > > > > > -- > Met vriendelijke groet, > > Martijn van Groningen >
Re: full-text indexing XML files
What kind of searches do you want to do? Do you want to do searches that match the XML tags? wunder On Dec 10, 2009, at 7:43 PM, Lance Norskog wrote: > Or CDATA (much easier to work with). > > On Wed, Dec 9, 2009 at 10:37 PM, Shalin Shekhar Mangar > wrote: >> On Thu, Dec 10, 2009 at 5:13 AM, Feroze Daud wrote: >> >>> Hi! >>> >>> >>> >>> I am trying to full text index an XML file. For various reasons, I >>> cannot use Tika or other technology to parse the XML file. The >>> requirement is to full-text index the XML file, including Tags and >>> everything. >>> >>> >>> >>> So, I created a input index spec like this: >>> >>> >>> >>> >>> >>> >>> >>> 1001 >>> >>> NASA Advanced Research Labs >>> >>> 1010 Main Street, Chattanooga, FL 32212 >>> >>> 1001< name > NASA Advanced >>> Research Labs 1010 main street, chattanooga, FL >>> 32212 >>> >>> >>> >>> >>> >>> >>> >> You need to XML encode the value of the "content" field. >> >> -- >> Regards, >> Shalin Shekhar Mangar. >> > > > > -- > Lance Norskog > goks...@gmail.com >
Re: UI for solr core admin?
There is a general problem here in that Solr has no configuration for content management. There is nowhere to throw in images and html pages as part of your velocity app either. On Wed, Dec 9, 2009 at 11:09 PM, Erik Hatcher wrote: > After I sent that, though, I realized that the core admin is special in that > it isn't within the context of a single core. I'll have to research this > and see, but I suspect there may be an issue with using VwR for this > particular handler, as it wouldn't have a solr-home/conf/velocity directory > to pull templates from. > > I'll look into it. > > Erik > > > > On Dec 10, 2009, at 7:51 AM, Mattmann, Chris A (388J) wrote: > >> Nice, Erik! >> >> Cheers, >> Chris >> >> >> >> On 12/9/09 10:39 PM, "Erik Hatcher" wrote: >> >>> Just a note about the hidden gem that I haven't taken as far as I'd >>> like... >>> >>> With the VelocityResponseWriter, it's as easy as creating a Velocity >>> template (and wiring in VwR in solrconfig, which I'll set up by >>> default in 1.5). It will even default to the template named after the >>> handler name, so all you have to do is &wt=velocity. >>> >>> Erik >>> >>> >>> >>> On Dec 10, 2009, at 7:33 AM, Mattmann, Chris A (388J) wrote: >>> Hi Jason, Patches welcome, though! :) Cheers, Chris On 12/9/09 10:31 PM, "Shalin Shekhar Mangar" wrote: > On Thu, Dec 10, 2009 at 11:52 AM, Jason Rutherglen < > jason.rutherg...@gmail.com> wrote: > >> I assume there isn't one? Anything in the works? >> > > Nope. > > -- > Regards, > Shalin Shekhar Mangar. > ++ Chris Mattmann, Ph.D. Senior Computer Scientist NASA Jet Propulsion Laboratory Pasadena, CA 91109 USA Office: 171-266B, Mailstop: 171-246 Email: chris.mattm...@jpl.nasa.gov WWW: http://sunset.usc.edu/~mattmann/ ++ Adjunct Assistant Professor, Computer Science Department University of Southern California, Los Angeles, CA 90089 USA ++ >>> >>> >> >> >> ++ >> Chris Mattmann, Ph.D. >> Senior Computer Scientist >> NASA Jet Propulsion Laboratory Pasadena, CA 91109 USA >> Office: 171-266B, Mailstop: 171-246 >> Email: chris.mattm...@jpl.nasa.gov >> WWW: http://sunset.usc.edu/~mattmann/ >> ++ >> Adjunct Assistant Professor, Computer Science Department University of >> Southern California, Los Angeles, CA 90089 USA >> ++ >> >> > > -- Lance Norskog goks...@gmail.com
Re: full-text indexing XML files
Or CDATA (much easier to work with). On Wed, Dec 9, 2009 at 10:37 PM, Shalin Shekhar Mangar wrote: > On Thu, Dec 10, 2009 at 5:13 AM, Feroze Daud wrote: > >> Hi! >> >> >> >> I am trying to full text index an XML file. For various reasons, I >> cannot use Tika or other technology to parse the XML file. The >> requirement is to full-text index the XML file, including Tags and >> everything. >> >> >> >> So, I created a input index spec like this: >> >> >> >> >> >> >> >> 1001 >> >> NASA Advanced Research Labs >> >> 1010 Main Street, Chattanooga, FL 32212 >> >> 1001< name > NASA Advanced >> Research Labs 1010 main street, chattanooga, FL >> 32212 >> >> >> >> >> >> >> > You need to XML encode the value of the "content" field. > > -- > Regards, > Shalin Shekhar Mangar. > -- Lance Norskog goks...@gmail.com
Re: Indexing content on Windows file shares?
If you are looking to index websites, Nutch would be a better alternative. However, it could be useful for indexing text files. There is documentation here for how to add data to the index http://lucene.apache.org/solr/tutorial.html#Indexing+Data http://wiki.apache.org/solr/#Search_and_Indexing There are some clients here to add data to the index programatically. http://wiki.apache.org/solr/IntegratingSolr On Thu, Dec 10, 2009 at 3:06 PM, Matt Wilkie wrote: > Hello, > > I'm new to Solr, I know nothing about it other than it's been touted in a > couple of places as a possible competitor to Google Search Appliance, which > is what brought me here. I'm looking for a search engine which can index > files on windows shares and websites, and, hopefully, integrate with Active > Directory to ensure results are not returned to users who don't have access > to those files(s). > > Can Solr do this? If so where is the documentation for it? Reconnaisance > searches of the mailing list and wiki have not turned up anything, so far. > > thanks, > > -- > matt wilkie > > Geomatics Analyst > Information Management and Technology > Yukon Department of Environment > 10 Burns Road * Whitehorse, Yukon * Y1A 4Y9 > 867-667-8133 Tel * 867-393-7003 Fax > http://environmentyukon.gov.yk.ca/geomatics/ > > -- "Good Enough" is not good enough. To give anything less than your best is to sacrifice the gift. Quality First. Measure Twice. Cut Once. http://www.israelekpo.com/
Re: Results after using Field Collapsing are not matching the results without using Field Collapsing
I tried to reproduce a similar situation here, but I got the expected and correct results. Those three documents that you saw in your first search result should be the first in your second search result (unless the index changes or the sort changes ) when fq on that specific category. I'm not sure what is causing this problem. Can you give me some more information like the field type configuration for the ctype field and how have configured field collapsing? I did find another problem to do with field collapse caching. The collapse.threshold or collapse.maxdocs parameters are not taken into account when caching, which is off course wrong because they do matter when collapsing. Based on the information you have given me this caching problem is not the cause of the situation you have. I will update the patch that fixes this problem shortly. Martijn 2009/12/10 Varun Gupta : > Hi Martijn, > > I am not sending the collapse parameters for the second query. Here are the > queries I am using: > > *When using field collapsing (searching over all categories):* > spellcheck=true&collapse.info.doc=true&facet=true&collapse.threshold=3&facet.mincount=1&spellcheck.q=weight+loss&collapse.facet=before&wt=xml&f.content.hl.snippets=2&hl=true&version=2.2&rows=20&collapse.field=ctype&fl=id,sid,title,image,ctype,score&start=0&q=weight+loss&collapse.info.count=false&facet.field=ctype&qt=contentsearch > > categories is represented as the field "ctype" above. > > *Without using field collapsing:* > spellcheck=true&facet=true&facet.mincount=1&spellcheck.q=weight+loss&wt=xml&hl=true&rows=10&version=2.2&fl=id,sid,title,image,ctype,score&start=0&q=weight+loss&facet.field=ctype&qt=contentsearch > > I append "&fq=ctype:1" to the above queries when trying to get results for a > particular category. > > -- > Thanks > Varun Gupta > > > On Thu, Dec 10, 2009 at 5:58 PM, Martijn v Groningen < > martijn.is.h...@gmail.com> wrote: > >> Hi Varun, >> >> Can you send the whole requests (with params), that you send to Solr >> for both queries? >> In your situation the collapse parameters only have to be used for the >> first query and not the second query. >> >> Martijn >> >> 2009/12/10 Varun Gupta : >> > Hi, >> > >> > I have documents under 6 different categories. While searching, I want to >> > show 3 documents from each category along with a link to see all the >> > documents under a single category. I decided to use field collapsing so >> that >> > I don't have to make 6 queries (one for each category). Currently I am >> using >> > the field collapsing patch uploaded on 29th Nov. >> > >> > Now, the results that are coming after using field collapsing are not >> > matching the results for a single category. For example, for category C1, >> I >> > am getting results R1, R2 and R3 using field collapsing, but after I see >> > results only from the category C1 (without using field collapsing) these >> > results are nowhere in the first 10 results. >> > >> > Am I doing something wrong or using the field collapsing for the wrong >> > feature? >> > >> > I am using the following field collapsing parameters while querying: >> > collapse.field=category >> > collapse.facet=before >> > collapse.threshold=3 >> > >> > -- >> > Thanks >> > Varun Gupta >> > >> >> >> >> -- >> Met vriendelijke groet, >> >> Martijn van Groningen >> > -- Met vriendelijke groet, Martijn van Groningen
Indexing content on Windows file shares?
Hello, I'm new to Solr, I know nothing about it other than it's been touted in a couple of places as a possible competitor to Google Search Appliance, which is what brought me here. I'm looking for a search engine which can index files on windows shares and websites, and, hopefully, integrate with Active Directory to ensure results are not returned to users who don't have access to those files(s). Can Solr do this? If so where is the documentation for it? Reconnaisance searches of the mailing list and wiki have not turned up anything, so far. thanks, -- matt wilkie Geomatics Analyst Information Management and Technology Yukon Department of Environment 10 Burns Road * Whitehorse, Yukon * Y1A 4Y9 867-667-8133 Tel * 867-393-7003 Fax http://environmentyukon.gov.yk.ca/geomatics/
RE: SolrPlugin Guidance
Ok, looks like I may not be taking the right approach here. I'm running a problem. Let's say a user is looking for all files in any directory 'foo' with a directory description 'bar' q:+directory_name:foo +directory_description:bar Our QParser plugin will perform queries against directory documents and return any file document that has the matching directory id(s). So the plugin transforms the query to something like q:+(directory_id:4 directory:10) +directory_id:(4) Note: directory_id is only in file documents. The query above assumes that two directories had the name 'foo' but only one had the description 'bar' Currently the parser plugin is doing the lookup queries via the standard request handler. The problem with this approach is that the look up queries are going to be analyzed twice. This only seems to be a problem because we're using stemming. For example, stemming 'franchise' gives 'franchis' and stemming it again gives 'franchi'. The second stemming will cause the query not to match anymore. So basically my questions are: 1. Should I not be passing my lookup queries back to the request handler, but instead to some lower level component? If so, which component would be good to look at? 2. Is there a way to tell the SolrQueryParser not to analyze or a different way to run the query so that they query analysis won't happen? Thanks again, Laurent Vauthrin -Original Message- From: solr-user-return-30170-laurent.vauthrin=disney@lucene.apache.org [mailto:solr-user-return-30170-laurent.vauthrin=disney@lucene.apache .org] On Behalf Of Vauthrin, Laurent Sent: Wednesday, December 09, 2009 2:53 PM To: solr-user@lucene.apache.org Subject: RE: SolrPlugin Guidance Thanks for the response. I went ahead and gave it a shot. In my case, the directory name may not be unique so if I get multiple ids back then I create a BooleanQuery (Occur.SHOULD) to substitute the directory name query. This seems to work at the moment so hopefully that's the right approach. Thanks, Laurent Vauthrin -Original Message- From: solr-user-return-30054-laurent.vauthrin=disney@lucene.apache.org [mailto:solr-user-return-30054-laurent.vauthrin=disney@lucene.apache .org] On Behalf Of Chris Hostetter Sent: Monday, December 07, 2009 12:17 PM To: solr-user@lucene.apache.org Subject: RE: SolrPlugin Guidance : e.g. For the following query that looks for a file in a directory: : q=+directory_name:"myDirectory" +file_name:"myFile" : : We'd need to decompose the query into the following two queries: : 1. q=+directory_name:"myDirectory"&fl=directory_id : 2. q=+file_name:"myFile" +directory_id:(results from query #1) : : I guess I'm looking for the following feedback: : - Does this sound crazy? it's a little crazy, but not absurd. : - Is the QParser the right place for this logic? If so, can I get a : little more guidance on how to decompose the queries there (filter : queries maybe)? a QParser could work. (and in general, if you can solve something with a QParser that's probably for the best, since it allows the most reuse). but exactly how to do it depends on how many results you expect from your first query: if you are going to structure things so they have to uniquely id a directory, and you'll have a singleID, then this is something that could easily make sense in a QParser (you are essentailly just rewriting part of the query from string to id -- you just happen to be using solr as a lookup table for those strings). but if you plan to support any arbitrary "N" directories, then you may need something more complicated ... straight filter queries won't help much because you'll want the union instead of hte intersection, so for every directoryId you find, use it as a query to get a DocSet and then maintain a running union of all those DocSets to use as your final filter (hmm... that may not actually be possible with the QParser API ... i haven't look at ti in a while, but for an approach like this you may beed to subclass QueryComponent instead) -Hoss
Re: Solr usage with Auctions/Classifieds?
On Dec 8, 2009, at 6:37 PM, regany wrote: > > hello! > > just wondering if anyone is using Solr as their search for an auction / > classified site, and if so how have you managed your setup in general? ie. > searching against listings that may have expired etc. I know several companies using Solr for classifieds/auctions. Some remove the old listings while others leave them in and filter them or even allow users to see old stuff (but often for reasons other than users finding them, i.e. SEO). For those that remove, it's typically a batch operation that takes place at night. -- Grant Ingersoll http://www.lucidimagination.com/ Search the Lucene ecosystem (Lucene/Solr/Nutch/Mahout/Tika/Droids) using Solr/Lucene: http://www.lucidimagination.com/search
Re: Custom Field sample?
Can you perhaps give a little more info on what problem you are trying to solve? FWIW, there are a lot of examples of custom FieldTypes in the Solr code. On Dec 10, 2009, at 11:46 AM, Antonio Zippo wrote: > Hi all, > > could you help me to create a custom field? > > I need to create a field structured like a Map > is it possible? how to define if the search string is on key or value (or > both)? > > A way could be to create a char separated multivalued string field... but it > isn't the best way. and with facets is the worst way > > could you give me a custom field sample? > > > Thanks in advance, > Revenge > > -- Grant Ingersoll http://www.lucidimagination.com/ Search the Lucene ecosystem (Lucene/Solr/Nutch/Mahout/Tika/Droids) using Solr/Lucene: http://www.lucidimagination.com/search
Re: Can we build complex filter queries in SOLR
Anyway i wrote a lucene syntax for this +((+category:audio +brand:sony)|(+category:video +brand:sony)) Thanks darniz Shalin Shekhar Mangar wrote: > > On Thu, Dec 10, 2009 at 2:50 AM, darniz wrote: > >> >> Can you please let me know how to describe that condition. >> >> >> For example lets say i want to give the following condition >> >> ((category:audio or category:video) AND (brand:sony OR brand:samsung OR >> brand:sanyo)) >> How would you represent this condition in fq paramenter of dismax >> >> > Are you saying that the above syntax does not work in an fq? Note, the > "or" > should be in capitals. > > -- > Regards, > Shalin Shekhar Mangar. > > -- View this message in context: http://old.nabble.com/Can-we-build-complex-filter-queries-in-SOLR-tp12735426p26731738.html Sent from the Solr - User mailing list archive at Nabble.com.
Custom Field sample?
Hi all, could you help me to create a custom field? I need to create a field structured like a Map is it possible? how to define if the search string is on key or value (or both)? A way could be to create a char separated multivalued string field... but it isn't the best way. and with facets is the worst way could you give me a custom field sample? Thanks in advance, Revenge
Re: Multiple Facet prefixes on the same facet field in one request?
Hey there Shalin, After scrolling through the Jira Issues I found, http://issues.apache.org/jira/browse/SOLR-1387, discussing this exact thing.Is anyone currently working on a patch for this issue still? If not has the syntax in the last post in the Jira issue been approved so a patch can be made? Thanks Robert. On 2009-12-10, at 2:19 AM, Shalin Shekhar Mangar wrote: > On Wed, Dec 9, 2009 at 5:25 AM, Robert Purdy wrote: > >> >> Hey all, >> >> Is there anyway in Solr 1.4/1.5 to perform multiple facet prefixes on the >> same facet field in one request? >> >> Ex. On field 'Foo' I want to perform a facet prefix of A* and B* so I can >> get a facet response of all terms prefixed with A and all terms prefixed >> with B, either grouped together in the same facet result list or seperate >> facet lists labeled by the prefix. >> >> Currently, I perform one request per facet prefix and I am hoping that >> there >> is some cryptic way using local params that I am missing that will allow me >> to do this. >> >> > No, I don't think there is a way to do request for more than one > facet.prefix on the same field in one request. > > -- > Regards, > Shalin Shekhar Mangar.
Re: Solr Configuration Management
Hi there, as i know there are two solr configuration properties files, dataimport.properties and solrcore.properties. What i want is to intercept solr app to use my own property definition files, is there any global bean (per core) which contains this data? Where in the code are the properties setted? Is there any open jira issue to allow global core configuration (dataimport and core)? Thx El 10 de diciembre de 2009 16:35, Licinio Fernández Maurelo < licinio.fernan...@gmail.com> escribió: > > > -- Mensaje reenviado -- > De: Licinio Fernández Maurelo > Fecha: 27 de octubre de 2009 09:50 > Asunto: Re: Solr Configuration Management > Para: solr-user@lucene.apache.org, noble.p...@gmail.com > > > > > are you referring to DIH? > yes > > 2009/10/27 Noble Paul നോബിള് नोब्ळ् > > 2009/10/26 Licinio Fernández Maurelo : >> > Hi there, >> > >> > i must enhance solr config deploys. >> > >> > I have a configuration file per environment and per role (Master-Slave) >> so i >> > want to separate DataSource definitions from the solrconfig.xml . Where >> can >> > i put them? >> are you referring to DIH? >> >> > >> > Same behaviour is desired for Master-Slave conf diffs. >> you can drop in all your custom properties in a solrcore.properties >> file (placed in conf dir) and can have different properties files for >> master and slave . These properties can be directly be referred from >> solrconfig >> > >> > Any help would be much appreciatted ... >> > >> > >> > -- >> > Lici >> > >> >> >> >> -- >> - >> Noble Paul | Principal Engineer| AOL | http://aol.com >> > > > > -- > Lici > > > > -- > Lici > ~Java Developer~ > -- Lici ~Java Developer~
Re: copyField question
Thanks. I will take a look at the TokenFilter. On Wed, Dec 9, 2009 at 11:44 PM, Shalin Shekhar Mangar < shalinman...@gmail.com> wrote: > On Wed, Dec 9, 2009 at 11:43 PM, P Franks wrote: > > > All, > > > > Can one use the copyField option and copy a TextField field into a > > longField field? I have some data that i want to extract (filter) out > all > > but the long and/or integer values. > > > > > No, that won't work. It'd be best to use a TokenFilter which remove > characters and just keeps the integer/long values. But you still won't be > able to use the LongField because that is not analyzed (so your token > filters will not be applied). > > -- > Regards, > Shalin Shekhar Mangar. >
Re: UpdateRequestProcessor to avoid documents of being indexed
Yes, it did Cheers Chris Male wrote: > > Hi, > > Yeah thats what I was suggesting. Did that work? > > On Thu, Dec 10, 2009 at 12:24 PM, Marc Sturlese > wrote: > >> >> Do you mean something like?: >> >>@Override >>public void processAdd(AddUpdateCommand cmd) throws IOException { >>boolean addDocToIndex >> =dealWithSolrDocFields(cmd.getSolrInputDocument()) ; >>if (next != null && addDocToIndex) { >>next.processAdd(cmd); >>} else { >> LOG.debug("Doc skipped!") ; >>} >>} >> >> Thanks in advance >> >> >> >> Chris Male wrote: >> > >> > Hi, >> > >> > If your UpdateRequestProcessor does not forward the AddUpdateCommand >> onto >> > the RunUpdateProcessor, I believe the document will not be indexed. >> > >> > Cheers >> > >> > On Thu, Dec 10, 2009 at 12:09 PM, Marc Sturlese >> > wrote: >> > >> >> >> >> Hey there, >> >> I need that once a document has been created be able to decide if I >> want >> >> it >> >> to be indexed or not. I have thought in implement an >> >> UpdateRequestProcessor >> >> to do that but don't know how to tell Solr in the processAdd void to >> skip >> >> the document. >> >> If I delete all the field would it be skiped or is there a better way >> to >> >> reach this goal? >> >> Thanks in advance. >> >> -- >> >> View this message in context: >> >> >> http://old.nabble.com/UpdateRequestProcessor-to-avoid-documents-of-being-indexed-tp26725534p26725534.html >> >> Sent from the Solr - User mailing list archive at Nabble.com. >> >> >> >> >> > >> > >> > -- >> > Chris Male | Software Developer | JTeam BV.| T: +31-(0)6-14344438 | >> > www.jteam.nl >> > >> > >> >> -- >> View this message in context: >> http://old.nabble.com/UpdateRequestProcessor-to-avoid-documents-of-being-indexed-tp26725534p26725698.html >> Sent from the Solr - User mailing list archive at Nabble.com. >> >> > > > -- > Chris Male | Software Developer | JTeam BV.| www.jteam.nl > > -- View this message in context: http://old.nabble.com/UpdateRequestProcessor-to-avoid-documents-of-being-indexed-tp26725534p26726566.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Results after using Field Collapsing are not matching the results without using Field Collapsing
Hi Martijn, I am not sending the collapse parameters for the second query. Here are the queries I am using: *When using field collapsing (searching over all categories):* spellcheck=true&collapse.info.doc=true&facet=true&collapse.threshold=3&facet.mincount=1&spellcheck.q=weight+loss&collapse.facet=before&wt=xml&f.content.hl.snippets=2&hl=true&version=2.2&rows=20&collapse.field=ctype&fl=id,sid,title,image,ctype,score&start=0&q=weight+loss&collapse.info.count=false&facet.field=ctype&qt=contentsearch categories is represented as the field "ctype" above. *Without using field collapsing:* spellcheck=true&facet=true&facet.mincount=1&spellcheck.q=weight+loss&wt=xml&hl=true&rows=10&version=2.2&fl=id,sid,title,image,ctype,score&start=0&q=weight+loss&facet.field=ctype&qt=contentsearch I append "&fq=ctype:1" to the above queries when trying to get results for a particular category. -- Thanks Varun Gupta On Thu, Dec 10, 2009 at 5:58 PM, Martijn v Groningen < martijn.is.h...@gmail.com> wrote: > Hi Varun, > > Can you send the whole requests (with params), that you send to Solr > for both queries? > In your situation the collapse parameters only have to be used for the > first query and not the second query. > > Martijn > > 2009/12/10 Varun Gupta : > > Hi, > > > > I have documents under 6 different categories. While searching, I want to > > show 3 documents from each category along with a link to see all the > > documents under a single category. I decided to use field collapsing so > that > > I don't have to make 6 queries (one for each category). Currently I am > using > > the field collapsing patch uploaded on 29th Nov. > > > > Now, the results that are coming after using field collapsing are not > > matching the results for a single category. For example, for category C1, > I > > am getting results R1, R2 and R3 using field collapsing, but after I see > > results only from the category C1 (without using field collapsing) these > > results are nowhere in the first 10 results. > > > > Am I doing something wrong or using the field collapsing for the wrong > > feature? > > > > I am using the following field collapsing parameters while querying: > > collapse.field=category > > collapse.facet=before > > collapse.threshold=3 > > > > -- > > Thanks > > Varun Gupta > > > > > > -- > Met vriendelijke groet, > > Martijn van Groningen >
Reverse sort facet query
Hello Forum, I've had a search in the mail archives and on the 'net, but I'm sure I wouldn't be the first to have a requirement for this: Does anyone know of a good way to perform a reverse-sorted facet query (i.e. rarest first)? As you know facet.sort toggles between sorting on count or field name, but there's no built-in method for reverse count. One way I've found to do this is to set facet.limit=-1 (and facet.mincount) to get the entire list, then take 'bottom-5' to get a 'rare' list. This works, but it's not great for very large lists. Does anyone know of a better way? Many thanks, Peter _ Add your Gmail and Yahoo! Mail email accounts into Hotmail - it's easy http://clk.atdmt.com/UKM/go/186394592/direct/01/
Re: Results after using Field Collapsing are not matching the results without using Field Collapsing
Hi Varun, Can you send the whole requests (with params), that you send to Solr for both queries? In your situation the collapse parameters only have to be used for the first query and not the second query. Martijn 2009/12/10 Varun Gupta : > Hi, > > I have documents under 6 different categories. While searching, I want to > show 3 documents from each category along with a link to see all the > documents under a single category. I decided to use field collapsing so that > I don't have to make 6 queries (one for each category). Currently I am using > the field collapsing patch uploaded on 29th Nov. > > Now, the results that are coming after using field collapsing are not > matching the results for a single category. For example, for category C1, I > am getting results R1, R2 and R3 using field collapsing, but after I see > results only from the category C1 (without using field collapsing) these > results are nowhere in the first 10 results. > > Am I doing something wrong or using the field collapsing for the wrong > feature? > > I am using the following field collapsing parameters while querying: > collapse.field=category > collapse.facet=before > collapse.threshold=3 > > -- > Thanks > Varun Gupta > -- Met vriendelijke groet, Martijn van Groningen
Re: Payloads with Phrase queries
> I was looking through some lucene > source codes and found the following class > org.apache.lucene.search.payloads.PayloadSpanUtil > > There is a function named queryToSpanQuery in this class. > Is this the > preferred way to convert a PhraseQuery to > PayloadNearQuery? queryToSpanQuery method does not return PayloadNearQuery type. You need to override getFieldQuery(String field, String queryText, int slop) of SolrQueryParser or QueryParser. This code is modified from Lucene In Action Book (2nd edition) Chapter 6.3.4 Allowing ordered phrase queries protected Query getFieldQuery(String field, String queryText, int slop) throws ParseException { Query orig = super.getFieldQuery(field, queryText, slop); if (!(orig instanceof PhraseQuery)) return orig; PhraseQuery pq = (PhraseQuery) orig; Term[] terms = pq.getTerms(); SpanQuery[] clauses = new SpanQuery[terms.length]; for (int i = 0; i < terms.length; i++) clauses[i] = new PayloadTermQuery(terms[i], new AveragePayloadFunction()); return new PayloadNearQuery(clauses, slop, true); } > Also, are there any performance considerations while using > a PayloadNearQuery instead of a PhraseQuery? I don't think there will be significant performance difference.
Re: Payloads with Phrase queries
On Dec 10, 2009, at 6:35 AM, Raghuveer Kancherla wrote: > I was looking through some lucene source codes and found the following class > org.apache.lucene.search.payloads.PayloadSpanUtil > > There is a function named queryToSpanQuery in this class. Is this the > preferred way to convert a PhraseQuery to PayloadNearQuery? I don't think that there is a conversion method. You will need to do this on your own. > > Also, are there any performance considerations while using a > PayloadNearQuery instead of a PhraseQuery? Yes, I would suspect PNQ to be slower, but likely not that much slower. That being said, I haven't benchmarked it. Then again, PNQ does stuff that PQ doesn't, so it's an apples and oranges debate. > > Thanks, > Raghu > > > > On Thu, Dec 10, 2009 at 4:40 PM, AHMET ARSLAN wrote: > >>> Hi, >>> I am looking for a way to use payloads in my search >>> application. Indexing >>> data with payloads into Solr is pretty straightforward. >>> However using the >>> payloads during search time is a bit confusing. Can anyone >>> point me in the >>> right direction to enable payloads on a *PhraseQuery*. I >>> looked at the >>> following resources and got payload on a TermQuery >>> working. >>> >>> 1. >>> >> http://www.lucidimagination.com/blog/2009/08/05/getting-started-with-payloads/ >>> 2. >> http://www.mail-archive.com/solr-user@lucene.apache.org/msg24863.html >>> 3. There is also a jira issue (SOLR-1485) >>> that gives a patch for using >>> Payload. >>> 4. Lucene-In-Action >>> >>> I am guessing that I should return a payload version of >>> PhraseQuery in >>> QueryParser's (package >>> org.apache.lucene.queryParser.queryParser.java) >>> newPhraseQuery function. If yes, what type should this >>> query be? >> >> Yes. PayloadNearQuery [1] >> >> [1] >> http://lucene.apache.org/java/3_0_0/api/core/org/apache/lucene/search/payloads/PayloadNearQuery.html >> >> >> >> -- Grant Ingersoll http://www.lucidimagination.com/ Search the Lucene ecosystem (Lucene/Solr/Nutch/Mahout/Tika/Droids) using Solr/Lucene: http://www.lucidimagination.com/search
Results after using Field Collapsing are not matching the results without using Field Collapsing
Hi, I have documents under 6 different categories. While searching, I want to show 3 documents from each category along with a link to see all the documents under a single category. I decided to use field collapsing so that I don't have to make 6 queries (one for each category). Currently I am using the field collapsing patch uploaded on 29th Nov. Now, the results that are coming after using field collapsing are not matching the results for a single category. For example, for category C1, I am getting results R1, R2 and R3 using field collapsing, but after I see results only from the category C1 (without using field collapsing) these results are nowhere in the first 10 results. Am I doing something wrong or using the field collapsing for the wrong feature? I am using the following field collapsing parameters while querying: collapse.field=category collapse.facet=before collapse.threshold=3 -- Thanks Varun Gupta
Re: UpdateRequestProcessor to avoid documents of being indexed
Hi, Yeah thats what I was suggesting. Did that work? On Thu, Dec 10, 2009 at 12:24 PM, Marc Sturlese wrote: > > Do you mean something like?: > >@Override >public void processAdd(AddUpdateCommand cmd) throws IOException { >boolean addDocToIndex > =dealWithSolrDocFields(cmd.getSolrInputDocument()) ; >if (next != null && addDocToIndex) { >next.processAdd(cmd); >} else { > LOG.debug("Doc skipped!") ; >} >} > > Thanks in advance > > > > Chris Male wrote: > > > > Hi, > > > > If your UpdateRequestProcessor does not forward the AddUpdateCommand onto > > the RunUpdateProcessor, I believe the document will not be indexed. > > > > Cheers > > > > On Thu, Dec 10, 2009 at 12:09 PM, Marc Sturlese > > wrote: > > > >> > >> Hey there, > >> I need that once a document has been created be able to decide if I want > >> it > >> to be indexed or not. I have thought in implement an > >> UpdateRequestProcessor > >> to do that but don't know how to tell Solr in the processAdd void to > skip > >> the document. > >> If I delete all the field would it be skiped or is there a better way to > >> reach this goal? > >> Thanks in advance. > >> -- > >> View this message in context: > >> > http://old.nabble.com/UpdateRequestProcessor-to-avoid-documents-of-being-indexed-tp26725534p26725534.html > >> Sent from the Solr - User mailing list archive at Nabble.com. > >> > >> > > > > > > -- > > Chris Male | Software Developer | JTeam BV.| T: +31-(0)6-14344438 | > > www.jteam.nl > > > > > > -- > View this message in context: > http://old.nabble.com/UpdateRequestProcessor-to-avoid-documents-of-being-indexed-tp26725534p26725698.html > Sent from the Solr - User mailing list archive at Nabble.com. > > -- Chris Male | Software Developer | JTeam BV.| www.jteam.nl
Re: Payloads with Phrase queries
I was looking through some lucene source codes and found the following class org.apache.lucene.search.payloads.PayloadSpanUtil There is a function named queryToSpanQuery in this class. Is this the preferred way to convert a PhraseQuery to PayloadNearQuery? Also, are there any performance considerations while using a PayloadNearQuery instead of a PhraseQuery? Thanks, Raghu On Thu, Dec 10, 2009 at 4:40 PM, AHMET ARSLAN wrote: > > Hi, > > I am looking for a way to use payloads in my search > > application. Indexing > > data with payloads into Solr is pretty straightforward. > > However using the > > payloads during search time is a bit confusing. Can anyone > > point me in the > > right direction to enable payloads on a *PhraseQuery*. I > > looked at the > > following resources and got payload on a TermQuery > > working. > > > >1. > > > http://www.lucidimagination.com/blog/2009/08/05/getting-started-with-payloads/ > >2. > http://www.mail-archive.com/solr-user@lucene.apache.org/msg24863.html > >3. There is also a jira issue (SOLR-1485) > > that gives a patch for using > >Payload. > >4. Lucene-In-Action > > > > I am guessing that I should return a payload version of > > PhraseQuery in > > QueryParser's (package > > org.apache.lucene.queryParser.queryParser.java) > > newPhraseQuery function. If yes, what type should this > > query be? > > Yes. PayloadNearQuery [1] > > [1] > http://lucene.apache.org/java/3_0_0/api/core/org/apache/lucene/search/payloads/PayloadNearQuery.html > > > >
Re: UpdateRequestProcessor to avoid documents of being indexed
Do you mean something like?: @Override public void processAdd(AddUpdateCommand cmd) throws IOException { boolean addDocToIndex =dealWithSolrDocFields(cmd.getSolrInputDocument()) ; if (next != null && addDocToIndex) { next.processAdd(cmd); } else { LOG.debug("Doc skipped!") ; } } Thanks in advance Chris Male wrote: > > Hi, > > If your UpdateRequestProcessor does not forward the AddUpdateCommand onto > the RunUpdateProcessor, I believe the document will not be indexed. > > Cheers > > On Thu, Dec 10, 2009 at 12:09 PM, Marc Sturlese > wrote: > >> >> Hey there, >> I need that once a document has been created be able to decide if I want >> it >> to be indexed or not. I have thought in implement an >> UpdateRequestProcessor >> to do that but don't know how to tell Solr in the processAdd void to skip >> the document. >> If I delete all the field would it be skiped or is there a better way to >> reach this goal? >> Thanks in advance. >> -- >> View this message in context: >> http://old.nabble.com/UpdateRequestProcessor-to-avoid-documents-of-being-indexed-tp26725534p26725534.html >> Sent from the Solr - User mailing list archive at Nabble.com. >> >> > > > -- > Chris Male | Software Developer | JTeam BV.| T: +31-(0)6-14344438 | > www.jteam.nl > > -- View this message in context: http://old.nabble.com/UpdateRequestProcessor-to-avoid-documents-of-being-indexed-tp26725534p26725698.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: UpdateRequestProcessor to avoid documents of being indexed
Hi, If your UpdateRequestProcessor does not forward the AddUpdateCommand onto the RunUpdateProcessor, I believe the document will not be indexed. Cheers On Thu, Dec 10, 2009 at 12:09 PM, Marc Sturlese wrote: > > Hey there, > I need that once a document has been created be able to decide if I want it > to be indexed or not. I have thought in implement an UpdateRequestProcessor > to do that but don't know how to tell Solr in the processAdd void to skip > the document. > If I delete all the field would it be skiped or is there a better way to > reach this goal? > Thanks in advance. > -- > View this message in context: > http://old.nabble.com/UpdateRequestProcessor-to-avoid-documents-of-being-indexed-tp26725534p26725534.html > Sent from the Solr - User mailing list archive at Nabble.com. > > -- Chris Male | Software Developer | JTeam BV.| T: +31-(0)6-14344438 | www.jteam.nl
Re: Payloads with Phrase queries
> Hi, > I am looking for a way to use payloads in my search > application. Indexing > data with payloads into Solr is pretty straightforward. > However using the > payloads during search time is a bit confusing. Can anyone > point me in the > right direction to enable payloads on a *PhraseQuery*. I > looked at the > following resources and got payload on a TermQuery > working. > > 1. > http://www.lucidimagination.com/blog/2009/08/05/getting-started-with-payloads/ > 2. http://www.mail-archive.com/solr-user@lucene.apache.org/msg24863.html > 3. There is also a jira issue (SOLR-1485) > that gives a patch for using > Payload. > 4. Lucene-In-Action > > I am guessing that I should return a payload version of > PhraseQuery in > QueryParser's (package > org.apache.lucene.queryParser.queryParser.java) > newPhraseQuery function. If yes, what type should this > query be? Yes. PayloadNearQuery [1] [1]http://lucene.apache.org/java/3_0_0/api/core/org/apache/lucene/search/payloads/PayloadNearQuery.html
UpdateRequestProcessor to avoid documents of being indexed
Hey there, I need that once a document has been created be able to decide if I want it to be indexed or not. I have thought in implement an UpdateRequestProcessor to do that but don't know how to tell Solr in the processAdd void to skip the document. If I delete all the field would it be skiped or is there a better way to reach this goal? Thanks in advance. -- View this message in context: http://old.nabble.com/UpdateRequestProcessor-to-avoid-documents-of-being-indexed-tp26725534p26725534.html Sent from the Solr - User mailing list archive at Nabble.com.
Payloads with Phrase queries
Hi, I am looking for a way to use payloads in my search application. Indexing data with payloads into Solr is pretty straightforward. However using the payloads during search time is a bit confusing. Can anyone point me in the right direction to enable payloads on a *PhraseQuery*. I looked at the following resources and got payload on a TermQuery working. 1. http://www.lucidimagination.com/blog/2009/08/05/getting-started-with-payloads/ 2. http://www.mail-archive.com/solr-user@lucene.apache.org/msg24863.html 3. There is also a jira issue (SOLR-1485) that gives a patch for using Payload. 4. Lucene-In-Action I am guessing that I should return a payload version of PhraseQuery in QueryParser's (package org.apache.lucene.queryParser.queryParser.java) newPhraseQuery function. If yes, what type should this query be? Thanks, Raghu
Re: atypical MLT use-case
solr 1.4 enterprise search server. it's on the left column of the solr homepage. http://www.packtpub.com/solr-1-4-enterprise-search-server?utm_source=http://lucene.apache.org/solr/&utm_medium=spons&utm_content=pod&utm_campaign=mdb_000275 On 09/12/09 19:14, Mike Anderson wrote: wow! exactly what i'm looking for. What solr1.4 book is this? thanks so much. If anybody knows the details of how to use this I'd love to hear your tips, experiences, or comments. -mike On Dec 9, 2009, at 12:55 PM, Andre Parodi wrote: the solr 1.4 book says you can do this. usages of mlt: "As a request handler with an external input document: What if you want similarity results based on something that isn't in the index? A final option that Solr supports is returning MLT results based on text data sent to the MLT handler (through HTTP POST). For example, if you were to send a text file to the handler, then Solr's MLT handler would return the documents in the index that are most similar to it. This is atypical but an interesting option nonetheless." not sure about the details of how though as i haven't used mlt myself. On 09/12/09 17:27, Mike Anderson wrote: This is somewhat of an odd use-case for MLT. Basically I'm using it for near-duplicate detection (I'm not using the built in dup detection for a variety of reasons). While this might sound like an okay idea, the problem lies in the order of which things happen. Ideally, duplicate detection would prevent me from adding a document to my index which is already there (or at least partially there). However, more like this only works on documents which are *already* in the index. Ideally what I would be able to do is: post an xml document to solr, and receive a MLT response (the same kind of MLT response I would recieve had the document been in Solr already, and queried with id=#{id}&mlt=true). Is anybody aware of how I could achieve this functionality leveraging existing handlers? If not I will bump over to solr-dev and see if this is a tractable problem. Thanks in advance, Mike
RE: Selection of returned fields - dynamic fields?
OK thanks for the reply, fortunately we have now found an approach which avoids storing the field. It would be nice to be able to search for dynamic fields in a way which is consistent with their definition, although I suppose there probably isn't demand for this. Regards, Ian. -Original Message- From: Chris Hostetter [mailto:hossman_luc...@fucit.org] Sent: 09 December 2009 19:36 To: solr-user@lucene.apache.org Cc: Gary Ratcliffe Subject: Re: Selection of returned fields - dynamic fields? : Unfortunately this does not seem to work for dynamic fields - you can definiltely ask for a field that exists because of a dynamicField by name, but you can't use wildcard style patterns in the "fl" param. : "fl=PREFIX*" does not return anything, and neither does "fl=*POSTFIX". : What seems to be missing from Solr is a removeField(FIELDNAME) method in : SolrJ, or a "fl=-FIELDNAME" query parameter to remove the fixed field. : : Is such a feature planned, or is there a workaround that I have missed? There's been a lot of discussion about it over the years, the crux of the problem is that it's hard to come up with a good way of dealing with field names using "meta characters" that doesn't make it hard for people to actaully use those metacharacters in their field names... http://wiki.apache.org/solr/FieldAliasesAndGlobsInParams -Hoss
Re: Logging
On Wed, Dec 9, 2009 at 8:19 PM, Lee Smith wrote: > Im trying to import data with DIH (mysql) > > All my SQL's are good having been tested manually. > > When I run full import ie: > http://localhost:8983/solr/dataimport?command=full-import > > I get my XML result but nothing is being imported and it Rolles back. > > In loggin I set DIH logging to fine and set them then re-run but I can seem > to find detailed logs. > > All DIH errors are logged with SEVERE level. You should be able to find them with your servlet container's log files. If you can post the stack traces we will be able to suggest solutions. -- Regards, Shalin Shekhar Mangar.
Re: Solr usage with Auctions/Classifieds?
On Wed, Dec 9, 2009 at 5:07 AM, regany wrote: > > hello! > > just wondering if anyone is using Solr as their search for an auction / > classified site, and if so how have you managed your setup in general? ie. > searching against listings that may have expired etc. > > We use Solr more many listings sites. Typically we index the expiry date in the document and exclude them through a filter e.g. fq=expiry_date:[NOW/DAY+1DAYS TO *] -- Regards, Shalin Shekhar Mangar.
Re: Multiple Facet prefixes on the same facet field in one request?
On Wed, Dec 9, 2009 at 5:25 AM, Robert Purdy wrote: > > Hey all, > > Is there anyway in Solr 1.4/1.5 to perform multiple facet prefixes on the > same facet field in one request? > > Ex. On field 'Foo' I want to perform a facet prefix of A* and B* so I can > get a facet response of all terms prefixed with A and all terms prefixed > with B, either grouped together in the same facet result list or seperate > facet lists labeled by the prefix. > > Currently, I perform one request per facet prefix and I am hoping that > there > is some cryptic way using local params that I am missing that will allow me > to do this. > > No, I don't think there is a way to do request for more than one facet.prefix on the same field in one request. -- Regards, Shalin Shekhar Mangar.
Re: Facet across second level of hierarchy - Dismax Request Handler - Best practice?
On Wed, Dec 9, 2009 at 8:23 PM, Daniel Knapp wrote: > Hello, > > i want a second level of hierarchy in my facets (as seen here: > http://www.lucidimagination.com/search/?q=) > > Although the facets are presented in a hierarchical fashion, I don't think it is the hierarchical faceting patch. They may just be having flag-fields for each level e.g. for a solr-user mail, they may index "Email", "user", "dev" in a multi-valued field called "Source" and display it in a hierarchical UI. -- Regards, Shalin Shekhar Mangar.