Re: Store input text after analyzers and token filters
For solr 1.4 Is basically the same but IndexSchema (org.apache.solr.schema.IndexSchema) needs to be updated to include the function getFieldTypeByName(String fieldTypeName) which is already in sorl1.5 /** * Given the name of a {...@link org.apache.solr.schema.FieldType} (not to be confused with {...@link #getFieldType(String)} which * takes in the name of a field), return the {...@link org.apache.solr.schema.FieldType}. * @param fieldTypeName The name of the {...@link org.apache.solr.schema.FieldType} * @return The {...@link org.apache.solr.schema.FieldType} or null. */ public FieldType getFieldTypeByName(String fieldTypeName){ return fieldTypes.get(fieldTypeName); } then the AnalyzedField is a bit different, but basicaclly is a copy of the TextField as is in 1.4 http://old.nabble.com/file/p27902273/AnalyzedField.java AnalyzedField.java -- View this message in context: http://old.nabble.com/Store-input-text-after-analyzers-and-token-filters-tp27792550p27902273.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Store input text after analyzers and token filters
Otis, I've been thinking on it, and trying to figure out the different solutions - Try to solve it doing a bridge between solr and clustering. - Try to solve it before/during indexing The second option, of course is better for performance, but how to do it?? I think a good option may be to create a new type derived type from the FieldType class like the SortableIntField which has the toInternal(String val) function. Then the problem is how to include the result of the analysis of anoter field type in the toInternal function So there would be a new type that can be used on copy fields , that takes the analysis of the source field and injects in the code. It takes as parameter the field from which takes the analysis . So, how can I get the result of the analysis of a given text by a given field using internal functions?? Otis Gospodnetic wrote: Hi Joan, You could use the FieldAnalysisRequestHandler: http://www.search-lucene.com/?q=FieldAnalysisRequestHandler Otis Sematext :: http://sematext.com/ :: Solr - Lucene - Nutch Hadoop ecosystem search :: http://search-hadoop.com/ -- View this message in context: http://old.nabble.com/Store-input-text-after-analyzers-and-token-filters-tp27792550p27840488.html Sent from the Solr - User mailing list archive at Nabble.com.
Store input text after analyzers and token filters
In an stored field, the content stored is the raw input text. But when the analyzers perform some cleaning or interesting transformation of the text, then it could be interesting to store the text after the tokenizer/Filter chain there is a way to do this? To be able to get back the text of the document after being processed?? thanks Joan -- View this message in context: http://old.nabble.com/Store-input-text-after-analyzers-and-token-filters-tp27792550p27792550.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Store input text after analyzers and token filters
Thanks, It can be useful as a workarrond, but I get a vector not a result that I may use wherever I could used the stored text. I'm thinking in clustering. Ahmet Arslan wrote: In an stored field, the content stored is the raw input text. But when the analyzers perform some cleaning or interesting transformation of the text, then it could be interesting to store the text after the tokenizer/Filter chain there is a way to do this? To be able to get back the text of the document after being processed?? You can get term vectors [1] of analyzed text. Also you can see analyzed text in solr/admin/analysis.jsp if you copy and paste sample text data. [1] http://wiki.apache.org/solr/TermVectorComponent -- View this message in context: http://old.nabble.com/Store-input-text-after-analyzers-and-token-filters-tp27792550p27794689.html Sent from the Solr - User mailing list archive at Nabble.com.
Clustering from anlayzed text instead of raw input
I'm trying to use carrot2 (now I started with the workbench) and I can cluster any field, but, the text used for clustering is the original raw text, the one that was indexed, without any of the processing performed by the tokenizer or filters. So I get stop words. I also did shingles (after filtering by POS) and I can not cluster using these multiwords. So my question is about how to get in a query answer the indexed text instead of the original one, because if I set stored to false, then the search does not return the content of the field. Tahnks in advance Joan -- View this message in context: http://old.nabble.com/Clustering-from-anlayzed-text-instead-of-raw-input-tp27765780p27765780.html Sent from the Solr - User mailing list archive at Nabble.com.
error in sum function
the sum function or the map one are not parsed correctly, doing this sort, works as a charm... sort=score+desc,sum(Num,map(Num,0,2000,42000))+asc but sort=score+desc,sum(map(Num,0,2000,42000),Num)+asc gives the following exception SEVERE: org.apache.solr.common.SolrException: Must declare sort field or function at org.apache.solr.search.QueryParsing.processSort(QueryParsing.java:376) at org.apache.solr.search.QueryParsing.parseSort(QueryParsing.java:281) at org.apache.solr.search.QParser.getSort(QParser.java:217) at org.apache.solr.handler.component.QueryComponent.prepare(QueryComponent.java:86) at org.apache.solr.handler.component.SearchHandler.handleRequestBody(SearchHandler.java:174) at org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:131) you can test it in here using these two url's http://varoitus.barcelonamedia.org:8180/fbm-ijec/select?indent=onfl=id,Num,scoresort=score%20asc,sum%28map%28Num,0,5000,42000%29,Num%29+ascq=+entities_org:%28%22Amena%22%29 http://varoitus.barcelonamedia.org:8180/fbm-ijec/select?indent=onwt=phpfl=id,Num,scorerows=50sort=score+desc,sum%28Num,map%28Num,0,2000,42000%29%29+ascq=+entities_org:Amena -- View this message in context: http://old.nabble.com/error-in-sum-function-tp27765881p27765881.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: error in sum function
Ok, solved!!! Joan Koji Sekiguchi-2 wrote: Can you try it latest trunk? I have just fixed it in a couple of days Koji Sekiguchi from mobile On 2010/03/03, at 18:18, JCodina joan.cod...@barcelonamedia.org wrote: the sum function or the map one are not parsed correctly, doing this sort, works as a charm... sort=score+desc,sum(Num,map(Num,0,2000,42000))+asc but sort=score+desc,sum(map(Num,0,2000,42000),Num)+asc gives the following exception SEVERE: org.apache.solr.common.SolrException: Must declare sort field or function at org.apache.solr.search.QueryParsing.processSort(QueryParsing.java:376) at org.apache.solr.search.QueryParsing.parseSort(QueryParsing.java:281) at org.apache.solr.search.QParser.getSort(QParser.java:217) at org.apache.solr.handler.component.QueryComponent.prepare (QueryComponent.java:86) at org.apache.solr.handler.component.SearchHandler.handleRequestBody (SearchHandler.java:174) at org.apache.solr.handler.RequestHandlerBase.handleRequest (RequestHandlerBase.java:131) you can test it in here using these two url's http://varoitus.barcelonamedia.org:8180/fbm-ijec/select?indent=onfl=id,Num,scoresort=score%20asc,sum%28map%28Num,0,5000,42000%29,Num%29+ascq=+entities_org:%28%22Amena%22%29 http://varoitus.barcelonamedia.org:8180/fbm-ijec/select?indent=onwt=phpfl=id,Num,scorerows=50sort=score+desc,sum%28Num,map%28Num,0,2000,42000%29%29+ascq=+entities_org:Amena -- View this message in context: http://old.nabble.com/error-in-sum-function-tp27765881p27765881.html Sent from the Solr - User mailing list archive at Nabble.com. -- View this message in context: http://old.nabble.com/error-in-sum-function-tp27765881p27768877.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Clustering from anlayzed text instead of raw input
Thanks Staszek I'll give a try to stopwords treatbment, but the problem is that we perform POS tagging and then use payloads to keep only Nouns and Adjectives, and we thought that could be interesting to perform clustering only with these elements, to avoid senseless words. Of course is a problem of clustering, but maybe is also a feature that could be interesting to have in solr: not to index the raw input text but the analyzed one, so stored could be False | Raw | analyzed Stanislaw Osinski-2 wrote: Hi Joan, I'm trying to use carrot2 (now I started with the workbench) and I can cluster any field, but, the text used for clustering is the original raw text, the one that was indexed, without any of the processing performed by the tokenizer or filters. So I get stop words. The easiest way to fix this is to update the stop words list used by Carrot2, see http://wiki.apache.org/solr/ClusteringComponent, Tuning Carrot2 clustering section at the bottom. If you want to get readable cluster labels, it's best to feed the raw text for clustering (cluster labels are phrases taken from the input text, if you remove stopwords and stem everything, the phrases will become unreadable). Cheers, Staszek -- View this message in context: http://old.nabble.com/Clustering-from-anlayzed-text-instead-of-raw-input-tp27765780p27769034.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Solr and UIMA
You can test our UIMA to Solr cas consumer is based on JulieLab Lucas and uses their CAS. but transformed to generate XML which can be saved to a file or posted direcly to solr In the map file you can define which information is generated for each token, and how its concatenaded, allowing the generation of thinks like the|AD car|NC which then can be processed using payloads. now you can get it from my page http://www.barcelonamedia.org/personal/joan.codina/en http://www.barcelonamedia.org/personal/joan.codina/en -- View this message in context: http://old.nabble.com/Solr-and-UIMA-tp24567504p27753399.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Solr and UIMA
Things are done :-) now we already have done the UIMA CAS consumer for Solr, we are making it public, more news soon. We have also been developing some filters based on payloads One of the filters is to remove words with the payloads in the list the other one maintains only these tokens with paylodas in the list. It works the same way than the stopsFilterFactory you can find it at my page: http://www.barcelonamedia.org/personal/joan.codina/en http://www.barcelonamedia.org/personal/joan.codina/en -- View this message in context: http://old.nabble.com/Solr-and-UIMA-tp24567504p27544646.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Solr and UIMA
On Jul 21, 2009, at 11:57 AM, JCodina wrote: Let me sintetize: We (well, I think Grant?) do changes in the DPTFF ( DelimitedPayloadTokenFilterFactory ) so that is able to index at the same position different tokes that may have payloads. 1. token delimiter (#) 2. payload delimiter (|) We (that's me) perform a SolCAS: a UIMA CAS consumer equivalent to LuCAS but that allows indexing using Solr. This SolCAS is able to manage generate different tokens at the same position and maybe payloads, the result is ready for the new DPTFF We (me again) develop some filtering utilities based on the payload that, something like the stopwords filter but instead of rejecting those tokens that are in the stopwords list would reject those that are in the payloads list. We will try also to develop an n-gram generator based on the payloads, like for example find the nouns followed by an adjective that are at less than 4 positions. For the moment searches can not be performed based on payloads, not even as a filter... but this is a matter of time. Problems to solve: Perform a nice processing of the N tokens that share the same position, as the tokenizer.Next() will not give them together (which is a pitty) .Write some utility tht would allow the tools that manage multitokens to have a similar front-end and back-end that does multiple Nexts in order to put toguether all the information at the same position, performs the treatment with a multitoken structure and then generates a multitoken that is sent to the backend that has the next again on single tokens... Joan -- View this message in context: http://www.nabble.com/Solr-and-UIMA-tp24567504p24639814.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Lemmatisation support in Solr
I think that to get the best results you need some kind of natural language processing I'm trying to do so using UIMA but i need to integrate it with SOLR as I explain in this post http://www.nabble.com/Solr-and-UIMA-tc24567504.html prerna07 wrote: Hi, I am implementing Lemmatisation in Solr, which means if user looks for Mouse then it should display results of Mouse and Mice both. I understand that this is something context search. I think of using synonym for this but then synonyms.txt will be having so many records and this will keep on adding. Please suggest how I can implement it in some other way. Thanks, Prerna -- View this message in context: http://www.nabble.com/Lemmatisation-support-in-Solr-tp24583655p24583841.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Solr and UIMA
Hello, Grant, there are two ways, to implement this, one is payloads, and the other one is multiple tokens at the same positions. Each of them can be useful, let me explain the way I thick they can be used. Payloads : every token has extra information that can be used in the processing , for example if I can add Part-of-speech then I can develop tokenizers that take into account the POS (or for example I can generate bigrams of Noum Adjective, or Noum prep Noum i can have a better stopwords algorithm) Multiple tokes in one position: If I can have different tokens at the same place, I can have different informations like: was #verb _be so I can do a search for you _be #adjective to find all the sentences that talk about you for example you were clever you are tall .. I have not understood the way that theDelimitedPayloadTokenFilterFactory may work in solr, which is the input format? so I was thinking in generating an xml where for each token a single string is generated like was#verb#be and then there is a tokenfilter that splits by # each white space separated string, in this case in three words and adds the trailing character that allows to search for the right semantic info. But gives them the same increment. Of course the full processing chain must be aware of this. But I must think on multiwords tokens Grant Ingersoll-6 wrote: On Jul 20, 2009, at 6:43 AM, JCodina wrote: D: Break things down. The CAS would only produce XML that solr can process. Then different Tokenizers can be used to deal with the data in the CAS. the main point is that the XML has the doc and field labels of solr. I just committed the DelimitedPayloadTokenFilterFactory, I suspect this is along the lines of what you are thinking, but I haven't done all that much with UIMA. I also suspect the Tee/Sink capabilities of Lucene could be helpful, but they aren't available in Solr yet. E: The set of capabilities to process the xml is defined in XML, similar to lucas to define the ouput and in the solr schema to define how this is processed. I want to use it in order to index something that is common but I can't get any tool to do that with sol: indexing a word and coding at the same position the syntactic and semantic information. I know that in Lucene this is evolving and it will be possible to include metadata but for the moment What does Lucas do with Lucene? Is it putting multiple tokens at the same position or using Payloads? -- Grant Ingersoll http://www.lucidimagination.com/ Search the Lucene ecosystem (Lucene/Solr/Nutch/Mahout/Tika/Droids) using Solr/Lucene: http://www.lucidimagination.com/search -- View this message in context: http://www.nabble.com/Solr-and-UIMA-tp24567504p24590509.html Sent from the Solr - User mailing list archive at Nabble.com.
Solr and UIMA
We are starting to use UIMA as a platform to analyze the text. The result of analyzing a document is a UIMA CAS. A Cas is a generic data structure that can contain different data. UIMA processes single documents, They get the documents from a CAS producer, process them using a PIPE that the user defines and finally sends the result to a CAS consumer, that saves or stores the result. The pipe is then a connection of different tools that annotate the text with different information. Different sets of tools are available out there, each of them deffining it's own data type's that are included in the CAS. To perform a PIPE output and input CAS of the elements to connect need to be compatible There is CAS consumer that feeds a LUCENE index, it's called LUCAS but I was looking to it, and I prefer to use UIMA connected to SOLR, why? A: I know solr ;-) and i like it B: I can configure the fields and their processing in solr using xml. Once done then I have it ready to use with a set of tools that allow me to easily explore the data C: Is easier to use SOLR as a web service that may receive docs from different UIMA's (Natural Language processing is CPU intensive ) D: Break things down. The CAS would only produce XML that solr can process. Then different Tokenizers can be used to deal with the data in the CAS. the main point is that the XML has a the doc and field labels of solr. E: The set of capabilities to process the xml is defined in XML, similar to lucas to define the ouput and in the solr schema to define how this is processed. I want to use it in order to index something that is common but I can't get any tool to do that with sol: indexing a word and coding at the same position the syntactic and semantic information. I know that in Lucene this is evolving and it will be possible to include metadata but for the moment So, my idea is first to produce a UIMA CAS consumer that performs the POST of an XML file containing the plain text text of the document to SOLR; then try to modify this in order to include multiple fields and start coding the semantic information. So, before starting, i would like to know your opinions and if anyone is interested to collaborate, or has some code that can be integrated into this. -- View this message in context: http://www.nabble.com/Solr-and-UIMA-tp24567504p24567504.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: facets and stopwords
hossman wrote: but are you sure that example would actually cause a problem? i suspect if you index thta exact sentence as is you wouldn't see the facet count for si or que increase at all. If you do a query for {!raw field=content}que you bypass the query parsers (which is respecting your stopwords file) and see all docs that contain the raw term que in the content field. if you look at some of the docs that match, and paste their content field into the analysis tool, i think you'll see that the problem comes from using the whitespace tokenizer, and is masked by using the WDF after the stop filter ... things like Que? are getting ignored by the stopfilter, but ultimately winding up in your index as que -Hoss Yes your are right, que? que, que... i need to change the analyzer. They are not detected by the stopwords analyzer because i use the whitespace tokenizer, I will use the StanadardTokenizer Thanks Hoss -- View this message in context: http://www.nabble.com/facets-and-stopwords-tp23952823p24390157.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: facets and stopwords
Sorry , I was too cryptic. I you follow this link http://projecte01.development.barcelonamedia.org/fonetic/ you will see a Top Words list (in Spanish and stemmed) in the list there is the word si which is in 20649 documents. If you click at this word, the system will perform the query (x) content:si, with no answers at all The same for la it is in 17881 documents, but the query content:la will give no answers at all the facets list is generated by the query http://projecte01.development.barcelonamedia.org/solr/select/?rows=0start=0q=*:*facet=truefacet.limit=-1facet.field=contentfacet.field=entities_miscwt=jsonjson.wrf=jsonp1246437157825jsoncallback=jsonp1246437157825_=1246437158023 but the question is why these two words (among others) are there if they are stop words? To see what's going on on the index I have tested with the analyzer http://projecte01.development.barcelonamedia.org/solr/admin/analysis.jsp If I select the field content and I write the text las cosas que si no pasan la proxima vez si que no veràs i get the following tokens at the end of the analyzer las cosapasan proxima vez sí verà where que, si, no, la are removed as treated as stop words. but... in the schema browser http://projecte01.development.barcelonamedia.org/solr/admin/schema.jsp in the field content que is the 3rd word no the 4th si and la are between the top 40 terms... the analyzer for the content can be seen in this page and has the following analyzers Tokenizer Class: org.apache.solr.analysis.WhitespaceTokenizerFactory Filters: 1. org.apache.solr.analysis.StopFilterFactory args:{enablePositionIncrements: true words: stopwords.txt ignoreCase: true } 2. org.apache.solr.analysis.WordDelimiterFilterFactory args:{catenateWords: 1 catenateNumbers: 1 splitOnCaseChange: 1 catenateAll: 0 generateNumberParts: 1 generateWordParts: 1 } 3. org.apache.solr.analysis.LowerCaseFilterFactory args:{} 4. org.apache.solr.analysis.SnowballPorterFilterFactory args:{languange: Spanish } 5. org.apache.solr.analysis.RemoveDuplicatesTokenFilterFactory args:{} The field is indexed, tokenized, stored and termvectors are stored. So, why the stopwords are in the index? -- View this message in context: http://www.nabble.com/facets-and-stopwords-tp23952823p24286283.html Sent from the Solr - User mailing list archive at Nabble.com.
Top tf_idf in TermVectorComponent
In order to perform any further study of the resultset, like clustering, the TermVectorComponent gives the list of words with the correspoing tf, idf, but this list can be huge for each document, and most of the terms may have a low tf or a too high df, maybe, it is usefull to compare the relative increment of DF to the collection in order to improve the facets (show only these terms that the relative DF in the query is higher than in the full collection) To perform this it could be interesting that the TermVectorComponent could sort the results by some of these options: *tf *DF * tf/df (to simplify) or tf*idf where idf is computed as log(total_docs/df) and truncate the list to a number of words or a given value or maybe there is another way to perform this? Joan -- View this message in context: http://www.nabble.com/Top-tf_idf-in-TermVectorComponent-tp24201076p24201076.html Sent from the Solr - User mailing list archive at Nabble.com.
version of lucene
I have the solr-nightly build of last week, and in the lib foloder i can find the lucene-core-2.9-dev.jar I need to do some changes to the shingle filter in order to remove stopwords from bigrams, but to do so I need to compile lucene, the problem is, lucene is in version 2.4 not 2.9 If I take, with subverison, version 2.4 then compiling solr I get the next error: .../apache-solr-nightly/src/java/org/apache/solr/search/DocSetHitCollector.java:21: cannot find symbol [javac] symbol : class Collector [javac] location: package org.apache.lucene.search [javac] import org.apache.lucene.search.Collector; any hints on the right version of lucene/solr to be able to use solr 1.4 Joan -- View this message in context: http://www.nabble.com/version-of-lucene-tp24036137p24036137.html Sent from the Solr - User mailing list archive at Nabble.com.
facets and stopwords
I have a text field from where I remove stop words, as a first approximation I use facets to see the most common words in the text, but.. stopwords are there, and if I search documents having the stopwords, then , there are no documents in the answer. You can test it in this address (using solrjs, the texts are in spanish but you can check in top words that que or en are there) but if you click on them to perform the search no results are given http://projecte01.development.barcelonamedia.org/fonetic/ or the administrator at http://projecte01.development.barcelonamedia.org/solr/admin so you can check wat's going on on the content field. I use the DataImportHandler to import the data, and Solr analyzer shows me how the stopwords are removed from both the query and the indexed text, but why facets show me these words? -- View this message in context: http://www.nabble.com/facets-and-stopwords-tp23952823p23952823.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Build Solr to run SolrJS
Yesterday I got it running, I thouhght I had post it, but I should not push the post button. The problem I had to solve to run it with the Velocity was to copy by hand the velocity jar files to /lib. Erik Hatcher wrote: I just got the client-side demo on trunk to work (with a few tweaks to make it work with the example core Solr data). On trunk follow these steps: * root directory: ant example * separate console, index data: cd example/exampledocs; java -jar post.jar *.xml * open contrib/javascript/example/testClientside.html in your browser The serverSide.html example is still not quite wired together properly to be an easy-to-run-demo, as what's on trunk doesn't include the reuters data and the velocity stuff is not wired in with it yet either. We'll get this working better/cleaner as we go, so we appreciate your early adopter help ironing out this stuff. Erik On Nov 20, 2008, at 5:44 PM, JCodina wrote: I could not manage, yet to use it. :confused: My doubts are: - must I download solr from svn - trunk? - then, must I apply the patches of solrjs and velocity and unzip the files? or is this already in trunk? because trunk contains velocity and javascript in contrib. but does not find the velocity - How do I edit/activate SolrJs to adapt it to my data, the wiki page says how to deploy the sample, and I looked at the sample page from the sample site, but I don't find how to manually install it on a tomcat server. PD. If I get how to do it, I promise I will introduce that information in the solrjs wiki page =) . Matthias Epheser wrote: Erik Hatcher schrieb: On Nov 16, 2008, at 1:40 PM, Matthias Epheser wrote: Matthias and Ryan - let's get SolrJS integrated into contrib/velocity. Any objections/reservations? As SolrJS may be used without velocity at all (using eg. ClientSideWidgets), is it possible to put it into contrib/ javascript and create a dependency to contrib/velocity for ServerSideWidgets? Sure, contrib/javascript sounds perfect. If that's ok, I'll have a look at the directory structure and the current ant build.xml to make them fit into the common solr structure and build. Awesome, thanks! Just uploaded solrjs.zip to https://issues.apache.org/jira/browse/SOLR-868. It is intended to be extracted in contrib/javascript and supports the following ant targets: * ant dist - creates a single js file and a jar that holds velocity templates. * ant docs - creates js docs. test in browser: doc/index.html * ant example-init - (depends ant dist on solr root) copies the current built of solr.war and solr-velocity.jar to example/testsolr/.. * ant example-start - starts the testsolr server on port 8983 * ant example-import - imports 3000 test data rows (requires a started testserver) Erik -- View this message in context: http://www.nabble.com/Build-Solr-to-run-SolrJS-tp20526644p20611635.html Sent from the Solr - User mailing list archive at Nabble.com. -- View this message in context: http://www.nabble.com/Build-Solr-to-run-SolrJS-tp20526644p20634725.html Sent from the Solr - User mailing list archive at Nabble.com.
DataImportHanler JDBC case problems
I tried to perform a DataImportHandler where the column name user and the field name User are the same but the case of the first letter, when performing a full import, I was getting different sorts of errors, on that field depending on the cases of the names, I tried the four possible combinations, and none worked. At the end, I changed column name in the database to be User and everthing was fine. dataConfig document entity name=item query=select * from DeNDocs field column=ID name=id / field column=User name=User / -- View this message in context: http://www.nabble.com/DataImportHanler-JDBC-case-problems-tp20617216p20617216.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Build Solr to run SolrJS
I could not manage, yet to use it. :confused: My doubts are: - must I download solr from svn - trunk? - then, must I apply the patches of solrjs and velocity and unzip the files? or is this already in trunk? because trunk contains velocity and javascript in contrib. but does not find the velocity - How do I edit/activate SolrJs to adapt it to my data, the wiki page says how to deploy the sample, and I looked at the sample page from the sample site, but I don't find how to manually install it on a tomcat server. PD. If I get how to do it, I promise I will introduce that information in the solrjs wiki page =) . Matthias Epheser wrote: Erik Hatcher schrieb: On Nov 16, 2008, at 1:40 PM, Matthias Epheser wrote: Matthias and Ryan - let's get SolrJS integrated into contrib/velocity. Any objections/reservations? As SolrJS may be used without velocity at all (using eg. ClientSideWidgets), is it possible to put it into contrib/javascript and create a dependency to contrib/velocity for ServerSideWidgets? Sure, contrib/javascript sounds perfect. If that's ok, I'll have a look at the directory structure and the current ant build.xml to make them fit into the common solr structure and build. Awesome, thanks! Just uploaded solrjs.zip to https://issues.apache.org/jira/browse/SOLR-868. It is intended to be extracted in contrib/javascript and supports the following ant targets: * ant dist - creates a single js file and a jar that holds velocity templates. * ant docs - creates js docs. test in browser: doc/index.html * ant example-init - (depends ant dist on solr root) copies the current built of solr.war and solr-velocity.jar to example/testsolr/.. * ant example-start - starts the testsolr server on port 8983 * ant example-import - imports 3000 test data rows (requires a started testserver) Erik -- View this message in context: http://www.nabble.com/Build-Solr-to-run-SolrJS-tp20526644p20611635.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Build Solr to run SolrJS
To give you more information. The error I get is this one: java.lang.NoClassDefFoundError: org/apache/solr/request/VelocityResponseWriter (wrong name: contrib/velocity/src/main/java/org/apache/solr/request/VelocityResponseWriter) at java.lang.ClassLoader.defineClass1(Native Method) at java.lang.ClassLoader.defineClass(ClassLoader.java:621) at java.security.SecureClassLoader.defineClass(SecureClassLoader.java:124) at org.apache.catalina.loader.WebappClassLoader.findClassInternal(WebappClassLoader.java:1847) at org.apache.catalina.loader.WebappClassLoader.findClass(WebappClassLoader.java:890) at org.apache.catalina.loader.WebappClassLoader.loadClass(WebappClassLoader.java:1354) at org.apache.catalina.loader.WebappClassLoader.loadClass(WebappClassLoader.java:1233) at java.lang.ClassLoader.loadClassInternal(ClassLoader.java:320) at java.lang.Class.forName0(Native Method) at java.lang.Class.forName(Class.java:247) at and in the build , if i do a build-contrib-dist i get these messages.. ... build: [jar] Building jar: /home/joan/workspace/solr/contrib/dataimporthandler/target/apache-solr-dataimporthandler-1.4-dev.jar dist: [copy] Copying 1 file to /home/joan/workspace/solr/build/web/WEB-INF/lib [copy] Copying 1 file to /home/joan/workspace/solr/dist init: init-forrest-entities: compile-common: compile: make-manifest: compile: [javac] Compiling 4 source files to /home/joan/workspace/solr/contrib/velocity/target/classes build: [jar] Building jar: /home/joan/workspace/solr/contrib/velocity/src/main/solr/lib/apache-solr-velocity-1.4-dev.jar dist: ... where the dataimporthanler seems that is copied in the dist folders but velocity is not. Hope this saves you some time.. Joan -- View this message in context: http://www.nabble.com/Build-Solr-to-run-SolrJS-tp20526644p20535777.html Sent from the Solr - User mailing list archive at Nabble.com.
Build Solr to run SolrJS
I downloaded solr/trunk and build it, everything seems to work except that the VelocityResponseWriter is not in the war file and tomcat , gives an error of configuration when using the conf.xml of the solrjs. Any suggestion on how to build the solr to work with solrjs?? Thanks Joan Codina -- View this message in context: http://www.nabble.com/Build-Solr-to-run-SolrJS-tp20526644p20526644.html Sent from the Solr - User mailing list archive at Nabble.com.