Need Features offered and comparison Chart for Solr 3.6 and Solr 4.6
Hi Team, I need a details around: 1. What are the features Apache Solr Offers. 2. Comparison Chart between Apache Solr 3.6 and Solr 4.6 Please provide as soon as possible so that I can suggest it in my company for implementation. Regards, Mayank Solanki
Re: using extract handler: data not extracted
I am working on Windows 7 -- View this message in context: http://lucene.472066.n3.nabble.com/using-extract-handler-data-not-extracted-tp4110850p4110993.html Sent from the Solr - User mailing list archive at Nabble.com.
background merge hit exception while optimizing index (SOLR 4.4.0)
Hi, I am currently running into merge-issues while optimizing an index. To give you some informations: We are running 4 SOLR Servers with identical OS, VM-Hardware, RAM etc. Only one Server by now is having issues, the others are fine. We are running SOLR 4.4.0 with Tomcat 6.0 It was running since October without any problems. The problems first occur after doing a minor change in the synonyms.txt, but I guess that was just a coincedence. We added `ulimit -v unlimited` to our tomcat init-script years ago. We have 4 Cores running on each SOLR Server, configuration, index-sizes of all 4 servers are identical (we are distributing cfgs via git). We did a rebuild of the index twice: First time without removing the old index files, second time deleting the data dir and starting from scratch. We are working with DIH, getting data from a MySQL DB. After an initial complete index-run, the optimize is working. The optimize fails one or two days later. We are doing one optimize-run a day, the index contains ~10 millions documents, the index size on disc is ~39GB while having 127G of free disc space. We have a mergeFactor of 3. The solr.log says: ERROR - 2014-01-12 22:47:11.062; org.apache.solr.common.SolrException; java.io.IOException: background merge hit exception: _dc8(4.4):C9876366/1327 _e8u(4.4):C4250/7 _f4a(4.4):C1553/13 _fj6(4.4 ):C1903/15 _ep3(4.4):C1217/42 _fle(4.4):C256/7 _flf(4.4):C11 into _flg [maxNumSegments=1] at org.apache.lucene.index.IndexWriter.forceMerge(IndexWriter.java:1714) at org.apache.lucene.index.IndexWriter.forceMerge(IndexWriter.java:1650) at org.apache.solr.update.DirectUpdateHandler2.commit(DirectUpdateHandler2.java:530) at org.apache.solr.update.processor.RunUpdateProcessor.processCommit(RunUpdateProcessorFactory.java:95) at org.apache.solr.update.processor.UpdateRequestProcessor.processCommit(UpdateRequestProcessor.java:64) at org.apache.solr.update.processor.DistributedUpdateProcessor.doLocalCommit(DistributedUpdateProcessor.java:1235) at org.apache.solr.update.processor.DistributedUpdateProcessor.processCommit(DistributedUpdateProcessor.java:1219) at org.apache.solr.update.processor.LogUpdateProcessor.processCommit(LogUpdateProcessorFactory.java:157) at org.apache.solr.handler.RequestHandlerUtils.handleCommit(RequestHandlerUtils.java:69) at org.apache.solr.handler.ContentStreamHandlerBase.handleRequestBody(ContentStreamHandlerBase.java:68) at org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:135) at org.apache.solr.core.SolrCore.execute(SolrCore.java:1904) at org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:659) at org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:362) at org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:158) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:235) at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:206) at org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:233) at org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:191) at org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:127) at org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:102) at org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:109) at org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:298) at org.apache.coyote.http11.Http11Processor.process(Http11Processor.java:857) at org.apache.coyote.http11.Http11Protocol$Http11ConnectionHandler.process(Http11Protocol.java:588) at org.apache.tomcat.util.net.JIoEndpoint$Worker.run(JIoEndpoint.java:489) at java.lang.Thread.run(Thread.java:735) Caused by: java.lang.NullPointerException at java.nio.ByteBuffer.get(ByteBuffer.java:661) at java.nio.DirectByteBuffer.get(DirectByteBuffer.java:245) at org.apache.lucene.store.ByteBufferIndexInput.readBytes(ByteBufferIndexInput.java:107) at org.apache.lucene.codecs.lucene41.ForUtil.readBlock(ForUtil.java:197) at org.apache.lucene.codecs.lucene41.Lucene41PostingsReader$BlockDocsAndPositionsEnum.refillDocs(Lucene41PostingsReader.java:748) at org.apache.lucene.codecs.lucene41.Lucene41PostingsReader$BlockDocsAndPositionsEnum.nextDoc(Lucene41PostingsReader.java:813) at org.apache.lucene.codecs.MappingMultiDocsAndPositionsEnum.nextDoc(MappingMultiDocsAndPositionsEnum.java:104) at org.apache.lucene.codecs.PostingsConsumer.merge(PostingsConsumer.java:101) at org.apache.lucene.codecs.TermsConsumer.merge(TermsConsumer.java:164) at org.apache.luc
Re: Correct to use to store urls (unicode)
On 13 January 2014 00:30, Hakim Benoudjit wrote: > > Yep sure. But is it good for me to store a link(http://...) in a solr > string field? knowing that this field isnt indexed, only stored. Yes, there is no issue. Not sure why they are not indexed, but if that is what you want, ... Regards, Gora
How Solr join query works?
Hi All, Can anyone please explain how solr join query works in solr4.2. we have 2 different documents.Both are in the same index. document1 contains the columns: docdate: 01-12-2012 previousmonthdate :01-11-2012 price:15 and some more fields. document2 contains: docdate :01-11-2012 previousmonthdate :01-10-2012 price:10 and some more fields. Here we have the same value in previousmonthdate (in document1) and docdate (in document2).So we want to make a join query based on this to retrieve these in a single document. the final document should look like this. docdate: 01-12-2012 previousmonthdate :01-11-2012 price:15 price:10(this is from document2) Is is possible using Solr join query??? Or do we have any other approach?. Please help.. Thanks. -- View this message in context: http://lucene.472066.n3.nabble.com/How-Solr-join-query-works-tp4110982.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: How to index data in muliValue field with key
thank you Mr. Steve. Now I understood and i figured out to separate field to title_th and title_en and it worked ;) -- View this message in context: http://lucene.472066.n3.nabble.com/How-to-index-data-in-muliValue-field-with-key-tp4110653p4110981.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: using extract handler: data not extracted
Not really sure...the issue seems related to text extraction so the first suspect is tika...SOLR is playing a secondary role here. If Tika is doing extraction good there should be an error, a warning on solr side (an exception, a content field too long warning or something like that) What about the option 3a above (finest org.apache + grep "tika")? On 12 Jan 2014 17:38, "sweety" wrote: > Sorry for the mistake. > im using solr 4.2, it has tika-1.3. > So now, java -jar tika-app-1.3.jar -v C:\Coding.pdf , parses pdf document > without error or msg. > Also, java -jar tika-app-1.3.jar -t C:\Coding.pdf, shows the entire > document. > Which means there is no problem in tika right?? > > > > -- > View this message in context: > http://lucene.472066.n3.nabble.com/using-extract-handler-data-not-extracted-tp4110850p4110957.html > Sent from the Solr - User mailing list archive at Nabble.com. >
Re: Correct to use to store urls (unicode)
Yep sure. But is it good for me to store a link(http://...) in a solr string field? knowing that this field isnt indexed, only stored. 2014/1/12 Gora Mohanty > On 12 January 2014 20:07, Hakim Benoudjit wrote: > > I have just forget the u'' next to a unicode string :\, > > Heh! Handling of Unicode in Python 2.x is annoying. 3.x > is better, but only a little. Off-topic on this list, so I will > shut up now. > > Regards, > Gora >
Re: using extract handler: data not extracted
Sorry for the mistake. im using solr 4.2, it has tika-1.3. So now, java -jar tika-app-1.3.jar -v C:\Coding.pdf , parses pdf document without error or msg. Also, java -jar tika-app-1.3.jar -t C:\Coding.pdf, shows the entire document. Which means there is no problem in tika right?? -- View this message in context: http://lucene.472066.n3.nabble.com/using-extract-handler-data-not-extracted-tp4110850p4110957.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: using extract handler: data not extracted
Sorry for the mistake. im using solr 4.2, it has tika-1.3. So now, java -jar tika-app-1.3.jar -v C:\Coding.pdf , parses pdf document without error or msg. Also, java -jar tika-app-1.3.jar -t C:\Coding.pdf, shows the entire document. Which means there is no problem in tika right?? -- View this message in context: http://lucene.472066.n3.nabble.com/using-extract-handler-data-not-extracted-tp4110850p4110954.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: using extract handler: data not extracted
Sorry for the mistake. im using solr 4.2, it has tika-1.3. So now, java -jar tika-app-1.3.jar -v C:Coding.pdf , parses pdf document without error or msg. Also, java -jar tika-app-1.4.jar* -t *C:Cloud.docx, shows the entire document. Which means there is no problem in tika right?? -- View this message in context: http://lucene.472066.n3.nabble.com/using-extract-handler-data-not-extracted-tp4110850p4110951.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Correct to use to store urls (unicode)
On 12 January 2014 20:07, Hakim Benoudjit wrote: > I have just forget the u'' next to a unicode string :\, Heh! Handling of Unicode in Python 2.x is annoying. 3.x is better, but only a little. Off-topic on this list, so I will shut up now. Regards, Gora
Re: Correct to use to store urls (unicode)
I have just forget the u'' next to a unicode string :\, 2014/1/12 Hakim Benoudjit > I can add this link using sunburnt (solr python client) so it can not be > related to solr. > I think you're right it might be a python issue. > > Thanks. > > > 2014/1/12 Gora Mohanty > >> On 12 January 2014 19:45, Hakim Benoudjit wrote: >> > hi, >> > >> > what's the correct type used to store urls, which can contain some >> > encoded unicode caracters in the form '%'. Because, the >> > string type returns an error when I try to store these urls. >> >> Please provide more details as that should not be the case. >> >> > Btw, I'm using a python client which gives me this error: "'ascii' >> > codec can't decode byte 0xc3". >> >> That is a different issue having to do with Python, and the proper >> handling of Unicode strings. Try searching Google, or asking on a >> Python list. >> >> Regards, >> Gora >> > >
Re: Correct to use to store urls (unicode)
I can add this link using sunburnt (solr python client) so it can not be related to solr. I think you're right it might be a python issue. Thanks. 2014/1/12 Gora Mohanty > On 12 January 2014 19:45, Hakim Benoudjit wrote: > > hi, > > > > what's the correct type used to store urls, which can contain some > > encoded unicode caracters in the form '%'. Because, the > > string type returns an error when I try to store these urls. > > Please provide more details as that should not be the case. > > > Btw, I'm using a python client which gives me this error: "'ascii' > > codec can't decode byte 0xc3". > > That is a different issue having to do with Python, and the proper > handling of Unicode strings. Try searching Google, or asking on a > Python list. > > Regards, > Gora >
Re: Correct to use to store urls (unicode)
On 12 January 2014 19:45, Hakim Benoudjit wrote: > hi, > > what's the correct type used to store urls, which can contain some > encoded unicode caracters in the form '%'. Because, the > string type returns an error when I try to store these urls. Please provide more details as that should not be the case. > Btw, I'm using a python client which gives me this error: "'ascii' > codec can't decode byte 0xc3". That is a different issue having to do with Python, and the proper handling of Unicode strings. Try searching Google, or asking on a Python list. Regards, Gora
Correct to use to store urls (unicode)
hi, what's the correct type used to store urls, which can contain some encoded unicode caracters in the form '%'. Because, the string type returns an error when I try to store these urls. Btw, I'm using a python client which gives me this error: "'ascii' codec can't decode byte 0xc3".
Re: using extract handler: data not extracted
Please stay on (or clarify) your issue: in the first example you told us the problem is with "Coding.pdf" file. What is that Cloud.docx? Why don't you try with Coding.pdf? And what is the result of the extraction from command line with Coding.pdf and the same tika version that is in your SOLR? I would use the tika version shipped with SOLR. Changing that version is obviously possible, don't know what Tika version are compatible with your SOLR (because dependencies and last but not least: you didn't tell us which SOLR version you are using) On Sun, Jan 12, 2014 at 1:39 PM, sweety wrote: > through command line(>java -jar tika-app-1.4.jar -v C:Cloud.docx) apache > tika is able to parse .docx files, so can i use this tika-app-1.4.jar in > solr?? how to do that?? > > > > -- > View this message in context: > http://lucene.472066.n3.nabble.com/using-extract-handler-data-not-extracted-tp4110850p4110938.html > Sent from the Solr - User mailing list archive at Nabble.com. >
Re: using extract handler: data not extracted
through command line(>java -jar tika-app-1.4.jar -v C:Cloud.docx) apache tika is able to parse .docx files, so can i use this tika-app-1.4.jar in solr?? how to do that?? -- View this message in context: http://lucene.472066.n3.nabble.com/using-extract-handler-data-not-extracted-tp4110850p4110938.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: using extract handler: data not extracted
A premise: as Erik explained, most probably this issue has nothing to do with SOLR. So, these are the options that, in my mind, you have *OPTION #1 : Using Tika as command line tool*a) Download Tika. Make sure the same version of your SOLR b) Read here: http://tika.apache.org/1.4/gettingstarted.html. There a section called "Using Tika as a command line utility", you can give it your file, set the verbose flag and see what is the output *OPTIONS #2 (if you are a Java dev)*a) create a simple Java project in your workspace and set in the build path the tika libs from your SOLR bundle. b) Read here: http://tika.apache.org/1.4/parser_guide.html. Creates and start a sample main parser. Here you should have a more deep control on what happens. *OPTIONS #3 Set the log level on SOLR* As far as I rememberr the old (pre 4.x) version of solr listed all packages found in classloader under the logging tab, so you could be able to set the appropiate level for each of them. Instead, I'm seeing that on 4.x (at least 4.4.0), after starting SOLR with tika libs, I don't see those packages in the log tree. You can a) if you are using linux, set the FINEST level on the *org.apache* package and grep the output log (otherwise you will get a lt of messages) b) directly change the logging.properties under $JETTY_HOME/etc (or your servlet engine log configuration files if you are not using jetty) More on (SOLR) logging http://wiki.apache.org/solr/SolrLogging http://wiki.apache.org/solr/LoggingInDefaultJettySetup Best, Andrea On Sun, Jan 12, 2014 at 10:13 AM, sweety wrote: > ya right all 3 points are right. > Let me solve the 1 first, there is some errror in tika level indexing, for > that i need to debug at tika level right?? > but how to do that?? Solr admin does not show package wise logging. > > > > -- > View this message in context: > http://lucene.472066.n3.nabble.com/using-extract-handler-data-not-extracted-tp4110850p4110922.html > Sent from the Solr - User mailing list archive at Nabble.com. >
Re: Problem with starting solr
Hello! If you want to use Morfologik you need to include it jar files, because they are in included in the standard distribution. Look at the downloaded Solr distribution, you should find the contrib/analysis-extras/lib directory there. Get those libraries, put it in a directory (ie: /var/solr/ext) add a section like this to your solrconfig.xml (you can find it in conf directory of Solr): In addition to that you'll need the lucene-analyzers-morfologik-4.6.0.jar from contrib/analysis-extras/lucene-lib. Just put it the same directory. After that restart Solr and you should be OK. Btw - you can find the instructions (although a bit old) at: http://solr.pl/2012/04/02/solr-4-0-i-mozliwosci-analizy-jezyka-polskiego/ -- Regards, Rafał Kuć Performance Monitoring * Log Analytics * Search Analytics Solr & Elasticsearch Support * http://sematext.com/ > Hi. > I'm total beginner to Solr. > When I'm trying to start it I get following errors: > lukasz@lukasz-VirtualBox:~/PycharmProjects/solr/solr-4.6.0/example$ java > -jar start.jar > 0[main] INFO org.eclipse.jetty.server.Server – jetty-8.1.10.v20130312 > 104 [main] INFO > org.eclipse.jetty.deploy.providers.ScanningAppProvider – > Deployment monitor > /home/lukasz/PycharmProjects/solr/solr-4.6.0/example/contexts at interval 0 > 214 [main] INFO org.eclipse.jetty.deploy.DeploymentManager – Deployable > added: > /home/lukasz/PycharmProjects/solr/solr-4.6.0/example/contexts/solr-jetty-context.xml > 9566 [main] INFO > org.eclipse.jetty.webapp.StandardDescriptorProcessor – NO > JSP Support for /solr, did not find > org.apache.jasper.servlet.JspServlet > 9746 [main] INFO org.apache.solr.servlet.SolrDispatchFilter – > SolrDispatchFilter.init() > 9805 [main] INFO org.apache.solr.core.SolrResourceLoader – JNDI not > configured for solr (NoInitialContextEx) > 9808 [main] INFO org.apache.solr.core.SolrResourceLoader – solr home > defaulted to 'solr/' (could not find system property or JNDI) > 9809 [main] INFO org.apache.solr.core.SolrResourceLoader – new > SolrResourceLoader for directory: 'solr/' > 10116 [main] INFO org.apache.solr.core.ConfigSolr – Loading container > configuration from > /home/lukasz/PycharmProjects/solr/solr-4.6.0/example/solr/solr.xml > 10406 [main] INFO org.apache.solr.core.CoreContainer – New CoreContainer > 30060911 > 10407 [main] INFO org.apache.solr.core.CoreContainer – Loading cores into > CoreContainer [instanceDir=solr/] > 10479 [main] INFO > org.apache.solr.handler.component.HttpShardHandlerFactory > – Setting socketTimeout to: 0 > 10479 [main] INFO > org.apache.solr.handler.component.HttpShardHandlerFactory > – Setting urlScheme to: http:// > 10480 [main] INFO > org.apache.solr.handler.component.HttpShardHandlerFactory > – Setting connTimeout to: 0 > 10481 [main] INFO > org.apache.solr.handler.component.HttpShardHandlerFactory > – Setting maxConnectionsPerHost to: 20 > 10483 [main] INFO > org.apache.solr.handler.component.HttpShardHandlerFactory > – Setting corePoolSize to: 0 > 10485 [main] INFO > org.apache.solr.handler.component.HttpShardHandlerFactory > – Setting maximumPoolSize to: 2147483647 > 10486 [main] INFO > org.apache.solr.handler.component.HttpShardHandlerFactory > – Setting maxThreadIdleTime to: 5 > 10487 [main] INFO > org.apache.solr.handler.component.HttpShardHandlerFactory > – Setting sizeOfQueue to: -1 > 10488 [main] INFO > org.apache.solr.handler.component.HttpShardHandlerFactory > – Setting fairnessPolicy to: false > 10702 [main] INFO org.apache.solr.logging.LogWatcher – SLF4J impl is > org.slf4j.impl.Log4jLoggerFactory > 10704 [main] INFO org.apache.solr.logging.LogWatcher – Registering Log > Listener [Log4j (org.slf4j.impl.Log4jLoggerFactory)] > 10839 [coreLoadExecutor-3-thread-1] INFO > org.apache.solr.core.CoreContainer > – Creating SolrCore 'content' using instanceDir: solr/content > 10839 [coreLoadExecutor-3-thread-1] INFO > org.apache.solr.core.SolrResourceLoader – new SolrResourceLoader for > directory: 'solr/content/' > 10850 [coreLoadExecutor-3-thread-1] INFO > org.apache.solr.core.SolrResourceLoader – Adding > 'file:/home/lukasz/PycharmProjects/solr/solr-4.6.0/example/solr/content/lib/lucene-analyzers-morfologik-4.6.0.jar' > to classloader > 10855 [coreLoadExecutor-3-thread-1] INFO > org.apache.solr.core.SolrResourceLoader – Adding > 'file:/home/lukasz/PycharmProjects/solr/solr-4.6.0/example/solr/content/lib/morfologik-fsa-1.8.1.jar' > to classloader > 10857 [coreLoadExecutor-3-thread-1] INFO > org.apache.solr.core.SolrResourceLoader – Adding > 'file:/home/lukasz/PycharmProjects/solr/solr-4.6.0/example/solr/content/lib/morfologik-polish-1.8.1.jar' > to classloader > 10859 [coreLoadExecutor-3-thread-1] INFO > org.apache.solr.core.SolrResourceLoader – Adding > 'file:/home/lukasz/PycharmProjects/solr/solr-4.6.0/example/solr/content/lib/morfologik-stemming-1.8.1.jar' > to classloader > 11005 [coreLoadExecutor-3-thread-1] INFO > org.apache.solr.core.S
Problem with starting solr
Hi. I'm total beginner to Solr. When I'm trying to start it I get following errors: lukasz@lukasz-VirtualBox:~/PycharmProjects/solr/solr-4.6.0/example$ java -jar start.jar 0[main] INFO org.eclipse.jetty.server.Server – jetty-8.1.10.v20130312 104 [main] INFO org.eclipse.jetty.deploy.providers.ScanningAppProvider – Deployment monitor /home/lukasz/PycharmProjects/solr/solr-4.6.0/example/contexts at interval 0 214 [main] INFO org.eclipse.jetty.deploy.DeploymentManager – Deployable added: /home/lukasz/PycharmProjects/solr/solr-4.6.0/example/contexts/solr-jetty-context.xml 9566 [main] INFO org.eclipse.jetty.webapp.StandardDescriptorProcessor – NO JSP Support for /solr, did not find org.apache.jasper.servlet.JspServlet 9746 [main] INFO org.apache.solr.servlet.SolrDispatchFilter – SolrDispatchFilter.init() 9805 [main] INFO org.apache.solr.core.SolrResourceLoader – JNDI not configured for solr (NoInitialContextEx) 9808 [main] INFO org.apache.solr.core.SolrResourceLoader – solr home defaulted to 'solr/' (could not find system property or JNDI) 9809 [main] INFO org.apache.solr.core.SolrResourceLoader – new SolrResourceLoader for directory: 'solr/' 10116 [main] INFO org.apache.solr.core.ConfigSolr – Loading container configuration from /home/lukasz/PycharmProjects/solr/solr-4.6.0/example/solr/solr.xml 10406 [main] INFO org.apache.solr.core.CoreContainer – New CoreContainer 30060911 10407 [main] INFO org.apache.solr.core.CoreContainer – Loading cores into CoreContainer [instanceDir=solr/] 10479 [main] INFO org.apache.solr.handler.component.HttpShardHandlerFactory – Setting socketTimeout to: 0 10479 [main] INFO org.apache.solr.handler.component.HttpShardHandlerFactory – Setting urlScheme to: http:// 10480 [main] INFO org.apache.solr.handler.component.HttpShardHandlerFactory – Setting connTimeout to: 0 10481 [main] INFO org.apache.solr.handler.component.HttpShardHandlerFactory – Setting maxConnectionsPerHost to: 20 10483 [main] INFO org.apache.solr.handler.component.HttpShardHandlerFactory – Setting corePoolSize to: 0 10485 [main] INFO org.apache.solr.handler.component.HttpShardHandlerFactory – Setting maximumPoolSize to: 2147483647 10486 [main] INFO org.apache.solr.handler.component.HttpShardHandlerFactory – Setting maxThreadIdleTime to: 5 10487 [main] INFO org.apache.solr.handler.component.HttpShardHandlerFactory – Setting sizeOfQueue to: -1 10488 [main] INFO org.apache.solr.handler.component.HttpShardHandlerFactory – Setting fairnessPolicy to: false 10702 [main] INFO org.apache.solr.logging.LogWatcher – SLF4J impl is org.slf4j.impl.Log4jLoggerFactory 10704 [main] INFO org.apache.solr.logging.LogWatcher – Registering Log Listener [Log4j (org.slf4j.impl.Log4jLoggerFactory)] 10839 [coreLoadExecutor-3-thread-1] INFO org.apache.solr.core.CoreContainer – Creating SolrCore 'content' using instanceDir: solr/content 10839 [coreLoadExecutor-3-thread-1] INFO org.apache.solr.core.SolrResourceLoader – new SolrResourceLoader for directory: 'solr/content/' 10850 [coreLoadExecutor-3-thread-1] INFO org.apache.solr.core.SolrResourceLoader – Adding 'file:/home/lukasz/PycharmProjects/solr/solr-4.6.0/example/solr/content/lib/lucene-analyzers-morfologik-4.6.0.jar' to classloader 10855 [coreLoadExecutor-3-thread-1] INFO org.apache.solr.core.SolrResourceLoader – Adding 'file:/home/lukasz/PycharmProjects/solr/solr-4.6.0/example/solr/content/lib/morfologik-fsa-1.8.1.jar' to classloader 10857 [coreLoadExecutor-3-thread-1] INFO org.apache.solr.core.SolrResourceLoader – Adding 'file:/home/lukasz/PycharmProjects/solr/solr-4.6.0/example/solr/content/lib/morfologik-polish-1.8.1.jar' to classloader 10859 [coreLoadExecutor-3-thread-1] INFO org.apache.solr.core.SolrResourceLoader – Adding 'file:/home/lukasz/PycharmProjects/solr/solr-4.6.0/example/solr/content/lib/morfologik-stemming-1.8.1.jar' to classloader 11005 [coreLoadExecutor-3-thread-1] INFO org.apache.solr.core.SolrConfig – Adding specified lib dirs to ClassLoader 11022 [coreLoadExecutor-3-thread-1] INFO org.apache.solr.core.SolrResourceLoader – Adding 'file:/home/lukasz/PycharmProjects/solr/solr-4.6.0/contrib/extraction/lib/apache-mime4j-core-0.7.2.jar' to classloader 11025 [coreLoadExecutor-3-thread-1] INFO org.apache.solr.core.SolrResourceLoader – Adding 'file:/home/lukasz/PycharmProjects/solr/solr-4.6.0/contrib/extraction/lib/apache-mime4j-dom-0.7.2.jar' to classloader 11027 [coreLoadExecutor-3-thread-1] INFO org.apache.solr.core.SolrResourceLoader – Adding 'file:/home/lukasz/PycharmProjects/solr/solr-4.6.0/contrib/extraction/lib/bcmail-jdk15-1.45.jar' to classloader 11030 [coreLoadExecutor-3-thread-1] INFO org.apache.solr.core.SolrResourceLoader – Adding 'file:/home/lukasz/PycharmProjects/solr/solr-4.6.0/contrib/extraction/lib/bcprov-jdk15-1.45.jar' to classloader 11044 [coreLoadExecutor-3-thread-1] INFO org.apache.solr.core.SolrResourceLoader – Adding 'file:/home/lukasz/PycharmProjects/solr/solr-4.6.0/contrib/ext
Re: using extract handler: data not extracted
ya right all 3 points are right. Let me solve the 1 first, there is some errror in tika level indexing, for that i need to debug at tika level right?? but how to do that?? Solr admin does not show package wise logging. -- View this message in context: http://lucene.472066.n3.nabble.com/using-extract-handler-data-not-extracted-tp4110850p4110922.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: using extract handler: data not extracted
Wait, don't confuse things...they should be three different issues: 1. with curl indexing happens but leaves the content field empty, so probably something occurs at tika level during the text extraction. That's the reason why I told you about the tika logging 2. with solrj ineexing doesn'happen because a syntax error. Probably there some usage mistake 3. I don't know solrnet but linkage error is something at low level (related to libraries and classes) so I adsume also here indexing doesn't happen. If you don't know about those "internalities" I would suggest to concentrate on 1) and 2). Best, Andres On 12 Jan 2014 08:21, "sweety" wrote: > this is the output i get when indexed through* solrj*, i followed the link > you suggested. > i tried indexing .doc file. > > > 400 > 17 > > > > org.apache.solr.search.SyntaxError: Cannot parse > 'id:C:\solr\document\src\new_index_doc\document_1.doc': Encountered " ":" > ": > "" at line 1, column 4. Was expecting one of: ... ... > > ... "+" ... "-" ... ... "(" ... "*" ... "^" ... ... > ... ... ... ... > ... > "[" ... "{" ... ... ... > > 400 > > > > Also when indexed with *solrnet*, i get this error: > Caused by: java.lang.LinkageError: loader constraint violation: loader > (instance of org/apache/catalina/loader/WebappClassLoader) previously > initiated loading for a different type with name > "org/apache/xmlbeans/XmlCursor" > why this linkage error?? > > Now *curl does not work, neither does solrj and solrnet.* > > > > -- > View this message in context: > http://lucene.472066.n3.nabble.com/using-extract-handler-data-not-extracted-tp4110850p4110915.html > Sent from the Solr - User mailing list archive at Nabble.com. >