Streaming Expressions (/stream) StreamHandler java.lang.NullPointerException
I've got an exception below running curl --data-urlencode 'expr=search(EventsAndDCF,q="*:*",fl="AccessPath",sort="AccessPath asc",qt="/export")' "http://localhost:8983/solr/EventsAndDCF/stream"; Solr responce: {"result-set":{"docs":[ {"EXCEPTION":null,"EOF":true}]}} My collection EventsAndDCF exists. and I succeed to run GET queries like: http://localhost:8983/solr/EventsAndDCF/export?fl=AccessPath&q=*:*&sort=AccessPath desc&wt=json Solr version: 6.0.1. Single node 2016-06-25 21:15:44.147 ERROR (qtp1514322932-16) [ x:EventsAndDCF] o.a.s.h.StreamHandler java.lang.NullPointerException at org.apache.solr.client.solrj.io.stream.expr.StreamExpressionParser.generateStreamExpression(StreamExpressionParser.java:46) at org.apache.solr.client.solrj.io.stream.expr.StreamExpressionParser.parse(StreamExpressionParser.java:37) at org.apache.solr.client.solrj.io.stream.expr.StreamFactory.constructStream(StreamFactory.java:178) at org.apache.solr.handler.StreamHandler.handleRequestBody(StreamHandler.java:164) at org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:155) at org.apache.solr.core.SolrCore.execute(SolrCore.java:2053) at org.apache.solr.servlet.HttpSolrCall.execute(HttpSolrCall.java:652) at org.apache.solr.servlet.HttpSolrCall.call(HttpSolrCall.java:460) at org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:229) at org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:184) at org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1668) at org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:581) at org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:143) at org.eclipse.jetty.security.SecurityHandler.handle(SecurityHandler.java:548) at org.eclipse.jetty.server.session.SessionHandler.doHandle(SessionHandler.java:226) at org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:1160) at org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:511) at org.eclipse.jetty.server.session.SessionHandler.doScope(SessionHandler.java:185) at org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:1092) at org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:141) at org.eclipse.jetty.server.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:213) at org.eclipse.jetty.server.handler.HandlerCollection.handle(HandlerCollection.java:119) at org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:134) at org.eclipse.jetty.server.Server.handle(Server.java:518) at org.eclipse.jetty.server.HttpChannel.handle(HttpChannel.java:308) at org.eclipse.jetty.server.HttpConnection.onFillable(HttpConnection.java:244) at org.eclipse.jetty.io.AbstractConnection$ReadCallback.succeeded(AbstractConnection.java:273) at org.eclipse.jetty.io.FillInterest.fillable(FillInterest.java:95) at org.eclipse.jetty.io.SelectChannelEndPoint$2.run(SelectChannelEndPoint.java:93) at org.eclipse.jetty.util.thread.strategy.ExecuteProduceConsume.produceAndRun(ExecuteProduceConsume.java:246) at org.eclipse.jetty.util.thread.strategy.ExecuteProduceConsume.run(ExecuteProduceConsume.java:156) at org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:654) at org.eclipse.jetty.util.thread.QueuedThreadPool$3.run(QueuedThreadPool.java:572) at java.lang.Thread.run(Unknown Source) 2016-06-25 21:15:44.147 INFO (qtp1514322932-16) [ x:EventsAndDCF] o.a.s.c.S.Request [EventsAndDCF] webapp=/solr path=/stream params={'expr=search(EventsAndDCF,q%3D*:*,fl%3DAccessPath,sort%3DAccessPath+asc,qt%3D/export)'} status=0 QTime=2
limit stored field size
Hi I have a field called content which I'm indexing and use for highlighting, which means it has to be stored as well. But this field may be too big, so I want to limit the stored size to X characters (it is fine to highlight only the first X characters). One solution is to create another field called content_snippet which will be a copy field of content field by maxchars of X (1 in my example), set content as non-stored and set content_snippet as stored and indexed. content_snippet must be indexed in order to highlight it. So as a result I have two indexed fields, which is redundant. My goal is to decrease index size. Is there a way to limit the stored size within one field without creating copy field? -- View this message in context: http://lucene.472066.n3.nabble.com/limit-stored-field-size-tp4284356.html Sent from the Solr - User mailing list archive at Nabble.com.
SolrCloud trying to upload documents and shards do not have storage anymore
Hi, I am currently using JSON Index Handler to upload documents to a specific collection on SolrCloud. Now what I need to know is: If I upload documents to SolrCloud collection and the machines hosting Shards for this collection have no storage left, will Solr reject the commit request? ?
Re: integrate SOLR with OSM
hi william, thx for your reply. is it any link article or documents that i can read for this? pls let me know if you've any suggestion. On 26/06/16 02:58, William Bell wrote: It depends on if your are using TIGERLINE data in OSM to increase the accuracy of home addresses. If you are using TIGERLINE, you are talking about > 500M rows or so + millions of OSM data. Which will is a LOT of data that you will be pumping into SOLR. You will need to use Solr Cloud since you will hit limits... With that amount of data I had issues with just USA OSM data with performance. If you get it working, please share. On Sat, Jun 25, 2016 at 12:49 PM, tkg_cangkul wrote: hi i wanna try to integrate SOLR with OpenStreetMap (OSM). well the plan is i wanna index some cordinaate (long & lat) to SOLR and then the OSM will try to showing the map of that coordinate. is there any article about that? pls help. i'm still confuse about this. thx before.
Could not load collection for SolrCloud
Hello, I am using solr 6.0.0 in SolCloud mode with 3 nodes, one zookeeper and 3 shard and 2 replica per collection. Getting below error for some insert/update when trying to insert documents to Solr. And it has been observed that few shard are in either recovery or fail recovery state. (Atleast one shard is up) org.apache.solr.common.SolrException: Could not load collection from ZK: MY_COLLECTION at org.apache.solr.common.cloud.ZkStateReader.getCollectionLive(ZkStateReader.java:969) ~[solr-solrj-6.0.0.jar:6.0.0 48c80f91b8e5cd9b3a9b48e6184bd53e7619e7e3 - nknize - 2016-04-01 14:41:50] at org.apache.solr.common.cloud.ZkStateReader$LazyCollectionRef.get(ZkStateReader.java:519) ~[solr-solrj-6.0.0.jar:6.0.0 48c80f91b8e5cd9b3a9b48e6184bd53e7619e7e3 - nknize - 2016-04-01 14:41:50] at org.apache.solr.common.cloud.ClusterState.getCollectionOrNull(ClusterState.java:189) ~[solr-solrj-6.0.0.jar:6.0.0 48c80f91b8e5cd9b3a9b48e6184bd53e7619e7e3 - nknize - 2016-04-01 14:41:50] at org.apache.solr.common.cloud.ClusterState.hasCollection(ClusterState.java:119) ~[solr-solrj-6.0.0.jar:6.0.0 48c80f91b8e5cd9b3a9b48e6184bd53e7619e7e3 - nknize - 2016-04-01 14:41:50] at org.apache.solr.client.solrj.impl.CloudSolrClient.getCollectionNames(CloudSolrClient.java:) ~[solr-solrj-6.0.0.jar:6.0.0 48c80f91b8e5cd9b3a9b48e6184bd53e7619e7e3 - nknize - 2016-04-01 14:41:50] at org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:833) ~[solr-solrj-6.0.0.jar:6.0.0 48c80f91b8e5cd9b3a9b48e6184bd53e7619e7e3 - nknize - 2016-04-01 14:41:50] at org.apache.solr.client.solrj.impl.CloudSolrClient.request(CloudSolrClient.java:806) ~[solr-solrj-6.0.0.jar:6.0.0 48c80f91b8e5cd9b3a9b48e6184bd53e7619e7e3 - nknize - 2016-04-01 14:41:50] at org.apache.solr.client.solrj.SolrRequest.process(SolrRequest.java:149) ~[solr-solrj-6.0.0.jar:6.0.0 48c80f91b8e5cd9b3a9b48e6184bd53e7619e7e3 - nknize - 2016-04-01 14:41:50] at org.apache.solr.client.solrj.SolrClient.add(SolrClient.java:106) ~[solr-solrj-6.0.0.jar:6.0.0 48c80f91b8e5cd9b3a9b48e6184bd53e7619e7e3 - nknize - 2016-04-01 14:41:50] at org.apache.solr.client.solrj.SolrClient.add(SolrClient.java:71) ~[solr-solrj-6.0.0.jar:6.0.0 48c80f91b8e5cd9b3a9b48e6184bd53e7619e7e3 - nknize - 2016-04-01 14:41:50] at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) [?:1.8.0_60] at java.util.concurrent.FutureTask.run(FutureTask.java:266) [?:1.8.0_60] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) [?:1.8.0_60] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) [?:1.8.0_60] at java.lang.Thread.run(Thread.java:745) [?:1.8.0_60] Caused by: org.apache.zookeeper.KeeperException$SessionExpiredException: KeeperErrorCode = Session expired for /collections/ MY_COLLECTION /state.json at org.apache.zookeeper.KeeperException.create(KeeperException.java:127) ~[zookeeper-3.4.6.jar:3.4.6-1569965] at org.apache.zookeeper.KeeperException.create(KeeperException.java:51) ~[zookeeper-3.4.6.jar:3.4.6-1569965] at org.apache.zookeeper.ZooKeeper.getData(ZooKeeper.java:1155) ~[zookeeper-3.4.6.jar:3.4.6-1569965] at org.apache.solr.common.cloud.SolrZkClient$7.execute(SolrZkClient.java:348) ~[solr-solrj-6.0.0.jar:6.0.0 48c80f91b8e5cd9b3a9b48e6184bd53e7619e7e3 - nknize - 2016-04-01 14:41:50] at org.apache.solr.common.cloud.SolrZkClient$7.execute(SolrZkClient.java:345) ~[solr-solrj-6.0.0.jar:6.0.0 48c80f91b8e5cd9b3a9b48e6184bd53e7619e7e3 - nknize - 2016-04-01 14:41:50] at org.apache.solr.common.cloud.ZkCmdExecutor.retryOperation(ZkCmdExecutor.java:60) ~[solr-solrj-6.0.0.jar:6.0.0 48c80f91b8e5cd9b3a9b48e6184bd53e7619e7e3 - nknize - 2016-04-01 14:41:50] at org.apache.solr.common.cloud.SolrZkClient.getData(SolrZkClient.java:345) ~[solr-solrj-6.0.0.jar:6.0.0 48c80f91b8e5cd9b3a9b48e6184bd53e7619e7e3 - nknize - 2016-04-01 14:41:50] at org.apache.solr.common.cloud.ZkStateReader.fetchCollectionState(ZkStateReader.java:980) ~[solr-solrj-6.0.0.jar:6.0.0 48c80f91b8e5cd9b3a9b48e6184bd53e7619e7e3 - nknize - 2016-04-01 14:41:50] at org.apache.solr.common.cloud.ZkStateReader.getCollectionLive(ZkStateReader.java:967) ~[solr-solrj-6.0.0.jar:6.0.0 48c80f91b8e5cd9b3a9b48e6184bd53e7619e7e3 - nknize - 2016-04-01 14:41:50] ... 16 more Regards, Roshan The information in this email is confidential and may be legally privileged. It is intended solely for the addressee. Access to this email by anyone else is unauthorised. If you are not the intended recipient, any disclosure, copying, distribution or any action taken or omitted to be taken in reliance on it, is prohibited and may be unlawful.
[ANNOUNCE] Apache Solr 5.5.2 released
25 June 2016, Apache Solr™ 5.5.2 available The Lucene PMC is pleased to announce the release of Apache Solr 5.5.2 Solr is the popular, blazing fast, open source NoSQL search platform from the Apache Lucene project. Its major features include powerful full-text search, hit highlighting, faceted search, dynamic clustering, database integration, rich document (e.g., Word, PDF) handling, and geospatial search. Solr is highly scalable, providing fault tolerant distributed search and indexing, and powers the search and navigation features of many of the world's largest internet sites. This release includes 38 bug fixes, documentation updates, etc., since the 5.5.1 release. The release is available for immediate download at: http://www.apache.org/dyn/closer.lua/lucene/solr/5.5.2 Please read CHANGES.txt for a detailed list of changes: https://lucene.apache.org/solr/5_5_2/changes/Changes.html Please report any feedback to the mailing lists (http://lucene.apache.org/solr/discussion.html) Note: The Apache Software Foundation uses an extensive mirroring network for distributing releases. It is possible that the mirror you are using may not have replicated the release yet. If that is the case, please try another mirror. This also goes for Maven access.
Re: integrate SOLR with OSM
It depends on if your are using TIGERLINE data in OSM to increase the accuracy of home addresses. If you are using TIGERLINE, you are talking about > 500M rows or so + millions of OSM data. Which will is a LOT of data that you will be pumping into SOLR. You will need to use Solr Cloud since you will hit limits... With that amount of data I had issues with just USA OSM data with performance. If you get it working, please share. On Sat, Jun 25, 2016 at 12:49 PM, tkg_cangkul wrote: > hi i wanna try to integrate SOLR with OpenStreetMap (OSM). well the plan > is i wanna index some cordinaate (long & lat) to SOLR and then the OSM will > try to showing the map of that coordinate. is there any article about that? > pls help. i'm still confuse about this. > > thx before. > -- Bill Bell billnb...@gmail.com cell 720-256-8076
integrate SOLR with OSM
hi i wanna try to integrate SOLR with OpenStreetMap (OSM). well the plan is i wanna index some cordinaate (long & lat) to SOLR and then the OSM will try to showing the map of that coordinate. is there any article about that? pls help. i'm still confuse about this. thx before.
Re: SolrCloud persisting data is very slow
On 6/25/2016 1:19 AM, Roshan Kamble wrote: > I am using Solr 6.0.0 in cloudMode (3 physical nodes + one zookeeper) > and have heavy insert/update/delete operations. I am using > CloudSolrClient and tried with all batch size from 100 to 1000. But it > has been observed that persist at Solr node is very slow. It takes > around 20 secords to store 50-100 records. Does anyone know how to > improve the speed for these operations? Is that 20 seconds the *index* time or the *commit* time? If it's the commit time, then see the "slow commits" section of the link that I provided below. You can see how long the last commit took by looking at the statistics in the admin UI for the searcher object. If it's the index time, how much data is in those records? What does the analysis in your schema do to that data? If you have no idea which process is taking the time, then you should decouple indexing from committing, so you can time both separately. Very slow indexing usually has one or more of these causes: 1) The data is very large and is heavily analyzed. 2) It is only being sent to Solr by a single thread. 3) Your Solr machine does not have enough memory for effective operation. That last item is a somewhat complex topic. It is one of the things discussed here: https://wiki.apache.org/solr/SolrPerformanceProblems There could be other problems, but these are the most common. The solutions for these issues are, in the same order: 1a) Reduce the amount of data per record. 1b) change the schema so analysis is not as heavy. 1c) Handle rich document processing in your indexing program, not Solr. 2) Use multiple threads/processes in your indexing program. 3) Add memory to the server, and sometimes increase the max heap size. Thanks, Shawn
SolrCloud persisting data is very slow
Hello, I am using Solr 6.0.0 in cloudMode (3 physical nodes + one zookeeper) and have heavy insert/update/delete operations. I am using CloudSolrClient and tried with all batch size from 100 to 1000. But it has been observed that persist at Solr node is very slow. It takes around 20 secords to store 50-100 records. Does anyone know how to improve the speed for these operations? Regards, Roshan The information in this email is confidential and may be legally privileged. It is intended solely for the addressee. Access to this email by anyone else is unauthorised. If you are not the intended recipient, any disclosure, copying, distribution or any action taken or omitted to be taken in reliance on it, is prohibited and may be unlawful.