Streaming Expressions (/stream) StreamHandler java.lang.NullPointerException

2016-06-25 Thread Peter Sh
I've got an exception below running
curl --data-urlencode
'expr=search(EventsAndDCF,q="*:*",fl="AccessPath",sort="AccessPath
asc",qt="/export")' "http://localhost:8983/solr/EventsAndDCF/stream";
Solr responce:
{"result-set":{"docs":[
{"EXCEPTION":null,"EOF":true}]}}


My collection EventsAndDCF exists. and I succeed to run GET queries like:
http://localhost:8983/solr/EventsAndDCF/export?fl=AccessPath&q=*:*&sort=AccessPath
desc&wt=json

Solr version: 6.0.1. Single node



2016-06-25 21:15:44.147 ERROR (qtp1514322932-16) [   x:EventsAndDCF]
o.a.s.h.StreamHandler java.lang.NullPointerException
at
org.apache.solr.client.solrj.io.stream.expr.StreamExpressionParser.generateStreamExpression(StreamExpressionParser.java:46)
at
org.apache.solr.client.solrj.io.stream.expr.StreamExpressionParser.parse(StreamExpressionParser.java:37)
at
org.apache.solr.client.solrj.io.stream.expr.StreamFactory.constructStream(StreamFactory.java:178)
at
org.apache.solr.handler.StreamHandler.handleRequestBody(StreamHandler.java:164)
at
org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:155)
at org.apache.solr.core.SolrCore.execute(SolrCore.java:2053)
at org.apache.solr.servlet.HttpSolrCall.execute(HttpSolrCall.java:652)
at org.apache.solr.servlet.HttpSolrCall.call(HttpSolrCall.java:460)
at
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:229)
at
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:184)
at
org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1668)
at
org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:581)
at
org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:143)
at
org.eclipse.jetty.security.SecurityHandler.handle(SecurityHandler.java:548)
at
org.eclipse.jetty.server.session.SessionHandler.doHandle(SessionHandler.java:226)
at
org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:1160)
at org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:511)
at
org.eclipse.jetty.server.session.SessionHandler.doScope(SessionHandler.java:185)
at
org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:1092)
at
org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:141)
at
org.eclipse.jetty.server.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:213)
at
org.eclipse.jetty.server.handler.HandlerCollection.handle(HandlerCollection.java:119)
at
org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:134)
at org.eclipse.jetty.server.Server.handle(Server.java:518)
at org.eclipse.jetty.server.HttpChannel.handle(HttpChannel.java:308)
at
org.eclipse.jetty.server.HttpConnection.onFillable(HttpConnection.java:244)
at
org.eclipse.jetty.io.AbstractConnection$ReadCallback.succeeded(AbstractConnection.java:273)
at org.eclipse.jetty.io.FillInterest.fillable(FillInterest.java:95)
at
org.eclipse.jetty.io.SelectChannelEndPoint$2.run(SelectChannelEndPoint.java:93)
at
org.eclipse.jetty.util.thread.strategy.ExecuteProduceConsume.produceAndRun(ExecuteProduceConsume.java:246)
at
org.eclipse.jetty.util.thread.strategy.ExecuteProduceConsume.run(ExecuteProduceConsume.java:156)
at
org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:654)
at
org.eclipse.jetty.util.thread.QueuedThreadPool$3.run(QueuedThreadPool.java:572)
at java.lang.Thread.run(Unknown Source)

2016-06-25 21:15:44.147 INFO  (qtp1514322932-16) [   x:EventsAndDCF]
o.a.s.c.S.Request [EventsAndDCF]  webapp=/solr path=/stream
params={'expr=search(EventsAndDCF,q%3D*:*,fl%3DAccessPath,sort%3DAccessPath+asc,qt%3D/export)'}
status=0 QTime=2


limit stored field size

2016-06-25 Thread asteiner
Hi

I have a field called content which I'm indexing and use for highlighting,
which means it has to be stored as well. 




But this field may be too big, so I want to limit the stored size to X
characters (it is fine to highlight only the first X characters).

One solution is to create another field called content_snippet which will be
a copy field of content field by maxchars of X (1 in my example), set
content as non-stored and set content_snippet as stored and indexed.
content_snippet must be indexed in order to highlight it. 






So as a result I have two indexed fields, which is redundant. My goal is to
decrease index size. Is there a way to limit the stored size within one
field without creating copy field?




--
View this message in context: 
http://lucene.472066.n3.nabble.com/limit-stored-field-size-tp4284356.html
Sent from the Solr - User mailing list archive at Nabble.com.


SolrCloud trying to upload documents and shards do not have storage anymore

2016-06-25 Thread Deeksha Sharma
Hi,


I am currently using JSON Index Handler to upload documents to a specific 
collection on SolrCloud. Now what I need to know is:


If I upload documents to SolrCloud collection and the machines hosting Shards 
for this collection have no storage left, will Solr reject the commit request?


?


Re: integrate SOLR with OSM

2016-06-25 Thread tkg_cangkul

hi william,

thx for your reply.
is it any link article or documents that i can read for this?
pls let me know if you've any suggestion.

On 26/06/16 02:58, William Bell wrote:

It depends on if your are using TIGERLINE data in OSM to increase the
accuracy of home addresses.

If you are using TIGERLINE, you are talking about > 500M rows or so +
millions of OSM data. Which will is a LOT of data that you will be pumping
into SOLR. You will need to use Solr Cloud since you will hit limits...

With that amount of data I had issues with just USA OSM data with
performance.

If you get it working, please share.

On Sat, Jun 25, 2016 at 12:49 PM, tkg_cangkul  wrote:


hi i wanna try to integrate SOLR with OpenStreetMap (OSM). well the plan
is i wanna index some cordinaate (long & lat) to SOLR and then the OSM will
try to showing the map of that coordinate. is there any article about that?
pls help. i'm still confuse about this.

thx before.








Could not load collection for SolrCloud

2016-06-25 Thread Roshan Kamble
Hello,

I am using solr 6.0.0 in SolCloud mode with 3 nodes, one zookeeper and 3 shard 
and 2 replica per collection.

Getting below error for some insert/update when trying to insert documents to 
Solr.

And it has been observed that few shard are in either recovery or fail recovery 
state. (Atleast one shard is up)


org.apache.solr.common.SolrException: Could not load collection from ZK: 
MY_COLLECTION
at 
org.apache.solr.common.cloud.ZkStateReader.getCollectionLive(ZkStateReader.java:969)
 ~[solr-solrj-6.0.0.jar:6.0.0 48c80f91b8e5cd9b3a9b48e6184bd53e7619e7e3 - nknize 
- 2016-04-01 14:41:50]
at 
org.apache.solr.common.cloud.ZkStateReader$LazyCollectionRef.get(ZkStateReader.java:519)
 ~[solr-solrj-6.0.0.jar:6.0.0 48c80f91b8e5cd9b3a9b48e6184bd53e7619e7e3 - nknize 
- 2016-04-01 14:41:50]
at 
org.apache.solr.common.cloud.ClusterState.getCollectionOrNull(ClusterState.java:189)
 ~[solr-solrj-6.0.0.jar:6.0.0 48c80f91b8e5cd9b3a9b48e6184bd53e7619e7e3 - nknize 
- 2016-04-01 14:41:50]
at 
org.apache.solr.common.cloud.ClusterState.hasCollection(ClusterState.java:119) 
~[solr-solrj-6.0.0.jar:6.0.0 48c80f91b8e5cd9b3a9b48e6184bd53e7619e7e3 - nknize 
- 2016-04-01 14:41:50]
at 
org.apache.solr.client.solrj.impl.CloudSolrClient.getCollectionNames(CloudSolrClient.java:)
 ~[solr-solrj-6.0.0.jar:6.0.0 48c80f91b8e5cd9b3a9b48e6184bd53e7619e7e3 - nknize 
- 2016-04-01 14:41:50]
at 
org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:833)
 ~[solr-solrj-6.0.0.jar:6.0.0 48c80f91b8e5cd9b3a9b48e6184bd53e7619e7e3 - nknize 
- 2016-04-01 14:41:50]
at 
org.apache.solr.client.solrj.impl.CloudSolrClient.request(CloudSolrClient.java:806)
 ~[solr-solrj-6.0.0.jar:6.0.0 48c80f91b8e5cd9b3a9b48e6184bd53e7619e7e3 - nknize 
- 2016-04-01 14:41:50]
at 
org.apache.solr.client.solrj.SolrRequest.process(SolrRequest.java:149) 
~[solr-solrj-6.0.0.jar:6.0.0 48c80f91b8e5cd9b3a9b48e6184bd53e7619e7e3 - nknize 
- 2016-04-01 14:41:50]
at org.apache.solr.client.solrj.SolrClient.add(SolrClient.java:106) 
~[solr-solrj-6.0.0.jar:6.0.0 48c80f91b8e5cd9b3a9b48e6184bd53e7619e7e3 - nknize 
- 2016-04-01 14:41:50]
at org.apache.solr.client.solrj.SolrClient.add(SolrClient.java:71) 
~[solr-solrj-6.0.0.jar:6.0.0 48c80f91b8e5cd9b3a9b48e6184bd53e7619e7e3 - nknize 
- 2016-04-01 14:41:50]
at 
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) 
[?:1.8.0_60]
at java.util.concurrent.FutureTask.run(FutureTask.java:266) [?:1.8.0_60]
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) 
[?:1.8.0_60]
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) 
[?:1.8.0_60]
at java.lang.Thread.run(Thread.java:745) [?:1.8.0_60]
Caused by: org.apache.zookeeper.KeeperException$SessionExpiredException: 
KeeperErrorCode = Session expired for /collections/ MY_COLLECTION /state.json
at 
org.apache.zookeeper.KeeperException.create(KeeperException.java:127) 
~[zookeeper-3.4.6.jar:3.4.6-1569965]
at org.apache.zookeeper.KeeperException.create(KeeperException.java:51) 
~[zookeeper-3.4.6.jar:3.4.6-1569965]
at org.apache.zookeeper.ZooKeeper.getData(ZooKeeper.java:1155) 
~[zookeeper-3.4.6.jar:3.4.6-1569965]
at 
org.apache.solr.common.cloud.SolrZkClient$7.execute(SolrZkClient.java:348) 
~[solr-solrj-6.0.0.jar:6.0.0 48c80f91b8e5cd9b3a9b48e6184bd53e7619e7e3 - nknize 
- 2016-04-01 14:41:50]
at 
org.apache.solr.common.cloud.SolrZkClient$7.execute(SolrZkClient.java:345) 
~[solr-solrj-6.0.0.jar:6.0.0 48c80f91b8e5cd9b3a9b48e6184bd53e7619e7e3 - nknize 
- 2016-04-01 14:41:50]
at 
org.apache.solr.common.cloud.ZkCmdExecutor.retryOperation(ZkCmdExecutor.java:60)
 ~[solr-solrj-6.0.0.jar:6.0.0 48c80f91b8e5cd9b3a9b48e6184bd53e7619e7e3 - nknize 
- 2016-04-01 14:41:50]
at 
org.apache.solr.common.cloud.SolrZkClient.getData(SolrZkClient.java:345) 
~[solr-solrj-6.0.0.jar:6.0.0 48c80f91b8e5cd9b3a9b48e6184bd53e7619e7e3 - nknize 
- 2016-04-01 14:41:50]
at 
org.apache.solr.common.cloud.ZkStateReader.fetchCollectionState(ZkStateReader.java:980)
 ~[solr-solrj-6.0.0.jar:6.0.0 48c80f91b8e5cd9b3a9b48e6184bd53e7619e7e3 - nknize 
- 2016-04-01 14:41:50]
at 
org.apache.solr.common.cloud.ZkStateReader.getCollectionLive(ZkStateReader.java:967)
 ~[solr-solrj-6.0.0.jar:6.0.0 48c80f91b8e5cd9b3a9b48e6184bd53e7619e7e3 - nknize 
- 2016-04-01 14:41:50]
... 16 more


Regards,

Roshan

The information in this email is confidential and may be legally privileged. It 
is intended solely for the addressee. Access to this email by anyone else is 
unauthorised. If you are not the intended recipient, any disclosure, copying, 
distribution or any action taken or omitted to be taken in reliance on it, is 
prohibited and may be unlawful.


[ANNOUNCE] Apache Solr 5.5.2 released

2016-06-25 Thread Steve Rowe
25 June 2016, Apache Solr™ 5.5.2 available

The Lucene PMC is pleased to announce the release of Apache Solr 5.5.2

Solr is the popular, blazing fast, open source NoSQL search platform
from the Apache Lucene project. Its major features include powerful
full-text search, hit highlighting, faceted search, dynamic
clustering, database integration, rich document (e.g., Word, PDF)
handling, and geospatial search. Solr is highly scalable, providing
fault tolerant distributed search and indexing, and powers the search
and navigation features of many of the world's largest internet sites.

This release includes 38 bug fixes, documentation updates, etc.,
since the 5.5.1 release.

The release is available for immediate download at:

  http://www.apache.org/dyn/closer.lua/lucene/solr/5.5.2

Please read CHANGES.txt for a detailed list of changes:

  https://lucene.apache.org/solr/5_5_2/changes/Changes.html

Please report any feedback to the mailing lists
(http://lucene.apache.org/solr/discussion.html)

Note: The Apache Software Foundation uses an extensive mirroring
network for distributing releases. It is possible that the mirror you
are using may not have replicated the release yet. If that is the
case, please try another mirror. This also goes for Maven access.



Re: integrate SOLR with OSM

2016-06-25 Thread William Bell
It depends on if your are using TIGERLINE data in OSM to increase the
accuracy of home addresses.

If you are using TIGERLINE, you are talking about > 500M rows or so +
millions of OSM data. Which will is a LOT of data that you will be pumping
into SOLR. You will need to use Solr Cloud since you will hit limits...

With that amount of data I had issues with just USA OSM data with
performance.

If you get it working, please share.

On Sat, Jun 25, 2016 at 12:49 PM, tkg_cangkul  wrote:

> hi i wanna try to integrate SOLR with OpenStreetMap (OSM). well the plan
> is i wanna index some cordinaate (long & lat) to SOLR and then the OSM will
> try to showing the map of that coordinate. is there any article about that?
> pls help. i'm still confuse about this.
>
> thx before.
>



-- 
Bill Bell
billnb...@gmail.com
cell 720-256-8076


integrate SOLR with OSM

2016-06-25 Thread tkg_cangkul
hi i wanna try to integrate SOLR with OpenStreetMap (OSM). well the plan 
is i wanna index some cordinaate (long & lat) to SOLR and then the OSM 
will try to showing the map of that coordinate. is there any article 
about that? pls help. i'm still confuse about this.


thx before.


Re: SolrCloud persisting data is very slow

2016-06-25 Thread Shawn Heisey
On 6/25/2016 1:19 AM, Roshan Kamble wrote:
> I am using Solr 6.0.0 in cloudMode (3 physical nodes + one zookeeper)
> and have heavy insert/update/delete operations. I am using
> CloudSolrClient and tried with all batch size from 100 to 1000. But it
> has been observed that persist at Solr node is very slow. It takes
> around 20 secords to store 50-100 records. Does anyone know how to
> improve the speed for these operations? 

Is that 20 seconds the *index* time or the *commit* time?  If it's the
commit time, then see the "slow commits" section of the link that I
provided below.  You can see how long the last commit took by looking at
the statistics in the admin UI for the searcher object.

If it's the index time, how much data is in those records?  What does
the analysis in your schema do to that data?

If you have no idea which process is taking the time, then you should
decouple indexing from committing, so you can time both separately.

Very slow indexing usually has one or more of these causes:

1) The data is very large and is heavily analyzed.
2) It is only being sent to Solr by a single thread.
3) Your Solr machine does not have enough memory for effective operation.

That last item is a somewhat complex topic.  It is one of the things
discussed here:

https://wiki.apache.org/solr/SolrPerformanceProblems

There could be other problems, but these are the most common.  The
solutions for these issues are, in the same order:

1a) Reduce the amount of data per record.
1b) change the schema so analysis is not as heavy.
1c) Handle rich document processing in your indexing program, not Solr.
2) Use multiple threads/processes in your indexing program.
3) Add memory to the server, and sometimes increase the max heap size.

Thanks,
Shawn



SolrCloud persisting data is very slow

2016-06-25 Thread Roshan Kamble
Hello,

I am using Solr 6.0.0 in cloudMode (3 physical nodes + one zookeeper)  and have 
heavy insert/update/delete operations.

I am using CloudSolrClient and tried with all batch size from 100 to 1000.

But it has been observed that persist at Solr node is very slow. It takes 
around 20 secords to store 50-100 records.

Does anyone know how to improve the speed for these operations?

Regards,
Roshan

The information in this email is confidential and may be legally privileged. It 
is intended solely for the addressee. Access to this email by anyone else is 
unauthorised. If you are not the intended recipient, any disclosure, copying, 
distribution or any action taken or omitted to be taken in reliance on it, is 
prohibited and may be unlawful.