Re: Corrupt Index error on Target cluster

2018-09-07 Thread Stephen Bianamara
Hmm, when this occurred for me I was also on 6.6 between minor releases. So
unclear if it's connected to 6.6 specifically.

If you want to resolve the problem, you should be able to use the
collection api delete that node from the collection, and then re-add it
which will trigger resync.


On Fri, Sep 7, 2018, 10:35 AM Susheel Kumar  wrote:

> No. The solr i have is 6.6.
>
> On Fri, Sep 7, 2018 at 10:51 AM Stephen Bianamara <
> sdl1tinsold...@gmail.com>
> wrote:
>
> > I've gotten incorrect checksums when upgrading solr versions across the
> > cluster. Or in other words, when indexing into a mixed version cluster.
> Are
> > you running mixed versions by chance?
> >
> > On Fri, Sep 7, 2018, 6:07 AM Susheel Kumar 
> wrote:
> >
> > > Anyone has  insight / have faced above errors ?
> > >
> > > On Thu, Sep 6, 2018 at 12:04 PM Susheel Kumar 
> > > wrote:
> > >
> > > > Hello,
> > > >
> > > > We had a running cluster with CDCR and there were some issues with
> > > > indexing on Source cluster which got resolved after restarting the
> > nodes
> > > > (in my absence...) and now I see  below errors on a shard at Target
> > > > cluster.  Any suggestions / ideas what could have caused this and
> whats
> > > the
> > > > best way to recover.
> > > >
> > > > Thnx
> > > >
> > > > Caused by: org.apache.solr.common.SolrException: Error opening new
> > > searcher
> > > > at
> > > > org.apache.solr.core.SolrCore.openNewSearcher(SolrCore.java:2069)
> > > > at
> > org.apache.solr.core.SolrCore.getSearcher(SolrCore.java:2189)
> > > > at
> > org.apache.solr.core.SolrCore.getSearcher(SolrCore.java:1926)
> > > > at
> > org.apache.solr.core.SolrCore.getSearcher(SolrCore.java:1826)
> > > > at
> > > >
> > >
> >
> org.apache.solr.request.SolrQueryRequestBase.getSearcher(SolrQueryRequestBase.java:127)
> > > > at
> > > >
> > >
> >
> org.apache.solr.handler.component.QueryComponent.process(QueryComponent.java:310)
> > > > at
> > > >
> > >
> >
> org.apache.solr.handler.component.SearchHandler.handleRequestBody(SearchHandler.java:296)
> > > > at
> > > >
> > >
> >
> org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:173)
> > > > at org.apache.solr.core.SolrCore.execute(SolrCore.java:2477)
> > > > at
> > > >
> > >
> >
> org.apache.solr.handler.PingRequestHandler.handlePing(PingRequestHandler.java:267)
> > > > ... 34 more
> > > > Caused by: org.apache.lucene.index.CorruptIndexException: Corrupted
> > > > bitsPerDocBase: 6033
> > > >
> > >
> >
> (resource=BufferedChecksumIndexInput(MMapIndexInput(path="/app/solr/data/COLL_shard8_replica1/data/index.20180903220548447/_9nsy.tvx")))
> > > > at
> > > >
> > >
> >
> org.apache.lucene.codecs.compressing.CompressingStoredFieldsIndexReader.(CompressingStoredFieldsIndexReader.java:89)
> > > > at
> > > >
> > >
> >
> org.apache.lucene.codecs.compressing.CompressingTermVectorsReader.(CompressingTermVectorsReader.java:126)
> > > > at
> > > >
> > >
> >
> org.apache.lucene.codecs.compressing.CompressingTermVectorsFormat.vectorsReader(CompressingTermVectorsFormat.java:91)
> > > > at
> > > >
> > >
> >
> org.apache.lucene.index.SegmentCoreReaders.(SegmentCoreReaders.java:128)
> > > > at
> > > > org.apache.lucene.index.SegmentReader.(SegmentReader.java:74)
> > > > at
> > > >
> > >
> >
> org.apache.lucene.index.ReadersAndUpdates.getReader(ReadersAndUpdates.java:145)
> > > > at
> > > >
> > >
> >
> org.apache.lucene.index.ReadersAndUpdates.getReadOnlyClone(ReadersAndUpdates.java:197)
> > > > at
> > > >
> > >
> >
> org.apache.lucene.index.StandardDirectoryReader.open(StandardDirectoryReader.java:103)
> > > > at
> > > > org.apache.lucene.index.IndexWriter.getReader(IndexWriter.java:467)
> > > > at
> > > >
> org.apache.lucene.index.DirectoryReader.open(DirectoryReader.java:103)
> > > > at
> > > > org.apache.lucene.index.DirectoryReader.open(DirectoryReader.java:79)
> > > > at
> > > >
> > >
> >
> org.apache.solr.core.StandardIndexReaderFactory.newReader(StandardIndexReaderFactory.java:39)
> > > > at
> > > > org.apache.solr.core.SolrCore.openNewSearcher(SolrCore.java:2033)
> > > > ... 43 more
> > > > Suppressed: org.apache.lucene.index.CorruptIndexException:
> > > > checksum failed (hardware problem?) : expected=e5bf0d15
> actual=21722825
> > > >
> > >
> >
> (resource=BufferedChecksumIndexInput(MMapIndexInput(path="/app/solr/data/COLL_shard8_replica1/data/index.20180903220548447/_9nsy.tvx")))
> > > > at
> > > > org.apache.lucene.codecs.CodecUtil.checkFooter(CodecUtil.java:419)
> > > > at
> > > > org.apache.lucene.codecs.CodecUtil.checkFooter(CodecUtil.java:462)
> > > > at
> > > >
> > >
> >
> org.apache.lucene.codecs.compressing.CompressingTermVectorsReader.(CompressingTermVectorsReader.java:131)
> > > > ... 54 more
> > > >
> > 

Re: parent/child rows in solr

2018-09-07 Thread Shawn Heisey

On 9/7/2018 7:44 PM, John Smith wrote:

Thanks Shawn, for your comments. The reason why I don't want to go flat
file structure, is due to all the wasted/duplicated data. If a department
has 100 employees, then it's very wasteful in terms of disk space to repeat
the header data over and over again, 100 times. In this example there is
only a few doc types, but my real-life data is much larger, and the problem
is a "scaling" problem; with just a little bit of data, no problem in
duplicating header fields, but with massive amounts of data it's a large
problem.


If your goal is data storage, then you are completely correct.  All that 
data duplication is something to avoid for a data storage situation.  
Normalizing your data so it's relational makes perfect sense, because 
most database software is designed to efficiently deal with those 
relationships.


Solr is not designed as a data storage platform, and does not handle 
those relationships efficiently.  Solr's design goals are all about 
*search*.  It often gets touted as filling a NoSQL role ... but it's not 
something I would personally use as a primary data repository.  Search 
is a space where data duplication is expected and completely normal.  
This is something that people often have a hard time accepting.


Thanks,
Shawn



Re: parent/child rows in solr

2018-09-07 Thread John Smith
Thanks Shawn, for your comments. The reason why I don't want to go flat
file structure, is due to all the wasted/duplicated data. If a department
has 100 employees, then it's very wasteful in terms of disk space to repeat
the header data over and over again, 100 times. In this example there is
only a few doc types, but my real-life data is much larger, and the problem
is a "scaling" problem; with just a little bit of data, no problem in
duplicating header fields, but with massive amounts of data it's a large
problem.

My understanding of both graph traversal and block joins, is that the
header data would only be present once, so that's why I'm gravitating
towards those solutions. I just can't seem to line up the "fq" and queries
correctly such that I am able to join 3+ document types together, filter on
them, and return my requested columns.

On Fri, Sep 7, 2018 at 9:32 PM Shawn Heisey  wrote:

> On 9/7/2018 3:06 PM, John Smith wrote:
> > Hi, I have a document structure like this (this is a made up schema, my
> > data has nothing to do with departments and employees, but the structure
> > holds true to my real data):
> >
> > department 1
> >  employee 11
> >  employee 12
> >  employee 13
> >  room 11
> >  room 12
> >  room 13
> >
> > department 2
> >  employee 21
> >  employee 22
> >  room 21
> >
> > ... etc
> >
> > I'm trying to figure out the best way to index this, and perform queries.
> > Due to the sheer volume of data, I cannot do a simple "flat file"
> approach,
> > repeating the header data for each child entry.
>
> Why not?
>
> For the precise use case you have outlined, Solr will work better if you
> only have the child documents and simply have every document contain a
> "department" field which contains an identifier for the department.
> Since this precise structure is not what you are doing, you'll need to
> adapt what I'm saying to your actual data.
>
> The volume of data should be irrelevant to this decision. Solr will
> always work best with a flat document structure.
>
> I have never used the parent/child document feature in Solr, so I cannot
> offer any advice on it.  Somebody else will need to help you if you
> choose to use that feature.
>
> Thanks,
> Shawn
>
>


Re: local "q.op=AND" ignored for edismax query

2018-09-07 Thread Shawn Heisey

On 9/7/2018 6:59 PM, dshih wrote:

Query:
/select?q={!q.op=AND}mysearchtext*&defType=edismax&debugQuery=true

Result:
"querystring": "{!q.op=AND}mysearchtext*",
"parsedquery": "+DisjunctionMaxQuery(((Synonym(text:q
text:qopandmysearchtext text:{!q.op=and}mysearchtext*) text:op text:and
text:mysearchtext)))"

In the above, I'm just trying to set the query operator to AND while leaving
the default as OR.  It looks like q.op is being completely ignored (and ends
up as part of the search terms).


Since 7.2, you can only use certain localparams if defType is lucene or 
func.  The lucene parser is the default.  Since you have changed defType 
to edismax, localparams will not work.   Just add q.op as a separate 
parameter.


/select?q=mysearchtext*&defType=edismax&debugQuery=true&q.op=AND

Or use mm=100% instead.  Which is effectively what q.op=AND does.

This is the issue that made the change in 7.2:

https://issues.apache.org/jira/browse/SOLR-11501

Thanks,
Shawn



Re: parent/child rows in solr

2018-09-07 Thread Shawn Heisey

On 9/7/2018 3:06 PM, John Smith wrote:

Hi, I have a document structure like this (this is a made up schema, my
data has nothing to do with departments and employees, but the structure
holds true to my real data):

department 1
 employee 11
 employee 12
 employee 13
 room 11
 room 12
 room 13

department 2
 employee 21
 employee 22
 room 21

... etc

I'm trying to figure out the best way to index this, and perform queries.
Due to the sheer volume of data, I cannot do a simple "flat file" approach,
repeating the header data for each child entry.


Why not?

For the precise use case you have outlined, Solr will work better if you 
only have the child documents and simply have every document contain a 
"department" field which contains an identifier for the department.  
Since this precise structure is not what you are doing, you'll need to 
adapt what I'm saying to your actual data.


The volume of data should be irrelevant to this decision. Solr will 
always work best with a flat document structure.


I have never used the parent/child document feature in Solr, so I cannot 
offer any advice on it.  Somebody else will need to help you if you 
choose to use that feature.


Thanks,
Shawn



local "q.op=AND" ignored for edismax query

2018-09-07 Thread dshih
SOLR 7.4.0

Apologies if this has been answered, but I can't find the answer for the
life of me if it has been.

Query:
/select?q={!q.op=AND}mysearchtext*&defType=edismax&debugQuery=true

Result:
"querystring": "{!q.op=AND}mysearchtext*",
"parsedquery": "+DisjunctionMaxQuery(((Synonym(text:q
text:qopandmysearchtext text:{!q.op=and}mysearchtext*) text:op text:and
text:mysearchtext)))"

In the above, I'm just trying to set the query operator to AND while leaving
the default as OR.  It looks like q.op is being completely ignored (and ends
up as part of the search terms).

I did read https://lucene.apache.org/solr/guide/7_2/solr-upgrade-notes.html
regarding switching the query parser using local parameters, but that is not
what is happening here.  Incidentally, if I set luceneMatchVersion=7.1.0, I
can workaround this issue by doing "{!edismax q.op=and}", but that doesn't
work with 7.4.0.

I also read a number of threads/bugs about changes to the "mm" option, but I
don't think that's relevant here either.

The above works in 4.10.3 (long time ago - I know).  What is the recommended
way to get the same functionality?



--
Sent from: http://lucene.472066.n3.nabble.com/Solr-User-f472068.html


Re: SolrCloud CDCR with 3+ DCs

2018-09-07 Thread Amrit Sarkar
Yeah, I am not sure about how the Authentication band aid feature will
work, the mentioned stackoverflow link. It is about time we include basic
authentication support in CDCR.

On Thu, 6 Sep 2018, 8:41 pm cdatta,  wrote:

> Hi Amrit, Thanks for your response.
>
> We wiped out our complete installation and started a fresh one. Now the
> multi-direction replication is working but we are seeing errors related to
> the authentication sporadically.
>
> Thanks & Regards,
> Chandi Datta
>
>
>
> --
> Sent from: http://lucene.472066.n3.nabble.com/Solr-User-f472068.html
>


unregistered

2018-09-07 Thread sirenfei
unregistered


Solr CDCR replication not working

2018-09-07 Thread Amrit Sarkar
Basic Authentication in clusters is not supported as of today in CDCR.

On Fri, 7 Sep 2018, 4:53 pm Mrityunjaya Pathak, 
wrote:

> I have setup two solr cloud instances in two different Datacenters Target
> solr cloud machine is copy of source machine with basicAuth enabled on
> them. I am unable to see any replication on target.
>
> Solr Version :6.6.3
>
> I have done config changes as suggested on
> https://lucene.apache.org/solr/guide/6_6/cross-data-center-replication-cdcr.html
>
> Source Config Changes
>
> 
> 
> ...
> 
>   
> serverIP:2181,serverIP:2182,serverIP:2183
> sitecore_master_index
> sitecore_master_index
>   
>
>   
> 8
> 1000
> 128
>   
>
>   
> 1000
>   
> 
>   
>
> 
>   ${solr.ulog.dir:}
>name="numVersionBuckets">${solr.ulog.numVersionBuckets:65536}
> 
>
>
> 
>   ${solr.autoCommit.maxTime:15000}
>   false
> 
>
> 
>   ${solr.autoSoftCommit.maxTime:-1}
> 
>   
>
>   ...
>   
>
> Target Config Changes
>
> 
> 
> ...
> 
>   
> disabled
>   
> 
> 
>   
>   
> 
> 
>   
> cdcr-proc-chain
>   
> 
>  
>
> 
>   ${solr.ulog.dir:}
>name="numVersionBuckets">${solr.ulog.numVersionBuckets:65536}
> 
>
> 
>   ${solr.autoCommit.maxTime:15000}
>   false
> 
>
> 
>   ${solr.autoSoftCommit.maxTime:-1}
> 
>
>   
>
>   ...
>   
>
> Below are logs from Source target.
>
> ERROR (zkCallback-4-thread-2-processing-n:sourceIP:8983_solr) [   ]
> o.a.s.c.s.i.CloudSolrClient Request to collection collection1 failed due to
> (510) org.apache.solr.common.SolrException: Could not find a healthy node
> to handle the request., retry? 5
> 2018-09-07 10:36:14.295 WARN
> (zkCallback-4-thread-2-processing-n:sourceIP:8983_solr) [   ]
> o.a.s.h.CdcrReplicatorManager Unable to instantiate the log reader for
> target collection collection1
> org.apache.solr.common.SolrException: Could not find a healthy node to
> handle the request.
> at
> org.apache.solr.client.solrj.impl.CloudSolrClient.sendRequest(CloudSolrClient.java:1377)
> at
> org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:1134)
> at
> org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:1237)
> at
> org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:1237)
> at
> org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:1237)
> at
> org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:1237)
> at
> org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:1237)
> at
> org.apache.solr.client.solrj.impl.CloudSolrClient.request(CloudSolrClient.java:1073)
> at
> org.apache.solr.client.solrj.SolrClient.request(SolrClient.java:1219)
> at
> org.apache.solr.handler.CdcrReplicatorManager.getCheckpoint(CdcrReplicatorManager.java:196)
> at
> org.apache.solr.handler.CdcrReplicatorManager.initLogReaders(CdcrReplicatorManager.java:159)
> at
> org.apache.solr.handler.CdcrReplicatorManager.stateUpdate(CdcrReplicatorManager.java:134)
> at
> org.apache.solr.handler.CdcrStateManager.callback(CdcrStateManager.java:36)
> at
> org.apache.solr.handler.CdcrLeaderStateManager.setAmILeader(CdcrLeaderStateManager.java:108)
> at
> org.apache.solr.handler.CdcrLeaderStateManager.checkIfIAmLeader(CdcrLeaderStateManager.java:95)
> at
> org.apache.solr.handler.CdcrLeaderStateManager.access$400(CdcrLeaderStateManager.java:40)
> at
> org.apache.solr.handler.CdcrLeaderStateManager$LeaderStateWatcher.process(CdcrLeaderStateManager.java:150)
> at
> org.apache.solr.common.cloud.SolrZkClient$3.lambda$process$0(SolrZkClient.java:269)
> at
> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
> at java.util.concurrent.FutureTask.run(FutureTask.java:266)
> at
> org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor.lambda$execute$0(ExecutorUtil.java:229)
> at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> at java.lang.Thread.run(Thread.java:748)
> 2018-09-07 10:36:14.310 INFO
> (coreLoadExecutor-8-thread-3-processing-n:sourceIP:8983_solr) [   ]
> o.a.s.c.SolrConfig Using Lucene MatchVersion: 6.6.3
> 2018-09-07 10:36:14.315 INFO
> (zkCallback-4-thread-1-processing-n:sourceIP:8983_solr) [   ]
> o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent
> state:SyncConnected type:NodeDataChanged
> path:/collections/collection1/state.json] for collection [sitecore] has
> occurred - updating... (live nodes size: [1])
> 2018-09-07 10:36:14.343 WARN
> (cdcr-replicator-211-thread-

parent/child rows in solr

2018-09-07 Thread John Smith
Hi, I have a document structure like this (this is a made up schema, my
data has nothing to do with departments and employees, but the structure
holds true to my real data):

department 1
employee 11
employee 12
employee 13
room 11
room 12
room 13

department 2
employee 21
employee 22
room 21

... etc

I'm trying to figure out the best way to index this, and perform queries.
Due to the sheer volume of data, I cannot do a simple "flat file" approach,
repeating the header data for each child entry.

So that leaves me with "graph traversal" or "block joins". I've played with
both of those, but I'm running into various issues with each approach.

I need to be able to run filters on any or all of the header + child rows
in the same query, at once (can't seem to get that working in either graph
or block join). One problem I had with graph is that I can't force solr to
return the header, then all the children for that header, then the next
header + all it's children, it just spits them out without keeping them
together. block join seems to return the children nested under the parents,
which is great, but then I can't seem to filter on parent + children in the
same query: I get the dreaded error message "Parent query must not match
any docs besides parent filter"

Kinda lost here, any tips/suggestions?


Re: Solr metric for finding number of merges in progress

2018-09-07 Thread Nawab Zada Asad Iqbal
Actually, i found this after posting;  (still if someone has more to offer,
please reply)

https://lucene.apache.org/solr/guide/7_0/metrics-reporting.html#index-merge-metrics

On Fri, Sep 7, 2018 at 12:13 PM Nawab Zada Asad Iqbal 
wrote:

> Hi,
>
> Does Solr expose a metric to find the number of segment merges which are
> in progress?
>
>
> Thanks
> Nawab
>


Solr metric for finding number of merges in progress

2018-09-07 Thread Nawab Zada Asad Iqbal
Hi,

Does Solr expose a metric to find the number of segment merges which are in
progress?


Thanks
Nawab


Re: Corrupt Index error on Target cluster

2018-09-07 Thread Susheel Kumar
No. The solr i have is 6.6.

On Fri, Sep 7, 2018 at 10:51 AM Stephen Bianamara 
wrote:

> I've gotten incorrect checksums when upgrading solr versions across the
> cluster. Or in other words, when indexing into a mixed version cluster. Are
> you running mixed versions by chance?
>
> On Fri, Sep 7, 2018, 6:07 AM Susheel Kumar  wrote:
>
> > Anyone has  insight / have faced above errors ?
> >
> > On Thu, Sep 6, 2018 at 12:04 PM Susheel Kumar 
> > wrote:
> >
> > > Hello,
> > >
> > > We had a running cluster with CDCR and there were some issues with
> > > indexing on Source cluster which got resolved after restarting the
> nodes
> > > (in my absence...) and now I see  below errors on a shard at Target
> > > cluster.  Any suggestions / ideas what could have caused this and whats
> > the
> > > best way to recover.
> > >
> > > Thnx
> > >
> > > Caused by: org.apache.solr.common.SolrException: Error opening new
> > searcher
> > > at
> > > org.apache.solr.core.SolrCore.openNewSearcher(SolrCore.java:2069)
> > > at
> org.apache.solr.core.SolrCore.getSearcher(SolrCore.java:2189)
> > > at
> org.apache.solr.core.SolrCore.getSearcher(SolrCore.java:1926)
> > > at
> org.apache.solr.core.SolrCore.getSearcher(SolrCore.java:1826)
> > > at
> > >
> >
> org.apache.solr.request.SolrQueryRequestBase.getSearcher(SolrQueryRequestBase.java:127)
> > > at
> > >
> >
> org.apache.solr.handler.component.QueryComponent.process(QueryComponent.java:310)
> > > at
> > >
> >
> org.apache.solr.handler.component.SearchHandler.handleRequestBody(SearchHandler.java:296)
> > > at
> > >
> >
> org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:173)
> > > at org.apache.solr.core.SolrCore.execute(SolrCore.java:2477)
> > > at
> > >
> >
> org.apache.solr.handler.PingRequestHandler.handlePing(PingRequestHandler.java:267)
> > > ... 34 more
> > > Caused by: org.apache.lucene.index.CorruptIndexException: Corrupted
> > > bitsPerDocBase: 6033
> > >
> >
> (resource=BufferedChecksumIndexInput(MMapIndexInput(path="/app/solr/data/COLL_shard8_replica1/data/index.20180903220548447/_9nsy.tvx")))
> > > at
> > >
> >
> org.apache.lucene.codecs.compressing.CompressingStoredFieldsIndexReader.(CompressingStoredFieldsIndexReader.java:89)
> > > at
> > >
> >
> org.apache.lucene.codecs.compressing.CompressingTermVectorsReader.(CompressingTermVectorsReader.java:126)
> > > at
> > >
> >
> org.apache.lucene.codecs.compressing.CompressingTermVectorsFormat.vectorsReader(CompressingTermVectorsFormat.java:91)
> > > at
> > >
> >
> org.apache.lucene.index.SegmentCoreReaders.(SegmentCoreReaders.java:128)
> > > at
> > > org.apache.lucene.index.SegmentReader.(SegmentReader.java:74)
> > > at
> > >
> >
> org.apache.lucene.index.ReadersAndUpdates.getReader(ReadersAndUpdates.java:145)
> > > at
> > >
> >
> org.apache.lucene.index.ReadersAndUpdates.getReadOnlyClone(ReadersAndUpdates.java:197)
> > > at
> > >
> >
> org.apache.lucene.index.StandardDirectoryReader.open(StandardDirectoryReader.java:103)
> > > at
> > > org.apache.lucene.index.IndexWriter.getReader(IndexWriter.java:467)
> > > at
> > > org.apache.lucene.index.DirectoryReader.open(DirectoryReader.java:103)
> > > at
> > > org.apache.lucene.index.DirectoryReader.open(DirectoryReader.java:79)
> > > at
> > >
> >
> org.apache.solr.core.StandardIndexReaderFactory.newReader(StandardIndexReaderFactory.java:39)
> > > at
> > > org.apache.solr.core.SolrCore.openNewSearcher(SolrCore.java:2033)
> > > ... 43 more
> > > Suppressed: org.apache.lucene.index.CorruptIndexException:
> > > checksum failed (hardware problem?) : expected=e5bf0d15 actual=21722825
> > >
> >
> (resource=BufferedChecksumIndexInput(MMapIndexInput(path="/app/solr/data/COLL_shard8_replica1/data/index.20180903220548447/_9nsy.tvx")))
> > > at
> > > org.apache.lucene.codecs.CodecUtil.checkFooter(CodecUtil.java:419)
> > > at
> > > org.apache.lucene.codecs.CodecUtil.checkFooter(CodecUtil.java:462)
> > > at
> > >
> >
> org.apache.lucene.codecs.compressing.CompressingTermVectorsReader.(CompressingTermVectorsReader.java:131)
> > > ... 54 more
> > >
> >
>


Re: Solr range faceting

2018-09-07 Thread Erick Erickson
Oh, indeed if you changed anything about the field definition and did
_not_ reindex everything from scratch your results are unreliable.

I'm overstating the case, but it'll do for now. For instance, your
facet counts should be find if you changed the stored parameter for
instance, but anything else

And it's best to index into a new collection (or core in stand-alone)
rather then re-use the current one just for paranoia's sake.

Best,
Erick
On Fri, Sep 7, 2018 at 12:18 AM Dwane Hall  wrote:
>
> Thanks Erick,
>
> The field is defined as a pfloat.
>
>  indexed="true" stored="true"/>
>
> I took your advice and tried smaller result range and the counts look good.  
> I might try an index rebuild I’m wondering if the data has somehow been 
> corrupted by a combination of old and new index mappings.
>
> Thanks again for your assistance.
>
>
> "responseHeader":{
> "zkConnected":true,
> "status":0,
> "QTime":7},
>   "response":{"numFound":3,"start":0,"docs":[
>   {
> "Value":34.0},
>   {
> "Value":34.0},
>   {
> "Value":34.0}]
>   },
>   "facet_counts":
> "facet_queries":{},
> "facet_ranges":{
>   "Value:{
> "counts":[
>   "0.0",3],
> "gap":100.0,
> "before":0,
> "after":0,
> "between":3,
> "start":0.0,
> "end":2000.0}}
> 
> From: Erick Erickson 
> Sent: Friday, 7 September 2018 12:54:48 PM
> To: solr-user
> Subject: Re: Solr range faceting
>
> Indeed this doesn't look right. By my count, you're missing 599 counts
> you'd expect in that range, although the after and between numbers
> total the numFound.
>
> What kind of a field is Value? Given the number of docs missing, I'd
> guess you could get the number of docs down really small and post
> them. Something like
> values 1, 2, 3, 4, 5, 
> and your range query so we could try it.
>
> What is the fieldType definition and field for Value?
>
> And finally, do you get different results if you use json faceting?
>
> Best,
> Erick
> On Thu, Sep 6, 2018 at 5:51 PM Dwane Hall  wrote:
> >
> > Thanks Jan that has fixed the bucket issue but I'm a little confused at why 
> > zero counts exist for some buckets when they appear to be values in them?
> >
> > "response":{"numFound":869,"start":0,"docs":[
> >   {
> > "Value":9475.08},
> >   {
> > "Value":780.0},
> >   {
> > "Value":9475.08},
> >   {
> > "Value":1000.0},
> >   {
> > "Value":50.0},
> >   {
> > "Value":50.0},
> >   {
> > "Value":0.0},
> >   {
> > "Value":800.0},
> >   {
> > "Value":0.0},
> >   {
> > "Value":1000.0},
> >   {
> > "Value":1000.0},
> >   {
> > "Value":5000.0},
> >   {
> > "Value":2000.0},
> >   {
> >"Value":4000.0},
> >   {
> > "Value":1500.0},
> >   {
> > "Value":0.0},
> >   {
> > "Value":1.0},
> >   {
> > "Value":5000.0},
> >   {
> > "Value":1000.0},
> >   {
> > "Value":0.0},
> >   {
> > "Value":1200.0},
> >   {
> > "Value":9000.0},
> >   {
> > "Value":1500.0},
> >   {
> > "Value":1.0},
> >   {
> > "Value":5000.0},
> >   {
> > "Value":4000.0},
> >   {
> > "Value":5000.0},
> >   {
> > "Value":5000.0},
> >   {
> > "Value":1.0},
> >   {
> > "Value":1000.0}]
> >   },
> >
> >   "facet_counts":{
> > "facet_queries":{},
> > "facet_ranges":{
> >   "Value":{
> > "counts":[
> >   "0.0",9,
> >   "100.0",0,
> >   "200.0",0,
> >   "300.0",0,
> >   "400.0",80,
> >   "500.0",0,
> >   "600.0",0,
> >   "700.0",69,
> >   "800.0",0,
> >   "900.0",0,
> >   "1000.0",0,
> >   "1100.0",0,
> >   "1200.0",0,
> >   "1300.0",0,
> >   "1400.0",0,
> >   "1500.0",0,
> >   "1600.0",0,
> >   "1700.0",0,
> >   "1800.0",0,
> >   "1900.0",9],
> > "gap":100.0,
> > "before":0,
> > "after":103,
> > "between":766,
> > "start":0.0,
> > "end":2000.0}
> >
> > Cheers,
> >
> > Dwane
> > 
> > From: Jan H?ydahl 
> > Sent: Friday, 7 September 2018 9:23:44 AM
> > To: solr-user@lucene.apache.org
> > Subject: Re: Solr range faceting
> >
> > Try facet.minCount=0
> >
> > Jan
> >
> > > 7. sep. 2018 kl. 01:07 skrev Dwane Hall :
> > >
> > > Good morning Solr community.  I'm having a few facet range issues for 
> > > which I'd appreciate some advice when somebody gets a spare couple of 
> > > minutes.
> > >
> > > Environment
> > > Solr Cloud (7.3.1)
> > > Single Shard Index, No replicas
> > >
> > > Facet Configuration (I'm using the request params API and useParams at 
>

Re: MLT in Cloud Mode - Not Returning Fields?

2018-09-07 Thread Erick Erickson
If this is still a problem in master/7x then a JIRA is in order.
On Fri, Sep 7, 2018 at 7:40 AM Doug Turnbull
 wrote:
>
> Looks like this is indeed a bug
>
> My colleague debugged this behavior and it turns out that Solr only
> requests id and score from the shards, and not the user-specified field
> list. You can see that on this line
>
> https://github.com/apache/lucene-solr/blob/branch_7_4/solr/core/src/java/org/apache/solr/handler/component/MoreLikeThisComponent.java#L342
>
> Happy to create a Jira ticket
>
> -Doug
>
> On Mon, Sep 3, 2018 at 5:23 AM Charlie Hull  wrote:
>
> > On 31/08/2018 19:36, Doug Turnbull wrote:
> > > Hello,
> > >
> > > We're working on a Solr More Like This project (Solr 6.6.2), using the
> > More
> > > Like This searchComponent. What we note is in standalone Solr, when we
> > > request MLT using the search component, we get every more like this
> > > document fully formed with complete fields in the moreLikeThis section.
> >
> > Hey Doug,
> >
> > IIRC there wasn't a lot of support for MLT in cloud mode a few years
> > ago, and there are certainly still a few open issues around cloud support:
> > https://issues.apache.org/jira/browse/SOLR-4414
> > https://issues.apache.org/jira/browse/SOLR-5480
> > Maybe there are some hints in the ticket comments about different ways
> > to do what you want.
> >
> > Cheers
> >
> > Charlie
> >
> > >
> > > In cloud, however, with the exact same query and config, we only get the
> > > doc ids under "moreLikeThis" requiring us to fetch the metadata
> > associated
> > > with each document.
> > >
> > > I can't easily share an example due to confidentiality, but I want to
> > check
> > > if we're missing something? Documentation doesn't mention any
> > limitations.
> > > The only interesting note I've found is this one which points to a
> > > potential difference in behavior
> > >
> > >>   The Cloud MLT Query Parser uses the realtime get handler to retrieve
> > the
> > > fields to be mined for keywords. Because of the way the realtime get
> > > handler is implemented, it does not return data for fields populated
> > using
> > > copyField.
> > >
> > > https://stackoverflow.com/a/46307140/8123
> > >
> > > Any thoughts?
> > >
> > > -Doug
> > >
> >
> >
> > --
> > Charlie Hull
> > Flax - Open Source Enterprise Search
> >
> > tel/fax: +44 (0)8700 118334 <+44%20870%20011%208334>
> > mobile:  +44 (0)7767 825828 <+44%207767%20825828>
> > web: www.flax.co.uk
> >
> --
> CTO, OpenSource Connections
> Author, Relevant Search
> http://o19s.com/doug


Re: Solr 7.4 and log4j2 JSONLayout

2018-09-07 Thread Shawn Heisey

On 9/6/2018 7:46 AM, Michael Aleythe, Sternwald wrote:

I'm trying to edit the log4j2 logging configuration for solr. The goal is to get a log 
file in json format. I configured the the JSONLayout for this purpose inside the 
rollingFile appender in the log4j2.xml. After this solr stops logging entirely. Solr.log 
file is empty. Only the solr-8983-console.log file contains 10 lines. The line 
"2018-09-06 13:22:25.378:INFO:oejs.Server:main: Started @2814ms" is the last 
one.
My first guess was that the jackson-core and jackson-databind jars were 
missing, but that did not fix the problem.


As Varun said, jackson is already included in Solr.  You won't need to 
add any jars for that.



Does anyone know where to find error-messages or exceptions that point me 
towards whats going wrong here?


Start Solr in the foreground, with the -f option.  This will cause Solr 
to log to the console.  When Solr is started in the background, it 
suppresses console logging.


I see that you have changed the logfile rollover size to 1MB.  If your 
Solr server sees much traffic, this is going to result in an extremely 
fast rollover, which may mean that you lose access to logged events VERY 
quickly.  This will especially be the case with JSON logging -- each 
event will take up a lot more space.


Thanks,
Shawn



Re: Expected mime type application/octet-stream but got text/html

2018-09-07 Thread Shawn Heisey

On 9/6/2018 12:29 PM, nalsrini wrote:

Here is the error message I am getting:

https://screencast.com/t/XwEjA22jX


Alexandre covered things quite well.  It looks like you're using an 
invalid URL.


Your original message shows this code:

SolrClient client = new HttpSolrClient("http://:/solr-master";);

That is not a valid URL.  It is missing the hostname, the port, and 
/solr-master is not the correct URL path.  The correct path would be 
/solr and MIGHT need to be /solr/indexname ... depending on exactly how 
the other code is written.  I prefer to use /solr and specify the 
collection name when making the requests.


Also, this is a Tomcat error page.  Is it Solr that's running in Tomcat, 
or are you running another application in Tomcat that is accessing Solr?


Running Solr in Tomcat became an unsupported configuration as of Solr 
5.0.  It is something you can do, but we don't support it.


https://wiki.apache.org/solr/WhyNoWar

Thanks,
Shawn



Re: Solr memory reqs for time-sorted data

2018-09-07 Thread Shawn Heisey

On 9/7/2018 8:39 AM, Pavel Micka wrote:

I found on wiki (https://wiki.apache.org/solr/SolrPerformanceProblems#RAM) that 
optimal amount of RAM for SOLR is equal to index size. This is lets say the 
ideal case to have everything in memory.

I wrote that page.


We plan to have small installation with 2 nodes and 8shards. We'll have inside 
the cluster 100M of documents. We expect that each document will take 5kB to 
index. With in-memory index this would mean that those two nodes would require 
~500GB RAM. This would mean 2x 256GB to have everything in memory. And those 
are really big machines... Is this calculation even correct in new Solr 
versions?

And we do have a bit restricted problem: Our data are time based logs and we 
generally have a restricted search for last 3 months. Which will match let's 
say 10M of documents. How will this affect SOLR memory requirements? Will we 
still need to have the whole inverted indexes in memory? Or is there some 
internal optimization, which will ensure that only some part will need to be in 
memory?

The questions:

1)  Is the 500GB of memory reqs correct assumption?


There are two things that Solr needs memory for.  One is Solr's heap, 
which is memory directly used by Solr itself.  The other is unused 
memory, which the operating system will use to cache data on disk.  Solr 
performance is helped dramatically by the latter kind of memory.


For *OPTIMAL* performance with a 500GB index, you need 500GB of memory 
for the OS to cache the data.  This is memory that is not used by 
programs, including Solr's heap.


For *good* performance, it's rare that you will need enough memory to 
cache the entire index.  But I cannot tell you with any reliability how 
much of the index you must be able to cache.  Some people are doing fine 
with only a few percent of their index cached.  Others see terrible 
performance unless they can get 75 percent of the index cached.



2)  Will the fact that we have time-based logs with majority of accesses to 
recent data only help?


Yes, it most likely will help, and reduce your memory requirements.


3)  Is there some best practice how to reduce required RAM in Solr?


The biggest thing you can do is to reduce the size of the index, so 
there is less data that must be accessed for a query. The page you 
referenced lists some things you might be able to do to reduce Solr's 
heap requirements.  If you reduce the heap requirements, then more of 
the server's memory available for caching.


Thanks,
Shawn



Re: Corrupt Index error on Target cluster

2018-09-07 Thread Stephen Bianamara
I've gotten incorrect checksums when upgrading solr versions across the
cluster. Or in other words, when indexing into a mixed version cluster. Are
you running mixed versions by chance?

On Fri, Sep 7, 2018, 6:07 AM Susheel Kumar  wrote:

> Anyone has  insight / have faced above errors ?
>
> On Thu, Sep 6, 2018 at 12:04 PM Susheel Kumar 
> wrote:
>
> > Hello,
> >
> > We had a running cluster with CDCR and there were some issues with
> > indexing on Source cluster which got resolved after restarting the nodes
> > (in my absence...) and now I see  below errors on a shard at Target
> > cluster.  Any suggestions / ideas what could have caused this and whats
> the
> > best way to recover.
> >
> > Thnx
> >
> > Caused by: org.apache.solr.common.SolrException: Error opening new
> searcher
> > at
> > org.apache.solr.core.SolrCore.openNewSearcher(SolrCore.java:2069)
> > at org.apache.solr.core.SolrCore.getSearcher(SolrCore.java:2189)
> > at org.apache.solr.core.SolrCore.getSearcher(SolrCore.java:1926)
> > at org.apache.solr.core.SolrCore.getSearcher(SolrCore.java:1826)
> > at
> >
> org.apache.solr.request.SolrQueryRequestBase.getSearcher(SolrQueryRequestBase.java:127)
> > at
> >
> org.apache.solr.handler.component.QueryComponent.process(QueryComponent.java:310)
> > at
> >
> org.apache.solr.handler.component.SearchHandler.handleRequestBody(SearchHandler.java:296)
> > at
> >
> org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:173)
> > at org.apache.solr.core.SolrCore.execute(SolrCore.java:2477)
> > at
> >
> org.apache.solr.handler.PingRequestHandler.handlePing(PingRequestHandler.java:267)
> > ... 34 more
> > Caused by: org.apache.lucene.index.CorruptIndexException: Corrupted
> > bitsPerDocBase: 6033
> >
> (resource=BufferedChecksumIndexInput(MMapIndexInput(path="/app/solr/data/COLL_shard8_replica1/data/index.20180903220548447/_9nsy.tvx")))
> > at
> >
> org.apache.lucene.codecs.compressing.CompressingStoredFieldsIndexReader.(CompressingStoredFieldsIndexReader.java:89)
> > at
> >
> org.apache.lucene.codecs.compressing.CompressingTermVectorsReader.(CompressingTermVectorsReader.java:126)
> > at
> >
> org.apache.lucene.codecs.compressing.CompressingTermVectorsFormat.vectorsReader(CompressingTermVectorsFormat.java:91)
> > at
> >
> org.apache.lucene.index.SegmentCoreReaders.(SegmentCoreReaders.java:128)
> > at
> > org.apache.lucene.index.SegmentReader.(SegmentReader.java:74)
> > at
> >
> org.apache.lucene.index.ReadersAndUpdates.getReader(ReadersAndUpdates.java:145)
> > at
> >
> org.apache.lucene.index.ReadersAndUpdates.getReadOnlyClone(ReadersAndUpdates.java:197)
> > at
> >
> org.apache.lucene.index.StandardDirectoryReader.open(StandardDirectoryReader.java:103)
> > at
> > org.apache.lucene.index.IndexWriter.getReader(IndexWriter.java:467)
> > at
> > org.apache.lucene.index.DirectoryReader.open(DirectoryReader.java:103)
> > at
> > org.apache.lucene.index.DirectoryReader.open(DirectoryReader.java:79)
> > at
> >
> org.apache.solr.core.StandardIndexReaderFactory.newReader(StandardIndexReaderFactory.java:39)
> > at
> > org.apache.solr.core.SolrCore.openNewSearcher(SolrCore.java:2033)
> > ... 43 more
> > Suppressed: org.apache.lucene.index.CorruptIndexException:
> > checksum failed (hardware problem?) : expected=e5bf0d15 actual=21722825
> >
> (resource=BufferedChecksumIndexInput(MMapIndexInput(path="/app/solr/data/COLL_shard8_replica1/data/index.20180903220548447/_9nsy.tvx")))
> > at
> > org.apache.lucene.codecs.CodecUtil.checkFooter(CodecUtil.java:419)
> > at
> > org.apache.lucene.codecs.CodecUtil.checkFooter(CodecUtil.java:462)
> > at
> >
> org.apache.lucene.codecs.compressing.CompressingTermVectorsReader.(CompressingTermVectorsReader.java:131)
> > ... 54 more
> >
>


Re: Keystore and Password is displayed in Dashboard

2018-09-07 Thread Steve Rowe
Hi,

Which version of Solr are you using?  AFAIK this was fixed in Solr 6.6: 
https://issues.apache.org/jira/browse/SOLR-10076 .

However, prior to Solr 7.0, sensitive property redaction was not enabled by 
default.  To enable, set system property solr.redaction.system.enabled=true .

--
Steve
www.lucidworks.com

> On Sep 7, 2018, at 9:16 AM, cyndefromva  wrote:
> 
> I have configured my solr instance to use ssl and it's working as expected.
> However, the keystore password is being displayed when I access the Solr
> Dashboard. Is there a way to hide this?
> 
> 
> 
> --
> Sent from: http://lucene.472066.n3.nabble.com/Solr-User-f472068.html



Re: MLT in Cloud Mode - Not Returning Fields?

2018-09-07 Thread Doug Turnbull
Looks like this is indeed a bug

My colleague debugged this behavior and it turns out that Solr only
requests id and score from the shards, and not the user-specified field
list. You can see that on this line

https://github.com/apache/lucene-solr/blob/branch_7_4/solr/core/src/java/org/apache/solr/handler/component/MoreLikeThisComponent.java#L342

Happy to create a Jira ticket

-Doug

On Mon, Sep 3, 2018 at 5:23 AM Charlie Hull  wrote:

> On 31/08/2018 19:36, Doug Turnbull wrote:
> > Hello,
> >
> > We're working on a Solr More Like This project (Solr 6.6.2), using the
> More
> > Like This searchComponent. What we note is in standalone Solr, when we
> > request MLT using the search component, we get every more like this
> > document fully formed with complete fields in the moreLikeThis section.
>
> Hey Doug,
>
> IIRC there wasn't a lot of support for MLT in cloud mode a few years
> ago, and there are certainly still a few open issues around cloud support:
> https://issues.apache.org/jira/browse/SOLR-4414
> https://issues.apache.org/jira/browse/SOLR-5480
> Maybe there are some hints in the ticket comments about different ways
> to do what you want.
>
> Cheers
>
> Charlie
>
> >
> > In cloud, however, with the exact same query and config, we only get the
> > doc ids under "moreLikeThis" requiring us to fetch the metadata
> associated
> > with each document.
> >
> > I can't easily share an example due to confidentiality, but I want to
> check
> > if we're missing something? Documentation doesn't mention any
> limitations.
> > The only interesting note I've found is this one which points to a
> > potential difference in behavior
> >
> >>   The Cloud MLT Query Parser uses the realtime get handler to retrieve
> the
> > fields to be mined for keywords. Because of the way the realtime get
> > handler is implemented, it does not return data for fields populated
> using
> > copyField.
> >
> > https://stackoverflow.com/a/46307140/8123
> >
> > Any thoughts?
> >
> > -Doug
> >
>
>
> --
> Charlie Hull
> Flax - Open Source Enterprise Search
>
> tel/fax: +44 (0)8700 118334 <+44%20870%20011%208334>
> mobile:  +44 (0)7767 825828 <+44%207767%20825828>
> web: www.flax.co.uk
>
-- 
CTO, OpenSource Connections
Author, Relevant Search
http://o19s.com/doug


Solr memory reqs for time-sorted data

2018-09-07 Thread Pavel Micka
Hi,

I found on wiki (https://wiki.apache.org/solr/SolrPerformanceProblems#RAM) that 
optimal amount of RAM for SOLR is equal to index size. This is lets say the 
ideal case to have everything in memory.

We plan to have small installation with 2 nodes and 8shards. We'll have inside 
the cluster 100M of documents. We expect that each document will take 5kB to 
index. With in-memory index this would mean that those two nodes would require 
~500GB RAM. This would mean 2x 256GB to have everything in memory. And those 
are really big machines... Is this calculation even correct in new Solr 
versions?

And we do have a bit restricted problem: Our data are time based logs and we 
generally have a restricted search for last 3 months. Which will match let's 
say 10M of documents. How will this affect SOLR memory requirements? Will we 
still need to have the whole inverted indexes in memory? Or is there some 
internal optimization, which will ensure that only some part will need to be in 
memory?

The questions:

1)  Is the 500GB of memory reqs correct assumption?

2)  Will the fact that we have time-based logs with majority of accesses to 
recent data only help?

3)  Is there some best practice how to reduce required RAM in Solr?



Thanks in advance!

Pavel


Side note:
We were thinking about DB partitioning based on Time Routed Aliases, but 
unfortunately we need to ensure disaster recovery through a bad network 
connection. And TRA and Cross Data Center Replication are not compatible. (CDCR 
requires static number of cores, while TRA creates cores dynamically).



RE: Re: Multi word searching is not working getting random search results

2018-09-07 Thread Muddapati, Jagadish
Hi Susheel,

Thanks for your response. Well If I use plural also it is giving the same 
results and the Solr is not finding the 2 different words that is in the same 
page. I am trying to figure out how to pass the query to find the results while 
doing multi word search in same page.

Thanks,
Jagadish M.

-Original Message-
From: Susheel Kumar [mailto:susheel2...@gmail.com] 
Sent: Thursday, September 6, 2018 3:01 PM
To: solr-user@lucene.apache.org
Subject: [EXTERNAL] Re: Multi word searching is not working getting random 
search results

How about you search with Intermodal Schedules (plural) & try phrase slop for 
better control on relevancy order

https://lucene.apache.org/solr/guide/6_6/the-extended-dismax-query-parser.html


On Thu, Sep 6, 2018 at 12:10 PM Muddapati, Jagadish < 
jagadish.muddap...@nscorp.com> wrote:

> Label: newbie
> Environment:
> I am currently running solr on Linux platform.
>
> NAME="Red Hat Enterprise Linux Server"
> VERSION="7.5"
>
> openjdk version "1.8.0_181"
>
> AEM version: 6.2
>
> I recently integrate solr to AEM and when i do search for multiple 
> words the search results are getting randomly.
>
> search words: Intermodal schedule
> Results: First solr displaying the search results related to 
> Intermodal and after few pages I am seeing the serch term schedule 
> related pages randomly. I am not getting the results related to multi words 
> on the page.
> For example: I am not seeing the results like [Terminals & Schedules | 
> Intermodal | Shipping Options ... page on starting and getting random 
> results and the  [Terminals & Schedules | Intermodal | Shipping Options ...
> page displaying after the 40 results.
>
> Here is the query on browser URL:
>
> http://secure-web.cisco.com/1YHJLnjxg38Ifakaan8ZFPHodNi4HbrQLcmhM5kTSr
> x5jOf2uH75_wUtu7AFAEP7KxUB5b6CrixqFq_S0f7HrOh3XWjXQ-Ulh0lCLiuuIfRXn0SG
> Cb4RFSBeyoQrKVcF6qdneCM8cn12qIiYp4k48VW-L2yHoLpFq3on--oSxJjw9n0ELt4iIo
> aT942NX3uB1ugWvTByJNd7dD2uv0FqAv3uLqPHCGs6apleaaXiLWjwnRO6f_yAMNTnfRy3
> QfUOyEi4KyrULJfGTS6Hs8hNewO86LHWjh6jbxMYJg_sbi8epNC0ptsjtDTj-6YOzHx7w/
> http%3A%2F%2Ftest-servername%2Fcontent%2Fnscorp%2Fen%2Fsearch-results.
> html%3Fstart%3D0%26q%3DIntermodal%2BSchedule
> <
> http://secure-web.cisco.com/1LoQdHzbU9XRxNqP3SAui7hMZDrw5utvOmhYY6FBfl
> yIR4Kn1pbTM5fxobgygd2KkgQ3MkCOlNCXtFBX48duxR6YmWb-pXqUKJavqTvE4V_9PpVm
> tJbE2B1bKJGFOXlEUqk2V4_k5NgmGIk4cxJHbOb0KWCxj9XD61t210oPjzOZUQnwzdZRnF
> jyAvNImN3Hmm8TIM_StSq5SnFeabqUcnhP73OIJF43s-Wt5tpPqobgdcAEEjf9Fy6oPivN
> 26n-AFiMOMveKIqiwxgX1NDgW1hDFIiuyM5bVr-JnXgXUuCrMHZxRZsDE4jthLDtJ44fb/
> http%3A%2F%2Fservername%2Fcontent%2Fnscorp%2Fen%2Fsearch-results.html%
> 3Fstart%3D0%26q%3DIntermodal%2BSchedule
> >
>
> I am using solr version 7.4
>
> Thanks,
> Jagadish M.
>
>
>


Keystore and Password is displayed in Dashboard

2018-09-07 Thread cyndefromva
I have configured my solr instance to use ssl and it's working as expected.
However, the keystore password is being displayed when I access the Solr
Dashboard. Is there a way to hide this?



--
Sent from: http://lucene.472066.n3.nabble.com/Solr-User-f472068.html


Re: Corrupt Index error on Target cluster

2018-09-07 Thread Susheel Kumar
Anyone has  insight / have faced above errors ?

On Thu, Sep 6, 2018 at 12:04 PM Susheel Kumar  wrote:

> Hello,
>
> We had a running cluster with CDCR and there were some issues with
> indexing on Source cluster which got resolved after restarting the nodes
> (in my absence...) and now I see  below errors on a shard at Target
> cluster.  Any suggestions / ideas what could have caused this and whats the
> best way to recover.
>
> Thnx
>
> Caused by: org.apache.solr.common.SolrException: Error opening new searcher
> at
> org.apache.solr.core.SolrCore.openNewSearcher(SolrCore.java:2069)
> at org.apache.solr.core.SolrCore.getSearcher(SolrCore.java:2189)
> at org.apache.solr.core.SolrCore.getSearcher(SolrCore.java:1926)
> at org.apache.solr.core.SolrCore.getSearcher(SolrCore.java:1826)
> at
> org.apache.solr.request.SolrQueryRequestBase.getSearcher(SolrQueryRequestBase.java:127)
> at
> org.apache.solr.handler.component.QueryComponent.process(QueryComponent.java:310)
> at
> org.apache.solr.handler.component.SearchHandler.handleRequestBody(SearchHandler.java:296)
> at
> org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:173)
> at org.apache.solr.core.SolrCore.execute(SolrCore.java:2477)
> at
> org.apache.solr.handler.PingRequestHandler.handlePing(PingRequestHandler.java:267)
> ... 34 more
> Caused by: org.apache.lucene.index.CorruptIndexException: Corrupted
> bitsPerDocBase: 6033
> (resource=BufferedChecksumIndexInput(MMapIndexInput(path="/app/solr/data/COLL_shard8_replica1/data/index.20180903220548447/_9nsy.tvx")))
> at
> org.apache.lucene.codecs.compressing.CompressingStoredFieldsIndexReader.(CompressingStoredFieldsIndexReader.java:89)
> at
> org.apache.lucene.codecs.compressing.CompressingTermVectorsReader.(CompressingTermVectorsReader.java:126)
> at
> org.apache.lucene.codecs.compressing.CompressingTermVectorsFormat.vectorsReader(CompressingTermVectorsFormat.java:91)
> at
> org.apache.lucene.index.SegmentCoreReaders.(SegmentCoreReaders.java:128)
> at
> org.apache.lucene.index.SegmentReader.(SegmentReader.java:74)
> at
> org.apache.lucene.index.ReadersAndUpdates.getReader(ReadersAndUpdates.java:145)
> at
> org.apache.lucene.index.ReadersAndUpdates.getReadOnlyClone(ReadersAndUpdates.java:197)
> at
> org.apache.lucene.index.StandardDirectoryReader.open(StandardDirectoryReader.java:103)
> at
> org.apache.lucene.index.IndexWriter.getReader(IndexWriter.java:467)
> at
> org.apache.lucene.index.DirectoryReader.open(DirectoryReader.java:103)
> at
> org.apache.lucene.index.DirectoryReader.open(DirectoryReader.java:79)
> at
> org.apache.solr.core.StandardIndexReaderFactory.newReader(StandardIndexReaderFactory.java:39)
> at
> org.apache.solr.core.SolrCore.openNewSearcher(SolrCore.java:2033)
> ... 43 more
> Suppressed: org.apache.lucene.index.CorruptIndexException:
> checksum failed (hardware problem?) : expected=e5bf0d15 actual=21722825
> (resource=BufferedChecksumIndexInput(MMapIndexInput(path="/app/solr/data/COLL_shard8_replica1/data/index.20180903220548447/_9nsy.tvx")))
> at
> org.apache.lucene.codecs.CodecUtil.checkFooter(CodecUtil.java:419)
> at
> org.apache.lucene.codecs.CodecUtil.checkFooter(CodecUtil.java:462)
> at
> org.apache.lucene.codecs.compressing.CompressingTermVectorsReader.(CompressingTermVectorsReader.java:131)
> ... 54 more
>


Re: SynonimGraphFilter expands wrong synonims

2018-09-07 Thread Andrea Gazzarini
And as you probably already checked, inserting the proper 
*tokenizerFactory* also expands the right synonym line:


q = (body:"Cytosolic 5'-nucleotidase II"  OR body:"EC 3.1.3.5")

parsedQuery = SpanOrQuery(spanOr([body:p49902, spanNear([body:cytosol, 
body:purin, body:5, body:nucleotidas], 0, true), spanNear([body:ec, 
body:3.1.3.5], 0, true), spanNear([body:cytosol, body:5, 
body:nucleotidas, body:ii], 0, true)])) SpanOrQuery(spanOr([body:p49902, 
spanNear([body:cytosol, body:purin, body:5, body:nucleotidas], 0, true), 
spanNear([body:cytosol, body:5, body:nucleotidas, body:ii], 0, true), 
spanNear([body:ec, body:3.1.3.5], 0, true)]))


Best,
Andrea

On 05/09/18 16:10, Andrea Gazzarini wrote:


You're right, my answer forgot to mention the *tokenizerFactory* 
parameter that you can add in the filter declaration. But, differently 
from what you think the default tokenizer used for parsing the 
synonyms _is not_ the tokenizer of the current analyzer 
(StandardTokenizer in your example) but WhitespaceTokenizer. See here 
[1] for a complete description of the filter capabilities.


So instead of switching the analyzer tokenizer you could also add a 
tokenizerFactory="solr.StandardTokenizerFactory" in the synonym filter 
declaration.


Best,
Andrea

[1] 
https://lucene.apache.org/solr/guide/6_6/filter-descriptions.html#FilterDescriptions-SynonymGraphFilter


On 05/09/2018 15:58, Danilo Tomasoni wrote:

Hi Andrea,

thank you for your answer.

About the second question: The standardTokenizer should be applied 
also to the phrase query, so the ' and - symbols should be removed 
even there, and this should allow a match in the synonim file isn't it?


With an example:


in phrase query:

"Cytosolic 5'-nucleotidase II" -> standardTokenizer -> Cytosolic, 5, 
nucleotidase, II



in synonym parsing:

...,Cytosolic 5'-nucleotidase II,... -> standardTokenizer -> 
Cytosolic, 5, nucleotidase, II



So the two graphs should match.. or I'm wrong?
Thank you
Danilo

ody:On 05/09/2018 13:23, Andrea Gazzarini wrote:

Hi Danilo,
let's see if this can help you (I'm sorry for the poor debugging, 
I'm reading & writing from my mobile): the first issue should have 
something to do with synonym overlapping and since I'm very curious 
about what it is happening, I will be more precise when I will be in 
front of a laptop.


The second: I guess the main problem is the StandardTokenizer, which 
removes the ' and - symbols. That should be the reason why you don't 
have any synonym detection. You should replace it with a 
WhitespaceTokenizer but, be aware that if you do that, the 
apostrophe in the document ( ′ ) is not the same symbol ( ' ) you've 
used in the query and in the synonyms file, so you need to replace 
it somewhere (in the document and/or in the query) otherwise you 
won't have any match.


HTH
Gazza

On 05/09/2018 12:19, Danilo Tomasoni wrote:

Hello to all,

I have an issue related to synonimgraphfilter expanding the wrong 
synonims for a phrase-term at query time.


I have a dictionary with the following lines

P49902,Cytosolic purine 5'-nucleotidase,EC 3.1.3.5,Cytosolic 
5'-nucleotidase II
A8K9N1,Glucosidase\, beta\, acid 3,Cytosolic,Glucosidase\, beta\, 
acid 3,Cytosolic\, isoform CRA_b,cDNA FLJ78196\, highly similar to 
Homo sapiens glucosidase\, beta\, acid 3,cytosolic,GBA3\, 
mRNA,cDNA\, FLJ93688\, Homo sapiens glucosidase\, beta\, acid 
3,cytosolic,GBA3\, mRNA


and two documents

{"body":"8. The method of claim 6 wherein said method inhibits at 
least one 5′-nucleotidase chosen from cytosolic 5′-nucleotidase II 
(cN-II), cytosolic 5′-nucleotidase IA (cN-IA), cytosolic 
5′-nucleotidase IB (cN-IB), cytosolic 5′-nucleotidase IMA 
(cN-IIIA), cytosolic 5′-nucleotidase NIB (cN-IIIB), 
ecto-5′-nucleotidase (eN, CD73), cytosolic 5′(3′)-deoxynucleotidase 
(cdN) and mitochondrial 5′(3′)-deoxynucleotidase (mdN)."}
{"body":"Trichomonosis caused by the flagellate protozoan 
Trichomonas vaginalis represents the most prevalent nonviral 
sexually transmitted disease worldwide (WHO-DRHR 2012). In women, 
the symptoms are cyclic and often worsen around the menstruation 
period. In men, trichomonosis is largely asymptomatic and these men 
are considered to be carriers of T. vaginalis (Petrin et al. 1998). 
This infection has been associated with birth outcomes (Klebanoff 
et al. 2001), infertility (Grodstein et al. 1993), cervical and 
prostate cancer (Viikki et al. 2000, Sutcliffe et al. 2012) and 
pelvic inflammatory disease (Cherpes et al. 2006). Importantly, T. 
vaginalis is a co-factor in human immunodeficiency virus 
transmission and acquisition (Sorvillo et al. 2001, Van Der Pol et 
al. 2008). Therefore, it is important to study the host-parasite 
relationship to understand T. vaginalis infection and pathogenesis. 
Colonisation of the mucosa by T. vaginalis is a complex multi-step 
process that involves distinct mechanisms (Alderete et al. 2004). 
The parasite interacts with mucin (Lehker & Sweeney 1999), adheres 
to vaginal

Solr CDCR replication not working

2018-09-07 Thread Mrityunjaya Pathak
I have setup two solr cloud instances in two different Datacenters Target solr 
cloud machine is copy of source machine with basicAuth enabled on them. I am 
unable to see any replication on target.

Solr Version :6.6.3

I have done config changes as suggested on 
https://lucene.apache.org/solr/guide/6_6/cross-data-center-replication-cdcr.html

Source Config Changes



...

  
serverIP:2181,serverIP:2182,serverIP:2183
sitecore_master_index
sitecore_master_index
  

  
8
1000
128
  

  
1000
  

  


  ${solr.ulog.dir:}
  ${solr.ulog.numVersionBuckets:65536}




  ${solr.autoCommit.maxTime:15000}
  false



  ${solr.autoSoftCommit.maxTime:-1}

  

  ...
  

Target Config Changes



...

  
disabled
  


  
  


  
cdcr-proc-chain
  

 


  ${solr.ulog.dir:}
  ${solr.ulog.numVersionBuckets:65536}



  ${solr.autoCommit.maxTime:15000}
  false



  ${solr.autoSoftCommit.maxTime:-1}


  

  ...
  

Below are logs from Source target.

ERROR (zkCallback-4-thread-2-processing-n:sourceIP:8983_solr) [   ] 
o.a.s.c.s.i.CloudSolrClient Request to collection collection1 failed due to 
(510) org.apache.solr.common.SolrException: Could not find a healthy node to 
handle the request., retry? 5
2018-09-07 10:36:14.295 WARN  
(zkCallback-4-thread-2-processing-n:sourceIP:8983_solr) [   ] 
o.a.s.h.CdcrReplicatorManager Unable to instantiate the log reader for target 
collection collection1
org.apache.solr.common.SolrException: Could not find a healthy node to handle 
the request.
at 
org.apache.solr.client.solrj.impl.CloudSolrClient.sendRequest(CloudSolrClient.java:1377)
at 
org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:1134)
at 
org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:1237)
at 
org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:1237)
at 
org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:1237)
at 
org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:1237)
at 
org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:1237)
at 
org.apache.solr.client.solrj.impl.CloudSolrClient.request(CloudSolrClient.java:1073)
at org.apache.solr.client.solrj.SolrClient.request(SolrClient.java:1219)
at 
org.apache.solr.handler.CdcrReplicatorManager.getCheckpoint(CdcrReplicatorManager.java:196)
at 
org.apache.solr.handler.CdcrReplicatorManager.initLogReaders(CdcrReplicatorManager.java:159)
at 
org.apache.solr.handler.CdcrReplicatorManager.stateUpdate(CdcrReplicatorManager.java:134)
at 
org.apache.solr.handler.CdcrStateManager.callback(CdcrStateManager.java:36)
at 
org.apache.solr.handler.CdcrLeaderStateManager.setAmILeader(CdcrLeaderStateManager.java:108)
at 
org.apache.solr.handler.CdcrLeaderStateManager.checkIfIAmLeader(CdcrLeaderStateManager.java:95)
at 
org.apache.solr.handler.CdcrLeaderStateManager.access$400(CdcrLeaderStateManager.java:40)
at 
org.apache.solr.handler.CdcrLeaderStateManager$LeaderStateWatcher.process(CdcrLeaderStateManager.java:150)
at 
org.apache.solr.common.cloud.SolrZkClient$3.lambda$process$0(SolrZkClient.java:269)
at 
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at 
org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor.lambda$execute$0(ExecutorUtil.java:229)
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
2018-09-07 10:36:14.310 INFO  
(coreLoadExecutor-8-thread-3-processing-n:sourceIP:8983_solr) [   ] 
o.a.s.c.SolrConfig Using Lucene MatchVersion: 6.6.3
2018-09-07 10:36:14.315 INFO  
(zkCallback-4-thread-1-processing-n:sourceIP:8983_solr) [   ] 
o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent 
state:SyncConnected type:NodeDataChanged 
path:/collections/collection1/state.json] for collection [sitecore] has 
occurred - updating... (live nodes size: [1])
2018-09-07 10:36:14.343 WARN  
(cdcr-replicator-211-thread-1-processing-n:sourceIP:8983_solr) [   ] 
o.a.s.h.CdcrReplicator Log reader for target collection1 is not initialised, it 
will be ignored.

I am unable to see anything on target. It will be great if someone can help me 
in it.
Can you help in this?
Regards
Mrityunjaya


Solr CDCR replication not working

2018-09-07 Thread Mrityunjaya Pathak
I have setup two solr cloud instances in two different Datacenters Target solr 
cloud machine is copy of source machine with basicAuth enabled on them. I am 
unable to see any replication on target.

Solr Version :6.6.3

I have done config changes as suggested on 
https://lucene.apache.org/solr/guide/6_6/cross-data-center-replication-cdcr.html

Source Config Changes



...

  
serverIP:2181,serverIP:2182,serverIP:2183
sitecore_master_index
sitecore_master_index
  

  
8
1000
128
  

  
1000
  

  


  ${solr.ulog.dir:}
  ${solr.ulog.numVersionBuckets:65536}




  ${solr.autoCommit.maxTime:15000}
  false



  ${solr.autoSoftCommit.maxTime:-1}

  

  ...
  

Target Config Changes



...

  
disabled
  


  
  


  
cdcr-proc-chain
  

 


  ${solr.ulog.dir:}
  ${solr.ulog.numVersionBuckets:65536}



  ${solr.autoCommit.maxTime:15000}
  false



  ${solr.autoSoftCommit.maxTime:-1}


  

  ...
  

Below are logs from Source target.

ERROR (zkCallback-4-thread-2-processing-n:sourceIP:8983_solr) [   ] 
o.a.s.c.s.i.CloudSolrClient Request to collection collection1 failed due to 
(510) org.apache.solr.common.SolrException: Could not find a healthy node to 
handle the request., retry? 5
2018-09-07 10:36:14.295 WARN  
(zkCallback-4-thread-2-processing-n:sourceIP:8983_solr) [   ] 
o.a.s.h.CdcrReplicatorManager Unable to instantiate the log reader for target 
collection collection1
org.apache.solr.common.SolrException: Could not find a healthy node to handle 
the request.
at 
org.apache.solr.client.solrj.impl.CloudSolrClient.sendRequest(CloudSolrClient.java:1377)
at 
org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:1134)
at 
org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:1237)
at 
org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:1237)
at 
org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:1237)
at 
org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:1237)
at 
org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:1237)
at 
org.apache.solr.client.solrj.impl.CloudSolrClient.request(CloudSolrClient.java:1073)
at org.apache.solr.client.solrj.SolrClient.request(SolrClient.java:1219)
at 
org.apache.solr.handler.CdcrReplicatorManager.getCheckpoint(CdcrReplicatorManager.java:196)
at 
org.apache.solr.handler.CdcrReplicatorManager.initLogReaders(CdcrReplicatorManager.java:159)
at 
org.apache.solr.handler.CdcrReplicatorManager.stateUpdate(CdcrReplicatorManager.java:134)
at 
org.apache.solr.handler.CdcrStateManager.callback(CdcrStateManager.java:36)
at 
org.apache.solr.handler.CdcrLeaderStateManager.setAmILeader(CdcrLeaderStateManager.java:108)
at 
org.apache.solr.handler.CdcrLeaderStateManager.checkIfIAmLeader(CdcrLeaderStateManager.java:95)
at 
org.apache.solr.handler.CdcrLeaderStateManager.access$400(CdcrLeaderStateManager.java:40)
at 
org.apache.solr.handler.CdcrLeaderStateManager$LeaderStateWatcher.process(CdcrLeaderStateManager.java:150)
at 
org.apache.solr.common.cloud.SolrZkClient$3.lambda$process$0(SolrZkClient.java:269)
at 
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at 
org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor.lambda$execute$0(ExecutorUtil.java:229)
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
2018-09-07 10:36:14.310 INFO  
(coreLoadExecutor-8-thread-3-processing-n:sourceIP:8983_solr) [   ] 
o.a.s.c.SolrConfig Using Lucene MatchVersion: 6.6.3
2018-09-07 10:36:14.315 INFO  
(zkCallback-4-thread-1-processing-n:sourceIP:8983_solr) [   ] 
o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent 
state:SyncConnected type:NodeDataChanged 
path:/collections/collection1/state.json] for collection [sitecore] has 
occurred - updating... (live nodes size: [1])
2018-09-07 10:36:14.343 WARN  
(cdcr-replicator-211-thread-1-processing-n:sourceIP:8983_solr) [   ] 
o.a.s.h.CdcrReplicator Log reader for target collection1 is not initialised, it 
will be ignored.

I am unable to see anything on target. It will be great if someone can help me 
in it.
Can you help in this?
Regards
Mrityunjaya


Re: regarding 'sharedlib' in solr

2018-09-07 Thread Vincenzo D'Amore
Usually sharedlib is a path to a directory that contains .jar files, then
added to the Solr classpath.

For example I've used sharedlib to deploy a jar where is a the
implementation of a custom plugin for Solr that extends
ExtendedDismaxQParserPlugin.



On Fri, Sep 7, 2018 at 12:57 PM Andrea Gazzarini 
wrote:

> Hi, please expand a bit. Specifically:
>
>   * what are those text files? Configuration files? You want something
> like a central point where to manage things like stopwords, synonyms?
>   * I don't think the shareLib folder has been created for this usage.
> However, please post the complete message you're receiving, together
> with the stacktrace, if available.
>
> Andrea
>
> On 07/09/18 12:02, Abhishek Agarwal wrote:
> > Hi,
> >
> > I want to share a folder containing text files in solr among different
> > cores so if the folder is updated ,so it would reflect in all the cores
> > having path specified but the problem I am facing is that , i am using
> > sharedlib in solr.xml and specifying the default path there.And also I am
> > updating schema.xml of my core but when I am loading core ,it is giving
> > error 'unsafe loading' .and not getting reload. Please help me in this.
> >
>
>

-- 
Vincenzo D'Amore


Re: regarding 'sharedlib' in solr

2018-09-07 Thread Andrea Gazzarini

Hi, please expand a bit. Specifically:

 * what are those text files? Configuration files? You want something
   like a central point where to manage things like stopwords, synonyms?
 * I don't think the shareLib folder has been created for this usage.
   However, please post the complete message you're receiving, together
   with the stacktrace, if available.

Andrea

On 07/09/18 12:02, Abhishek Agarwal wrote:

Hi,

I want to share a folder containing text files in solr among different
cores so if the folder is updated ,so it would reflect in all the cores
having path specified but the problem I am facing is that , i am using
sharedlib in solr.xml and specifying the default path there.And also I am
updating schema.xml of my core but when I am loading core ,it is giving
error 'unsafe loading' .and not getting reload. Please help me in this.





regarding 'sharedlib' in solr

2018-09-07 Thread Abhishek Agarwal
Hi,

I want to share a folder containing text files in solr among different
cores so if the folder is updated ,so it would reflect in all the cores
having path specified but the problem I am facing is that , i am using
sharedlib in solr.xml and specifying the default path there.And also I am
updating schema.xml of my core but when I am loading core ,it is giving
error 'unsafe loading' .and not getting reload. Please help me in this.

-- 
Thanks,
Abhishek Agarwal


RE: Replicas do not come up after nodes are restarted in SOLR cloud

2018-09-07 Thread Sudip Mukherjee
Hi Shawn,

Thanks for the insights. 

I verified the points you mentioned,
Socket timeout defaults weren't changed.

Both nodes (two different hosts, windows OS ) are given heap space of 4GB. They 
are having two Collections as of now.
One is without replicas but with 8 shards on each node.
One is with replicas and 1 shard. (both replicas are down)

I monitored the application for some time and I do not see obvious memory 
issues or prolonged garbage collection.

Also, the nodes are registered with their hostnames in the collection.


Thanks,
Sudip

-Original Message-
From: Shawn Heisey [mailto:elyog...@elyograg.org] 
Sent: Wednesday, September 05, 2018 7:23 PM
To: solr-user@lucene.apache.org
Subject: Re: Replicas do not come up after nodes are restarted in SOLR cloud

On 9/5/2018 2:55 AM, Sudip Mukherjee wrote:
> I have a 2 node SOLR (7.x) cloud cluster on which I have collection 
> with replicas ( replicationFactor = 2, shard = 1 ). I am seeing that the 
> replicas do not come up ( state is "down")  when both nodes are restarted. 
> From the "legend" in Graph section, I see that the replicas are in "recovery 
> failed" state.

> Caused by: java.net.SocketTimeoutException: Read timed out
>  at java.net.SocketInputStream.socketRead0(Native Method)
>  at 
> java.net.SocketInputStream.socketRead(SocketInputStream.java:116)

Have you changed the socket timeout in Solr's config?

The socket timeout for internode requests defaults to 60 seconds.  If something 
happened that prevented a Solr server from responding within
60 seconds, then there's something *REALLY* wrong.

My best guess is that your Solr heap is too small, causing Java to spend almost 
all of its time doing garbage collection.  Or that a too-small heap has caused 
one of your servers to experience an OutOfMemoryError, which on non-Windows 
systems will result in the Solr process being killed.

Some questions in case that's not it:

How many collections do you have on this setup?

In the admin UI (Cloud tab), what hostname do your nodes show they are 
registered as?  If it's localhost, that's going to be a problem for a 2-node 
cluster.

Thanks,
Shawn

***Legal Disclaimer***
"This communication may contain confidential and privileged material for the
sole use of the intended recipient. Any unauthorized review, use or distribution
by others is strictly prohibited. If you have received the message by mistake,
please advise the sender by reply email and delete the message. Thank you."
**


Solr upgrade issues

2018-09-07 Thread Srinivas Kashyap
Hi,

We are in the process of upgrading Solr from solr 5.2.1 to solr 7.4.0 and I'm 
facing below issues. Please help me in resolving.

1)HttpSolrClient tempClient = new 
HttpSolrClient.Builder("http://localhost:8983/solr";).build();

ModifiableSolrParams params = new 
ModifiableSolrParams();
   params.set("qt", "/dataimport");
   params.set("command", "delta-import");
   params.set("clean", "false");
   params.set("commit", "true");
   params.set("entity", "collection1");

GenericSolrRequest req = new 
GenericSolrRequest(SolrRequest.METHOD.POST, "/select" ,params);
   tempClient.request(req," collection1");


In the back end I have schedulers which call delta-import on my collection1. 
This used to work before in 5.2.1 and when this code gets executed, I could see 
the delta-import was being run. Now in 7.4.0, I'm not able to see the 
delta-import running. Does solr restricts taking external requests in 7.4.0?

2) Also in Jetty.xml , I have configured the datasource as below


  
 java:comp/env/jdbc/tssindex
 
 
 thin
 jdbc:oracle:thin:@localhost:1521:xe
 X
 X
 
 
 

How to fetch the data source configured in this xml 
(java:comp/env/jdbc/tssindex)

Earlier I used to fetch from tomcat context xml as below:

Context initContext = new InitialContext();
DataSource ds = null;
Context webContext = 
(Context)initContext.lookup("java:/comp/env");
ds = (DataSource) webContext.lookup("jdbc/tssindex");

How to fetch it in jetty.

Thanks in advance,
Srinivas kashyap



Re: Solr range faceting

2018-09-07 Thread Dwane Hall
Thanks Erick,

The field is defined as a pfloat.



I took your advice and tried smaller result range and the counts look good.  I 
might try an index rebuild I’m wondering if the data has somehow been corrupted 
by a combination of old and new index mappings.

Thanks again for your assistance.


"responseHeader":{
"zkConnected":true,
"status":0,
"QTime":7},
  "response":{"numFound":3,"start":0,"docs":[
  {
"Value":34.0},
  {
"Value":34.0},
  {
"Value":34.0}]
  },
  "facet_counts":
"facet_queries":{},
"facet_ranges":{
  "Value:{
"counts":[
  "0.0",3],
"gap":100.0,
"before":0,
"after":0,
"between":3,
"start":0.0,
"end":2000.0}}

From: Erick Erickson 
Sent: Friday, 7 September 2018 12:54:48 PM
To: solr-user
Subject: Re: Solr range faceting

Indeed this doesn't look right. By my count, you're missing 599 counts
you'd expect in that range, although the after and between numbers
total the numFound.

What kind of a field is Value? Given the number of docs missing, I'd
guess you could get the number of docs down really small and post
them. Something like
values 1, 2, 3, 4, 5, 
and your range query so we could try it.

What is the fieldType definition and field for Value?

And finally, do you get different results if you use json faceting?

Best,
Erick
On Thu, Sep 6, 2018 at 5:51 PM Dwane Hall  wrote:
>
> Thanks Jan that has fixed the bucket issue but I'm a little confused at why 
> zero counts exist for some buckets when they appear to be values in them?
>
> "response":{"numFound":869,"start":0,"docs":[
>   {
> "Value":9475.08},
>   {
> "Value":780.0},
>   {
> "Value":9475.08},
>   {
> "Value":1000.0},
>   {
> "Value":50.0},
>   {
> "Value":50.0},
>   {
> "Value":0.0},
>   {
> "Value":800.0},
>   {
> "Value":0.0},
>   {
> "Value":1000.0},
>   {
> "Value":1000.0},
>   {
> "Value":5000.0},
>   {
> "Value":2000.0},
>   {
>"Value":4000.0},
>   {
> "Value":1500.0},
>   {
> "Value":0.0},
>   {
> "Value":1.0},
>   {
> "Value":5000.0},
>   {
> "Value":1000.0},
>   {
> "Value":0.0},
>   {
> "Value":1200.0},
>   {
> "Value":9000.0},
>   {
> "Value":1500.0},
>   {
> "Value":1.0},
>   {
> "Value":5000.0},
>   {
> "Value":4000.0},
>   {
> "Value":5000.0},
>   {
> "Value":5000.0},
>   {
> "Value":1.0},
>   {
> "Value":1000.0}]
>   },
>
>   "facet_counts":{
> "facet_queries":{},
> "facet_ranges":{
>   "Value":{
> "counts":[
>   "0.0",9,
>   "100.0",0,
>   "200.0",0,
>   "300.0",0,
>   "400.0",80,
>   "500.0",0,
>   "600.0",0,
>   "700.0",69,
>   "800.0",0,
>   "900.0",0,
>   "1000.0",0,
>   "1100.0",0,
>   "1200.0",0,
>   "1300.0",0,
>   "1400.0",0,
>   "1500.0",0,
>   "1600.0",0,
>   "1700.0",0,
>   "1800.0",0,
>   "1900.0",9],
> "gap":100.0,
> "before":0,
> "after":103,
> "between":766,
> "start":0.0,
> "end":2000.0}
>
> Cheers,
>
> Dwane
> 
> From: Jan H?ydahl 
> Sent: Friday, 7 September 2018 9:23:44 AM
> To: solr-user@lucene.apache.org
> Subject: Re: Solr range faceting
>
> Try facet.minCount=0
>
> Jan
>
> > 7. sep. 2018 kl. 01:07 skrev Dwane Hall :
> >
> > Good morning Solr community.  I'm having a few facet range issues for which 
> > I'd appreciate some advice when somebody gets a spare couple of minutes.
> >
> > Environment
> > Solr Cloud (7.3.1)
> > Single Shard Index, No replicas
> >
> > Facet Configuration (I'm using the request params API and useParams at 
> > runtime)
> > "facet":"true",
> > "facet.mincount":1,
> > "facet.missing":"false",
> > "facet.range":"Value"
> > "f.Value.facet.range.start":0.0,
> > "f.Value.facet.range.end":2000.0,
> > "f.Value.facet.range.gap":100,
> > "f.Value.facet.range.include":"edge",
> > "f.Value.facet.range.other":"all",
> >
> > My problem
> > With my range facet configuration I'm expecting to see a facet range entry 
> > for every 'step' (100 in my case) between my facet.range.start and 
> > facet.range.end settings. Something like the following 0.0,100.0,200.0, 
> > ...2000.0 with a sum of the number of values that occur between each range 
> > step.  This does not appear to be the case and in some instances I don't 
> > appear to get counts for some range steps (800.0 and 1000.0 for example are 
> > present in my result set range below but I don't get a range value facets 
> > for these values?)
> >