Re: Solr AutoSuggest Configuration Issue

2018-10-08 Thread Zheng Lin Edwin Yeo
The link that you are referring to is for Solr 6.6, but you are using Solr
5.1 which is quite an old version, so there could be some differences.
You can refer this guide for Solr 5.1:
http://archive.apache.org/dist/lucene/solr/ref-guide/apache-solr-ref-guide-5.1.pdf

The current version of Solr is already Solr 7.5, and it is recommended to
upgrade to the new version so that you can use the new features and and
also things like better memory consumption and better authentication

Regards,
Edwin

On Mon, 8 Oct 2018 at 20:20, Manu Nair  wrote:

> Hi,
>
> I am using Solr 5.1 for my application.
> I am trying to use the autoSuggest feature of Solr.
> I want to do context filtering on the results returned by Solr suggest.
>
> Please help me know if this feature is supported in the version that I am
> using(5.1).
> Also if it works with multivalued field. I tried multiple times but it is
> not working.
>
> I am referring the following link for details :
> https://lucene.apache.org/solr/guide/6_6/suggester.html
>
> Please find the configuration in my solrconfig.xml as below
> 
>   
> mySuggester
> AnalyzingInfixLookupFactory
> DocumentDictionaryFactory
> name
> price
> text_en
> false
> countries
>   
> 
>
> Thanks alot for your help in advance.
>
> Regards,
> Manu Nair.
>


Re: not able to use deleteDocByQuery: please provide detailed example

2018-10-08 Thread Erick Erickson
Chris:

Please follow the instructions here:
http://lucene.apache.org/solr/community.html#mailing-lists-irc. You
must use the _exact_ same e-mail as you used to subscribe.

If the initial try doesn't work and following the suggestions at the
"problems" link doesn't work for you, let us know. But note you need
to show us the _entire_ return header to allow anyone to diagnose the
problem.

Best,
Erick
On Mon, Oct 8, 2018 at 2:48 PM Chris Gerke  wrote:
>
> UNSUBSCRIBE
>
> On Tue, 9 Oct 2018 at 4:36 am, Sudheer Shinde 
> wrote:
>
> > solr 6.2.1:
> >
> > my schema:
> >
> > id is primary key here:
> >
> >   > required="true" multiValued="false"/>
> >  > name="TC_0Y0_ItemRevision_0Y0_awp0Item_item_id" required="false"
> > stored="false" type="string"/>
> >  > name="TC_0Y0_Item_uid" required="false" stored="false" type="string"/>
> >  > name="TC_0Y0_ProductScope_0Y0_product_uid" required="false" stored="true"
> > type="string"/>
> >
> >
> >
> > I want to delete using field TC_0Y0_ProductScope_0Y0_product_uid:
> >
> > my db-data config file:
> >  > transformer="TemplateTransformer"
> > query="Select pproduct from PQSEARCHINDEXEDPRODUCT t1 where t1.pis_deleted
> > = 1" >
> >  >
> > template="TC_0Y0_ProductScope_0Y0_product_uid:${onlydeleteByQuery.pproduct}"/>
> > 
> > 
> > 
> > 
> > 
> >


Re: not able to use deleteDocByQuery: please provide detailed example

2018-10-08 Thread Chris Gerke
UNSUBSCRIBE

On Tue, 9 Oct 2018 at 4:36 am, Sudheer Shinde 
wrote:

> solr 6.2.1:
>
> my schema:
>
> id is primary key here:
>
>   required="true" multiValued="false"/>
>  name="TC_0Y0_ItemRevision_0Y0_awp0Item_item_id" required="false"
> stored="false" type="string"/>
>  name="TC_0Y0_Item_uid" required="false" stored="false" type="string"/>
>  name="TC_0Y0_ProductScope_0Y0_product_uid" required="false" stored="true"
> type="string"/>
>
>
>
> I want to delete using field TC_0Y0_ProductScope_0Y0_product_uid:
>
> my db-data config file:
>  transformer="TemplateTransformer"
> query="Select pproduct from PQSEARCHINDEXEDPRODUCT t1 where t1.pis_deleted
> = 1" >
> 
> template="TC_0Y0_ProductScope_0Y0_product_uid:${onlydeleteByQuery.pproduct}"/>
> 
> 
> 
> 
> 
>


Re: checksum failed (hardware problem?)

2018-10-08 Thread Stephen Bianamara
Hi Susheel,

Yes, I believe you are correct on fixing a node in place. My org actually
just cycles instances rather than repair broken ones.

It's too bad that there's nothing conclusive we can look for to help
investigate the scope. We'd love to pin this down so that we could take
something concrete to investigate to AWS if it's a hardware failure (e.g.,
we found a log indicating ). I haven't been able to find anything which
might clarify that matter outside of SOLR either. Perhaps it's just not
realistic at this time.

I'm also curious about another aspect, which is that the nodes don't report
as unhealthy. Currently a node with a bad checksum will just stay in the
collection forever. Shouldn't the node go to "down" if it has an
irreparable checksum?

On Fri, Oct 5, 2018 at 5:25 AM Susheel Kumar  wrote:

> My understanding is once the index is corrupt, the only way to fix is using
> checkindex utility which will remove some bad segments and then only we can
> use it.
>
> This is bit scary that you see similar error on 6.6.2 though in our case we
> know we are going thru some hardware problem which likely would have caused
> the corruption but there is no concrete evidence which can be used to
> confirm if it is hardware or Solr/Lucene.  Are you able to use another AWS
> instance similar to Simon's case.
>
> Thanks,
> Susheel
>
> On Thu, Oct 4, 2018 at 7:11 PM Stephen Bianamara 
> wrote:
>
> > To be more concrete: Is the definitive test of whether or not a core's
> > index is corrupt to copy it onto a new set of hardware and attempt to
> write
> > to it? If this is a definitive test, we can run the experiment and update
> > the report so you have a sense of how often this happens.
> >
> > Since this is a SOLR cloud node, which is already removed but whose data
> > dir was preserved, I believe I can just copy the data directory to a
> fresh
> > machine and start a regular non-cloud solr node hosting this core. Can
> you
> > please confirm that this will be a definitive test, or whether there is
> > some aspect needed to make it definitive?
> >
> > Thanks!
> >
> > On Wed, Oct 3, 2018 at 2:10 AM Stephen Bianamara  >
> > wrote:
> >
> > > Hello All --
> > >
> > > As it would happen, we've seen this error on version 6.6.2 very
> recently.
> > > This is also on an AWS instance, like Simon's report. The drive doesn't
> > > show any sign of being unhealthy, either from cursory investigation.
> > FWIW,
> > > this occurred during a collection backup.
> > >
> > > Erick, is there some diagnostic data we can find to help pin this down?
> > >
> > > Thanks!
> > > Stephen
> > >
> > > On Sun, Sep 30, 2018 at 12:48 PM Susheel Kumar 
> > > wrote:
> > >
> > >> Thank you, Simon. Which basically points that something related to env
> > and
> > >> was causing the checksum failures than any lucene/solr issue.
> > >>
> > >> Eric - I did check with hardware folks and they are aware of some
> VMware
> > >> issue where the VM hosted in HCI environment is coming into some halt
> > >> state
> > >> for minute or so and may be loosing connections to disk/network.  So
> > that
> > >> probably may be the reason of index corruption though they have not
> been
> > >> able to find anything specific from logs during the time Solr run into
> > >> issue
> > >>
> > >> Also I had again issue where Solr is loosing the connection with
> > zookeeper
> > >> (Client session timed out, have not heard from server in 8367ms for
> > >> sessionid 0x0)  Does that points to similar hardware issue, Any
> > >> suggestions?
> > >>
> > >> Thanks,
> > >> Susheel
> > >>
> > >> 2018-09-29 17:30:44.070 INFO
> > >> (searcherExecutor-7-thread-1-processing-n:server54:8080_solr
> > >> x:COLL_shard4_replica2 s:shard4 c:COLL r:core_node8) [c:COLL s:shard4
> > >> r:core_node8 x:COLL_shard4_replica2] o.a.s.c.SolrCore
> > >> [COLL_shard4_replica2] Registered new searcher
> > >> Searcher@7a4465b1[COLL_shard4_replica2]
> > >>
> > >>
> >
> main{ExitableDirectoryReader(UninvertingDirectoryReader(Uninverting(_7x3f(6.6.2):C826923/317917:delGen=2523)
> > >> Uninverting(_83pb(6.6.2):C805451/172968:delGen=2957)
> > >> Uninverting(_3ywj(6.6.2):C727978/334529:delGen=2962)
> > >> Uninverting(_7vsw(6.6.2):C872110/385178:delGen=2020)
> > >> Uninverting(_8n89(6.6.2):C741293/109260:delGen=3863)
> > >> Uninverting(_7zkq(6.6.2):C720666/101205:delGen=3151)
> > >> Uninverting(_825d(6.6.2):C707731/112410:delGen=3168)
> > >> Uninverting(_dgwu(6.6.2):C760421/295964:delGen=4624)
> > >> Uninverting(_gs5x(6.6.2):C540942/138952:delGen=1623)
> > >> Uninverting(_gu6a(6.6.2):c75213/35640:delGen=1110)
> > >> Uninverting(_h33i(6.6.2):c131276/40356:delGen=706)
> > >> Uninverting(_h5tc(6.6.2):c44320/11080:delGen=380)
> > >> Uninverting(_h9d9(6.6.2):c35088/3188:delGen=104)
> > >> Uninverting(_h80h(6.6.2):c11927/3412:delGen=153)
> > >> Uninverting(_h7ll(6.6.2):c11284/1368:delGen=205)
> > >> Uninverting(_h8bs(6.6.2):c11518/2103:delGen=149)
> > >> Uninverting(_h9r3(6.6.2):c16439/1018:delGen=52)
> > >> Uninver

not able to use deleteDocByQuery: please provide detailed example

2018-10-08 Thread Sudheer Shinde
my schema:

id is primary key here:

 






I want to delete using field TC_0Y0_ProductScope_0Y0_product_uid:

my db-data config file:









not able to use deleteDocByQuery: please provide detailed example

2018-10-08 Thread Sudheer Shinde
solr 6.2.1:

my schema:

id is primary key here:

 






I want to delete using field TC_0Y0_ProductScope_0Y0_product_uid:

my db-data config file:









ComplexPhraseQParser vs phrase slop

2018-10-08 Thread David Santamauro
Hi, quick question. Should

  1) {!complexphrase inOrder=false}f: ( "cat jump"~2 )

... and

  2) f: ( "cat jump"~2 )

... yield the same results? I'm trying to diagnose a more complicated 
discrepancy that I've boiled down to this simple case. I understand #1 creates 
a SpanQuery and #2 a PhraseQuery but I would have thought without wildcards and 
with the attribute inOrder=false that both would/should yield the exact same 
results. If they should ( and they aren't for me ) what could the problem be? 
If the shouldn't, could someone explain why?

Thanks



Re: Indexing documents from S3 bucket

2018-10-08 Thread ☼ R Nair
S3 gives listeners. So tap those listeners when objects are added, updated
or deleted and use Solr API to push. That's high level, but I believe
doable.

I worked on Minio, an open source object storage supporting S3 and could do
this because Minio gives me good and stable listeners.

Best, Ravion

On Mon, Oct 8, 2018, 11:26 AM marotosg  wrote:

> Hi,
>
> At the moment I have a SolrCloud Cluster with a documents collection being
> populated indexing documents coming from a DFS server. Linux boxes are
> mounting that DFS server using samba.
>
> There is a request to move that DFS server to a AWS S3 bucket.
> Does anyone have previous experience about it? Speed etc.
>
> Thanks for the advice.
>
>
>
>
> --
> Sent from: http://lucene.472066.n3.nabble.com/Solr-User-f472068.html
>


Indexing documents from S3 bucket

2018-10-08 Thread marotosg
Hi,

At the moment I have a SolrCloud Cluster with a documents collection being
populated indexing documents coming from a DFS server. Linux boxes are
mounting that DFS server using samba. 

There is a request to move that DFS server to a AWS S3 bucket.
Does anyone have previous experience about it? Speed etc.

Thanks for the advice.




--
Sent from: http://lucene.472066.n3.nabble.com/Solr-User-f472068.html


Solr AutoSuggest Configuration Issue

2018-10-08 Thread Manu Nair
Hi,

I am using Solr 5.1 for my application.
I am trying to use the autoSuggest feature of Solr.
I want to do context filtering on the results returned by Solr suggest.

Please help me know if this feature is supported in the version that I am
using(5.1).
Also if it works with multivalued field. I tried multiple times but it is
not working.

I am referring the following link for details :
https://lucene.apache.org/solr/guide/6_6/suggester.html

Please find the configuration in my solrconfig.xml as below

  
mySuggester
AnalyzingInfixLookupFactory
DocumentDictionaryFactory
name
price
text_en
false
countries
  


Thanks alot for your help in advance.

Regards,
Manu Nair.


Re: Using function in fiter query

2018-10-08 Thread Emir Arnautović
Hi Skanth,
You can use FunctionRangeQueryParser to do that:
https://lucene.apache.org/solr/guide/6_6/other-parsers.html#OtherParsers-FunctionRangeQueryParser
 


Let us know if you are having troubles forming query. You have examples in 
https://lucene.apache.org/solr/guide/6_6/function-queries.html#FunctionQueries-UsingFunctionQuery
 


HTH,
Emir
--
Monitoring - Log Management - Alerting - Anomaly Detection
Solr & Elasticsearch Consulting Support Training - http://sematext.com/



> On 7 Oct 2018, at 10:07, skanth2...@gmail.com wrote:
> 
> Hi,
> 
> I need help on using a custom function in filter query. Can anyone help on
> how to get it wokring. Below is the problem statement.
> 
> Have a date field in long and a buffer time in milliseconds in the documents
> which can vary.
> 
> startTime: 153886680
> bufferTime: 86400
> 
> Need to query for docs who's startTime is currentTime - bufferTime
> 
> like:
> 
> fq=startTime:[* TO sub(NOW, bufferTime)]
> 
> Thanks,
> skanth
> 
> 
> 
> 
> 
> --
> Sent from: http://lucene.472066.n3.nabble.com/Solr-User-f472068.html



Re: Index size issue in SOLR-6.5.1

2018-10-08 Thread Dominique Bejean
HI,

In the Solr Admin console, you can access for each core to the "Segment
info" page. You can see if there are more deleted documents in segments on
server X.

Dominique

Le lun. 8 oct. 2018 à 07:29, SOLR4189  a écrit :

> About which details do you ask? Yesterday we restarted all our solr
> services
> and index size in serverX descreased from 82Gb to 60Gb, and in serverY
> index
> size didn't change (49Gb).
>
>
>
> --
> Sent from: http://lucene.472066.n3.nabble.com/Solr-User-f472068.html
>


Re: Nutch+Solr

2018-10-08 Thread Bineesh
This is solved.

Nutch 1.15 have index-writers.xml file wherein we can pass the UN/PWD for
indexing to solr.



--
Sent from: http://lucene.472066.n3.nabble.com/Solr-User-f472068.html