Re: Could not find configName error

2017-09-05 Thread Wael Kader
i am using  SOLR  4.10.3

I am not sure I have them in source control. I don't actually know what
that is.
I am using SOLR on a pre-setup VM.

On Tue, Sep 5, 2017 at 5:26 PM, Erick Erickson 
wrote:

> What version of Solr?
>
> bin/solr zk -help
>
> In particular upconfig can be used to move configsets up to Zookeeper
> (or back down or whatever) in relatively recent versions of Solr. Yo
> are keeping them in source control right? ;)
>
> Best,
> Erick
>
> On Mon, Sep 4, 2017 at 11:27 PM, Wael Kader  wrote:
> > Hi,
> >
> > I had some issues in SOLR shutting down on a single node application on
> > Hadoop.
> >
> > After starting up i got the error:
> > Could not find configName for collection XXX found.
> >
> > I know the issue is that the configs has issues in Zookeeper but I would
> > like to know how I can push this configuration back to get the index
> > running.
> >
> > --
> > Regards,
> > Wael
>



-- 
Regards,
Wael


Phrase boosting on multiple fields

2017-09-05 Thread ritesh kumar
I have a situation where I have to apply phrase boosting on multiple fields. I 
am using Edismax as query parser.
For instance, I am accepting a keyword (phrase) from the user and I want the 
doc with the exact phrase to be on the top of the resultant list. The problem 
is, there are multiple fields to be queried upon.
I could find some assistance from the code below but again it works on a single 
field only.

q=java+design+patterns&defType=edismax&qf=name&pf2=name^30&ps=0

Any kind of suggestion would be helpful.

Best,

Ritesh Kumar




Solr6.6 Issue/Bug

2017-09-05 Thread Kasim Jinwala
Dear team,
  I am using solr 5.0 last 1 year, now we are planning to upgrade
solr 6.6.
 While trying to start solr using root user, we need to pass -force
parameter to start solr forcefully,
please help to start solr using root user without -force command.

Regards
Kasim J.


Re: Could not find configName error

2017-09-05 Thread Erick Erickson
What version of Solr?

bin/solr zk -help

In particular upconfig can be used to move configsets up to Zookeeper
(or back down or whatever) in relatively recent versions of Solr. Yo
are keeping them in source control right? ;)

Best,
Erick

On Mon, Sep 4, 2017 at 11:27 PM, Wael Kader  wrote:
> Hi,
>
> I had some issues in SOLR shutting down on a single node application on
> Hadoop.
>
> After starting up i got the error:
> Could not find configName for collection XXX found.
>
> I know the issue is that the configs has issues in Zookeeper but I would
> like to know how I can push this configuration back to get the index
> running.
>
> --
> Regards,
> Wael


Re: Solr Poc -Help Needed

2017-09-05 Thread Erick Erickson
This might be useful:
 
https://lucene.apache.org/solr/6_6_0//solr-core/org/apache/solr/update/processor/UUIDUpdateProcessorFactory.html

On Tue, Sep 5, 2017 at 4:13 AM, Rick Leir  wrote:
> Harshal,
> Look in solrconfig.xml, this needs to be configured based on one or more 
> fields.
> Cheers -- Rick
>
> On September 4, 2017 11:40:29 PM EDT, "Agrawal, Harshal (GE Digital)" 
>  wrote:
>>Hello All,
>>
>>I am looking to index csv file/ PDF documents. I  want solr to generate
>>unique key itself.
>>I tried to index one file. I am able to configure data config, Solr
>>config and Schema file. But I am getting following error while
>>indexing:
>>
>>org.apache.solr.common.SolrException: Document is missing mandatory
>>uniqueKey field: id.
>>
>>Please find attached config files and data files which I am using.
>>
>>Kindly help me in fixing this issue:
>>
>>Regards
>>Harshal
>
> --
> Sorry for being brief. Alternate email is rickleir at yahoo dot com


Re: unordered autocomplete search

2017-09-05 Thread Niraj Aswani
Hi Mikhail,

Thank you so much for your inpu. I am looking into this and let you know
when I've tried this.

@Walter: indeed, it should be included :). If my code works, I'll post it
here.

Many thanks,
Niraj

On Mon, Sep 4, 2017 at 5:27 PM, Walter Underwood 
wrote:

> This should probably be a feature of the analyzing infix suggester.
>
> Right now, the fuzzy suggester is broken with the file dictionary, so we
> can’t use fuzzy suggestions at all.
>
> wunder
> Walter Underwood
> wun...@wunderwood.org
> http://observer.wunderwood.org/  (my blog)
>
>
> > On Sep 4, 2017, at 4:50 AM, Mikhail Khludnev  wrote:
> >
> > You probably can override AnalyzingInfixSuggester.finishQuery(Builder,
> > boolean) where you can rip clauses from builder and change TermQueries/
> > PrefixQuery to FuzzyQueries. Let me know how it works, please.
> >
> > On Mon, Sep 4, 2017 at 2:27 PM, Niraj Aswani 
> wrote:
> >
> >> Hi Mikhali,
> >>
> >> Thank you very much for your quick response.
> >>
> >> This does seem to fix the issue of unordered words in the autocomplete,
> >> however, then I loose the capability of matching misspelled words.
> >>
> >> Is there anything that can be done to combine the functionality of the
> two?
> >>
> >> Regards,
> >> Niraj
> >>
> >>
> >>
> >> On Mon, Sep 4, 2017 at 11:13 AM, Mikhail Khludnev 
> wrote:
> >>
> >>> The first question is, "sourceLocation">test.dict makes sense?
> >>> Then, for me, it looks like a case for AnalyzingInfixLookupFactory
> >>> https://lucidworks.com/2015/03/04/solr-suggester/ rather than
> >>>
> >>>   - The FuzzyLookupFactory that creates suggestions for misspelled
> words
> >>>   in fields.
> >>>
> >>>
> >>>
> >>>
> >>> On Mon, Sep 4, 2017 at 11:57 AM, Niraj Aswani 
> >>> wrote:
> >>>
>  Hi,
> 
>  I am using solr 5.5.4.
> 
>  I would like to perform an unordered autocomplete search however it
> >> seems
>  that it only suggests phrases that start with my term in the query.
> 
>  For example, in my dictionary I have,
> 
>  *lamp desk*
> 
>  When searching for the term "lamp", it is able to show the the
> >> suggestion
>  "lamp desk" as the suggestion is starting with the word "lamp".
> >> However,
>  when I search for the term "desk", it doesn't show any suggestion.
> 
>  My field in the *managed-schema* is as below:
> 
>    positionIncrementGap="100">
>   
> 
> 
>   
>   
> 
>   
>  
> 
>  My controller in *solconfig.xml* is defined as the following:
> 
>  
>  
>   
> default
> FuzzyLookupFactory
> test.dict
> suggest_field
>   
>  
> 
>   >>> startup="lazy">
>   
> true
> 10
>   
>   
> suggest
>   
>  
> 
>  The *query* I fire is:
> 
>  http://localhost:8983/solr/core1/suggest?q=desk&suggest.build=true
> 
>  is there anyway, I can get the above scenario working?
> 
>  Many thanks,
>  Niraj
> 
> >>>
> >>>
> >>>
> >>> --
> >>> Sincerely yours
> >>> Mikhail Khludnev
> >>>
> >>
> >
> >
> >
> > --
> > Sincerely yours
> > Mikhail Khludnev
>
>


Re: slow solr facet processing

2017-09-05 Thread Ere Maijala

Toke Eskildsen kirjoitti 5.9.2017 klo 13.49:

On Mon, 2017-09-04 at 11:03 -0400, Yonik Seeley wrote:

It's due to this (see comments in UnInvertedField):


I have read that. What I don't understand is the difference between 4.x
and 6.x. But as you say, Ere seems to be in the process of verifying
whether this is simply due to more segments in 6.x.


During my testing I never optimized the 4.x index, so unless it 
maintains a minimal number of segments automatically, there's something 
else too.



There's probably a number of ways we can speed this up somewhat:
- optimize how much memory is used to store the term index and use
the savings to store more than every 128th term
- store the terms contiguously in block(s)


I'm considering taking a shot at that. A fairly easy optimization would
be to replace the BytesRef[] indexedTermsArray with a BytesRefArray.


I'd be happy to try out any patches.. :)

--Ere

--
Ere Maijala
Kansalliskirjasto / The National Library of Finland


RE: ERR_SSL_VERSION_OR_CIPHER_MISMATCH

2017-09-05 Thread Younge, Kent A - Norman, OK - Contractor
The java.security files are the same.  I even copied over the files from a 
machine that is working and renamed the security files and it still did not 
work.. I am getting the same error.







-Original Message-
From: Younge, Kent A - Norman, OK - Contractor 
[mailto:kent.a.you...@usps.gov.INVALID] 
Sent: Tuesday, September 05, 2017 6:54 AM
To: solr-user@lucene.apache.org
Subject: RE: ERR_SSL_VERSION_OR_CIPHER_MISMATCH

The new box is a clone of all the boxes so nothing should have changed other 
than the certificates and the keystore.  That is why I am at such a loss on 
this issue.   Java is the same across five servers all settings are the same 
across five servers.  I will look into the JVM security and see if it is the 
same across all the boxes.





-Original Message-
From: Chris Hostetter [mailto:hossman_luc...@fucit.org] 
Sent: Friday, September 01, 2017 5:46 PM
To: solr-user@lucene.apache.org
Subject: Re: ERR_SSL_VERSION_OR_CIPHER_MISMATCH


all of the low level SSL code used by Solr comes from the JVM.

double check which version of java you are using and make sure it's consistent 
on all of your servers -- if you disable SSL on the affected server you can use 
the Solr Admin UI to be 100% certain of exactly which version of java is being 
used...

https://lucene.apache.org/solr/guide/6_6/overview-of-the-solr-admin-ui.html

If the JVM Runtime *versions* are identicle, the next thing to check would be 
the the JVM security settings which control which ciphers are used.  
For Oracle JVMs this file is named "java.security" -- compare that file between 
your functional/non-functional servers.

There are lots of docs out there on SSL protocol and cipher configuration in 
java's java.security file, here's a quick one that links deep into the details 
of enabling/disabling protocols...

http://docs.oracle.com/javase/8/docs/technotes/guides/security/SunProviders.html#SunJSSE_Protocols

...but the bottomline is: you probably want to fix your broken server to match 
your working servers, and unless the JVM versions are different, that means 
someone/thing must have modified the JVM security settings on one of your 
servers -- find out who & why.


-Hoss
http://www.lucidworks.com/


Re: slow solr facet processing

2017-09-05 Thread Ere Maijala

Yonik Seeley kirjoitti 4.9.2017 klo 18.03:

It's due to this (see comments in UnInvertedField):
*   To further save memory, the terms (the actual string values) are
not all stored in
*   memory, but a TermIndex is used to convert term numbers to term values only
*   for the terms needed after faceting has completed.  Only every
128th term value
*   is stored, along with its corresponding term number, and this is used as an
*   index to find the closest term and iterate until the desired number is hit

There's probably a number of ways we can speed this up somewhat:
- optimize how much memory is used to store the term index and use the
savings to store more than every 128th term
- store the terms contiguously in block(s)
- don't store the whole term, only store what's needed to seek to the
Nth term correctly
- when retrieving many terms, sort them first and convert from ord->str in order


For what it's worth, I've now tested on our production servers that can 
hold the full index in memory, and the results are in line with the 
previous ones (47 million records, 1785 buckets in the tested facet):


1.) index with docValues="true":

- unoptimized: ~6000ms if facet.method is not specified
- unoptimized: ~7000ms with facet.method=uif
- optimized: ~7800ms if facet.method is not specified
- optimized: ~7700ms with facet.method=uif

Note that optimization took its time and other activity varies 
throughout the day, so the numbers between optimized and unoptimized 
cannot be directly compared. Still bugs me a bit that the optimized 
index seems to be a bit slower here.


2.) index with docValues="false":

- unoptimized: ~2600ms if facet.method is not specified
- unoptimized ~1200ms with facet.method=uif
- optimized: ~2600ms if facet.method is not specified
- optimized: ~17ms with facet.method=uif

--Ere

--
Ere Maijala
Kansalliskirjasto / The National Library of Finland


RE: ERR_SSL_VERSION_OR_CIPHER_MISMATCH

2017-09-05 Thread Younge, Kent A - Norman, OK - Contractor
The new box is a clone of all the boxes so nothing should have changed other 
than the certificates and the keystore.  That is why I am at such a loss on 
this issue.   Java is the same across five servers all settings are the same 
across five servers.  I will look into the JVM security and see if it is the 
same across all the boxes.






Thank you,

Kent Younge
Systems Engineer
USPS MTSC IT Support
600 W. Rock Creek Rd, Norman, OK  73069-8357
O:405 573 2273


-Original Message-
From: Chris Hostetter [mailto:hossman_luc...@fucit.org] 
Sent: Friday, September 01, 2017 5:46 PM
To: solr-user@lucene.apache.org
Subject: Re: ERR_SSL_VERSION_OR_CIPHER_MISMATCH


all of the low level SSL code used by Solr comes from the JVM.

double check which version of java you are using and make sure it's consistent 
on all of your servers -- if you disable SSL on the affected server you can use 
the Solr Admin UI to be 100% certain of exactly which version of java is being 
used...

https://lucene.apache.org/solr/guide/6_6/overview-of-the-solr-admin-ui.html

If the JVM Runtime *versions* are identicle, the next thing to check would be 
the the JVM security settings which control which ciphers are used.  
For Oracle JVMs this file is named "java.security" -- compare that file between 
your functional/non-functional servers.

There are lots of docs out there on SSL protocol and cipher configuration in 
java's java.security file, here's a quick one that links deep into the details 
of enabling/disabling protocols...

http://docs.oracle.com/javase/8/docs/technotes/guides/security/SunProviders.html#SunJSSE_Protocols

...but the bottomline is: you probably want to fix your broken server to match 
your working servers, and unless the JVM versions are different, that means 
someone/thing must have modified the JVM security settings on one of your 
servers -- find out who & why.


-Hoss
http://www.lucidworks.com/


Re: slow solr facet processing

2017-09-05 Thread Yonik Seeley
The number-of-segments noise probably swamps this... but one
optimization around deep-facet-paging that didn't get carried forward
is
https://issues.apache.org/jira/browse/SOLR-2092

-Yonik


On Tue, Sep 5, 2017 at 6:49 AM, Toke Eskildsen  wrote:
> On Mon, 2017-09-04 at 11:03 -0400, Yonik Seeley wrote:
>> It's due to this (see comments in UnInvertedField):
>
> I have read that. What I don't understand is the difference between 4.x
> and 6.x. But as you say, Ere seems to be in the process of verifying
> whether this is simply due to more segments in 6.x.
>
>> There's probably a number of ways we can speed this up somewhat:
>> - optimize how much memory is used to store the term index and use
>> the savings to store more than every 128th term
>> - store the terms contiguously in block(s)
>
> I'm considering taking a shot at that. A fairly easy optimization would
> be to replace the BytesRef[] indexedTermsArray with a BytesRefArray.
>
> - Toke Eskildsen, Royal Danish Library
>


Re: Solr Poc -Help Needed

2017-09-05 Thread Rick Leir
Harshal,
Look in solrconfig.xml, this needs to be configured based on one or more 
fields. 
Cheers -- Rick

On September 4, 2017 11:40:29 PM EDT, "Agrawal, Harshal (GE Digital)" 
 wrote:
>Hello All,
>
>I am looking to index csv file/ PDF documents. I  want solr to generate
>unique key itself.
>I tried to index one file. I am able to configure data config, Solr
>config and Schema file. But I am getting following error while
>indexing:
>
>org.apache.solr.common.SolrException: Document is missing mandatory
>uniqueKey field: id.
>
>Please find attached config files and data files which I am using.
>
>Kindly help me in fixing this issue:
>
>Regards
>Harshal

-- 
Sorry for being brief. Alternate email is rickleir at yahoo dot com 

Re: slow solr facet processing

2017-09-05 Thread Toke Eskildsen
On Mon, 2017-09-04 at 11:03 -0400, Yonik Seeley wrote:
> It's due to this (see comments in UnInvertedField):

I have read that. What I don't understand is the difference between 4.x
and 6.x. But as you say, Ere seems to be in the process of verifying
whether this is simply due to more segments in 6.x.

> There's probably a number of ways we can speed this up somewhat:
> - optimize how much memory is used to store the term index and use
> the savings to store more than every 128th term
> - store the terms contiguously in block(s)

I'm considering taking a shot at that. A fairly easy optimization would
be to replace the BytesRef[] indexedTermsArray with a BytesRefArray.

- Toke Eskildsen, Royal Danish Library