Very long young generation stop the world GC pause

2016-12-07 Thread forest_soup
Hi Shawn, 

Thanks a lot for your response! 

I'll use this mail thread on tracking the issue in JIRA
https://issues.apache.org/jira/browse/SOLR-9828 .



--
View this message in context: 
http://lucene.472066.n3.nabble.com/Very-long-young-generation-stop-the-world-GC-pause-tp4308911.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: Solr Issue

2016-12-07 Thread Greg Harris
Hi Monti,

As pointed out there is a huge gap of no information. There are two primary
possibilities. One is that something about your resources is depleted. As
Shawn has pointed out... watch them as you start up. Two, Solr is somehow
locked or waiting on something. Since there is no information at the INFO
level you can either throw your logs into DEBUG level and pray to god you
see something relevant in the noise. Or, you can run a thread dump after
you've hit your gap in time and see what the server is doing at that
moment. My recommendation is to do the latter. I'd also wonder about that 0
cores found entry as well.

Greg



On Dec 7, 2016 3:02 AM, "Monti Chandra"  wrote:

> Hello team,
> I am working on solr version to 6.2.1. It was working so nice for the first
> 20 days and now the server is restarting very slow(15-20 min).
> Please get the hardware specs of my system below:
> Linux version 3.10.0-327.el7.x86_64 (buil...@kbuilder.dev.centos.org) (gcc
> version 4.8.3 20140911 (Red Hat 4.8.3-9) (GCC) )
> kernel-3.10.0-327.el7.x86_64
> It is working fine when i am taking solr directory to another server of
> same configuartion.
>
> As I am using 6.2.1 basically,  So i have downloaded the fresh
> solr-6.2.1(without any collection) and try to start the same but the server
> is not  able to start effectively. please get the log below:
>
> 0INFO  (main) [   ] o.e.j.u.log Logging initialized @257ms
> 131  INFO  (main) [   ] o.e.j.s.Server jetty-9.3.8.v20160314
> 146  INFO  (main) [   ] o.e.j.d.p.ScanningAppProvider Deployment monitor
> [file:///root/temp/solr-6.2.1/server/contexts/] at interval 0
> 311  INFO  (main) [   ] o.e.j.w.StandardDescriptorProcessor NO JSP Support
> for /solr, did not find org.apache.jasper.servlet.JspServlet
> 320  WARN  (main) [   ] o.e.j.s.SecurityHandler
> ServletContext@o.e.j.w.WebAppContext@5383967b{/solr,
> file:///root/temp/solr-6.2.1/server/solr-webapp/webapp/,
> STARTING}{/root/temp/solr-6.2.1/server/solr-webapp/webapp}
> has uncovered http methods for path: /
> 327  INFO  (main) [   ] o.a.s.s.SolrDispatchFilter
> SolrDispatchFilter.init(): WebAppClassLoader=1465085305@57536d79
> 340  INFO  (main) [   ] o.a.s.c.SolrResourceLoader JNDI not configured for
> solr (NoInitialContextEx)
> 340  INFO  (main) [   ] o.a.s.c.SolrResourceLoader using system property
> solr.solr.home: /root/temp/solr-6.2.1/server/solr
> 340  INFO  (main) [   ] o.a.s.c.SolrResourceLoader new SolrResourceLoader
> for directory: '/root/temp/solr-6.2.1/server/solr'
> 340  INFO  (main) [   ] o.a.s.c.SolrResourceLoader JNDI not configured for
> solr (NoInitialContextEx)
> 340  INFO  (main) [   ] o.a.s.c.SolrResourceLoader using system property
> solr.solr.home: /root/temp/solr-6.2.1/server/solr
> 345  INFO  (main) [   ] o.a.s.c.SolrXmlConfig Loading container
> configuration from /root/temp/solr-6.2.1/server/solr/solr.xml
> 402  INFO  (main) [   ] o.a.s.c.CorePropertiesLocator Config-defined core
> root directory: /root/temp/solr-6.2.1/server/solr
> 424  INFO  (main) [   ] o.a.s.c.CoreContainer New CoreContainer 1663619914
> 425  INFO  (main) [   ] o.a.s.c.CoreContainer Loading cores into
> CoreContainer [instanceDir=/root/temp/solr-6.2.1/server/solr]
> 425  WARN  (main) [   ] o.a.s.c.CoreContainer Couldn't add files from
> /root/temp/solr-6.2.1/server/solr/lib to classpath:
> /root/temp/solr-6.2.1/server/solr/lib
> 434  INFO  (main) [   ] o.a.s.h.c.HttpShardHandlerFactory created with
> socketTimeout : 60,connTimeout : 6,maxConnectionsPerHost :
> 20,maxConnections : 1,corePoolSize : 0,maximumPoolSize :
> 2147483647,maxThreadIdleTime : 5,sizeOfQueue : -1,fairnessPolicy :
> false,useRetries : false,connectionsEvictorSleepDelay :
> 5000,maxConnectionIdleTime : 4,
> 592  INFO  (main) [   ] o.a.s.u.UpdateShardHandler Creating
> UpdateShardHandler HTTP client with params:
> socketTimeout=60&connTimeout=6&retry=true
> 594  INFO  (main) [   ] o.a.s.l.LogWatcher SLF4J impl is
> org.slf4j.impl.Log4jLoggerFactory
> 594  INFO  (main) [   ] o.a.s.l.LogWatcher Registering Log Listener [Log4j
> (org.slf4j.impl.Log4jLoggerFactory)]
> 596  INFO  (main) [   ] o.a.s.c.CoreContainer Security conf doesn't exist.
> Skipping setup for authorization module.
> 596  INFO  (main) [   ] o.a.s.c.CoreContainer No authentication plugin
> used.
> 127902 INFO  (main) [   ] o.a.s.c.CorePropertiesLocator Looking for core
> definitions underneath /root/temp/solr-6.2.1/server/solr
> 127906 INFO  (main) [   ] o.a.s.c.CorePropertiesLocator Found 0 core
> definitions
> 127909 INFO  (main) [   ] o.a.s.s.SolrDispatchFilter
> user.dir=/root/temp/solr-6.2.1/server
> 127909 INFO  (main) [   ] o.a.s.s.SolrDispatchFilter
> SolrDispatchFilter.init() done
> 127918 INFO  (main) [   ] o.e.j.s.h.ContextHandler Started
> o.e.j.w.WebAppContext@5383967b
> {/solr,file:///root/temp/solr-6.2.1/server/solr-webapp/
> webapp/,AVAILABLE}{/root/temp/solr-6.2.1/server/solr-webapp/webapp}
> 127926 INFO  (main) [   ] o.e.j.s.ServerConnector Sta

Re: Solr node not found in ZK live_nodes

2016-12-07 Thread Mark Miller
That already happens. The ZK client itself will reconnect when it can and
trigger everything to be setup like when the cluster first starts up,
including a live node and leader election, etc.

You may have hit a bug or something else missing from this conversation,
but reconnecting after losing the ZK connection is a basic feature from day
one.

Mark
On Wed, Dec 7, 2016 at 12:34 AM Manohar Sripada 
wrote:

> Thanks Erick! Should I create a JIRA issue for the same?
>
> Regarding the logs, I have changed the log level to WARN. That may be the
> reason, I couldn't get anything from it.
>
> Thanks,
> Manohar
>
> On Tue, Dec 6, 2016 at 9:58 PM, Erick Erickson 
> wrote:
>
> > Most likely reason is that the Solr node in question,
> > was not reachable thus it was removed from
> > live_nodes. Perhaps due to temporary network
> > glitch, long GC pause or the like. If you're rolling
> > your logs over it's quite possible that any illuminating
> > messages were lost. The default 4M size for each
> > log is quite lo at INFO level...
> >
> > It does seem possible for a Solr node to periodically
> > check its status and re-insert itself into live_nodes,
> > go through recovery and all that. So far most of that
> > registration logic is baked into startup code. What
> > do others think? Worth a JIRA?
> >
> > Erick
> >
> > On Tue, Dec 6, 2016 at 3:53 AM, Manohar Sripada 
> > wrote:
> > > We have a 16 node cluster of Solr (5.2.1) and 5 node Zookeeper (3.4.6).
> > >
> > > All the Solr nodes were registered to Zookeeper (ls /live_nodes) when
> > setup
> > > was done 3 months back. Suddenly, few days back our search started
> > failing
> > > because one of the solr node(consider s16) was not seen in Zookeeper,
> > i.e.,
> > > when we checked for *"ls /live_nodes"*, *s16 *solr node was not found.
> > > However, the corresponding Solr process was up and running.
> > >
> > > To my surprise, I couldn't find any errors or warnings in solr or
> > zookeeper
> > > logs related to this. I have few questions -
> > >
> > > 1. Is there any reason why this registration to ZK was lost? I know
> logs
> > > should provide some information, but, it didn't. Did anyone encountered
> > > similar issue, if so, what can be the root cause?
> > > 2. Shouldn't Solr be clever enough to detect that the registration to
> ZK
> > > was lost (for some reason) and should try to re-register again?
> > >
> > > PS: The issue is resolved by restarting the Solr node. However, I am
> > > curious to know why it happened in the first place.
> > >
> > > Thanks
> >
>
-- 
- Mark
about.me/markrmiller


Re: Difference between currency fieldType and float fieldType

2016-12-07 Thread Zheng Lin Edwin Yeo
Thanks for all the replies.

Probably will have to go for long or currency fieldType. Int is still
32-bit, and there will be error in indexing if the amount is larger than
2,147,483,647. Since we are storing it as cents, it will hit the limit with
just $21.4 million.

Regards,
Edwin


On 7 December 2016 at 22:16, Esther-Melaine Quansah <
esther.quan...@lucidworks.com> wrote:

> Cool, that makes sense!
>
> Esther Quansah
> > On Dec 7, 2016, at 9:13 AM, Dorian Hoxha  wrote:
> >
> > Yeah, you always *100 when you store,query,facet, and you always /100
> when
> > displaying.
> >
> > On Wed, Dec 7, 2016 at 3:07 PM,  wrote:
> >
> >> I think Edwin might be concerned that in storing it as a long type,
> there
> >> will be no distinguishing between, in example, $1234.56 and $123456.
> >> But correct me if I'm wrong - the latter would be stored as 12345600.
> >>
> >> When sending in a search for all values less than $100,000 on a long
> >> field, will there be a need to send in that value in cents? (That is,
> >> q=*:*&fq=long_field[* TO 1000] )
> >>
> >> Thanks,
> >>
> >> Esther Quansah
> >>
> >>> Le 7 déc. 2016 à 07:26, Dorian Hoxha  a écrit
> :
> >>>
> >>> Come on dude, just use the int/long.
> >>> Source: double is still a float.
> >>>
> >>> On Wed, Dec 7, 2016 at 1:17 PM, Zheng Lin Edwin Yeo <
> >> edwinye...@gmail.com>
> >>> wrote:
> >>>
>  Thanks for the reply.
> 
>  How about using the double fieldType?
>  I tried that it works, as it is 64-bit, as compared to 32-bit for
> float.
>  But will it hit the same issue again if the amount exceeds 64-bit?
> 
>  Regards,
>  Edwin
> 
> 
> > On 7 December 2016 at 15:28, Dorian Hoxha 
> >> wrote:
> >
> > Yeah, you'll have to do the conversion yourself (or something
> internal,
> > like the currencyField).
> >
> > Think about it as datetimes. You store everything in utc (cents), but
> > display to each user in it's own timezone (different currency, or
> just
>  from
> > cents to full dollars).
> >
> > On Wed, Dec 7, 2016 at 8:23 AM, Zheng Lin Edwin Yeo <
>  edwinye...@gmail.com>
> > wrote:
> >
> >> But if I index $1234.56 as "123456", won't it affect the search or
>  facet
> > if
> >> I do a query directly to Solr?
> >>
> >> Say if I search for index with amount that is lesser that $2000, it
>  will
> >> not match, unless when we do the search, we have to pass "20" to
> > Solr?
> >>
> >> Regards,
> >> Edwin
> >>
> >>
> >> On 7 December 2016 at 07:44, Chris Hostetter <
> >> hossman_luc...@fucit.org
> >
> >> wrote:
> >>
> >>> : Thanks for your reply.
> >>> :
> >>> : That means the best fieldType to use for money is currencyField,
>  and
> >> not
> >>> : any other fieldType?
> >>>
> >>> The primary use case for CurrencyField is when you want to do
> dynamic
> >>> currency fluctuations between multiple currency types at query time
>  --
> >> but
> >>> to do that you either need to use the FileExchangeRateProvider and
>  have
> >>> your owne backend system to update the exchange rates, or you have
> to
> >> have
> >>> an openexchangerates.org account, or implement some other provider
> > (with
> >>> custom solr java code)
> >>>
> >>>
> >>> If you only care about a single type of currency -- for example, if
>  all
> >>> you care about is is US Dollars -- then just use either
> >>> TrieIntField or TrieLongField and represent in the smallest
> possible
> >>> increment you need to measure -- for US Dollars this would be
> cents.
> > ie:
> >>> $1234.56 would be put in your index as "123456"
> >>>
> >>>
> >>>
> >>> -Hoss
> >>> http://www.lucidworks.com/
> >>>
> >>
> >
> 
> >>
>
>


Re: Partial Field Update "removeregex" Command

2016-12-07 Thread Karthik Ramachandran
if you want to remove all the data in the then use "null" in set

curl . . . -d '[{"id":"docId","someField":{"set",null}}]'

-Karthik

On Wed, Dec 7, 2016 at 1:31 PM, Richard Bergmann 
wrote:

> Hello,
>
> I am new to this and have found no examples or guidance on how to use
> "removeregex" to remove (in my case) all entries in a multi-valued field.
>
> The following curl commands work just fine:
>
> curl . . . -d '[{"id":"docId","someField":{"add",["val1","val2"]}}]'
>
> and
>
> curl . . . -d '[{"id":"docId","someField":{"remove",["val1","val2"]}}]'
>
>
> None of the following have any effect, however:
>
> curl . . . -d '[{"id":"docId","someField":{"removeregex","val1"}}]'
>
> curl . . . -d '[{"id":"docId","someField":{"removeregex",".*"}}]'
>
>
> I appreciate your help and hope your answer makes it out on the cloud
> somewhere so others can find the solution.
>
> Regards,
>
> Rich Bergmann
>


Re: Solr Issue

2016-12-07 Thread Erick Erickson
It's possibly you have autosuggest configured and it's rebuilding on startup.
See "buildOnStartup" here:
https://cwiki.apache.org/confluence/display/solr/Suggester

Depending on the suggester, this will re-read _all_ documents
from the index to build the internal autosuggest structures which
can take many minutes. That would also explain Shawn's observation
that there is quite a gap in the timestamps.

Although I am puzzled by a couple of things:
1> There should be a message in your log
explicitly referencing building autosuggest. I don't see it.
2> this message:
o.a.s.c.CorePropertiesLocator Found 0 core
definitions

seems to indicate that there are no replicas on the server.

Best,
Erick



On Wed, Dec 7, 2016 at 9:01 AM, Shawn Heisey  wrote:
> On 12/7/2016 3:24 AM, Monti Chandra wrote:
>> I am working on solr version to 6.2.1. It was working so nice for the first
>> 20 days and now the server is restarting very slow(15-20 min).
>> Please get the hardware specs of my system below:
>> Linux version 3.10.0-327.el7.x86_64 (buil...@kbuilder.dev.centos.org) (gcc
>> version 4.8.3 20140911 (Red Hat 4.8.3-9) (GCC) )
>> kernel-3.10.0-327.el7.x86_64
>
> These are not hardware specs.  They are software specs, and really don't
> contain anything useful.  From this, we can tell that you are running
> CentOS 7 64-bit with a 3.10 kernel, and that's pretty much it.
>
>> It is working fine when i am taking solr directory to another server of
>> same configuartion.
>>
>> As I am using 6.2.1 basically,  So i have downloaded the fresh
>> solr-6.2.1(without any collection) and try to start the same but the server
>> is not  able to start effectively. please get the log below:
> 
>> 596  INFO  (main) [   ] o.a.s.c.CoreContainer No authentication plugin used.
>> 127902 INFO  (main) [   ] o.a.s.c.CorePropertiesLocator Looking for core
>> definitions underneath /root/temp/solr-6.2.1/server/solr
>> 127906 INFO  (main) [   ] o.a.s.c.CorePropertiesLocator Found 0 core
>> definitions
>> 127909 INFO  (main) [   ] o.a.s.s.SolrDispatchFilter
>> user.dir=/root/temp/solr-6.2.1/server
>> 127909 INFO  (main) [   ] o.a.s.s.SolrDispatchFilter
>> SolrDispatchFilter.init() done
>> 127918 INFO  (main) [   ] o.e.j.s.h.ContextHandler Started
>> o.e.j.w.WebAppContext@5383967b
>> {/solr,file:///root/temp/solr-6.2.1/server/solr-webapp/webapp/,AVAILABLE}{/root/temp/solr-6.2.1/server/solr-webapp/webapp}
>> 127926 INFO  (main) [   ] o.e.j.s.ServerConnector Started
>> ServerConnector@1573f9fc{HTTP/1.1,[http/1.1]}{0.0.0.0:}
>> 127926 INFO  (main) [   ] o.e.j.s.Server Started @128185ms
>>
>>
>> Is there any H/w, OS or kernel level threat caused by running solr? Please
>> help me I am wondering for a very long time.
>
> This log shows no errors.  It says that Jetty is listening on port
> .  It does look like it took a couple of minutes for Solr to finish
> starting, but other than that, I cannot see any problems here.
>
> Troubleshooting performance issues like this requires careful
> observation, and may be very difficult to do via a mailing list.  We
> have pretty much zero information to go on, and the logs you have sent
> do not even contain any relevant warnings.
>
> There is only one odd thing.  The log jumps from 596 milliseconds to
> 127902 milliseconds -- over two minutes with absolutely no information
> logged, after which it completes startup.  When I start a
> freshly-downloaded Solr 6.2.1 in foreground mode, the time difference
> between these two log entries on my system is only 32 milliseconds, and
> then Solr is fully started less than a second later.
>
> 4271 INFO  (main) [   ] o.a.s.c.CoreContainer No authentication plugin used.
> 4303 INFO  (main) [   ] o.a.s.c.CorePropertiesLocator Looking for core
> definitions underneath C:\Users\sheisey\Downloads\solr-6.2.1\server\solr
> 4552 INFO  (main) [   ] o.a.s.c.CorePropertiesLocator Found 0 core
> definitions
> 4552 INFO  (main) [   ] o.a.s.s.SolrDispatchFilter
> user.dir=C:\Users\sheisey\Downloads\solr-6.2.1\server
> 4552 INFO  (main) [   ] o.a.s.s.SolrDispatchFilter
> SolrDispatchFilter.init() done
> 4553 INFO  (main) [   ] o.e.j.s.h.ContextHandler Started
> o.e.j.w.WebAppContext@a67c67e{/solr,file:///C:/Users/sheisey/Downloads/solr-6.2.1/server/solr-webapp/webapp/,AVAILABLE}{C:\Users\sheisey\Downloads\solr-6.2.1\server/solr-webapp/webapp}
> 4559 INFO  (main) [   ] o.e.j.s.ServerConnector Started
> ServerConnector@4c12331b{HTTP/1.1,[http/1.1]}{0.0.0.0:8983}
> 4560 INFO  (main) [   ] o.e.j.s.Server Started @5548ms
>
> Do you have any more information you can share?  Something I'd like to
> see is a screenshot of "top" running in your terminal, after starting
> the fresh Solr with no collections.  Before getting the screenshot,
> press shift-M to sort the list by memory usage.
>
> Thanks,
> Shawn
>


Partial Field Update "removeregex" Command

2016-12-07 Thread Richard Bergmann
Hello,

I am new to this and have found no examples or guidance on how to use 
"removeregex" to remove (in my case) all entries in a multi-valued field.

The following curl commands work just fine:

curl . . . -d '[{"id":"docId","someField":{"add",["val1","val2"]}}]'

and

curl . . . -d '[{"id":"docId","someField":{"remove",["val1","val2"]}}]'


None of the following have any effect, however:

curl . . . -d '[{"id":"docId","someField":{"removeregex","val1"}}]'

curl . . . -d '[{"id":"docId","someField":{"removeregex",".*"}}]'


I appreciate your help and hope your answer makes it out on the cloud somewhere 
so others can find the solution.

Regards,

Rich Bergmann


Re: Configuration folder for the collection generated with instead of just

2016-12-07 Thread Erik Hatcher
Nicole - 

Since this is probably off-topic for the solr-user list, let’s take this 
offline and over to your Lucidworks support.  But while we’re here, here’s an 
example of using the Fusion API to create a collection and then the Solr API to 
configure the schema.  In this example, it’s not using the zkcli to upconfig, 
though I’ve done that technique as well with no problems.   Something is fishy, 
obviously, with the discrepancy of your config set and your collection name, 
but I’m not sure where that is coming from.  Feel free to reply to me offline 
at "erik  hatcher @ lucidworks.com" with the exact commands that you’ve 
issued and I’ll get a support ticket opened for you.

Erik

> On Dec 7, 2016, at 6:06 AM, Nicole Bilić  wrote:
> 
> Good suggestion, but unfortunately it does not address this issue as we are
> not using the time-based partitioning in this project.
> 
> It would be useful to know in which case is the configuration created with
>  in Solr, what scenario does lead to that so we can
> investigate further. Any other suggestions?
> 
> Thanks
> 
> On Wed, Dec 7, 2016 at 1:53 PM, Erik Hatcher  wrote:
> 
>> Looks best to file that as a Lucidworks support ticket.
>> 
>> But are you using the time-based sharding feature of Fusion?   If that's
>> the case that might explain it as that creates collections for each time
>> partition.
>> 
>>Erik
>> 
>>> On Dec 7, 2016, at 00:31, Nicole Bilić  wrote:
>>> 
>>> Hi all,
>>> 
>>> 
>>> We are using Lucidworks Fusion on top of Solr and recently we’ve
>>> encountered an unexpected behavior. We’ve created bash scripts which we
>> use
>>> to create collections in Solr using the Fusion API and upload the
>>> collection configuration (with bash $ZKCLIENT -cmd upconfig -confdir
>> $path
>>> -confname $collection -z $HOST:$ZKPORT).
>>> 
>>> We had issues that the index fields of our custom configurations were not
>>> available (ie. our custom field types and fields we’ve defined in
>>> managed-schema). So we checked in Solr admin and found out that the
>>> configuration folder for the collection was generated with
>>>  instead of just  (eg.
>>> https://drive.google.com/file/d/0B9Nu5vSE4ep8OUdyckZoV2FMdlE/
>> view?usp=sharing).
>>> 
>>> 
>>> Our collection configurations are under the one with the id in the end.
>> So
>>> the configuration set that was uploaded by our scripts was never used and
>>> therefore the index fields were not existing. Also, we did not manage to
>>> discover a pattern (or cause) for when does this happen (because
>> sometimes
>>> the issue does not occur and sometimes it does and to us it seems pretty
>>> nondeterministic).
>>> 
>>> Furthermore, in the screenshot above, it’s possible to see the same thing
>>> happened with system_metrics collection configs, which we do not create
>> or
>>> modify with our scripts. Therefore, the scripts should not be the cause
>> of
>>> this behavior.
>>> 
>>> What we are trying to determine is if this issue is coming from Fusion,
>>> Solr or Zookeeper. Does someone know in which case the collection
>>> configuration folder is generated with the id in the end and how to avoid
>>> it? Thank you!
>>> 
>>> 
>>> Nicole
>> 



Re: Solr Issue

2016-12-07 Thread Shawn Heisey
On 12/7/2016 3:24 AM, Monti Chandra wrote:
> I am working on solr version to 6.2.1. It was working so nice for the first
> 20 days and now the server is restarting very slow(15-20 min).
> Please get the hardware specs of my system below:
> Linux version 3.10.0-327.el7.x86_64 (buil...@kbuilder.dev.centos.org) (gcc
> version 4.8.3 20140911 (Red Hat 4.8.3-9) (GCC) )
> kernel-3.10.0-327.el7.x86_64

These are not hardware specs.  They are software specs, and really don't
contain anything useful.  From this, we can tell that you are running
CentOS 7 64-bit with a 3.10 kernel, and that's pretty much it.

> It is working fine when i am taking solr directory to another server of
> same configuartion.
>
> As I am using 6.2.1 basically,  So i have downloaded the fresh
> solr-6.2.1(without any collection) and try to start the same but the server
> is not  able to start effectively. please get the log below:

> 596  INFO  (main) [   ] o.a.s.c.CoreContainer No authentication plugin used.
> 127902 INFO  (main) [   ] o.a.s.c.CorePropertiesLocator Looking for core
> definitions underneath /root/temp/solr-6.2.1/server/solr
> 127906 INFO  (main) [   ] o.a.s.c.CorePropertiesLocator Found 0 core
> definitions
> 127909 INFO  (main) [   ] o.a.s.s.SolrDispatchFilter
> user.dir=/root/temp/solr-6.2.1/server
> 127909 INFO  (main) [   ] o.a.s.s.SolrDispatchFilter
> SolrDispatchFilter.init() done
> 127918 INFO  (main) [   ] o.e.j.s.h.ContextHandler Started
> o.e.j.w.WebAppContext@5383967b
> {/solr,file:///root/temp/solr-6.2.1/server/solr-webapp/webapp/,AVAILABLE}{/root/temp/solr-6.2.1/server/solr-webapp/webapp}
> 127926 INFO  (main) [   ] o.e.j.s.ServerConnector Started
> ServerConnector@1573f9fc{HTTP/1.1,[http/1.1]}{0.0.0.0:}
> 127926 INFO  (main) [   ] o.e.j.s.Server Started @128185ms
>
>
> Is there any H/w, OS or kernel level threat caused by running solr? Please
> help me I am wondering for a very long time.

This log shows no errors.  It says that Jetty is listening on port
.  It does look like it took a couple of minutes for Solr to finish
starting, but other than that, I cannot see any problems here.

Troubleshooting performance issues like this requires careful
observation, and may be very difficult to do via a mailing list.  We
have pretty much zero information to go on, and the logs you have sent
do not even contain any relevant warnings.

There is only one odd thing.  The log jumps from 596 milliseconds to
127902 milliseconds -- over two minutes with absolutely no information
logged, after which it completes startup.  When I start a
freshly-downloaded Solr 6.2.1 in foreground mode, the time difference
between these two log entries on my system is only 32 milliseconds, and
then Solr is fully started less than a second later.

4271 INFO  (main) [   ] o.a.s.c.CoreContainer No authentication plugin used.
4303 INFO  (main) [   ] o.a.s.c.CorePropertiesLocator Looking for core
definitions underneath C:\Users\sheisey\Downloads\solr-6.2.1\server\solr
4552 INFO  (main) [   ] o.a.s.c.CorePropertiesLocator Found 0 core
definitions
4552 INFO  (main) [   ] o.a.s.s.SolrDispatchFilter
user.dir=C:\Users\sheisey\Downloads\solr-6.2.1\server
4552 INFO  (main) [   ] o.a.s.s.SolrDispatchFilter
SolrDispatchFilter.init() done
4553 INFO  (main) [   ] o.e.j.s.h.ContextHandler Started
o.e.j.w.WebAppContext@a67c67e{/solr,file:///C:/Users/sheisey/Downloads/solr-6.2.1/server/solr-webapp/webapp/,AVAILABLE}{C:\Users\sheisey\Downloads\solr-6.2.1\server/solr-webapp/webapp}
4559 INFO  (main) [   ] o.e.j.s.ServerConnector Started
ServerConnector@4c12331b{HTTP/1.1,[http/1.1]}{0.0.0.0:8983}
4560 INFO  (main) [   ] o.e.j.s.Server Started @5548ms

Do you have any more information you can share?  Something I'd like to
see is a screenshot of "top" running in your terminal, after starting
the fresh Solr with no collections.  Before getting the screenshot,
press shift-M to sort the list by memory usage.

Thanks,
Shawn



Cannot use Properties in a CDCR Request Handler Definition

2016-12-07 Thread Webster Homer
I have been testing and setting up CDCR replication between Solrcloud
instances.

We are currently using Solr 6.2

We have a lot of collections and a number of environments for testing and
deployment. It seemed that using properties in the cdcrRequestHandler would
help a lot. Since we have a naming convention and collection names should
be the same between clouds I tried using property names for the source and
target collection, and for the target Zookeeper.

We also use the Config-API to create and change the request handler.
{"responseHeader": {"status": 0,"QTime": 2},"overlay": {"znodeVersion": 17,"
requestHandler": {"/cdcr": {"name": "/cdcr","class": "
solr.CdcrRequestHandler","replica": {"zkHost": "${targetZk}","source":
"${solr.core.collection}","target": "${solr.core.collection}"},"replicator
": {"threadPoolSize": 2,"schedule": 500,"batchSize": 128},"
updateLogSynchronizer": {"schedule": 6}}},"userProps": {"targetZk": "
stldeepx06.sial.com:2181/solr"}}}

I cannot get this to work unless I remove all the property definitions and
use the real names:
This works:
{"responseHeader": {"status": 0,"QTime": 2},"overlay": {"znodeVersion": 17,"
requestHandler": {"/cdcr": {"name": "/cdcr","class": "
solr.CdcrRequestHandler","replica": {"zkHost": "
stldeepx06.sial.com:2181/solr","source": "sial-catalog-product","target": "
sial-catalog-product"},"replicator": {"threadPoolSize": 2,"schedule": 500,"
batchSize": 128},"updateLogSynchronizer": {"schedule": 6}

replication fails when any of the parameters are properties. Even though
the documentation has an example that uses properties.
https://cwiki.apache.org/confluence/pages/viewpage.action?pageId=62687462#CrossDataCenterReplication(CDCR)-InitialStartup

This seems like a defect, at least a problem with the docs, but I would
expect properties to work, or is it just that it doesn't work with the
Config-API?

Thanks,
Web

-- 


This message and any attachment are confidential and may be privileged or 
otherwise protected from disclosure. If you are not the intended recipient, 
you must not copy this message or attachment or disclose the contents to 
any other person. If you have received this transmission in error, please 
notify the sender immediately and delete the message and any attachment 
from your system. Merck KGaA, Darmstadt, Germany and any of its 
subsidiaries do not accept liability for any omissions or errors in this 
message which may arise as a result of E-Mail-transmission or for damages 
resulting from any unauthorized changes of the content of this message and 
any attachment thereto. Merck KGaA, Darmstadt, Germany and any of its 
subsidiaries do not guarantee that this message is free of viruses and does 
not accept liability for any damages caused by any virus transmitted 
therewith.

Click http://www.merckgroup.com/disclaimer to access the German, French, 
Spanish and Portuguese versions of this disclaimer.


Re: solr audit logging

2016-12-07 Thread Jeff Courtade
Thanks very much every one.

They will probably pursue custom code to see if they can get this data and
log it.

J

--
Thanks,

Jeff Courtade
M: 240.507.6116

On Tue, Dec 6, 2016 at 7:07 PM, John Bickerstaff 
wrote:

> You know - if I had to build this, I would consider slurping up the
> relevant log entries (if they exist) and feeding them to Kafka - then your
> people who want to analyze what happened can get those entries again and
> again (Think of Kafka kind of like a persistent messaging store that can
> store log entries or anything you want...)
>
> Of course, how much work you'd have to put into that depends on the
> technical skill of whoever is going to consume this stuff...
>
> Also, a plain old relational database can easily hold these things as well
> -and the code to parse the log messages into some simple tables wouldn't be
> that difficult...  There are probably existing examples / projects...
>
> Naturally - if the standard log entries do NOT get you what you need, then
> it gets to be more of an effort, although adding an extension to Solr isn't
> too hard once you understand the process...
>
> Ping back and let us know what you find in the logs and if you want more
> "advice" -- which you should always take with a grain of salt...
>
> On Tue, Dec 6, 2016 at 3:56 PM, John Bickerstaff  >
> wrote:
>
> > If you can identify currently-logged messages that give you what you need
> > (even if you have to modify or process them afterwards) you can easily
> make
> > a custom log4j config that grabs ONLY what you want and dumps it into a
> > separate file...
> >
> > I'm pretty sure I've seen all the request coming through in my SOLR log
> > files...
> >
> > In case that helps...
> >
> > On Tue, Dec 6, 2016 at 2:08 PM, Alexandre Rafalovitch <
> arafa...@gmail.com>
> > wrote:
> >
> >> There is also Jetty level access log which shows the requests, though
> >> it may not show the HTTP PUT bodies.
> >>
> >> Finally, various online monitoring services probably have agents that
> >> integrate with Solr to show what's happening. Usually costs money
> >> though.
> >>
> >> Regards,
> >> Alex.
> >> 
> >> http://www.solr-start.com/ - Resources for Solr users, new and
> >> experienced
> >>
> >>
> >> On 6 December 2016 at 14:34, Jeff Courtade 
> >> wrote:
> >> > Thanks very much the trace idea is a brilliant way to dig into it. Did
> >> not
> >> > occur to me.
> >> >
> >> > I had another coworker suggest the custom
> >> >
> >> > http://lucene.apache.org/solr/6_3_0/solr-core/org/apache/sol
> >> r/update/processor/LogUpdateProcessorFactory.html
> >> >
> >> >
> >> > this is beyond my litmited abilites.
> >> >
> >> >
> >> > I will see what we can dig up out of the logs...
> >> >
> >> >
> >> > the original request was this...
> >> >
> >> >
> >> > "Is there any configuration, plugin, or application that will create
> an
> >> > audit trail for Solr requests? We have teams that would like to be
> able
> >> to
> >> > pull back changes/requests to documents in solr given a time period.
> The
> >> > information they would like to retrieve is the request to solr, where
> it
> >> > came from, and what the request did."
> >> >
> >> >
> >> > I am starting to think there is not a simple solution to this. I was
> >> hoping
> >> > there was an UpdateAudit class or something I could flip a switch on
> or
> >> > some such...
> >> >
> >> >
> >> >
> >> > On Tue, Dec 6, 2016 at 2:20 PM, Alexandre Rafalovitch <
> >> arafa...@gmail.com>
> >> > wrote:
> >> >
> >> >> You could turn the trace mode for everything in the Admin UI (under
> >> >> logs/levels) and see if any of the existing information is sufficient
> >> >> for your needs. If yes, then you change log level in the
> configuration
> >> >> just for that class/element.
> >> >>
> >> >> Alternatively, you could do a custom UpdateRequestProcessor in the
> >> >> request handler(s) that deal with update. Or perhaps
> >> >> LogUpdateProcessor (that's in every standard chain) is sufficient:
> >> >> http://www.solr-start.com/javadoc/solr-lucene/org/
> >> >> apache/solr/update/processor/LogUpdateProcessorFactory.html
> >> >>
> >> >> But it is also possible that the audit.log is something that has a
> >> >> specific format that other tools use. So, you could start from asking
> >> >> how that file would be used and then working backwards into Solr.
> >> >> Which would most likely be a custom URP, as I mentioned earlier.
> >> >>
> >> >> Regards,
> >> >>Alex.
> >> >> P.s. Remember that there are full document updates and partial
> >> >> updates. What you want to log about that is your business level
> >> >> decision.
> >> >>
> >> >
> >> >
> >> >
> >> > --
> >> > Thanks,
> >> >
> >> > Jeff Courtade
> >> > M: 240.507.6116
> >>
> >
> >
>


Re: Difference between currency fieldType and float fieldType

2016-12-07 Thread Esther-Melaine Quansah
Cool, that makes sense!

Esther Quansah
> On Dec 7, 2016, at 9:13 AM, Dorian Hoxha  wrote:
> 
> Yeah, you always *100 when you store,query,facet, and you always /100 when
> displaying.
> 
> On Wed, Dec 7, 2016 at 3:07 PM,  wrote:
> 
>> I think Edwin might be concerned that in storing it as a long type, there
>> will be no distinguishing between, in example, $1234.56 and $123456.
>> But correct me if I'm wrong - the latter would be stored as 12345600.
>> 
>> When sending in a search for all values less than $100,000 on a long
>> field, will there be a need to send in that value in cents? (That is,
>> q=*:*&fq=long_field[* TO 1000] )
>> 
>> Thanks,
>> 
>> Esther Quansah
>> 
>>> Le 7 déc. 2016 à 07:26, Dorian Hoxha  a écrit :
>>> 
>>> Come on dude, just use the int/long.
>>> Source: double is still a float.
>>> 
>>> On Wed, Dec 7, 2016 at 1:17 PM, Zheng Lin Edwin Yeo <
>> edwinye...@gmail.com>
>>> wrote:
>>> 
 Thanks for the reply.
 
 How about using the double fieldType?
 I tried that it works, as it is 64-bit, as compared to 32-bit for float.
 But will it hit the same issue again if the amount exceeds 64-bit?
 
 Regards,
 Edwin
 
 
> On 7 December 2016 at 15:28, Dorian Hoxha 
>> wrote:
> 
> Yeah, you'll have to do the conversion yourself (or something internal,
> like the currencyField).
> 
> Think about it as datetimes. You store everything in utc (cents), but
> display to each user in it's own timezone (different currency, or just
 from
> cents to full dollars).
> 
> On Wed, Dec 7, 2016 at 8:23 AM, Zheng Lin Edwin Yeo <
 edwinye...@gmail.com>
> wrote:
> 
>> But if I index $1234.56 as "123456", won't it affect the search or
 facet
> if
>> I do a query directly to Solr?
>> 
>> Say if I search for index with amount that is lesser that $2000, it
 will
>> not match, unless when we do the search, we have to pass "20" to
> Solr?
>> 
>> Regards,
>> Edwin
>> 
>> 
>> On 7 December 2016 at 07:44, Chris Hostetter <
>> hossman_luc...@fucit.org
> 
>> wrote:
>> 
>>> : Thanks for your reply.
>>> :
>>> : That means the best fieldType to use for money is currencyField,
 and
>> not
>>> : any other fieldType?
>>> 
>>> The primary use case for CurrencyField is when you want to do dynamic
>>> currency fluctuations between multiple currency types at query time
 --
>> but
>>> to do that you either need to use the FileExchangeRateProvider and
 have
>>> your owne backend system to update the exchange rates, or you have to
>> have
>>> an openexchangerates.org account, or implement some other provider
> (with
>>> custom solr java code)
>>> 
>>> 
>>> If you only care about a single type of currency -- for example, if
 all
>>> you care about is is US Dollars -- then just use either
>>> TrieIntField or TrieLongField and represent in the smallest possible
>>> increment you need to measure -- for US Dollars this would be cents.
> ie:
>>> $1234.56 would be put in your index as "123456"
>>> 
>>> 
>>> 
>>> -Hoss
>>> http://www.lucidworks.com/
>>> 
>> 
> 
 
>> 



Re: Difference between currency fieldType and float fieldType

2016-12-07 Thread Dorian Hoxha
Yeah, you always *100 when you store,query,facet, and you always /100 when
displaying.

On Wed, Dec 7, 2016 at 3:07 PM,  wrote:

> I think Edwin might be concerned that in storing it as a long type, there
> will be no distinguishing between, in example, $1234.56 and $123456.
> But correct me if I'm wrong - the latter would be stored as 12345600.
>
> When sending in a search for all values less than $100,000 on a long
> field, will there be a need to send in that value in cents? (That is,
> q=*:*&fq=long_field[* TO 1000] )
>
> Thanks,
>
> Esther Quansah
>
> > Le 7 déc. 2016 à 07:26, Dorian Hoxha  a écrit :
> >
> > Come on dude, just use the int/long.
> > Source: double is still a float.
> >
> > On Wed, Dec 7, 2016 at 1:17 PM, Zheng Lin Edwin Yeo <
> edwinye...@gmail.com>
> > wrote:
> >
> >> Thanks for the reply.
> >>
> >> How about using the double fieldType?
> >> I tried that it works, as it is 64-bit, as compared to 32-bit for float.
> >> But will it hit the same issue again if the amount exceeds 64-bit?
> >>
> >> Regards,
> >> Edwin
> >>
> >>
> >>> On 7 December 2016 at 15:28, Dorian Hoxha 
> wrote:
> >>>
> >>> Yeah, you'll have to do the conversion yourself (or something internal,
> >>> like the currencyField).
> >>>
> >>> Think about it as datetimes. You store everything in utc (cents), but
> >>> display to each user in it's own timezone (different currency, or just
> >> from
> >>> cents to full dollars).
> >>>
> >>> On Wed, Dec 7, 2016 at 8:23 AM, Zheng Lin Edwin Yeo <
> >> edwinye...@gmail.com>
> >>> wrote:
> >>>
>  But if I index $1234.56 as "123456", won't it affect the search or
> >> facet
> >>> if
>  I do a query directly to Solr?
> 
>  Say if I search for index with amount that is lesser that $2000, it
> >> will
>  not match, unless when we do the search, we have to pass "20" to
> >>> Solr?
> 
>  Regards,
>  Edwin
> 
> 
>  On 7 December 2016 at 07:44, Chris Hostetter <
> hossman_luc...@fucit.org
> >>>
>  wrote:
> 
> > : Thanks for your reply.
> > :
> > : That means the best fieldType to use for money is currencyField,
> >> and
>  not
> > : any other fieldType?
> >
> > The primary use case for CurrencyField is when you want to do dynamic
> > currency fluctuations between multiple currency types at query time
> >> --
>  but
> > to do that you either need to use the FileExchangeRateProvider and
> >> have
> > your owne backend system to update the exchange rates, or you have to
>  have
> > an openexchangerates.org account, or implement some other provider
> >>> (with
> > custom solr java code)
> >
> >
> > If you only care about a single type of currency -- for example, if
> >> all
> > you care about is is US Dollars -- then just use either
> > TrieIntField or TrieLongField and represent in the smallest possible
> > increment you need to measure -- for US Dollars this would be cents.
> >>> ie:
> > $1234.56 would be put in your index as "123456"
> >
> >
> >
> > -Hoss
> > http://www.lucidworks.com/
> >
> 
> >>>
> >>
>


Re: Difference between currency fieldType and float fieldType

2016-12-07 Thread esther . quansah
I think Edwin might be concerned that in storing it as a long type, there will 
be no distinguishing between, in example, $1234.56 and $123456. 
But correct me if I'm wrong - the latter would be stored as 12345600. 

When sending in a search for all values less than $100,000 on a long field, 
will there be a need to send in that value in cents? (That is, 
q=*:*&fq=long_field[* TO 1000] )

Thanks,

Esther Quansah

> Le 7 déc. 2016 à 07:26, Dorian Hoxha  a écrit :
> 
> Come on dude, just use the int/long.
> Source: double is still a float.
> 
> On Wed, Dec 7, 2016 at 1:17 PM, Zheng Lin Edwin Yeo 
> wrote:
> 
>> Thanks for the reply.
>> 
>> How about using the double fieldType?
>> I tried that it works, as it is 64-bit, as compared to 32-bit for float.
>> But will it hit the same issue again if the amount exceeds 64-bit?
>> 
>> Regards,
>> Edwin
>> 
>> 
>>> On 7 December 2016 at 15:28, Dorian Hoxha  wrote:
>>> 
>>> Yeah, you'll have to do the conversion yourself (or something internal,
>>> like the currencyField).
>>> 
>>> Think about it as datetimes. You store everything in utc (cents), but
>>> display to each user in it's own timezone (different currency, or just
>> from
>>> cents to full dollars).
>>> 
>>> On Wed, Dec 7, 2016 at 8:23 AM, Zheng Lin Edwin Yeo <
>> edwinye...@gmail.com>
>>> wrote:
>>> 
 But if I index $1234.56 as "123456", won't it affect the search or
>> facet
>>> if
 I do a query directly to Solr?
 
 Say if I search for index with amount that is lesser that $2000, it
>> will
 not match, unless when we do the search, we have to pass "20" to
>>> Solr?
 
 Regards,
 Edwin
 
 
 On 7 December 2016 at 07:44, Chris Hostetter >> 
 wrote:
 
> : Thanks for your reply.
> :
> : That means the best fieldType to use for money is currencyField,
>> and
 not
> : any other fieldType?
> 
> The primary use case for CurrencyField is when you want to do dynamic
> currency fluctuations between multiple currency types at query time
>> --
 but
> to do that you either need to use the FileExchangeRateProvider and
>> have
> your owne backend system to update the exchange rates, or you have to
 have
> an openexchangerates.org account, or implement some other provider
>>> (with
> custom solr java code)
> 
> 
> If you only care about a single type of currency -- for example, if
>> all
> you care about is is US Dollars -- then just use either
> TrieIntField or TrieLongField and represent in the smallest possible
> increment you need to measure -- for US Dollars this would be cents.
>>> ie:
> $1234.56 would be put in your index as "123456"
> 
> 
> 
> -Hoss
> http://www.lucidworks.com/
> 
 
>>> 
>> 


Re: Configuration folder for the collection generated with instead of just

2016-12-07 Thread Nicole Bilić
Good suggestion, but unfortunately it does not address this issue as we are
not using the time-based partitioning in this project.

It would be useful to know in which case is the configuration created with
 in Solr, what scenario does lead to that so we can
investigate further. Any other suggestions?

Thanks

On Wed, Dec 7, 2016 at 1:53 PM, Erik Hatcher  wrote:

> Looks best to file that as a Lucidworks support ticket.
>
> But are you using the time-based sharding feature of Fusion?   If that's
> the case that might explain it as that creates collections for each time
> partition.
>
> Erik
>
> > On Dec 7, 2016, at 00:31, Nicole Bilić  wrote:
> >
> > Hi all,
> >
> >
> > We are using Lucidworks Fusion on top of Solr and recently we’ve
> > encountered an unexpected behavior. We’ve created bash scripts which we
> use
> > to create collections in Solr using the Fusion API and upload the
> > collection configuration (with bash $ZKCLIENT -cmd upconfig -confdir
> $path
> > -confname $collection -z $HOST:$ZKPORT).
> >
> > We had issues that the index fields of our custom configurations were not
> > available (ie. our custom field types and fields we’ve defined in
> > managed-schema). So we checked in Solr admin and found out that the
> > configuration folder for the collection was generated with
> >  instead of just  (eg.
> > https://drive.google.com/file/d/0B9Nu5vSE4ep8OUdyckZoV2FMdlE/
> view?usp=sharing).
> >
> >
> > Our collection configurations are under the one with the id in the end.
> So
> > the configuration set that was uploaded by our scripts was never used and
> > therefore the index fields were not existing. Also, we did not manage to
> > discover a pattern (or cause) for when does this happen (because
> sometimes
> > the issue does not occur and sometimes it does and to us it seems pretty
> > nondeterministic).
> >
> > Furthermore, in the screenshot above, it’s possible to see the same thing
> > happened with system_metrics collection configs, which we do not create
> or
> > modify with our scripts. Therefore, the scripts should not be the cause
> of
> > this behavior.
> >
> > What we are trying to determine is if this issue is coming from Fusion,
> > Solr or Zookeeper. Does someone know in which case the collection
> > configuration folder is generated with the id in the end and how to avoid
> > it? Thank you!
> >
> >
> > Nicole
>


Re: Configuration folder for the collection generated with instead of just

2016-12-07 Thread Erik Hatcher
Looks best to file that as a Lucidworks support ticket.  

But are you using the time-based sharding feature of Fusion?   If that's the 
case that might explain it as that creates collections for each time partition. 
 

Erik

> On Dec 7, 2016, at 00:31, Nicole Bilić  wrote:
> 
> Hi all,
> 
> 
> We are using Lucidworks Fusion on top of Solr and recently we’ve
> encountered an unexpected behavior. We’ve created bash scripts which we use
> to create collections in Solr using the Fusion API and upload the
> collection configuration (with bash $ZKCLIENT -cmd upconfig -confdir $path
> -confname $collection -z $HOST:$ZKPORT).
> 
> We had issues that the index fields of our custom configurations were not
> available (ie. our custom field types and fields we’ve defined in
> managed-schema). So we checked in Solr admin and found out that the
> configuration folder for the collection was generated with
>  instead of just  (eg.
> https://drive.google.com/file/d/0B9Nu5vSE4ep8OUdyckZoV2FMdlE/view?usp=sharing).
> 
> 
> Our collection configurations are under the one with the id in the end. So
> the configuration set that was uploaded by our scripts was never used and
> therefore the index fields were not existing. Also, we did not manage to
> discover a pattern (or cause) for when does this happen (because sometimes
> the issue does not occur and sometimes it does and to us it seems pretty
> nondeterministic).
> 
> Furthermore, in the screenshot above, it’s possible to see the same thing
> happened with system_metrics collection configs, which we do not create or
> modify with our scripts. Therefore, the scripts should not be the cause of
> this behavior.
> 
> What we are trying to determine is if this issue is coming from Fusion,
> Solr or Zookeeper. Does someone know in which case the collection
> configuration folder is generated with the id in the end and how to avoid
> it? Thank you!
> 
> 
> Nicole


Re: Difference between currency fieldType and float fieldType

2016-12-07 Thread Dorian Hoxha
Come on dude, just use the int/long.
Source: double is still a float.

On Wed, Dec 7, 2016 at 1:17 PM, Zheng Lin Edwin Yeo 
wrote:

> Thanks for the reply.
>
> How about using the double fieldType?
> I tried that it works, as it is 64-bit, as compared to 32-bit for float.
> But will it hit the same issue again if the amount exceeds 64-bit?
>
> Regards,
> Edwin
>
>
> On 7 December 2016 at 15:28, Dorian Hoxha  wrote:
>
> > Yeah, you'll have to do the conversion yourself (or something internal,
> > like the currencyField).
> >
> > Think about it as datetimes. You store everything in utc (cents), but
> > display to each user in it's own timezone (different currency, or just
> from
> > cents to full dollars).
> >
> > On Wed, Dec 7, 2016 at 8:23 AM, Zheng Lin Edwin Yeo <
> edwinye...@gmail.com>
> > wrote:
> >
> > > But if I index $1234.56 as "123456", won't it affect the search or
> facet
> > if
> > > I do a query directly to Solr?
> > >
> > > Say if I search for index with amount that is lesser that $2000, it
> will
> > > not match, unless when we do the search, we have to pass "20" to
> > Solr?
> > >
> > > Regards,
> > > Edwin
> > >
> > >
> > > On 7 December 2016 at 07:44, Chris Hostetter  >
> > > wrote:
> > >
> > > > : Thanks for your reply.
> > > > :
> > > > : That means the best fieldType to use for money is currencyField,
> and
> > > not
> > > > : any other fieldType?
> > > >
> > > > The primary use case for CurrencyField is when you want to do dynamic
> > > > currency fluctuations between multiple currency types at query time
> --
> > > but
> > > > to do that you either need to use the FileExchangeRateProvider and
> have
> > > > your owne backend system to update the exchange rates, or you have to
> > > have
> > > > an openexchangerates.org account, or implement some other provider
> > (with
> > > > custom solr java code)
> > > >
> > > >
> > > > If you only care about a single type of currency -- for example, if
> all
> > > > you care about is is US Dollars -- then just use either
> > > > TrieIntField or TrieLongField and represent in the smallest possible
> > > > increment you need to measure -- for US Dollars this would be cents.
> > ie:
> > > > $1234.56 would be put in your index as "123456"
> > > >
> > > >
> > > >
> > > > -Hoss
> > > > http://www.lucidworks.com/
> > > >
> > >
> >
>


Re: Difference between currency fieldType and float fieldType

2016-12-07 Thread Zheng Lin Edwin Yeo
Thanks for the reply.

How about using the double fieldType?
I tried that it works, as it is 64-bit, as compared to 32-bit for float.
But will it hit the same issue again if the amount exceeds 64-bit?

Regards,
Edwin


On 7 December 2016 at 15:28, Dorian Hoxha  wrote:

> Yeah, you'll have to do the conversion yourself (or something internal,
> like the currencyField).
>
> Think about it as datetimes. You store everything in utc (cents), but
> display to each user in it's own timezone (different currency, or just from
> cents to full dollars).
>
> On Wed, Dec 7, 2016 at 8:23 AM, Zheng Lin Edwin Yeo 
> wrote:
>
> > But if I index $1234.56 as "123456", won't it affect the search or facet
> if
> > I do a query directly to Solr?
> >
> > Say if I search for index with amount that is lesser that $2000, it will
> > not match, unless when we do the search, we have to pass "20" to
> Solr?
> >
> > Regards,
> > Edwin
> >
> >
> > On 7 December 2016 at 07:44, Chris Hostetter 
> > wrote:
> >
> > > : Thanks for your reply.
> > > :
> > > : That means the best fieldType to use for money is currencyField, and
> > not
> > > : any other fieldType?
> > >
> > > The primary use case for CurrencyField is when you want to do dynamic
> > > currency fluctuations between multiple currency types at query time --
> > but
> > > to do that you either need to use the FileExchangeRateProvider and have
> > > your owne backend system to update the exchange rates, or you have to
> > have
> > > an openexchangerates.org account, or implement some other provider
> (with
> > > custom solr java code)
> > >
> > >
> > > If you only care about a single type of currency -- for example, if all
> > > you care about is is US Dollars -- then just use either
> > > TrieIntField or TrieLongField and represent in the smallest possible
> > > increment you need to measure -- for US Dollars this would be cents.
> ie:
> > > $1234.56 would be put in your index as "123456"
> > >
> > >
> > >
> > > -Hoss
> > > http://www.lucidworks.com/
> > >
> >
>


Solr Issue

2016-12-07 Thread Monti Chandra
Hello team,
I am working on solr version to 6.2.1. It was working so nice for the first
20 days and now the server is restarting very slow(15-20 min).
Please get the hardware specs of my system below:
Linux version 3.10.0-327.el7.x86_64 (buil...@kbuilder.dev.centos.org) (gcc
version 4.8.3 20140911 (Red Hat 4.8.3-9) (GCC) )
kernel-3.10.0-327.el7.x86_64
It is working fine when i am taking solr directory to another server of
same configuartion.

As I am using 6.2.1 basically,  So i have downloaded the fresh
solr-6.2.1(without any collection) and try to start the same but the server
is not  able to start effectively. please get the log below:

0INFO  (main) [   ] o.e.j.u.log Logging initialized @257ms
131  INFO  (main) [   ] o.e.j.s.Server jetty-9.3.8.v20160314
146  INFO  (main) [   ] o.e.j.d.p.ScanningAppProvider Deployment monitor
[file:///root/temp/solr-6.2.1/server/contexts/] at interval 0
311  INFO  (main) [   ] o.e.j.w.StandardDescriptorProcessor NO JSP Support
for /solr, did not find org.apache.jasper.servlet.JspServlet
320  WARN  (main) [   ] o.e.j.s.SecurityHandler
ServletContext@o.e.j.w.WebAppContext@5383967b{/solr,file:///root/temp/solr-6.2.1/server/solr-webapp/webapp/,STARTING}{/root/temp/solr-6.2.1/server/solr-webapp/webapp}
has uncovered http methods for path: /
327  INFO  (main) [   ] o.a.s.s.SolrDispatchFilter
SolrDispatchFilter.init(): WebAppClassLoader=1465085305@57536d79
340  INFO  (main) [   ] o.a.s.c.SolrResourceLoader JNDI not configured for
solr (NoInitialContextEx)
340  INFO  (main) [   ] o.a.s.c.SolrResourceLoader using system property
solr.solr.home: /root/temp/solr-6.2.1/server/solr
340  INFO  (main) [   ] o.a.s.c.SolrResourceLoader new SolrResourceLoader
for directory: '/root/temp/solr-6.2.1/server/solr'
340  INFO  (main) [   ] o.a.s.c.SolrResourceLoader JNDI not configured for
solr (NoInitialContextEx)
340  INFO  (main) [   ] o.a.s.c.SolrResourceLoader using system property
solr.solr.home: /root/temp/solr-6.2.1/server/solr
345  INFO  (main) [   ] o.a.s.c.SolrXmlConfig Loading container
configuration from /root/temp/solr-6.2.1/server/solr/solr.xml
402  INFO  (main) [   ] o.a.s.c.CorePropertiesLocator Config-defined core
root directory: /root/temp/solr-6.2.1/server/solr
424  INFO  (main) [   ] o.a.s.c.CoreContainer New CoreContainer 1663619914
425  INFO  (main) [   ] o.a.s.c.CoreContainer Loading cores into
CoreContainer [instanceDir=/root/temp/solr-6.2.1/server/solr]
425  WARN  (main) [   ] o.a.s.c.CoreContainer Couldn't add files from
/root/temp/solr-6.2.1/server/solr/lib to classpath:
/root/temp/solr-6.2.1/server/solr/lib
434  INFO  (main) [   ] o.a.s.h.c.HttpShardHandlerFactory created with
socketTimeout : 60,connTimeout : 6,maxConnectionsPerHost :
20,maxConnections : 1,corePoolSize : 0,maximumPoolSize :
2147483647,maxThreadIdleTime : 5,sizeOfQueue : -1,fairnessPolicy :
false,useRetries : false,connectionsEvictorSleepDelay :
5000,maxConnectionIdleTime : 4,
592  INFO  (main) [   ] o.a.s.u.UpdateShardHandler Creating
UpdateShardHandler HTTP client with params:
socketTimeout=60&connTimeout=6&retry=true
594  INFO  (main) [   ] o.a.s.l.LogWatcher SLF4J impl is
org.slf4j.impl.Log4jLoggerFactory
594  INFO  (main) [   ] o.a.s.l.LogWatcher Registering Log Listener [Log4j
(org.slf4j.impl.Log4jLoggerFactory)]
596  INFO  (main) [   ] o.a.s.c.CoreContainer Security conf doesn't exist.
Skipping setup for authorization module.
596  INFO  (main) [   ] o.a.s.c.CoreContainer No authentication plugin used.
127902 INFO  (main) [   ] o.a.s.c.CorePropertiesLocator Looking for core
definitions underneath /root/temp/solr-6.2.1/server/solr
127906 INFO  (main) [   ] o.a.s.c.CorePropertiesLocator Found 0 core
definitions
127909 INFO  (main) [   ] o.a.s.s.SolrDispatchFilter
user.dir=/root/temp/solr-6.2.1/server
127909 INFO  (main) [   ] o.a.s.s.SolrDispatchFilter
SolrDispatchFilter.init() done
127918 INFO  (main) [   ] o.e.j.s.h.ContextHandler Started
o.e.j.w.WebAppContext@5383967b
{/solr,file:///root/temp/solr-6.2.1/server/solr-webapp/webapp/,AVAILABLE}{/root/temp/solr-6.2.1/server/solr-webapp/webapp}
127926 INFO  (main) [   ] o.e.j.s.ServerConnector Started
ServerConnector@1573f9fc{HTTP/1.1,[http/1.1]}{0.0.0.0:}
127926 INFO  (main) [   ] o.e.j.s.Server Started @128185ms


Is there any H/w, OS or kernel level threat caused by running solr? Please
help me I am wondering for a very long time.


-- 

*Thanks & Regards*

MONTI CHANDRA


Configuration folder for the collection generated with instead of just

2016-12-07 Thread Nicole Bilić
Hi all,


We are using Lucidworks Fusion on top of Solr and recently we’ve
encountered an unexpected behavior. We’ve created bash scripts which we use
to create collections in Solr using the Fusion API and upload the
collection configuration (with bash $ZKCLIENT -cmd upconfig -confdir $path
-confname $collection -z $HOST:$ZKPORT).

We had issues that the index fields of our custom configurations were not
available (ie. our custom field types and fields we’ve defined in
managed-schema). So we checked in Solr admin and found out that the
configuration folder for the collection was generated with
 instead of just  (eg.
https://drive.google.com/file/d/0B9Nu5vSE4ep8OUdyckZoV2FMdlE/view?usp=sharing).


Our collection configurations are under the one with the id in the end. So
the configuration set that was uploaded by our scripts was never used and
therefore the index fields were not existing. Also, we did not manage to
discover a pattern (or cause) for when does this happen (because sometimes
the issue does not occur and sometimes it does and to us it seems pretty
nondeterministic).

Furthermore, in the screenshot above, it’s possible to see the same thing
happened with system_metrics collection configs, which we do not create or
modify with our scripts. Therefore, the scripts should not be the cause of
this behavior.

What we are trying to determine is if this issue is coming from Fusion,
Solr or Zookeeper. Does someone know in which case the collection
configuration folder is generated with the id in the end and how to avoid
it? Thank you!


Nicole


Re: Memory leak in Solr

2016-12-07 Thread William Bell
What do you mean by JVM level? Run Solr on different ports on the same
machine? If you have a 32 core box would you run 2,3,4 JVMs?

On Sun, Dec 4, 2016 at 8:46 PM, Jeff Wartes  wrote:

>
> Here’s an earlier post where I mentioned some GC investigation tools:
> https://mail-archives.apache.org/mod_mbox/lucene-solr-user/
> 201604.mbox/%3c8f8fa32d-ec0e-4352-86f7-4b2d8a906...@whitepages.com%3E
>
> In my experience, there are many aspects of the Solr/Lucene memory
> allocation model that scale with things other than documents returned.
> (such as cardinality, or simply index size) A single query on a large index
> might consume dozens of megabytes of heap to complete. But that heap should
> also be released quickly after the query finishes.
> The key characteristic of a memory leak is that the software is allocating
> memory that it cannot reclaim. If it’s a leak, you ought to be able to
> reproduce it at any query rate - have you tried this? A run with, say, half
> the rate, over twice the duration?
>
> I’m inclined to agree with others here, that although you’ve correctly
> attributed the cause to GC, it’s probably less an indication of a leak, and
> more an indication of simply allocating memory faster than it can be
> reclaimed, combined with the long pauses that are increasingly unavoidable
> as heap size goes up.
> Note that in the case of a CMS allocation failure, the fallback full-GC is
> *single threaded*, which means it’ll usually take considerably longer than
> a normal GC - even for a comparable amount of garbage.
>
> In addition to GC tuning, you can address these by sharding more, both at
> the core and jvm level.
>
>
> On 12/4/16, 3:46 PM, "Shawn Heisey"  wrote:
>
> On 12/3/2016 9:46 PM, S G wrote:
> > The symptom we see is that the java clients querying Solr see
> response
> > times in 10s of seconds (not milliseconds).
> 
> > Some numbers for the Solr Cloud:
> >
> > *Overall infrastructure:*
> > - Only one collection
> > - 16 VMs used
> > - 8 shards (1 leader and 1 replica per shard - each core on separate
> VM)
> >
> > *Overview from one core:*
> > - Num Docs:193,623,388
> > - Max Doc:230,577,696
> > - Heap Memory Usage:231,217,880
> > - Deleted Docs:36,954,308
> > - Version:2,357,757
> > - Segment Count:37
>
> The heap memory usage number isn't useful.  It doesn't cover all the
> memory used.
>
> > *Stats from QueryHandler/select*
> > - requests:78,557
> > - errors:358
> > - timeouts:0
> > - totalTime:1,639,975.27
> > - avgRequestsPerSecond:2.62
> > - 5minRateReqsPerSecond:1.39
> > - 15minRateReqsPerSecond:1.64
> > - avgTimePerRequest:20.87
> > - medianRequestTime:0.70
> > - 75thPcRequestTime:1.11
> > - 95thPcRequestTime:191.76
>
> These times are in *milliseconds*, not seconds .. and these are even
> better numbers than you showed before.  Where are you seeing 10 plus
> second query times?  Solr is not showing numbers like that.
>
> If your VM host has 16 VMs on it and each one has a total memory size
> of
> 92GB, then if that machine doesn't have 1.5 terabytes of memory, you're
> oversubscribed, and this is going to lead to terrible performance...
> but
> the numbers you've shown here do not show terrible performance.
>
> > Plus, on every server, we are seeing lots of exceptions.
> > For example:
> >
> > Between 8:06:55 PM and 8:21:36 PM, exceptions are:
> >
> > 1) Request says it is coming from leader, but we are the leader:
> > update.distrib=FROMLEADER&distrib.from=HOSTB_ca_1_
> 1456430020/&wt=javabin&version=2
> >
> > 2) org.apache.solr.common.SolrException: Request says it is coming
> from
> > leader, but we are the leader
> >
> > 3) org.apache.solr.common.SolrException:
> > org.apache.solr.client.solrj.SolrServerException: Tried one server
> for read
> > operation and it timed out, so failing fast
> >
> > 4) null:org.apache.solr.common.SolrException:
> > org.apache.solr.client.solrj.SolrServerException: Tried one server
> for read
> > operation and it timed out, so failing fast
> >
> > 5) org.apache.solr.common.SolrException:
> > org.apache.solr.client.solrj.SolrServerException: Tried one server
> for read
> > operation and it timed out, so failing fast
> >
> > 6) null:org.apache.solr.common.SolrException:
> > org.apache.solr.client.solrj.SolrServerException: Tried one server
> for read
> > operation and it timed out, so failing fast
> >
> > 7) org.apache.solr.common.SolrException:
> > org.apache.solr.client.solrj.SolrServerException: No live
> SolrServers
> > available to handle this request. Zombie server list:
> > [HOSTA_ca_1_1456429897]
> >
> > 8) null:org.apache.solr.common.SolrException:
> > org.apache.solr.client.solrj.SolrServerException: No live
> SolrServers
> > available to handle t