Hi guys, I'd like to get your tips on how to run a Solr fork at my
company. I know Yonik has a "heliosearch" fork, and I'm sure many others
have a fork. There have been times where I want to add features to an
existing core plugin, and subclassing isn't possible so I end up copying
the source
Yes, I'm aware that building an index is expensive and I will remove
"buildOnStartup" once I have it working. The field I added was an
attempt to get it working...
I have attached my latest version of solrconfig.xml and schema.xml (both
are in the same attachment), except that I have removed
On 10/13/2015 9:30 AM, Dyer, James wrote:
Mark,
The older spellcheck implementations create an n-gram sidecar index, which is
why you're seeing your name split into 2-grams like this. See the IR Book by
Manning et al, section 3.3.4 for more information. Based on the results you're
getting,
Ryan,
>From a "solr-user" perspective :) I would advise against forking Solr. Some
of our consulting business is "people who forked Solr, need to upgrade, and
now have gotten themselves into hot water."
I would try, in the following order
1. Creating a plugin (sounds like you can't do this)
2.
Thanks for the feedback, forking lucene/solr is my last resort indeed.
1) It's not about creating fresh new plugins. It's about modifying
existing ones or core solr code.
2) I want to submit the patch to modify core solr or lucene code, but I
also want to run it in prod before its accepted and
I suspect these questions should go the Lucene Dev list instead. This
one is more for those who build on top of standard Solr.
Regards,
Alex.
Solr Analyzers, Tokenizers, Filters, URPs and even a newsletter:
http://www.solr-start.com/
On 16 October 2015 at 12:07, Ryan Josal
I am looking at using child documents and noticed that if I specify a child
and parent with the same key solr indexes this fine and I can retrieve both
documents separately. Is this expected to work?
-Jamie
Hi,
I have a data model where I would store and index a lot of integer values
with a very restricted range (e.g. 0-255), so theoretically the 32 bits of
Solr's integer fields are complete overkill. I want to be able to to things
like vector distance calculations on those fields. Should I worry
On 15 October 2015 at 23:54, Matteo Grolla wrote:
> Don't think so,
> the default behaviour at 4), to my knowledge,is to wait 3 minutes
> (leaderVoteWait) for all replicas to come up to avoid electing a leader
> with stale data.
So the observed behaviour is unexpected
Thanks for pointing out as I am using Solr cloud 5.3. However, it looks
like they are talking about boolean operation in context field and not the
support of context field itself. Are you sure that context filtering is not
supported with any lookup prior to 5.4?
On Oct 16, 2015 12:47 PM,
Hi Robert,
current Solr compression will work really well , both for Stored and
DocValues contents.
Related the index term dictionaries, I ask for some help to other experts
as I never checked how the actual compression works in there, but I assume
it is quite efficient.
Usually the field type
This will sound silly, but which version of Solr are you using ?
According to :
https://issues.apache.org/jira/browse/SOLR-7888
This new cool feature will be included in solr 5.4 .
Cheers
On 15 October 2015 at 22:53, Salman Ansari wrote:
> Hi guys,
>
> I am working
Yes, as Jan confirmed, I am sure it was not there in 5.3 :)
Cheers
On 16 October 2015 at 12:10, Jan Høydahl wrote:
> Yesm
>
> Context filtering is a new feature in yet-to-be-released Solr5.4.
> So you have to build branch_5x from source yourself to try it out.
>
> --
>
Yesm
Context filtering is a new feature in yet-to-be-released Solr5.4.
So you have to build branch_5x from source yourself to try it out.
--
Jan Høydahl, search solution architect
Cominvent AS - www.cominvent.com
> 16. okt. 2015 kl. 12.35 skrev Salman Ansari :
>
>
Hi Matthias,
you should use . in your expressions. So for
example, here
WHERE fb.EBI_NR='${firma.firma_ebi_nr}'
should be
WHERE fb.EBI_NR='${firma.EBI_NR}'
Best,
Andrea
2015-10-16 13:40 GMT+02:00 Matthias Fischer :
> Hello everybody,
>
> I am trying to
SolrJ does not have any file crawler built in.
But you are free to steal code from SimplePostTool.java related to directory
traversal,
and then index each document found using SolrJ.
Note that SimplePostTool.java tries to be smart with what endpoint to post
files to,
xml, csv and json content
Hello everybody,
I am trying to import from an Oracle DB 11g2 via DIH using SOLR 5.3.1.
In my relational DB there are company addresses (table tb_firmen_adressen) and
branches (table tb_branchen). They have an n:m relationship using the join
table tb_firmen_branchen.
Now I would like to find
Also, check this link for SolrJ example code (including the recursion):
https://lucidworks.com/blog/2012/02/14/indexing-with-solrj/
Geraint
Geraint Duck
Data Scientist
Toxicology and Health Sciences
Syngenta UK
Email: geraint.d...@syngenta.com
-Original Message-
From: Jan Høydahl
Thanks for your reply Scott.
I tried
bs.language=de=de
Unfortunately the problem still occurs.
I have just discovered that the problem does not only affect "ß" but
also "æ" (which is mapped to "ae"
at query and index time)
q=hae --> hæna
So it seems to me that the problem is related to
Thank you, Andrea, for answering so quickly.
However I got further errors. I also had to change
"firma_ebi_nr" to "id". But it
still does not work properly. It seems that an id is auto generated for the
company documents but not for the nested ones (the business branches). Any
ideas how to
Hi Matthias,
I guess the company.id field is not unique so you need a "compound"
uniqueKey on Solr, which is not strctly possible. As consequence of that
(company) UUID is probably created before the index phase by an
UpdateRequestProcessor [1] so you should check your solrconfig.xml and, if
I'm
Hi,
I am aware of Solr’s geospatial capabilities. However, wondering what is the
best way to visualise Solr geospatial data. Is there any native support in
Solritas or is there any other mechanism that suits thes requirement best? For
example, if my problem is to find the best possible route
Under the covers, Lucene stores ints in a packed format, so I'd just count
on that for a first pass.
What is "a lot of integer values"? Hundreds of millions? Billions? Trillions?
Unless you give us some indication of scale, it's hard to say anything
helpful. But unless you have some evidence
One more observation made is that tomcat's acceptor thread for http
disappears (http-bio-8080-acceptor thread) and due to this no incoming
connections could be opened on http. During this time ZK potentially
thinks node is up and shows green from leader.
On 10/13/15 9:17 AM, Erick Erickson
Thanks for your advice. I also found this method which so far has been able
to traverse all the documents in the folder and index them in Solr.
public static void showFiles(File[] files) {
for (File file : files) {
if (file.isDirectory()) {
System.out.println("Directory: "
On Fri, Oct 16, 2015, at 04:00 PM, Ryan Josal wrote:
> Thanks for the feedback, forking lucene/solr is my last resort indeed.
>
> 1) It's not about creating fresh new plugins. It's about modifying
> existing ones or core solr code.
> 2) I want to submit the patch to modify core solr or lucene
26 matches
Mail list logo