I'm having the same issue with 4.10.3
I'm performing various task on clusterstate API and getting random timeouts
throguhout the day.
--
View this message in context:
http://lucene.472066.n3.nabble.com/CLUSTERSTATE-timeout-tp4199367p4199501.html
Sent from the Solr - User mailing list archive
Hi, it worked!
The issue was originally on WAS 7, but has somehow regressed to WebSphere
8.5.
Thanks.
On Thu, Feb 19, 2015 at 10:13 PM, Chris Hostetter hossman_luc...@fucit.org
wrote:
: Hi! Solr is starting up dormant for me, until a client wake it up with
a
: REST request, or I open admin
On 4/13/2015 10:11 PM, vsilgalis wrote:
just a couple of notes:
this a 2 shard setup with 2 nodes per shard.
Currently these are on VMs with 8 cores and 8GB of ram each (java max heap
is ~5588mb but we usually never even get that high) backed by a NFS file
store which we store the indexes
Dears,
Hi,
I have strange problem with Solr 4.10.x. My problem is when I do searching
on solr Zero date which is 0002-11-30T00:00:00Z if more than one filter
be considered, the results became invalid. For example consider this
scenario:
When I search for a document with
Hi
I might misunderstand you, but if you are talking about securing the
actual files/folders of the index, I do not think this is a Solr/Lucene
concern. Use standard mechanisms of your OS. E.g. on linux/unix use
chown, chgrp, chmod, sudo, apparmor etc - e.g. allowing only root to
write the
Hi,
I am using external field for price field since it changes frequently.
generate facets using external field? how?
I understand that faceting requires indexing and external fields fields are
not actually indexed.
--
Thanks Regards,
Jainam Vora
Hi All,
we recently migrated from SOLR 3.6 to SOLR 4, while indexing in SOLR 4
we are getting below exception.
Apr 1, 2015 9:22:57 AM org.apache.solr.common.SolrException log
SEVERE: null:org.apache.solr.common.SolrException: Exception writing
document id 932684555 to the index; possible
Right now index size is about 10GB on each shard (yes I could use more RAM),
but I'm looking more for a step up then step down approach. I will try
adding more RAM to these machines as my next step.
1. Zookeeper is external to these boxes in a three node cluster with more
than enough RAM to keep
Hi,
Here are the solr-config xml and the error log from Solr logs for your
reference. As mentioned earlier, I didnt make any changes to the
solr-config.xml as I am using the xml file out of the box one that came
with the default installation.
Please let me know your thoughts on why these issues
Andrea,
Yes, I am using the stock schema.xml that comes with the example server of
Solr-4.10.2 Hence not sure why the PDF content is not getting extracted and
put into the content field in the index.
Please find the log information for the Parsing error below.
It seems something like https://issues.apache.org/jira/browse/TIKA-1251.
I see you're using Solr 4.10.2 which uses Tika 1.5 and that issue seems
to be fixed in Tika 1.6.
I agree with Erik: you should try with another version of Tika.
Best,
Andrea
On 04/14/2015 06:44 PM, Vijaya Narayana Reddy
Hello,
I have a pretty basic question: how can I sort by a copyfield?
My schema conf is:
field name=name type=text_general_edge_ngram indexed=true
stored=true omitNorms=true termVectors=true/
field name=name_sort type=string indexed=true stored=false/
copyField source=name
14 April 2015 - The Lucene PMC is pleased to announce the release of
Apache Solr 5.1.0.
Solr 5.1.0 is available for immediate download at:
http://www.apache.org/dyn/closer.cgi/lucene/solr/5.1.0
Solr 5.1.0 includes 39 new features, 40 bug fixes, and 36 optimizations
/ other changes from over 60
looks like this is just a file that Tika can't handle, based on this line:
bq: org.apache.tika.exception.TikaException: Unexpected
RuntimeException from org.apache.tika.parser.microsoft.OfficeParser
You might be able to get some joy from parsing this from Java and see
if a more recent Tika would
Hi,
I am trying to index PDF and Microsoft Office files (.doc, .docx, .ppt,
.pptx, .xlx, and .xlx) files into Solr. I am facing the following issues.
Request to please let me know what is going wrong with the indexing
process.
I am using solr 4.10.2 and using the default example server
Hi Vijay,
Please paste an extract of your schema, where the content field (the
field where the PDF text shoudl be) and its type are declared.
For the other issue, please paste the whole stacktrace because
org.apache.tika.parser.microsoft.OfficeParser*
says nothing. The complete stacktrace (or
What does your main query look like? Normally we don't speak of searching
with the fq parameter - it filters the results, but the actual searching is
done via the main query with the q parameter.
-- Jack Krupansky
On Tue, Apr 14, 2015 at 4:17 AM, Ali Nazemian alinazem...@gmail.com wrote:
Elisabeth,
Currently ConjunctionSolrSpellChecker only supports adding
WordBreakSolrSpellchecker to IndexBased- FileBased- or DirectSolrSpellChecker.
In the future, it would be great if it could handle other Spell Checker
combinations. For instance, if you had a (e)dismax query that searches
Hi,
solrconfig.xml (especially if you didn't touch it) should be good. What
about the schema? Are you using the one that comes with the download
bundle, too?
I don't see the stacktrace..did you forget to paste it?
Best,
Andrea
On 04/14/2015 06:06 PM, Vijaya Narayana Reddy Bhoomi Reddy
Thanks for your answer!
I didn't realize this what not supposed to be done (conjunction of
DirectSolrSpellChecker and FileBasedSpellChecker). I got this idea in the
mailing list while searching for a solution to get a list of words to
ignore for the DirectSolrSpellChecker.
Well well well, I'll
Hi all -
I've just upgraded my dev install of Solr (cloud) from 4.10 to 5.0. Our
client is written in Go, for which I am not aware of a client, so we wrote
our own. One tricky bit for this was the routing logic; if a document has
routing prefix X and belong to collection Y, we need to know
That's a good point - if he's talking about securing the Solr filesystem, he
can use standard mechanisms.
You can also go beyond user/group/other permissions if your filesystem supports
it. You can use Posix ACLs on many local linux filesystems.
-Original Message-
From: Per
Try doing a manual extraction request directly to Solr (not via SolrJ) and
use the extractOnly option to see if the content is actually extracted.
See:
https://cwiki.apache.org/confluence/display/solr/Uploading+Data+with+Solr+Cell+using+Apache+Tika
Also, some PDF files actually have the content
Vijay,
You could try different excel files with different formats to rule out the
issue is with TIKA version being used.
Thanks
Murthy
On Wed, Apr 15, 2015 at 9:35 AM, Terry Rhodes trhodes...@gmail.com wrote:
Perhaps the PDF is protected and the content can not be extracted?
i have an
Perhaps the PDF is protected and the content can not be extracted?
i have an unverified suspicion that the tika shipped with solr 4.10.2
may not support some/all office 2013 document formats.
On 4/14/2015 8:18 PM, Jack Krupansky wrote:
Try doing a manual extraction request directly to
We ran into this during our indexing process running on 4.10.3. After
increasing zookeeper timeouts, client timeouts, socket timeouts,
implementing retry logic on our loading process the thing that worked was
to change the Hard Commit timing. We were performing a Hard Commit every 5
minutes and
Hi Joe,
This should help you:
http://lucene.apache.org/solr/5_1_0/changes/Changes.html#v5.1.0.upgrading_from_solr_5.0
On Tue, Apr 14, 2015 at 12:47 PM, Joseph Obernberger
j...@lovehorsepower.com wrote:
Great news!
Any tips on how to do an upgrade from 5.0.0 to 5.1.0?
Thank you!
-Joe
On
On 4/14/2015 11:32 AM, Pedro Figueiredo wrote:
And when I try to sort by name_sort the following error is raised:
error: {
msg: sort param field can't be found: name_sort,
code: 400
}
What was the exact sort parameter you sent to Solr?
Did you reload the core or restart Solr
Hi Pedro
Please post the request that produces that error
Andrea
On 14 Apr 2015 19:33, Pedro Figueiredo pjlfigueir...@criticalsoftware.com
wrote:
Hello,
I have a pretty basic question: how can I sort by a copyfield?
My schema conf is:
field name=name type=text_general_edge_ngram
Hi Ian,
As per my understanding, Solrj does not use Zookeeper watches but instead
caches the information (along with a TTL). You can find more information
here,
https://issues.apache.org/jira/browse/SOLR-5473
https://issues.apache.org/jira/browse/SOLR-5474
Regards
Hrishikesh
On Tue, Apr 14,
Thank you.. This really helps.
--
View this message in context:
http://lucene.472066.n3.nabble.com/Disable-or-limit-the-size-of-Lucene-field-cache-tp4198798p4199646.html
Sent from the Solr - User mailing list archive at Nabble.com.
Hi Hrishikesh,
Thanks for the pointers - I had not looked at SOLR-5474
https://issues.apache.org/jira/browse/SOLR-5474 previously. Interesting
approach... I think we will stick with trying to keep zk watches open from
all clients to all collections for now, but if that starts to be a
bottleneck
Great news!
Any tips on how to do an upgrade from 5.0.0 to 5.1.0?
Thank you!
-Joe
On 4/14/2015 2:39 PM, Timothy Potter wrote:
I apologize - Yonik prepared these nice release notes for 5.1 and I
neglected to include them:
Solr 5.1 Release Highlights:
* The new Facet Module, including the
Folks, there's a new JSON Facet API in the just released Solr 5.1
(actually, a new facet module under the covers too).
It's marked as experimental so we have time to change the API based on
your feedback. So let us know what you like, what you would change,
what's missing, or any other ideas you
Hello,
I am using Solr 4.10.1 and trying to use DirectSolrSpellChecker and
FileBasedSpellchecker in same request.
I've applied change from patch 135.patch (cf Solr-6271). I've tried running
the command patch -p1 -i 135.patch --dry-run but it didn't work, maybe
because the patch was a fix to Solr
I apologize - Yonik prepared these nice release notes for 5.1 and I
neglected to include them:
Solr 5.1 Release Highlights:
* The new Facet Module, including the JSON Facet API.
This module is currently marked as experimental to allow for
further API feedback and improvements.
* A new JSON
36 matches
Mail list logo