Hi folks,
I am currently migrating our Solr servers from a 4.0.0 nightly build
(aprox. November 2011, which worked very well) to the newly released
4.0.0 and am running into some issues concerning the existing
DataImportHandler configuratiions. Maybe you have an idea where I am
going wrong here.
Can you please share some information on Setting up Solr 4.0 as a singleCore.
I tried doing it and keep seeing ClassNotFound Exception for
KeywordTokenizerFactory. on server start up.
I see the jar files being loaded in the logs but its unable to find the
class.
Can you let me know what jars
Hello everyone.
I have two questions. I am considering using Solr 4.0 to perform full
searches on the data output in real-time by a Storm cluster
(http://storm-project.net/).
1. In particular, I'm concerned whether Solr would be able to keep up
with the 2000-message-per-second throughput of the
On 19 October 2012 12:07, Romita Saha romita.s...@sg.panasonic.com wrote:
[...]
My data-config file is :
entity name=camera
query=SELECT id FROM camera
field column=id name=id/
field column=data name=data/
/entity
The
status shows that all your 4 records were not indexed.
str name=Total Documents Failed4/str
On Fri, Oct 19, 2012 at 12:22 PM, Romita Saha
romita.s...@sg.panasonic.comwrote:
Hi,
Even after indexing a mysql table,in solr am not able to retrieve data
after querying. Here is the status after
Hi,
your question is not easy to answer. It depends on so many things, that
there is no standard way to realize an enterprise solution and time planning
aspects are depending on so much things.
I can try to give you some brief notes about our solution, but there are
some differences in target
Thanks Mark!
Cheers, Jeeva
On Oct 19, 2012, at 8:35 AM, Mark Miller markrmil...@gmail.com wrote:
Yes, those exceptions are fine. These are cases where we try to delete the
node if it's there, but don't care if it's not there - things like that. In
some of these cases, ZooKeeper logs
Hello SOLR expert,
yesterday in our group we realized that a danger we may need to face is that a
search result includes very similar results.
Of course, one would expect skimming so that duplicates that show almost the
same results in a search result would be avoided but we fear that this is
Hi Alexandre,
Yes it is active. ManifoldCF 1.0.1 is released yesterday :)
You can index content of SharePoint 2010 to Solr 4.0.0 .
'End user documentation' and 'in action book' are two main resources.
http://manifoldcf.apache.org/release/release-1.0.1/en_US/end-user-documentation.html
Hi Paul,
yes that`s a typical problem in configuring a search engine. A solution
depends on your data. Sometimes you can overcome this problem by fine tuning
your search engine on boosting level. Thats not easy and always based on
trail and error tests.
Another thing you can do is to try to
Hi,
I made a Solr XML data source in lucidworks enterprise v2.1. When I search in
Solr Admin for text. I am unable to get the result.
Could you help me in this?
Thanks Regards,
Leena Jawale
Software Engineer Trainee
BFS BU
Phone No. - 9762658130
Email -
I will be out of the office starting 18/10/2012 and will not return until
23/10/2012.
Please email to itsta...@actionimages.com for any urgent issues.
Action Images is a division of Reuters Limited and your data will therefore be
protected
in accordance with the Reuters Group Privacy / Data
Fetching the same records using a raw Http-Request works fine and
characters are OK. I am actually considering to fetch the data in Java
via raw Http-Requests + XSLTResponsWriter as a workaround, but I want to
try it first using the 'native' way with SolrJ.
Andreas
Jack Krupansky
I wonder if you're getting hit by the browser caching the admin page and
serving up the old version? What happens if you try from a different
browser or purge the browser cache?
Of course you have to refresh the master admin page, there's no
automatic update but I assume you did that.
Best
Erick
I have found that segment flush is controlled by
DocumentWriterFlushControl, and indexing is implemented by
DocumentWriterPerThread. DocumentWriterFlushControl has information about
number of doc and size of RAM buffer, but this seemed be shared by
all DocumentWriterPerThread. Is that RAM limit is
Hi
does anybody could give some direction / suggestion on how to correctly
configure and use the SimpleTextCodec?
http://lucene.apache.org/core/4_0_0-BETA/core/org/apache/lucene/codecs/simpletext/SimpleTextCodec.html
i'd like to do some test for debugging purpose, but i'm not shure on how to
Thanks for the quiz. It is refreshing. Do you plan on covering other parts
of SOLR management, like various handlers, scoring, plugins, sharding etc?
Dmitry
On Wed, Oct 17, 2012 at 7:12 PM, Yulia Crowder yulia.crow...@gmail.comwrote:
I love Solr!
I have searched for a quiz about Solr and
Leena -
It's best to ask LucidWorks related questions at http://support.lucidworks.com
rather than in this e-mail list.
As for your issue more information is needed in order to assist. Did you
start the Solr XML crawler? Does your data source show that there are
documents in the index?
Leena,
Please ask on Lucid fora. You'll get better and faster help there.
Otis
--
Performance Monitoring - http://sematext.com/spm
On Oct 19, 2012 5:54 AM, Leena Jawale leena.jaw...@lntinfotech.com
wrote:
Hi,
I made a Solr XML data source in lucidworks enterprise v2.1. When I search
in
Hello,
Looking to get all documents with empty geolocalisation field, I have not
found any way to do it, with ['' to *],
geodata being a specific field, do you have any solution ?
Thanks,
Jul
--
View this message in context:
Hi all,
Congrats on the 4.0.0 delivery, it's a pleasure to work with!
I have a small problem that I am trying to elegantly resolve: while using
multi-select faceting it might happen that a facet is selected which is not
part of the facet list (due to limit for example). When executing the
sorry, I mean this field called geodata in my schema
fieldType name=location class=solr.LatLonType
subFieldSuffix=_coordinate/
field name=geodata type=location indexed=true stored=true/
--
View this message in context:
On Fri, Oct 19, 2012 at 2:50 AM, higashihara_hdk
higashihara_...@es-planning.jp wrote:
Hello everyone.
I have two questions. I am considering using Solr 4.0 to perform full
searches on the data output in real-time by a Storm cluster
(http://storm-project.net/).
1. In particular, I'm
Hello,
Did you try q=-geodata:[* TO *] ? (Note the '-' (minus))
This reads as documents without any value for field named geodata.
Also if you plan to use this intensively, you'd better declare a boolean
field telling if geodata are set or not and set a value to each doc,
because the
Did you look think of using 'facet.query' ?
Adding 'facet.query=category:Article' to your url should return what
you expected.
Franck Brisbart
Le vendredi 19 octobre 2012 à 15:18 +0200, Stephane Gamard a écrit :
Hi all,
Congrats on the 4.0.0 delivery, it's a pleasure to work with!
I
Hi all,
I know there have been many posts about this already and I have done
my best to read through them but one lingering question remains. When
doing performance testing on a Solr instance (under normal production
like circumstances, not the ones where commits are happening more
frequently
Hello all,
I am having problems indexing xml attributes using the DIH.
I have the following xml:
root
Stuff attr1=some attr attr2=another attr
...
/Stuff
/root
I am using the following XPath for my fields:
field column=attr1 xpath=/root/Stuff/@attr1 /
field column=attr2
What about querying on the dynamic lat/long field to see if there are
documents that do not have the dynamic _latlon0 or whatever defined?
On Fri, Oct 19, 2012 at 8:17 AM, darul daru...@gmail.com wrote:
I have already tried but get a nice exception because of this field type :
--
View this
Thanks Chris for your reply. I really need some help here.
1) If I put the apache-solr-dataimporthandler-*.jar files in solr/lib
folder, the jar files are loading. I see that in the tomcat logs. But in the
end it says 'ClassNotFoundException DataImportHandler'.
2) So If I remove
Your idea looks great but with this schema info :
fieldType name=point class=solr.PointType dimension=2
subFieldSuffix=_d/
fieldType name=location class=solr.LatLonType
subFieldSuffix=_coordinate/
fieldtype name=geohash class=solr.GeoHashField/
.
field name=geodata type=location
Hi, all.
The content I'm trying to index contains dollar signs that should be
indexed and matched, e.g., $1.
I've set up my schema to index the dollar sign, and am able to successfully
match it with the query analyzer; searching for $1 matches $1.
However, the highlighter doesn't seem to
very often when we try to shutdown tomcat, we got following error in
catalina.out indicating a solr thread can not be stopped, the tomcat results
hanging, we have to kill -9, which we think lead to some core corruptions in
our production environment. please help ...
catalina.out:
... ...
Oct
by the way, I am running tomcat 6, solr 3.5 on redhat 2.6.18-274.el5 #1 SMP
Fri Jul 8 17:36:59 EDT 2011 x86_64 x86_64 x86_64 GNU/Linux
--
View this message in context:
http://lucene.472066.n3.nabble.com/solr-memory-leak-prevent-tomcat-shutdown-tp4014788p4014792.html
Sent from the Solr - User
So here is my spec for lat/long (similar to yours except I explicitly
define the sub-field names for clarity)
fieldType name=latLon class=solr.LatLonType subFieldSuffix=_latLon/
field name=location type=latLon indexed=true stored=true/
!-- Could use dynamic fields here but prefer explicitly
I have a document with name ABC 102030 XYZ and if i search for this document
with ABC and -10 then i dont get this document (which is correct behavior)
but when i do ABC and -10 i don't get the correct result back. Any
explanation around this.
--
View this message in context:
It's not the browser cache. I have tried reloading the admin page and
accessing the admin page from another machine. Both show the older index
version and generation. On the slave, replication did kicked in and show
the new index version and generation for the slave. But the slave admin
page
What exactly is the precise symptom - give us an example with field names of
source and dest and what precise value is in fact being indexed. Is the
entire field value being indexed as a single term/string (if analyzer is not
being applied)? Or, what?
-- Jack Krupansky
-Original
Because you used solr.StandardTokenizerFactory which will tokenize terms at
some delimiters - such as the hyphens that surround your errant 404 case.
Try solr.WhitespaceTokenizerFactory or solr.KeywordTokenizerFactory.
And maybe rename your field type from text_general_trim to text_exact
hello jack,
thank you very much for the reply - i will re-test and let you know.
really appreciate it ;)
thx
mark
--
View this message in context:
http://lucene.472066.n3.nabble.com/need-help-with-exact-match-search-tp4014832p4014848.html
Sent from the Solr - User mailing list archive at
Lance,
I have seen this error when the Solr process hit the maximum file
descriptors (because the commit triggered an optimize). Make sure your
maxfds is set as high as possible. In my case, 1024 was not nearly
sufficient.
--Casey
On 10/19/12 6:20 PM, Lance Norskog wrote:
When a transient
Do other fields get added?
Do these fields have type problems? I.e. is 'attr1' a number and you are adding
a string?
There is a logging EP that I think shows the data found- I don't know how to
use it.
Is it possible to post the whole DIH script?
- Original Message -
| From: Billy
Hi Amit,
I'm not sure I follow what you are after...
Yes, seeing how queries that result in cache misses perform is
valuable (esp. if you have low cache hit rate in production)
But figuring out if you chose a bad field type or bad faceting method
or doesn't require profiling - you can review
Hi Paul,
We've done this for a client in the past via a custom SearchComponent
and it worked well. Yes, it involved some post-processing, but on the
server, not client. I *think* we saw 10% performance degradation.
Otis
--
Search Analytics - http://sematext.com/search-analytics/index.html
If it worked before and does not work now, I don't think you are doing anything
wrong :)
Do you have a different version of your JDBC driver?
Can you make a unit test with a minimal DIH script and schema?
Or, scan through all of the JIRA issues against the DIH from your old Solr
capture date.
44 matches
Mail list logo