Thanks all for your response.
I presume this conversation concludes that indexing around 1Billion
documents per shard won't be a problem, as I have 10 Billion docs to index,
so approx 10 shards with 1 Billion each should be fine with it and how
about Memory, what size of RAM should be fine for
On 6/4/2014 12:45 AM, Vineet Mishra wrote:
Thanks all for your response.
I presume this conversation concludes that indexing around 1Billion
documents per shard won't be a problem, as I have 10 Billion docs to index,
so approx 10 shards with 1 Billion each should be fine with it and how
about
Hi,
I am new in solr and i am trying to create the custom filter, to create
that filter i just copied the lowercasefilter and making all the changes in
the increment token, but to make sure that my changes are applying
properly, i am also printing some debugging info in log.
public final boolean
Hi all,
Ahment was suggesting to eventually use UIMA integration because OpenNLP
has already an integration with Apache UIMA and so you would just have to
use that [1].
And that's one of the main reason UIMA integration was done: it's a
framework that you can easily hook into in order to plug
Hi Aman,
What you see is normal. If you want to convert it to a string use
this.termAttribute.toString();
Please see source code of org.apache.lucene.analysis.br.BrazilianStemFilter for
an example.
Ahmet
On Wednesday, June 4, 2014 10:21 AM, Aman Tandon amantandon...@gmail.com
wrote:
Hi,
Hi,
Long time ago I defined a field in schema with space(e.g 'Movie Name').
Things were going very cool till I need to use edismax query parser and want
to give 'Movie Name' in qf. But as we all know qf consider space as field
delimiter. I tried 'Movie\ Name' 'Movie\+Name' nothing is working.
good.
--
View this message in context:
http://lucene.472066.n3.nabble.com/DataImportHandler-while-Replication-tp4138763p4139774.html
Sent from the Solr - User mailing list archive at Nabble.com.
Hi Tommaso,
Yes, you are right. 4.4 version will work.. I'm able to compile now. I'm
trying to apply named recognition(person name) token but im not seeing any
change. my schema.xml looks like this:
field name=text type=text_opennlp_pos_ner indexed=true stored=true
multiValued=true/
fieldType
Thanks Ahmet that worked
Can anybody help me here to how should i start to develop and learn the
solr internals, so that i can make these custom solr developments
efficiently with proper understanding for all these classes.
With Regards
Aman Tandon
On Wed, Jun 4, 2014 at 1:30 PM, Ahmet
Thanks David, yeah i want to contribute can you please suggest me that how
should i start to learn deeply about solr spatial, i am new in solr and i
really want to contribute here :)
Any help will be really appreciated.
@David Sorry for the late reply.
With Regards
Aman Tandon
On Tue, May 27,
Hi,
I am working on Solr using DataImortHander for indexing rich documents
like pdf,word,image etc
I am using TikaEntityProcessor for extracting contents from the files.
I have one small issue regarding setting value to 'url' entry.
My data-config.xml file is like so:
dataConfig
Hi,
I am using Block in my index structure as I have many variations of
documents, which have the same content.
This means my parent document has the content I am searching in and I am
filtering and returning on the child documents:
parent doc
content
child1
filter
/child
mark.
--
View this message in context:
http://lucene.472066.n3.nabble.com/Solr-High-GC-issue-tp4138570p4139785.html
Sent from the Solr - User mailing list archive at Nabble.com.
Unfortunately, field aliasing works above the level of the qf parameter
values.
Maybe the Lucene guys could suggest a way to forcibly rename a field on
disk.
-- Jack Krupansky
-Original Message-
From: devraj.jaiman
Sent: Wednesday, June 4, 2014 6:27 AM
To:
Did my suggestion work out?
RE contributing — most people start out with making improvements needed for
their application. Alternatively you could look at some of the open issues
in JIRA that have the “spatial” or “modules/spatial” component (for Solr or
Lucene, respectively). Most of the real
How many documents was in that 20GB index?
I'm skeptical that a 1 billion document shard won't be a problem. I mean
technically it is possible, but as you are already experiencing, it may take
a long time and a very powerful machine to do so. 100 million (or 250
million max) would be a more
Hi,
We are currently using Solr 4.3 and have highlighting activated on three
different fields using FVH.
Is it possible with Solr to prioritize highlighting for these fields ?
I mean, how to configure Solr, when it's possible highlight the keywords
from the first field, and highlight the
Hi Folks,
I recently started using the spellchecker in my solrconfig.xml. I am able to
build up an index in Solr.
But,if I ever shutdown tomcat I am not able to restart it.The server never
spits out the server startup time in seconds in the logs,nor does it print any
error messages in the
Hi all,
Someone posted this problem over a year ago but I did not see a clear
resolution in the thread.
Intermittently--i.e. for some searches, not others--the
'suggest/spellcheck' component throws a n NullPointerException (NPE) when a
user executes a search. It fails on FSTCompletionLookup
Hi,
We are currently using Solr 4.3 and have highlighting activated on three
different fields using FVH.
Is it possible with Solr to prioritize highlighting for these fields ?
I mean, how to configure Solr, when it’s possible highlight the keywords
from the first field, and highlight the
If you have access to the solr admin screen you have access to how it was
analyzed through the analysis page. You have to hover over the little
abbreviations to see the class in the analysis chain.
Likewise, the admin screen should have access to the raw schema.xml file
which _also_ has the
There's some pinging going on between ZK and registered nodes, and when the
timeout is exceeded there ZK marks the node as down and broadcasts messages
to all the _other_ nodes that the node is down. Then each Solr node knows
not to use the downed node until a message is received indicating it's
You shouldn't have to do anything, assuming that instance3 is a replica of
instance1 or instance2, it should be automatic. You do have to wait for the
synchronization to happen, and you should be seeing messages in the various
Solr logs (particularly instance3 and the leader of the shard). What do
I guess if you try to copy the index and then kill the process of tomcat
then it might help. If still the index need to be delete you would have the
back up. Next time always make back up.
On Jun 4, 2014 7:55 PM, S.L simpleliving...@gmail.com wrote:
Hi Folks,
I recently started using the
right, if that line is uncommented, then _anything_ you throw at Solr will
be processed just fine. You've essentially told Solr there's no input
that's wrong.
Perhaps confusingly, the ignored field type has stored=false and
indexed=false so the effect at indexing time is for the input to be,
The CollapsingQParserPlugin can be resource intensive so you'll want to be
careful about how it's used. Particularly with autowarming in the
queryResultCache. If you autowarm lots of queries while using the
CollapsingQParserPlugin, your be running lots of CPU and memory intensive
queries after
Hi,
This is not a case of accidental deletion , the only way I can restart the
tomcat is by deleting the data directory for the index that was created
earlier, this started happening after I started using spellcheckers in my
solrconfig.xml. As long as the Tomcat is running its fine.
Any help
All, Can anyone help me on what is going wrong in my tomcat. When I restart the
tomcat after schema update, the Cores are removed.
I need to add the cores manually to get back them on work.
Is there anything someone experience..
Thanks
Ravi
Any updates on this? Any help will be greatly appreciated. :)
-
Thanks,
Kashish
--
View this message in context:
http://lucene.472066.n3.nabble.com/Solr-cloud-nodes-falling-tp4139390p4139856.html
Sent from the Solr - User mailing list archive at Nabble.com.
The following two queries are doing the same thing, one using a normal fq
range query and another using a parent query. The cache is warm (these are
both hits) but the normal ones takes ~6 to 7.5sec while the parent query
hack takes ~1.2sec.
Is this expected? Is there anything wrong with my
Any chance you don't have a persistent=true attribute in your solr.xml?
Michael Della Bitta
Applications Developer
o: +1 646 532 3062
appinions inc.
“The Science of Influence Marketing”
18 East 41st Street
New York, NY 10017
t: @appinions https://twitter.com/Appinions | g+:
Is there a JMX metric for measuring the cache request time?
I can see the avg request times, but I'm assuming this includes the cache and
non-cache values.
http://wiki.apache.org/solr/SolrPerformanceFactors
This e-mail may contain Sprint proprietary
http://lucene.472066.n3.nabble.com/file/n4139908/1.png
http://lucene.472066.n3.nabble.com/file/n4139908/2.png
http://lucene.472066.n3.nabble.com/file/n4139908/3.png
http://lucene.472066.n3.nabble.com/file/n4139908/4.png
PS: I am using the same default solrconfig.xml file without any
I have not come across one. Is your question directed to the queryResultCache?
My understanding is that the queryResultCache is the only cache that contains
full query results that could be used to compare against non-cached results
times. I believe the other caches can participate in
I can't seem to find a solution to do wild card search on a multiValued
field.
For Eg consider a multiValued field called Name with 3 values -
Name : [
[[\Ethan\, \G\, \\],[\Steve\, \Wonder\, \\]],
[],
[[\hifte\, \Grop\, \\]]
]
For a multiValued like above, I want search like-
q=***[\Steve\,
Wildcard, fuzzy, and regex query operate on a single term of a single
tokenized field value or a single string field value.
-- Jack Krupansky
-Original Message-
From: Ethan
Sent: Wednesday, June 4, 2014 6:59 PM
To: solr-user
Subject: Multivalue wild card search
I can't seem to find
Thanks Jack -
The following keyword search, based on the previous synonym definition,
actually runs in SOLR and produces a HTTP 500 error (attempted to create too
many clauses error)
asurion device protection has tep, tep plus, erp, esrp programs
HTTP/1.1 500 Internal Server Error
Server:
Hi Jeremy,
Nothing in Solr tracks that time. Caches are pluggable. If you really
want this info you could write your own cache that is just a proxy for the
real cache and then you can time it.
But why do you need this info? Do you suspect that is slow?
Otis
--
Performance Monitoring * Log
Hi all,
I know this probably seems like an uninteresting problem and smells, even
to me, like a stupid/newbie mis-configuration [Yes. I am reading the
excellent solr in action and trying my hand at applying the suggestion
examples], but I looked a bit into this tonight, fired up the debugger,
39 matches
Mail list logo