called. Is this
expected behavior? The events I observed are:
1. Cache init() method called
2. Searcher created and registered
3. Replication recovery
Thanks!
Li
Thank you for creating the JIRA! Will follow
On 12/19/19, 11:09 AM, "Andrzej Białecki" wrote:
Hi,
Thanks for the data. I see the problem now - it’s a bug in the simulator. I
filed a Jira issue to track and fix it: SOLR-14122.
> On 16 Dec 2019, at 19:13, C
ou’re only allowed to use #ANY or a single specific node name (with optional
“!" NOT operand), or a JSON array containing node names to indicate the IN
operand.
The Ref Guide indeed is not very clear on that…
> On 17 Dec 2019, at 21:20, Cao, Li wrote:
>
> Hi!
$ExceptionWithErrObject"],
"details":[{
"set-cluster-policy":[{
"cores":"<3",
"nodeset":{"sysprop.rex.node.type":"tlog"}}],
"errorMessages":["No value present"]}],
"msg":"Error in command payload",
"code":400}}
However, this works:
{ "set-cluster-policy":[{"cores":"<3","node":"#ANY"}]}
I read the autoscaling policy documentations and cannot figure out why. Could
someone help me on this?
Thanks!
Li
e it needs the zookeeper inside the Kubernetes.
Thanks!
Li
On 12/15/19, 5:13 PM, "Andrzej Białecki" wrote:
Could you please provide the exact command-line? It would also help if you
could provide an autoscaling snapshot of the cluster (bin/solr autoscaling
-save ) or at l
"method":"POST",
"path":"/c/entity/shards",
"command":{"add-replica":{
"shard":"shard2",
"node":"my_node:8983_solr",
"type":"TLOG",
"replicaInfo":null}}},
Does anyone know how to fix this? Is this a bug?
Thanks!
Li
? favoritePlace/f85333c1-c444-4cfb-afd7-37281a07b0f7
On Mon, Oct 1, 2018 at 3:37 PM Li, Yi wrote:
> Hi,
>
> cursorMark appears as something like
> AoE/E2Zhdm9yaXRlUGxhY2UvZjg1MzMzYzEtYzQ0NC00Y2ZiLWFmZDctMzcyODFhMDdiMGY3
>
> and the do
Hi,
cursorMark appears as something like
AoE/E2Zhdm9yaXRlUGxhY2UvZjg1MzMzYzEtYzQ0NC00Y2ZiLWFmZDctMzcyODFhMDdiMGY3
and the document says it is “Base64 encoded serialized representation of the
sort values encapsulated by this object”
I like to know if I can decode and what content I will see in
Hi,
Currently we are running Solr 5.3.1 with JDK8 and we are trying to run Solr
5.3.1 with JDK10. Initially we got a few errors complaining some JVM options
are removed since JDK9. We removed those options in solr.in.sh:
UseConcMarkSweepGC
UseParNewGC
PrintHeapAtGC
PrintGCDateStamps
PrintGCTimeS
To whom it may concern,
I am running Solr 7.1.0 and encountered a problem starting Solr after I
killed the Java process running Solr without proper cleanup. The error
message that I received is as following:
solr-7.1.0 liyifan$ bin/solr run
dyld: Library not loaded: /usr/local/opt/mpfr/lib/libm
This happened when the second time I'm performing restart. But after that,
every time this collection is stuck at here. If I restart the leader node
as well, the core can get out of the recovering state
On Mon, May 16, 2016 at 5:00 PM, Li Ding wrote:
> Hi Anshum,
>
> This is fo
heckLive=true&core=test_collection_112_shard1_replica1
&wt=javabin&onlyIfLeader=true&version=2} status=0 QTime=4001
Is there any known bug? all collections are empty.
Thanks,
Li
On Mon, May 16, 2016 at 12:50 PM, Anshum Gupta
wrote:
> I think you are approaching the problem al
, our
solr restart will be more robust.
Any suggestions will be appreciated.
Thanks,
Li
r calls succeeds and the next zk ping should
bring the core back to normal? right? We have an active monitor running at
the same time querying every core in distrib=false mode and every query
succeeds.
Thanks,
Li
On Tue, Apr 26, 2016 at 6:20 PM, Erick Erickson
wrote:
> One of the reasons this
But I don't think it will affect the core status.
Do you guys have any idea about why this particular core is not published
as active since from the log, most steps are done except the very last one
to publish info to ZK.
Thanks,
Li
On Thu, Apr 21, 2016 at 7:08 AM, Rajesh Hazari
wrote
on its own?
Thanks,
Li
Thanks for your help. I figured it out. Just as you said. Appreciate your
help. Somehow forgot to reply your post.
On Wed, Apr 29, 2015 at 9:24 AM, Chris Hostetter
wrote:
>
> : We did two SOLR qeries and they supposed to return the same results but
> : did not:
>
> the short answer is: if you wa
I figured it out now. It works. "cats" just a name, right? It does not
matter what is used.
Really appreciate your help. This is going to be really useful. I meant
"json.facet".
On Sun, May 10, 2015 at 12:13 AM, Frank li wrote:
> Here is our SOLR query:
>
>
> h
Here is our SOLR query:
http://qa-solr:8080/solr/select?q=type:PortalCase&json.facet={categories:{terms:{field:campaign_id_ls,sort:%27count+asc%27}}}&rows=0
I replaced "cats" with "categories". It is still not working.
On Sun, May 10, 2015 at 12:10 AM, Frank li
ely easier to use "-d" with curl...
>
> curl "http://localhost:8983/solr/techproducts/query"; -d
> 'q=*:*&json.facet={cats:{terms:{field:cat,sort:"count asc"}}}'
>
> That also allows you to format it nicer for reading as well:
>
> c
Hi Yonik,
Any update for the question?
Thanks in advance,
Frank
On Thu, May 7, 2015 at 2:49 PM, Frank li wrote:
> Is there any book to read so I won't ask such dummy questions? Thanks.
>
> On Thu, May 7, 2015 at 2:32 PM, Frank li wrote:
>
>> This one does not hav
Is there any book to read so I won't ask such dummy questions? Thanks.
On Thu, May 7, 2015 at 2:32 PM, Frank li wrote:
> This one does not have problem, but how do I include "sort" in this facet
> query. Basically, I want to write a solr query which can sort the fa
}
<http://localhost:8983/solr/demo/query?q=apple&json.facet=%7Bx:%27avg%28price%29%27%7D>
I really appreciate your help.
Frank
<http://localhost:8983/solr/demo/query?q=apple&json.facet=%7Bx:%27avg%28price%29%27%7D>
On Thu, May 7, 2015 at 2:24 PM, Yonik Seeley wrote:
> On
Hi Yonik,
I am reading your blog. It is helpful. One question for you, for following
example,
curl http://localhost:8983/solr/query -d 'q=*:*&rows=0&
json.facet={
categories:{
type : terms,
field : cat,
sort : { x : desc},
facet:{
x : "avg(price)",
y : "sum(p
We did two SOLR qeries and they supposed to return the same results but
didnot:
Query 1: all_text:(US 4,568,649 A)
"parsedquery": "(+((all_text:us ((all_text:4 all_text:568 all_text:649
all_text:4568649)~4))~2))/no_coord",
Result: "numFound": 0,
Query 2: all_text:(US 4568649)
"parsedquery": "(
r, you need to use
> edismax or explicitly create the multiple clauses.
>
> I'm not quite sure what the join parser is doing with the df
> parameter. So my first question is "what happens if you just use a
> single field for df?".
>
> Best,
> Erick
>
> On Mon,
The error message was from the query with "debug=query".
On Mon, Apr 6, 2015 at 11:49 AM, Frank li wrote:
> Hi Erick,
>
>
> Thanks for your response.
>
> Here is the query I am sending:
>
> http://dev-solr:8080/solr/collection1/select?q={!join+from=litigation_
ng else that would help. You might review:
>
> http://wiki.apache.org/solr/UsingMailingLists
>
> Best,
> Erick
>
> On Fri, Apr 3, 2015 at 10:58 AM, Frank li wrote:
> > Hi,
> >
> > I am starting using join parser with our solr. We have some default
>
Hi,
I am starting using join parser with our solr. We have some default fields.
They are defined in solrconfig.xml:
edismax
explicit
10
all_text number party name all_code ent_name
all_text number^3 name^5 party^3 all_code^2
ent_name^7
id descripti
We have a query which has both sort and group.sort. What we are expecting
is that we can use sort to sort groups but inside the group we have a
different sort.
However, looks like sort is over-writting the sorting order inside groups.
Can any one of you help us on this?
Basically we want to sort
Hi Guys,
Just some update.
I've tried with Solr 4.10 (same code for Solr 4.9). And that has the same index
speed as 4.0. The only problem left now is that Solr 4.10 takes more memory
than 4.0 so I'm trying to figure out what is the best number for Java heap size.
I think that proves there is s
n at https://issues.apache.org/jira/browse/LUCENE-5914.
Best,
Erick
____
From: Li, Ryan
Sent: Friday, September 05, 2014 3:28 PM
To: solr-user@lucene.apache.org
Subject: Re: Solr add document over 20 times slower after upgrade from 4.0 to
4.9
HI Shawn,
Thanks for y
HI Shawn,
Thanks for your reply.
The memory setting of my Solr box is
12G physically memory.
4G for java (-Xmx4096m)
The index size is around 4G in Solr 4.9, I think it was over 6G in Solr 4.0.
I do think the RAM size of java is one of the reasons for this slowness. I'm
doing one big commit an
I have a Solr server indexes 2500 documents (up to 50MB each, ave 3MB) to Solr
server. When running on Solr 4.0 I managed to finish index in 3 hours.
However after we upgrade to Solr 4.9, the index need 3 days to finish.
I've done some profiling, numbers I get are:
size figure of document,t
We got different results for these two queries. The first one returned 115
records and the second returns 179 records.
Thanks,
Fudong
Hi Jack,
Do you have a date for the new version of your book:
solr_4x_deep_dive_early_access?
Thanks,
Fudong
On Mon, Oct 21, 2013 at 10:39 AM, Jack Krupansky wrote:
> Take a look at the unit tests for various "value sources", and find a Jira
> that added some value source and look at the patc
Hi,
I don't seem to be able to find any info on the possibility to get stats on
dynamic fields. stats=true&states.field=xyz_* appears to literally treat
"xyz_*" as the field name with a star. Is there a way to get stats on
dynamic fields without explicitly listing them in the query?
Thanks!
Li
How to shut down the SolrCloud? Just kill all nodes?
Regards,
Ivan
This email message and any attachments are for the sole use of the intended
recipients and may contain proprietary and/or confidential information which
may be privileged or otherwise protected from disclosure. Any unauthorized
-user@lucene.apache.org
Subject: Re: How to share config files in SolrCloud between multiple
cores(collections)
To share configs in SolrCloud you just upload a single config set and then link
it to multiple collections. You don't actually use solr.xml to do it.
- Mark
On Mar 19, 2013, at 10:43 AM,
We have multiple cores with the same configurations, before using SolrCloud, we
can use relative path in solr.xml. But with Solr4, is seems denied for using
relative path for the schema and config in solr.xml.
Regards,
Ivan
This email message and any attachments are for the sole use of the inte
; A colleague of mine when I was working for Sourcesense made a CMIS
> plugin for Solr. It was one way, and we used it to index stuff out of
> Alfresco into Solr. I can't search for it now, let me know if you can't
> find it.
>
> Upayavira
>
> On Fri, Jan 18, 2013, at 05:35 AM
I want to make something like Alfresco, but not having that many features.
And I'd like to utilise the searching ability of Solr.
On Fri, Jan 18, 2013 at 4:11 PM, Gora Mohanty wrote:
> On 18 January 2013 10:36, Nicholas Li wrote:
> > hi
> >
> > I am new to solr and I
hi
I am new to solr and I would like to use Solr as my document server, plus
search engine. But solr is not CMIS compatible( While it shoud not be, as
it is not build as a pure document management server). In that sense, I
would build another layer beyond Solr so that the exposed interface would
hi there,
I am quite new to Solr and have a very basic question about storing and
indexing the document.
I am trying with the Solr example, and when I run command like 'java -jar
post.jar foo/test.xml', it gives me the feeling that solr will index the
given file, no matter where it is store, and
Hi,
I ran into an issue lately with Index version & generation for Solr 3.5.
In Solr 1.4., the index version of slave service increments upon each
replication. However, I noticed it's not the case for Solr 3.5; the
index version would increase 20, or 30 after replication. Does anyone
know why and
http://zookeeper.apache.org/doc/r3.3.6/recipes.html#sc_recipes_twoPhasedCommit
On Thu, Aug 16, 2012 at 7:41 AM, Nicholas Ball
wrote:
>
> Haven't managed to find a good way to do this yet. Does anyone have any
> ideas on how I could implement this feature?
> Really need to move docs across from on
do you really need this?
distributed transaction is a difficult problem. in 2pc, every node could
fail, including coordinator. something like leader election needed to make
sure it works. you maybe try zookeeper.
but if the transaction is not very very important like transfer money in
bank, you can
在 2012-7-2 傍晚6:37,"Nicholas Ball" 写道:
>
>
> That could work, but then how do you ensure commit is called on the two
> cores at the exact same time?
that may needs something like two phrase commit in relational dB. lucene
has prepareCommit, but to implement 2pc, many things need to do.
> Also, any w
create an field for exact match. it is a optional boolean clause
在 2012-8-11 下午1:42,"abhayd" 写道:
> hi
>
> I have documents like
> iphone 4 - white
> iphone 4s - black
> ipone4 - black
>
> when user searches for iphone 4 i would like to show iphone 4 docs first
> and
> iphone 4s after that.
> Simil
I have used Solr 3.4 for a long time. Recently, when I upgrade to Solr 4.0
and reindex the whole data, I find that the fields which are specified as
string type can not be searched by q parameter. If I just change the type
to text_general, it works. So my question is for Solr 4.0, must I set the
Hi All,
I want to search some keywords like "Non-taxable", which has a "-" in the word.
Can I make it working in Solr by some configuration? Or any other ways?
Thanks & Regards,
Ivan
This email message and any attachments are for the sole use of the intended
recipients and may contain propriet
hu, Jun 28, 2012 at 3:51 PM, ZHANG Liang F
wrote:
> Thanks a lot, but the precisionStep is still very vague to me! Could you give
> me a example?
>
> -Original Message-
> From: Li Li [mailto:fancye...@gmail.com]
> Sent: 2012年6月28日 11:25
> To: solr-user@lucene.ap
could you please use jstack to dump the call stacks?
On Thu, Jun 28, 2012 at 2:53 PM, Arkadi Colson wrote:
> It now hanging for 15 hour and nothing changes in the index directory.
>
> Tips for further debugging?
>
>
> On 06/27/2012 03:50 PM, Arkadi Colson wrote:
>>
>> I'm sending files to solr wi
I think they are logically the same. but 1 may be a little bit faster than 2
On Thu, Jun 28, 2012 at 5:59 AM, Rublex wrote:
> Hi,
>
> Can someone explain to me please why these two queries return different
> results:
>
> 1. -PaymentType:Finance AND -PaymentType:Lease AND -PaymentType:Cash *(700
>
1. precisionStep is used for ranging query of Numeric Fields. see
http://lucene.apache.org/core/old_versioned_docs/versions/3_5_0/api/all/org/apache/lucene/search/NumericRangeQuery.html
2. positionIncrementGap is used for phrase query of multi-value fields
e.g. doc1 has two titles.
title1: ab cd
seems that the indexwriter wants to flush but need to wait others become
idle. but i see you the n gram filter is working. is your field's value too
long? you sould also tell us average load the system. the free memory and
memory used by jvm
在 2012-6-27 晚上7:51,"Arkadi Colson" 写道:
> Anybody an idea
ul approach
> http://lucene.472066.n3.nabble.com/High-response-time-after-being-idle-tp3616599p3617604.html.
>
> On Mon, Jun 11, 2012 at 3:02 PM, Toke Eskildsen
> wrote:
>
>> On Mon, 2012-06-11 at 11:38 +0200, Li Li wrote:
>> > yes, I need average query time less than
t;
> http://en.wikipedia.org/wiki/Swappiness
>
> -Kuli
>
> Am 11.06.2012 10:38, schrieb Li Li:
>
>> I have roughly read the codes of RAMDirectory. it use a list of 1024
>> byte arrays and many overheads.
>> But as far as I know, using MMapDirectory, I can't prev
persist your index,
> you'll need to live with disk IO anyway.
>
> Greetings,
> Kuli
>
> Am 11.06.2012 11:20, schrieb Li Li:
>
>> I am sorry. I make a mistake. even use RAMDirectory, I can not
>> guarantee they are not swapped out.
>>
>> On Mon,
ss
>
> -Kuli
>
> Am 11.06.2012 10:38, schrieb Li Li:
>
>> I have roughly read the codes of RAMDirectory. it use a list of 1024
>> byte arrays and many overheads.
>> But as far as I know, using MMapDirectory, I can't prevent the page
>> faults. OS will swap less
d a "small" segment. Every night I will
merge them. new added documents will flush into a new segment and I
will merge the new generated segment and the small one.
Our update operations are not very frequent.
On Mon, Jun 11, 2012 at 4:59 PM, Paul Libbrecht wrote:
> Li Li,
>
> have yo
at 4:45 PM, Michael Kuhlmann wrote:
> Set the swapiness to 0 to avoid memory pages being swapped to disk too
> early.
>
> http://en.wikipedia.org/wiki/Swappiness
>
> -Kuli
>
> Am 11.06.2012 10:38, schrieb Li Li:
>
>> I have roughly read the codes of RAMDirectory. it
This sounds wrong, but it is true. With
> RAMDirectory, Java has to work harder doing garbage collection.
>
> On Fri, Jun 8, 2012 at 1:30 AM, Li Li wrote:
>> hi all
>> I want to use lucene 3.6 providing searching service. my data is
>> not very large, raw data is le
yes, I am also interested in good performance with 2 billion docs. how
many search nodes do you use? what's the average response time and qps
?
another question: where can I find related paper or resources of your
algorithm which explains the algorithm in detail? why it's better than
google site(b
you should find some clues from tomcat log
在 2012-5-22 晚上7:49,"Spadez" 写道:
> Hi,
>
> This is the install process I used in my shell script to try and get Tomcat
> running with Solr (debian server):
>
>
>
> I swear this used to work, but currently only Tomcat works. The Solr page
> just comes up wi
you should define your search first.
if the site is www.google.com. how do you match it. full string
matching or partial matching. e.g. is "google" should match? if it
does, you should write your own analyzer for this field.
On Tue, May 22, 2012 at 2:03 PM, Shameema Umer wrote:
> Sorry,
> Please
query=parser.parse(q);
System.out.println(query);
On Thu, May 10, 2012 at 8:20 AM, Li Li wrote:
> + before term is correct. in lucene term includes field and value.
>
> Query ::= ( Clause )*
>
> Clause ::= ["+", "-"] [ ":"] ( | "
+ before term is correct. in lucene term includes field and value.
Query ::= ( Clause )*
Clause ::= ["+", "-"] [ ":"] ( | "(" Query ")" )
<#_TERM_CHAR: ( <_TERM_START_CHAR> | <_ESCAPED_CHAR> | "-" | "+" ) >
<#_ESCAPED_CHAR: "\\" ~[] >
in lucene query syntax, you can't express a term value i
don't score by relevance and score by document id may speed it up a little?
I haven't done any test of this. may be u can give it a try. because
scoring will consume
some cpu time. you just want to match and get total count
On Wed, May 2, 2012 at 11:58 PM, vybe3142 wrote:
> I can achieve this by
for this version, you may consider using payload for position boost.
you can save boost values in payload.
I have used it in lucene api where anchor text should weigh more than
normal text. but I haven't used it in solr.
some searched urls:
http://wiki.apache.org/solr/Payloads
http://digitalpebble.
as for version below 4.0, it's not possible because lucene's score
model. position information is stored, but only used to support phrase
query. it just tell us whether a document is matched, but we can boost
a document. The similar problem is : how to implement proximity boost.
for 2 search terms,
you should reverse your sort algorithm. maybe you can override the tf
method of Similarity and return -1.0f * tf(). (I don't know whether
default collector allow score smaller than zero)
Or you can hack this by add a large number or write your own
collector, in its collect(int doc) method, you can
hi
I checked out the trunk and played with its new soft commit
feature. it's cool. But I've got a few questions about it.
By reading some introductory articles and wiki, and hasted code
reading, my understand of it's implementation is:
For normal commit(hard commit), we should flush all in
http://wiki.apache.org/solr/SolrCaching
On Fri, Apr 13, 2012 at 2:30 PM, Kashif Khan wrote:
> Does anyone explain what does the following parameters mean in SOLR cache
> statistics?
>
> *name*: queryResultCache
> *class*: org.apache.solr.search.LRUCache
> *version*: 1.0
> *description*: LRU
another way is to use payload http://wiki.apache.org/solr/Payloads
the advantage of payload is that you only need one field and can make frq
file smaller than use two fields. but the disadvantage is payload is stored
in prx file, so I am not sure which one is fast. maybe you can try them
both.
On
houldMatch parameter'. Also
> norms can be used as a source for dynamics mm values.
>
> Wdyt?
>
> On Wed, Apr 11, 2012 at 10:08 AM, Li Li wrote:
>
> > it's not possible now because lucene don't support this.
> > when doing disjunction query, it onl
it's not possible now because lucene don't support this.
when doing disjunction query, it only record how many terms match this
document.
I think this is a common requirement for many users.
I suggest lucene should divide scorer to a matcher and a scorer.
the matcher just return which doc is matche
According to my knowledge, Solr cannot support this.
In my case, I get data by keyword-matching from Solr and then rank the data
by PageRank after that.
Thanks,
Bing
On Wed, Apr 4, 2012 at 6:37 AM, Manuel Antonio Novoa Proenza <
mano...@estudiantes.uci.cu> wrote:
> Hello,
>
> I have in my Solr
gt;> Classpath entry /solr3_5/ssrc/solr/lib/easymock-2.2.jar will not be
>> exported or published. Runtime ClassNotFoundExceptions may result.
>> solr3_5P/solr3_5Classpath Dependency Validator Message
>> Classpath entry
>> /solr3_5/ssrc/solr/lib/geronimo-stax
here is my method.
1. check out latest source codes from trunk or download tar ball
svn checkout http://svn.apache.org/repos/asf/lucene/dev/trunklucene_trunk
2. create a dynamic web project in eclipse and close it.
for example, I create a project name lucene-solr-trunk in my
workspace.
it's not the right place.
when you use java -Durl=http://... -jar post.jar data.xml
the data.xml file must be a valid xml file. you shoud escape special chars
in this file.
I don't know how you generate this file.
if you use java program(or other scripts) to generate this file, you should
use xml t
ag solved a real problem we were
having. Whoever wrote the JRocket book you refer to no doubt had other
scenarios in mind...
On Thu, Mar 15, 2012 at 3:02 PM, C.Yunqin <345804...@qq.com> wrote:
> why should enable pointer compression?
>
>
>
>
> -- Original -
ver with exactly same system and solr configuration &
> memory it is working fine?
>
>
> -Original Message-
> From: Li Li [mailto:fancye...@gmail.com]
> Sent: Thursday, March 15, 2012 11:11 AM
> To: solr-user@lucene.apache.org
> Subject: Re: Solr out of memory excep
how many memory are allocated to JVM?
On Thu, Mar 15, 2012 at 1:27 PM, Husain, Yavar wrote:
> Solr is giving out of memory exception. Full Indexing was completed fine.
> Later while searching maybe when it tries to load the results in memory it
> starts giving this exception. Though with the sam
no, it's nothing to do with schema.xml
post.jar just post a file, it don't parse this file.
solr will use xml parser to parse this file. if you don't escape special
characters, it's not a valid xml file and solr will throw exceptions.
On Thu, Mar 15, 2012 at 12:33 AM, neosky wrote:
> Thanks!
> D
There is a class org.apache.solr.common.util.XML in solr
you can use this wrapper:
public static String escapeXml(String s) throws IOException{
StringWriter sw=new StringWriter();
XML.escapeCharData(s, sw);
return sw.getBuffer().toString();
}
On Wed, Mar 14, 2012 at
it should be indexed by not analyzed. it don't need stored.
reading field values from stored fields is extremely slow.
So lucene will use StringIndex to read fields for sort. so if you want to
sort by some field, you should index this field and don't analyze it.
On Wed, Mar 14, 2012 at 6:43 PM, Fi
optimize will generate new segments and delete old ones. if your master
also provides searching service during indexing, the old files may be
opened by old SolrIndexSearcher. they will be deleted later. So when
indexing, the index size may double. But a moment later, old indexes will
be deleted.
what do u mean "programmatically"? modify codes of solr? becuase solr is
not like lucene, it only provide http interfaces for its users other than
java api
if you want to modify solr, you can find codes in SolrCore
private final LinkedList> _searchers = new
LinkedList>();
and _searcher is current
lucene will never modify old segment files, it just flushes into a new
segment or merges old segments into new one. after merging, old segments
will be deleted.
once a file(such as fdt and fdx) is generated. it will never be
re-generated. the only possible is that in the generating stage, there is
nt schema and index the
> rank too for range queries and such. is my understanding of your scenario
> wrong?
>
> thanks
>
>
> On Wed, Feb 22, 2012 at 9:51 AM, Bing Li wrote:
>
>> Mr Gupta,
>>
>> Thanks so much for your reply!
>>
>> In my use
>> It's on our road map.
>>
>> FYI
>>
>> On Wed, Feb 22, 2012 at 9:28 AM, Bing Li wrote:
>>
>> > Jacques,
>> >
>> > Yes. But I still have questions about that.
>> >
>> > In my system, when users search with a keywor
Dear all,
I wonder how data in HBase is indexed? Now Solr is used in my system
because data is managed in inverted index. Such an index is suitable to
retrieve unstructured and huge amount of data. How does HBase deal with the
issue? May I replaced Solr with HBase?
Thanks so much!
Best regards,
you can fool the lucene scoring fuction. override each function such as idf
queryNorm lengthNorm and let them simply return 1.0f.
I don't lucene 4 will expose more details. but for 2.x/3.x, lucene can only
score by vector space model and the formula can't be replaced by users.
On Fri, Feb 17, 2012
w have a shrunk index with specified terms removed.
>
> Implementation uses separate thread for each segment, so it re-writes
> them in parallel. Took about 15 minutes to do 770,000 doc index on my
> macbook.
>
>
> On Tue, Feb 14, 2012 at 10:12 PM, Li Li wrote:
> > I have rough
nd Terms(...) it might work.
>
> Something like:
>
> HashSet ignoredTerms=...;
>
> FilteringIndexReader wrapper=new FilterIndexReader(reader);
>
> SegmentMerger merger=new SegmentMerger(writer);
>
> merger.add(wrapper);
>
> merger.Merge();
>
>
>
>
>
for method 2, delete is wrong. we can't delete terms.
you also should hack with the tii and tis file.
On Tue, Feb 14, 2012 at 2:46 PM, Li Li wrote:
> method1, dumping data
> for stored fields, you can traverse the whole index and save it to
> somewhere else.
> for index
method1, dumping data
for stored fields, you can traverse the whole index and save it to
somewhere else.
for indexed but not stored fields, it may be more difficult.
if the indexed and not stored field is not analyzed(fields such as id),
it's easy to get from FieldCache.StringIndex.
But for
available after adding to the index.
>
> What I don't understand is why new segment files are created so often.
> Are the commit calls triggering new segment files being created? I don't
> see this behavior in another environment of the same version of solr.
>
>
ts be available after adding to the index.
>
> What I don't understand is why new segment files are created so often.
> Are the commit calls triggering new segment files being created? I don't
> see this behavior in another environment of the same version of solr.
>
>
1 - 100 of 284 matches
Mail list logo