Have you recently change its type from unstored to stored?
if so you have to reindex .
-
Thanx:
Grijesh
www.gettinhahead.co.in
--
View this message in context:
http://lucene.472066.n3.nabble.com/Solr-1-3-highlighting-problem-tp2918089p2922117.html
Sent from the Solr - User mailing list
See this
http://www.gettingahead.co.in/blogwp/2011/03/technical/solr/shalini/some-info-about-extended-dismax-query-parser/
-
Thanx:
Grijesh
www.gettinhahead.co.in
--
View this message in context:
http://lucene.472066.n3.nabble.com/edismax-available-in-solr-3-1-tp2910613p2922128.html
Sent
Hallo..
I search a Way to indexing e-Mails fast and easy to safe in Solr
A lot of users do this too, but i cant find a howTo or something like that.
when you have or know where i can find a HowTo..
it will nice if you can help me..
King
Hi,
On 09.05.11 11:04, Martin Jansen wrote:
I just attempted to set up an instance of Solr 3.1 in Tomcat 5.5
running in Java 1.5. It fails with the following exception on start-up:
java.lang.AssertionError: Analyzer implementation classes or at least
their tokenStream() and
Ravi,
as far as I remember, this is how the replication logic works (see
SnapPuller class, fetchLatestIndex method):
1. Does the Slave get the whole index every time during replication or
just the delta since the last replication happened ?
It look at the index version AND the index
On Tue, May 10, 2011 at 1:36 PM, Jörg Agatz joerg.ag...@googlemail.com wrote:
Hallo..
I search a Way to indexing e-Mails fast and easy to safe in Solr
A lot of users do this too, but i cant find a howTo or something like that.
http://wiki.apache.org/solr/DataImportHandler#Indexing_Emails
will the E-Mail ID, and the recent E-Mail Ids, indext too?
and witch fiels i have to create in schema.xml?
I have a schema which has field Polarity which is of type text and it
can have three values 0,1 or -1 and CreatedAt which is of type date.
*How can I get count of polarity based on dates. For example, it gives
the output that on 5/1/2011 there were 10 counts of 0, 10 counts of 1
and 10 counts
Yes you can but one doc will contain one type of data other wise it will be
overwritten by latest one
-
Thanx:
Grijesh
www.gettinhahead.co.in
--
View this message in context:
http://lucene.472066.n3.nabble.com/How-to-set-a-common-field-to-several-values-types-tp2922192p2922206.html
Sent
I indeed changed the type from unstored to stored. But I did do a reindex
(several times already) and a commit, restarted Tomcat, restarted PC,
nothing works. When I do a search the text field shows up in the results.
But that is the only thing that has changed. lst name=highlighting tag
only
You're using the string fieldType to get highlighting to work but it must be
tokenized. Use a fieldType with a tokenizer.
See also: http://wiki.apache.org/solr/FieldOptionsByUseCase
Hi,
I'm using the old 1.3 Solr version on one of my sites and I decided to
add a highlighting feature.
Oh!
I did't see the string
-
Thanx:
Grijesh
www.gettinhahead.co.in
--
View this message in context:
http://lucene.472066.n3.nabble.com/Solr-1-3-highlighting-problem-tp2918089p2922532.html
Sent from the Solr - User mailing list archive at Nabble.com.
Have you looked at Pivot Faceting
http://wiki.apache.org/solr/HierarchicalFaceting
http://wiki.apache.org/solr/SimpleFacetParameters#Pivot_.28ie_Decision_Tree.29_Faceting-1
-
Thanx:
Grijesh
www.gettinhahead.co.in
--
View this message in context:
Look at http://wiki.apache.org/solr/MailEntityProcessor
http://wiki.apache.org/solr/MailEntityProcessor it will give you the all
informations
You may have to create the fields as
single valued fields :
messageId
subject
from
sentDate
xMailer
multi valued fields
allTo
flags : possible flags
I think OS will manage it when required OS will use other core of your CPU
also. Also you can use some memory management(GC Options) settings with
java.
You can see about memory management a blog posted by Mark Miller at
Lucidimagination.com
Thanks for your answer!
It is pretty weird because all my tries are wrong.
If I try to index data with this field in the schema :
field name=content type=text indexed=true required=true
stored=true multiValued=true /
It works perfectly (in the case of indexing extract from file (via solr
I want to add score for documents dynamically. For
this, the method that I know right now is to use add. Using this
method, everytime I need to get the contents of that document, and then
add another record with the boost field. Is there a way in which I can
just specify the boost attribute
Which way you are using to index the Plain text data?
As I see your both the fields are same then why not use same field for both
the data.
What is the need of extracting request handler if you are not going to
extract any document?
-
Thanx:
Grijesh
www.gettinhahead.co.in
--
View this
NO, You have to update the whole document .There is no way to update only
some of fields or any field attribute in solr/Lucene currently.
-
Thanx:
Grijesh
www.gettinhahead.co.in
--
View this message in context:
Any other workaround? Maybe by fiddling around with the source?
From: kvs_cheezy_...@hotmail.com
To: solr-user@lucene.apache.org
Subject: Boosting score of a document without deleting and adding another
document
Date: Tue, 10 May 2011 16:17:25 +0530
I want to add score for documents
1. Do we have to reindex all content again to use Solr 3.1 ?
2. If we don't reindex all content are there any potential issues ? (I
read somewhere that first commit would change the 1.4.1 format to 3.1.
have the analyzer's behavior changed which warrants reindexing ?)
3. Apart from deploying
You can add the boost at the time of indexing the doc at first time.
You can see the function queries at query time to boost the documents
dynamically.
You can see around DefaultSimilarityClass to tweak around scores
-
Thanx:
Grijesh
www.gettinhahead.co.in
--
View this message in context:
What I want to do is : everytime an item is searched, I get the ID of that
document, and now I want to increase this document's score.Function queries at
query time won't really help with this. Any way to do this other than adding
the entire document again?
From: kvs_cheezy_...@hotmail.com
Perhaps the query elevation component is what you are looking for :
http://wiki.apache.org/solr/QueryElevationComponent
Ludovic.
-
Jouve
France.
--
View this message in context:
Nope can't do,
you can keep the counter at application end and update the time-stamp of
documents which have been searched then update/index the document with
counter or assign the boost to document being updated.
if keeping counter then by using function query you can influence the score
of
Hi Markus,
thanks for the tip. I replaced the string field with a tokenizer field and
now it works. Thank you!
Kind regards,
Nick
--
View this message in context:
http://lucene.472066.n3.nabble.com/Solr-1-3-highlighting-problem-tp2918089p2922785.html
Sent from the Solr - User mailing list
Grijesh wrote:
Which way you are using to index the Plain text data?
I'm using Solr in a Php application, and have to index documents (with tika)
and plain text data.
The php interface uses curl to send xml data to the solr server.
Grijesh wrote:
As I see your both the fields are same
Is it possible to create hierarchies of query operator objects and pass that
to solr instead of constructing a long string query?
This would enable client applications to construct complex queries and have
them verified by the compiler, example (psuedo code):
q = BooleanQuery(
Hallo,
I have a problem using threads option in entity in DIH it just does not work,
it either hangs it self or fails to import anything.
Does this feature even work ?
Without threads the import works fine, just too slow
**
sorry, this was not the target I used (this one should work too, but...),
Can we expand on the but...?
$ wget http://apache.panu.it//lucene/solr/3.1.0/apache-solr-3.1.0-src.tgz
$ tar xf apache-solr-3.1.0-src.tgz
$ cd apache-solr-3.1.0
$ ant generate-maven-artifacts
*generate-maven-artifacts:
The WIKI has a loose interpretation of how to set-up Jetty securely.
Please take a look at the article I wrote here:
http://anthonyw.net/2011/04/securing-jetty-and-solr-with-php-authentication/.
Even if PHP is not your language that sits on top of Solr you can still
use the first part of
On Tue, May 10, 2011 at 5:58 PM, Jamroz Marcin
marcin.jam...@it2media.de wrote:
Hallo,
I have a problem using threads option in entity in DIH it just does not work,
it either hangs it self or fails to import anything.
Does this feature even work ?
Without threads the import works fine,
now its working gr8,
thanks stefan,scott
--
View this message in context:
http://lucene.472066.n3.nabble.com/org-apache-solr-common-SolrException-Error-loading-class-org-apache-solr-handler-dataimport-DataImpo-tp2865625p2922985.html
Sent from the Solr - User mailing list archive at Nabble.com.
Hello all,
indexing with dataimporthandler runs every hour (new records will be added,
some records will be updated) note :large data
requirement is when indexing is in progress, searching (on already indexed
data) should not affect
so should i use multicore-with merge and swap or delta query
What I want to do is : everytime an item is searched, I get
the ID of that document, and now I want to increase this
document's score.Function queries at query time won't really
help with this. Any way to do this other than adding the
entire document again?
Looks like you need
One approach is to use Solr's replication features. Index to a 'master',
periodically replicate to 'slave' on which all the searching is done.
That's what I do; my master and slave are in fact on the same server (one with
a bunch of CPUs and RAM however), although not alternate cores in a
Hi Gabriele,
There are some Maven instructions here (not in Lucene/Solr 3.1 because I just
wrote the file a couple of days ago):
http://svn.apache.org/repos/asf/lucene/dev/branches/lucene_solr_3_1/dev-tools/maven/README.maven
My recommendation, since the Solr 3.1 source tarball does not include
Under the hood, Lucene can support this by keeping multiple commit
points in the index.
So you'd make a new commit whenever you finish indexing the updates
from each hour, and record that this is the last searchable commit.
Then you are free to commit while indexing the next hour's worth of
On Tue, May 10, 2011 at 3:50 PM, Steven A Rowe sar...@syr.edu wrote:
Hi Gabriele,
There are some Maven instructions here (not in Lucene/Solr 3.1 because I
just wrote the file a couple of days ago):
Very nice Steve ! Thanks again. (I'm building from svn so that's perfect for
me)
Is this file referenced somewhere in the wiki ?
Ludovic.
-
Jouve
France.
--
View this message in context:
Hi Ludovic,
On 5/10/2011 at 10:02 AM, lboutros wrote:
Very nice Steve ! Thanks again. (I'm building from svn so that's perfect
for me)
Is this file referenced somewhere in the wiki ?
Not yet, no. Probably should be linked from the HowToContribute pages for
Lucene and Solr. Feel free to add
On 5/10/2011 at 9:57 AM, Gabriele Kahlout wrote:
On Tue, May 10, 2011 at 3:50 PM, Steven A Rowe sar...@syr.edu wrote:
http://svn.apache.org/repos/asf/lucene/dev/branches/lucene_solr_3_1/dev-tools/maven/README.maven
[...]
svn co http://svn.apache.org/repos/asf/lucene/dev/tags/lucene_solr_3_1
I recently set up a solrj application that uses Solr Trunk and grouping. I
didn't see where there was any explicit support in solrj for grouping (in
Trunk...Maybe there is in the old SOLR-236 version). But you can set any
parameters on the request like this:
SolrQuery query = new
Hello,
I'm having trouble getting Solr 3.1 to work with nutch-1.3. I'm not sure
where the problem is, but I'm wondering why does the solrHome path end with
/./.
cwd=/Applications/NetBeans/apache-tomcat-7.0.6/bin
SolrHome=/Users/simpatico/apache-solr-3.1.0/solr/./
In the web.xml of solr:
Steve,
I'm not used to update wikis, but I've added a small part after the IntelliJ
part here :
http://wiki.apache.org/solr/HowToContribute
Ludovic.
-
Jouve
France.
--
View this message in context:
It apparently is normal, and my issue is indeed with nutch.
I've modified post.sh from the example docs to use the solr in
http://localhost:8080/apache-solr-3.1-SNAPSHOT and now finally data made it
to the index.
$ post.sh solr.xml monitor.xml
With nutch I'm at:
$ svn info
Path: .
URL:
From solr logs:
May 10, 2011 4:33:20 PM org.apache.solr.common.SolrException log
*SEVERE: org.apache.solr.common.SolrException: ERROR:unknown field 'content'
*
at
org.apache.solr.update.DocumentBuilder.toDocument(DocumentBuilder.java:321)
at
There is a working example schema in Nutch' conf directory.
On Tuesday 10 May 2011 16:40:02 Gabriele Kahlout wrote:
From solr logs:
May 10, 2011 4:33:20 PM org.apache.solr.common.SolrException log
*SEVERE: org.apache.solr.common.SolrException: ERROR:unknown field
'content' *
at
I don't get you, are you talking about conf/schema.xml? That's what I'm
referring to. Am i supposed to do something with the nutch's
conf/schema.xml?
On Tue, May 10, 2011 at 4:46 PM, Markus Jelsma
markus.jel...@openindex.iowrote:
There is a working example schema in Nutch' conf directory.
On
Thanks Grijesh for responding. I meant that I will use the Lucene 3.1
jars for indexing also from now on. My current index already has a
million docs indexed with solr 1.4.1 version, I read somewhere that
once server is upgraded to 3.1, it is said that the first commit will
change the indexes to
You need to use the schema.xml shipped with Nutch in Solr. It provides most
fields that you need.
On Tuesday 10 May 2011 17:31:33 Gabriele Kahlout wrote:
I don't get you, are you talking about conf/schema.xml? That's what I'm
referring to. Am i supposed to do something with the nutch's
You mean that I should copy it from nutch into solr?
$ cp $NUTCH_HOME/conf/schema.xml $SOLR_HOME/conf/schema.xml
After restarting tomcat, and re-executing the script nothing changed.
On Tue, May 10, 2011 at 5:35 PM, Markus Jelsma
markus.jel...@openindex.iowrote:
You need to use the schema.xml
actually something changed, I managed to crawl and index some pages (the
other must have to do with regex-urls). Thank you!
Was this always necessary? Any pointer discussing why it's needed?
On Tue, May 10, 2011 at 5:40 PM, Gabriele Kahlout
gabri...@mysimpatico.comwrote:
You mean that I should
Yes, that schema is configured for the fields Nutch can generate. It's
recommended to remove the Solr index after changing the schema. Then you must
reindex. The
On Tuesday 10 May 2011 17:40:39 Gabriele Kahlout wrote:
You mean that I should copy it from nutch into solr?
$ cp
Great posts all. I will give these a look and come up with something based
on these recommendations. I'm sure as I begin implementing something, I will
have more questions arise.
On Tue, May 10, 2011 at 9:00 AM, Anthony Wlodarski
anth...@tinkertownlabs.com wrote:
The WIKI has a loose
I'm actually using php but I get what you're saying. I think I understand
what I need to do. Thanks a lot man!
--
View this message in context:
http://lucene.472066.n3.nabble.com/SolrQuery-API-for-adding-group-filter-tp2921539p2923701.html
Sent from the Solr - User mailing list archive at
Just a heads up on a solution.
copyField wasn't need, but a new fieldType and a non-indexed, non-stored
field was added.
Within a new Synonym processor that executes right before the
AnalyzerQueryNodeProcessor, I was able to modify the field name for each
node to point at the new field.
Hello Mr. Kanarsky,
Thank you very much for the detailed explanation,
probably the best explanation I found regarding replication. Just to
be sure, I wanted to test solr 3.1 to see if it alleviates the
problems...I dont think it helped. The master index version and
generation are
Very nice! Good job! :)
-Original Message-
From: mtraynham [mailto:mtrayn...@digitalsmiths.com]
Sent: Tuesday, May 10, 2011 9:44 AM
To: solr-user@lucene.apache.org
Subject: RE: Synonym Filter disable at query time
Just a heads up on a solution.
copyField wasn't need, but a new
Lucene also lets you change the norms directly, from an IndexReader.
In theory this could do what you're seeking, I believe. But it's
rather expert... and I don't think it's exposed through to Solr (?).
Mike
http://blog.mikemccandless.com
On Tue, May 10, 2011 at 9:39 AM, Ahmet Arslan
Hi,
I am trying to use solrj for my application, my code is given below,
query.add(q, simplify360);
query.add(facet, true);
query.add(facet.range, createdOnGMTDate);
query.add(facet.range.start,
2010-08-01T00:00:00Z+330MINUTES);
It's already much better than the old dismax, and probably more stable.
If you do not require the ability to lock down what fields people should be
allowed to search for, I suggest you start using it right away. If you require
to lock that down, I suggest you help out finishing that part of
Jan, is there already an issue for that?
It's already much better than the old dismax, and probably more stable.
If you do not require the ability to lock down what fields people should be
allowed to search for, I suggest you start using it right away. If you
require to lock that down, I
OK, then your log is probably just coming out to the console. you can
start it as java -jar start.jar file.log 21 and keep a permanent
record of the log if you're on windows
Best
Erick
On Mon, May 9, 2011 at 7:32 AM, Rohit ro...@in-rev.com wrote:
Hi Erick,
Thats exactly how I am starting
: It's already much better than the old dismax, and probably more stable.
: If you do not require the ability to lock down what fields people should
: be allowed to search for, I suggest you start using it right away. If
: you require to lock that down, I suggest you help out finishing that
:
I actually have another question unrelated to this (but related to grouping).
I'm wondering if I can do a more complex grouping, such as grouping by a
field and also making sure it matches some other criteria (such as date).
For example, currently it might group 5 items from some field, but the
: Thanks Grijesh for responding. I meant that I will use the Lucene 3.1
: jars for indexing also from now on. My current index already has a
: million docs indexed with solr 1.4.1 version, I read somewhere that
: once server is upgraded to 3.1, it is said that the first commit will
: change the
: The reason why this assert hits you in one of your tomcat installations
: could also be related to some instrumentation tools you have enabled in this
: tomcat. Lot's of instrumentation tools may dynamically change class bytecode
: and e.g. make them unfinal. In that case the assertion of
: Subject: Total Documents Failed : How to find out why
: References: 96637.83834...@web121717.mail.ne1.yahoo.com
: 1304627391.29391.156.ca...@lysdexic.healthline.com
: In-Reply-To: 1304627391.29391.156.ca...@lysdexic.healthline.com
http://people.apache.org/~hossman/#threadhijack
Thread
: 2) Use the usual Solr servlet apparatus, but defer its startup until
: other code in the webapp makes up its mind about configuration and
: calls System.setProperty to locate the solr home and data directories.
I would suggest that route ... basically just create your own war, and
subclass
Hello,
I am using solr search application, try to search text with wildcard
character getting some weird result.
Case :- 1 q=tes*
Result: test, test1, test2
Case:- 1 q=test*
Result: NONE
Kindly help me to solve above problem.
Thanks Regards,
Satyam
: Subject: Solrj Questions
: References: banlktikz0roj7eiccm-tx9q7u7j6fg0...@mail.gmail.com
: 19898929-d0b9-47fb-b62a-5fd0437a8...@cominvent.com
: 4dc936de.9050...@tinkertownlabs.com
: BANLkTi=j3rfo3_ivwfmhzadfgusdpfd...@mail.gmail.com
: In-Reply-To:
Hoss,
Thank you very much for clearly delineating the difference.
Just to be clear - My intent to move to 3.1 was driven by my desire to
improve my replication performance - Deducing from your explanation, I
believe the replication/indexing related changes/bug fixes like the
following
I think the key question here is what's the best way to perform indexing
without affecting search performance, or without affecting it much. If
you have a batch of documents to index (say a daily batch that takes an
hour to index and merge), you'd like to do that on an offline system,
and
I am using solr search application, try to search text with
wildcard
character getting some weird result.
Case :- 1 q=tes*
Result: test, test1, test2
Case:- 1 q=test*
Result: NONE
Kindly help me to solve above problem.
Do you have stemming filter defined in your analyzer
Master/slave replication does this out of the box, easily. Just set the slave
to update on Optimize only. Then you can update the master as much as you
want. When you are ready to update the slave (the search instance), just
optimize the master. On the slave's next cycle check it will refresh
Hi, all.
Sorry for the 'spam' - I'm just testing that my posts are actually being seen.
I've sent a few queries over the past couple of weeks and haven't had a single
response :(
Anyways - if one or two would respond to this, I'd appreciate it - just to let
me know that I'm being ignored, vs
Hi Bob
I just joined Solr Mailing List and I can see your post :)
Hope your query get solved.
Regards
Jasneet Sabharwal
On 11-05-2011 01:59, Bob Sandiford wrote:
Hi, all.
Sorry for the 'spam' -- I'm just testing that my posts are actually
being seen. I've sent a few queries over the past
Thanks - that sounds like what I was hoping for. So the I/O during
replication will have *some* impact on search performance, but
presumably much less than reindexing and merging/optimizing?
-Mike
Master/slave replication does this out of the box, easily. Just set the slave
to update on
Replication large files can be bad for OS page cache as files being written are
also written to the page cache. Search latency can grow due to I/O for getting
the current index version back into memory. Also, Solr cache warming can casue
a doubling of your heap usage.
Frequent replication in
Hello everyone,
My Query
q=google.com
should give the ouput as
http://www.googe.com
In my search engine when i put q=google.com, it does not give the above
output, rather it shows no result.
but putting
q=url:google.com gives the correct output as http://www.google.com
Now my question how can i
Hello.
One option is to specify a default search field in your schema.xml. If your
query doesn´t include an specific field the query parser will use the
default one to launch the query. You should include in your schema.xml
something like this:
!-- field for the QueryParser to use when an
thanks! , it worked.
Can i mention
defaultSearchFieldurl,content/
defaultSearchField
to inlcude two default fields.?
On Wed, May 11, 2011 at 3:02 AM, Rakudten [via Lucene]
ml-node+2924686-576776982-146...@n3.nabble.com wrote:
Hello.
One option is to specify a default search field in your
OK let me rephrase.
In solrconfig.xml there is a setting called mergeFactor. The default is
usually 10.
Practically it means there are 10 segments. If you are doing fast delta
indexing (adding a couple documents, then committing),
You will cycle through all 10 segments pretty fast.
It appears
This is fine. Solr needs lots of memory though... :) It may not be as fast
as you need, but you can test that.
On 5/9/11 2:33 AM, solr_begin...@onet.pl solr_begin...@onet.pl wrote:
Hi,
Is that possible that solr on tomcat on windows 2008 is using only one
core of processor? Do I need configure
85 matches
Mail list logo