Thanks for reply. Please find my comments below. although after all the
basic R&D, i wrote the issue.
On Fri, Dec 9, 2011 at 1:37 AM, Erick Erickson wrote:
> If you look at the admin/stats page, does it show anything
> for numDocs and maxDocs?
>
> i did look at stats. there are around 71 thousand
Hi guys,
I'm looking for NRT functionality or similar in Solr 3.5. Is that possible?
>From what I understand there's NRT in Solr 4, but I can't figure out
whether or not 3.5 can do it as well?
If not, is it feasible to use an autoCommit every 1000ms? We don't
currently process *that* much data so
[?], thanks. Try in a minute.
Spark
2011/12/9 Mark Miller
> pseudo code:
>
>SolrServer client = ...
>
>UpdateRequest ureq = new UpdateRequest();
>ureq.add(doc);
>ureq.setParam("param", "value");
>ureq.setAction( UpdateRequest.ACTION.COMMIT, waitFlush, waitSearcher);
>ure
pseudo code:
SolrServer client = ...
UpdateRequest ureq = new UpdateRequest();
ureq.add(doc);
ureq.setParam("param", "value");
ureq.setAction( UpdateRequest.ACTION.COMMIT, waitFlush, waitSearcher);
ureq.process(client);
It will say setAction is deprecated, but internal So
I will try to use your methods.
2011/12/9 Sharath Jagannath
> Kinda off topic why not directly use /solr/update/json?commitWithin=3000,
> ofcourse with autocommit.maxtime and autoSoftCommit in configs.
> I am not using the java client but send the docs using the http client.
> It is doing fairly
Hi Mark,
My question inline.
2011/12/9 Mark Miller
> I don't know about DataImportHandler.
>
> But the whole point of AutoCommit and AutoSoftCommit is that they happen
> 'Auto'matically. You couldn't measure how long they take from solrj, and if
> you are using solrj to do the commit, it has no
Kinda off topic why not directly use /solr/update/json?commitWithin=3000,
ofcourse with autocommit.maxtime and autoSoftCommit in configs.
I am not using the java client but send the docs using the http client.
It is doing fairly good for me.
Zoie is another option worth looking at. Had very good e
I don't know about DataImportHandler.
But the whole point of AutoCommit and AutoSoftCommit is that they happen
'Auto'matically. You couldn't measure how long they take from solrj, and if
you are using solrj to do the commit, it has nothing to do with auto commit.
If you want to commit from solrj
That sounds like a good idea. I will check my schema configuration, and see
what I can do.
Thanks Mark and Siva for all the information, and see what I can do, and
update the result here.
Spark
2011/12/9 Siva Kommuri
> Hi Spark,
>
> Is it possible to store a modified time on the database recor
Hi Spark,
Is it possible to store a modified time on the database record which would help
you to get the updated documents?
That way you can avoid reindexing the documents that have not been updated.
Best wishes,
Siva on 3GS
On Dec 8, 2011, at 15:40, yu shen wrote:
> Hi Mark,
>
> I did not
On Thu, Dec 8, 2011 at 6:21 PM, Tom Lianza wrote:
> We're seeing the same thing (though we're not using replication). Based on
> the trace, it looks like it would happen when Solr's response is too slow
> for the client, and it's trying to send a response back to someone who's no
> longer listeni
Hi Mark,
I did not fully catch you.
I have a dataImportHandler which import data from a database. Are you
suggesting autoCommit/autoSoftCommit can automaticall pull data from
database using dataImportHandler?
Otherwise, how can I commit user modification of search result from my web
page back to
We're seeing the same thing (though we're not using replication). Based
on the trace, it looks like it would happen when Solr's response is too
slow for the client, and it's trying to send a response back to someone
who's no longer listening for one.
So, I think this isn't an error in itself
: I've a "long" field defined in my schema:
:
:
:
:
:
: Before r1201855 I could use "stats.facet=ts" which allowed me to have a
: timeseries of sorts, now I get an error:
Since your field type doesn't explicitly state "multiValued="true|false"
Solr's behavior in parsing the schema.xml is dri
Hello,
I've a "long" field defined in my schema:
omitNorms="true" positionIncrementGap="0" />
Before r1201855 I could use "stats.facet=ts" which allowed me to have a
timeseries of sorts, now I get an error:
"Stats can only facet on single-valued fields, not:
ts[long{class=org.apache.s
Greetings,
I see that we can query multiple facets for a search with a syntax like
"fq=grade:A OR grade:B". However, I only know how to do this by modifying the
URL parameter. Is there a UI component that allows you to select multiple
facet values? I'm thinking something like a checkbox next
Hello,
It should do the job, thanks!
Marc.
On Thu, Dec 8, 2011 at 2:06 PM, Erick Erickson wrote:
> What about q=my_field1:my_search_string~2 my_search_string
> ?
> True, the clause edismax created would search for your
> my_search_string in my_field1 twice, but it might suffice..
>
> Best
> Eri
I believe all you need to do is add a ?clean=false to your query string.
If you have a unique key setup as your ID in solr then it should update
the existing documents instead of delete and re-indexing.
Cody
-Original Message-
From: P Williams [mailto:williams.tricia.l...@gmail.com]
Sen
: *pk*: The primary key for the entity. It is*optional*and only needed when
: using delta-imports. It has no relation to the uniqueKey defined in schema.xml
: but they both can be the same.
:
: When using in a nested entity is the PK the primary key column of the join
: table or the key used for
These are really horrible numbers:
hitratio : 0.18
inserts : 5488
evictions : 5011
Your cache isn't doing you much good, part of the problem
is you're faceting on so many values and it's chewing through
your cache. If your full data set has significantly more terms,
you're going to have trouble sc
OK, migrating to Solr 3.5 from what?
But what it *looks* like is that you are running a 3.5
config (see luceneMatchVersion in solrconfig.xml)
against a 3.3 Solr installation. In other words
it might be that you're running the old code against
a new schema.
But I'll freely admit that the luceneMat
If you look at the admin/stats page, does it show anything
for numDocs and maxDocs?
Have you tried looking at the index with Luke to see what's
in there?
Have you tried just looking at your data/index directory and
seeing if there are files there?
Are you sure you commit the changes?
Did you re
Hi,
Is there a specific reason why it is hard-coded to use the "lucene"
QParser? I was looking at JoinQParserPlugin.java and here it is in
createParser:
QParser fromQueryParser = subQuery(v, "lucene");
I could pass another param named "fromQueryParser" and use it instead of
"lucene". But again,
: 3) When the user clicks into a single video in the search result,
: retrieve from the corresponding doc in Solr the timestamps of all
: words matching the keyword(s) (including stemming).
...
: Ok, so now for the harder part. For #3 it would seem I need something
: roughly like the high
Thanks Robert. I'll watch them all. Any others that are good to keep track of?
On Thu, Dec 8, 2011 at 1:25 PM, Robert Muir wrote:
> On Thu, Dec 8, 2011 at 12:55 PM, Jamie Johnson wrote:
>> Thanks Andrzej. I'll continue to follow the portable format JIRA
>> along with 3622, are there any other
: But there is a workaround:
: 1) Do a normal query without facets (you only need to request doc ids
: at this point)
: 2) Collect all the IDs of the documents returned
: 3) Do a second query for all fields and facets, adding a filter to
: restrict result to those IDs collected in step 2.
an easi
: If I check in the solr.admin.analyzer, I get the same analysis for the two
: different requests. But it seems, if fact, that the lacking space after
: coma prevents name and number from matching.
query analysis is only part of hte picture ... Did you look at the
debuqQuery output? ... i belie
Hi Bernd,
Check this:
"SPM for Solr is the enterprise-class, cloud-based, System/OS and Solr
Performance Monitoring SaaS."
So it's a SaaS - you simply sign up for it. During the signup you'll get to
download a small agent that works on RedHat, CentOS, Debian, Ubuntu, and maybe
other OSes.
Ah. Thanks Erick.
I see now that my question is different from sabman's.
Is there a way to use the DataImportHandler's "full-import" command so that
it does not delete the existing material before it begins?
Thanks,
Tricia
On Thu, Dec 8, 2011 at 6:35 AM, Erick Erickson wrote:
> This is all co
On Thu, Dec 8, 2011 at 12:55 PM, Jamie Johnson wrote:
> Thanks Andrzej. I'll continue to follow the portable format JIRA
> along with 3622, are there any others that you're aware of that are
> blockers that would be useful to watch?
>
There is a lot to be done, particularly norms and deleted doc
Thanks Andrzej. I'll continue to follow the portable format JIRA
along with 3622, are there any others that you're aware of that are
blockers that would be useful to watch?
On Thu, Dec 8, 2011 at 10:49 AM, Andrzej Bialecki wrote:
> On 08/12/2011 14:50, Jamie Johnson wrote:
>>
>> Mark,
>>
>> Agre
On 08/12/2011 14:50, Jamie Johnson wrote:
Mark,
Agreed that Replication wouldn't help, I was dreaming that there was
some intermediate format used in replication.
Ideally you are right, I could just reindex the data and go on with
life, but my case is not so simple. Currently we have some set
Have you looked at this page?
http://wiki.apache.org/solr/SolrTomcat
In particular the "installing Solr instances under Tomcat".
You haven't told us much at all about how you start up Tomcat,
what configurations you have set, not even what error you get
when you try to hit the admin page. Please
Hello!!!
I have a trouble with searching russian words in Solr. And also i'm used
alfresco with solr. I'm add a content in my custom field on the russian
language. But when i search this words in Solr - i don't found anyone. I'm
search topics with this problems and did the following things, but t
Thanks Robert. I'll continue to watch the Jira and try not to bother
folks about this. Again greatly appreciate the insight.
On Thu, Dec 8, 2011 at 11:31 AM, Robert Muir wrote:
> On Thu, Dec 8, 2011 at 10:46 AM, Mark Miller wrote:
>>
>> On Dec 8, 2011, at 8:50 AM, Jamie Johnson wrote:
>>
>>> I
On Thu, Dec 8, 2011 at 10:46 AM, Mark Miller wrote:
>
> On Dec 8, 2011, at 8:50 AM, Jamie Johnson wrote:
>
>> Isn't the codec stuff merged with trunk now?
>
> Robert merged this recently AFAIK.
>
true but that issue only moved the majority of the rest of the index
(stored fields, term vectors, fi
On Thu, Dec 8, 2011 at 11:01 AM, Jay Luker wrote:
> Hi,
>
> I am trying to provide a means to search our corpus of nearly 2
> million fulltext astronomy and physics articles using regular
> expressions. A small percentage of our users need to be able to
> locate, for example, certain types of iden
Hi,
I am trying to provide a means to search our corpus of nearly 2
million fulltext astronomy and physics articles using regular
expressions. A small percentage of our users need to be able to
locate, for example, certain types of identifiers that are present
within the fulltext (grant numbers, d
Now when i start my solr it gives an exception :
org.apache.solr.common.SolrException: Error loading class
'org.apache.solr.update.processor.LangDetectLanguageIdentifierUpdateProcessorFactory'
Caused by: java.lang.ClassNotFoundException:
org.apache.solr.update.processor.LangDetectLanguageIdentifi
On Dec 8, 2011, at 8:50 AM, Jamie Johnson wrote:
> Isn't the codec stuff merged with trunk now?
Robert merged this recently AFAIK.
- Mark Miller
lucidimagination.com
Lets please keep the conversation on list.
If you are using auto commit and soft auto commit, it makes no sense to be
committing with solrj. The commits happen automatically. You couldn't measure
them with solrj.
- mark
On Dec 8, 2011, at 12:44 AM, yu shen wrote:
> Thanks for the help.
>
> F
This would seem to indicate that you are using a whitespace analyzer on
the default search field. I believe other analyzers will properly tokenize
around the comma.
> same problem with Solr 4.0
>
> 2011/12/8 elisabeth benoit
>
>>
>>
>> Hello,
>>
>> I'm using Solr 3.4, and I'm having a problem wit
I read this response, but it lacks the quoted text so I have no clue what
your advice is in reference to. This makes it hard for others to benefit
from the advice. Just a thought.
> Go ahead with SOLR based text search. Thats what it is meant for and does
> it
> great.
>
> Regards
> Pravesh
>
> --
Thanks for the response Erick. I actually turned up logging yesterday and
noticed spellchecker builds were causing the delays. Setting buildOnCommit to
false solved the problem. Our plan is to schedule a nightly timer task that
sends a 'spellcheck.build=true' to trigger it.
Mike
> Date: Th
Good Afternoon,
Im looking at Deltas via a DeltaImportHandler. I was running Solr 1.4.1
but just upgraded to 3.5. Previously I was able to run debug and verbose
from:
http://localhost:8080/solr/admin/dataimport.jsp?handler=/advert
But since upgrading when choosing these options the right
Mark,
Agreed that Replication wouldn't help, I was dreaming that there was
some intermediate format used in replication.
Ideally you are right, I could just reindex the data and go on with
life, but my case is not so simple. Currently we have some set of
processes which is run against the raw ar
This is all controlled by Solr via the field in your schema. Just
remove that entry.
But then it's all up to you to handle the fact that there will be multiple
documents with the same ID all returned as a result of querying. And
it won't matter what program adds data, *nothing* will be overwritte
Nope, they're the same. The original name was Field Collapsing,
but it was changed to "Grouping" later.
But note that the functionality has changed over time, so you might
be seeing documents from different incarnations of the code.
Best
Erick
On Wed, Dec 7, 2011 at 10:41 AM, Kissue Kissue wrot
Hi,
I was just testing field collapsing in my solr admin on solr 3.5.0. I have
observed that the results of field collapsing are not being cached unlike
other solr query results. Am doing the same query multiple times and the
time taken still remains approximately the same. Is there something i n
Have you, for instance, used the admin/browse schema or Luke
to examine your indes and see if you're indexing your text data
at all?
How are you indexing with SolrJ? Using the extracting request
handler or parsing the docs with the Tika libs yourself?
Details matter.
Best
Erick
On Wed, Dec 7, 20
Hmmm, this is unusual. Can we see the code you use to delete?
And your solrconfig file? You're not doing something odd like
optimizing on commit or anything, right?
You shouldn't have to commit after deletes. The fact that you're
hanging is very odd (BTW, does "hanging" mean you're system
is locke
I have increased the filtercache values as to as and and but still it performs as in seconds. I have got
shards along with an average of 6 gb of data is indexed including the
shards. But in production servers the data including shards has average of
80Gb data. it has 4,273 unique terms after do
What about q=my_field1:my_search_string~2 my_search_string
?
True, the clause edismax created would search for your
my_search_string in my_field1 twice, but it might suffice..
Best
Erick
On Wed, Dec 7, 2011 at 3:15 AM, Marc SCHNEIDER
wrote:
> Hello,
>
> I'm using edismax and Solr 4.0 and I'd lik
I have apache-solr-langid-3.5.jar and also the jars in contrib/langid/lib
Ankita
On 8 December 2011 18:20, O. Klein wrote:
> Did you also install apache-solr-langid-4.0.jar ?
>
> And the jars in contrib/langid/lib ?
>
> --
> View this message in context:
> http://lucene.472066.n3.nabble.com/Lan
Did you also install apache-solr-langid-4.0.jar ?
And the jars in contrib/langid/lib ?
--
View this message in context:
http://lucene.472066.n3.nabble.com/Language-Detection-in-SOLR-3-5-tp3570177p3570202.html
Sent from the Solr - User mailing list archive at Nabble.com.
Hi,
I tried to implement language detection in SOLR 3.5.
I added the processor class for langDetect in solrconfig.xml :
cContent
language
en
langid
I declared the filed "language" in the schema as the output field for the
detect
Thanks Marcus. To resolve this problem i just added a shared lib folder for
my cores and added the velocity jars in this folder and that resolved the
error. I hope it was the right thing to do though.
Thanks.
On Thu, Dec 8, 2011 at 11:20 AM, Markus Jelsma
wrote:
> From the changelog:
>
>
> 187
Oh. That's bad to me.
Thanks anyway.
--
View this message in context:
http://lucene.472066.n3.nabble.com/How-to-set-default-query-operator-in-surround-query-parser-tp3570034p3570088.html
Sent from the Solr - User mailing list archive at Nabble.com.
> I'm using surround query parser.
> The request "A B" returns ParseException.
> But "A OR B" returns correct results.
> I think this is the problem of default query operator.
> Anyone know how to set?
There is no room for default operator in surround query parser. This is a
limitation of surroun
How many unique terms do you have in the faceting field
Since there are lot of evictions, consider increasing the size of the
filtercache. Try to keep evictions to min.
BTW how much is your index size (GB/MB??) How much RAM is allocated?
Above All: Have you benchmarked your search? Is search
From the changelog:
187
* SOLR-2588: Moved VelocityResponseWriter back to contrib module in order to
188
remove it as a mandatory core dependency. (Erik Hatcher)
http://svn.apache.org/viewvc/lucene/dev/branches/branch_3x/solr/CHANGES.txt?view=markup
> I just migrated to Solr 3.5 and whene
change the default operator in schema
Thanks and Regards,
S SYED ABDUL KATHER
On Thu, Dec 8, 2011 at 4:44 PM, Jason, Kim [via Lucene] <
ml-node+s472066n3570034...@n3.nabble.com> wrote:
> Hi, all
>
> I'm using surround query parser.
> The request "A B" returns ParseException.
Hi, all
I'm using surround query parser.
The request "A B" returns ParseException.
But "A OR B" returns correct results.
I think this is the problem of default query operator.
Anyone know how to set?
Thanks,
Jason
--
View this message in context:
http://lucene.472066.n3.nabble.com/How-to-set-de
I just migrated to Solr 3.5 and whenever i start it up i get the error
below. Any ideas what might be wrong? Previously i didn't have to do
anything special to get it to work. HAs anything changed in solr 3.5?
08-Dec-2011 10:45:03 org.apache.solr.common.SolrException log
SEVERE: org.apache.solr.c
filed type is either long or string.
Solr Statistics are:
*name: * filterCache *class: * org.apache.solr.search.FastLRUCache *
version: * 1.0 *description: * Concurrent LRU Cache(maxSize=512,
initialSize=512, minSize=460, acceptableSize=486, cleanupThread=false) *
stats: * lookups : 6679
h
After migrating to Solr 3.5, i restart tomcat and i get the error below.
Any ideas what i am doing wrong?
SEVERE: org.apache.solr.common.SolrException: Invalid luceneMatchVersion
'LUCENE_35', valid values are: [LUCENE_20, LUCENE_21, LUCENE_22, LU
CENE_23, LUCENE_24, LUCENE_29, LUCENE_30, LUCENE_31
On 08/12/2011 05:00, Mark Miller wrote:
Replication just copies the index, so I'm not sure how this would help offhand?
With SolrCloud this is a breeze - just fire up another replica for a shard and
the current index will replicate to it.
If you where willing to export the data to some portabl
What is the type of the field on which you are getting facets (string, Text,
int, date etc.). Is it multivalued or not?
How many unique values do you have for the field?
What is your filtercache setting in your solrconfig.xml?
Regards
Pravesh
--
View this message in context:
http://lucene.47206
Hi
I newly upgraded to Solr 3.2 from Solr 1.4. I was using DIH in 1.4 & also
in 3.2. I built my index & it doesn't gave me any issues. The index
directory & it's files are properly in place
Now, when i tried to look the data through admin, it is giving me ZERO
results. does it due to "HTTP Cachin
I have a query and faceting on one field. i want to improve the speed of
getting this facet please advise me the solutions what are all the options
available to get this facet faster.
--
View this message in context:
http://lucene.472066.n3.nabble.com/How-to-improve-facet-search-tp3569910p3569910
Can u share more info: like what is your H/W infra, CPU, RAM, HDD??
>From where you pick the records/documents to index; RDBMS, Files, Network??
Regards
Pravesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/Solr-using-very-high-I-O-tp3567076p3569903.html
Sent from the Solr
Go ahead with SOLR based text search. Thats what it is meant for and does it
great.
Regards
Pravesh
--
View this message in context:
http://lucene.472066.n3.nabble.com/Solr-or-SQL-fultext-search-tp3566654p3569894.html
Sent from the Solr - User mailing list archive at Nabble.com.
I have a text field, using stopwords...
Index and query analysers setup as follows:
SynonymFilterFactory
StopFilterFactory
WordDelimiterFilterFactory
LowerCaseFilterFactory
SnowballPorterFilterFactory
Searching for "front of house" brings back perfect matches, but
doesn't highlight the "of".
same problem with Solr 4.0
2011/12/8 elisabeth benoit
>
>
> Hello,
>
> I'm using Solr 3.4, and I'm having a problem with a request returning
> different results if I have or not a space after a coma.
>
> The request "name, number rue taine paris" returns results with 4 words
> out of 5 matching
Hello,
I'm using Solr 3.4, and I'm having a problem with a request returning
different results if I have or not a space after a coma.
The request "name, number rue taine paris" returns results with 4 words out
of 5 matching ("name", "number", "rue", "paris")
The request "name,number rue taine pa
Hi All,
My autocommit settings are
max docs - 1000
max time - 86 secs
We have put newrelic agent so as to monitor our solr performance.
In that we see a continous curve for autocommit. It is as good as
autocommit is continuously being fired.
Is it that if Autocommit for certain documents takes s
Hello Yavar,
thanks for your reply.
I have two entries in the log files.
one entry is in the localhost_access_log.txt:
my ip adress - - [08/Dec/2011:08:59:50 +] "GET
/solr/select/?q=sicherheitsschuh&version=2.2&start=0&rows=10&indent=on
HTTP/1.1" 200 388
and the other entry is in catalina.t
Hi Otis,
I can't find the download for the free SPM.
What Hardware and OS do I need for installing SPM to monitor my servers?
Regards
Bernd
Am 07.12.2011 18:47, schrieb Otis Gospodnetic:
Hi Dmitry,
You should use SPM for Solr - it exposes all Solr metrics and more (JVM, system
info, etc.)
PL
Hello,
Usually, when such an error occur, there are some good hints of what's
wrong with your new configuration in solr logs.
Depending on how you setup your solr instance and configured logging for
solr (http://wiki.apache.org/solr/SolrLogging), log files may be located
at different places.
Hi,
I have a solr core named collection1 running in a tomcat instance
modified with a DIH.
It's working fine and everything seems to be in order, but there was
discovered some minor flaws in
the DIH setup, so I have corrected it in the data-config.xml file that
is used by the DIH requestHandler.
When I start solr indexing RAM taken by MS SQL Server 2008 R2 also keeps on
increasing & initially from some 1GB it went on to 3.6 GB (when indexing was
completed for just 1 Million records/5GB). I have set the responseBuffering
parameter to adaptive in data-config.xml however it didn't help me
I tried that, didn't seem to affect anything.
I think it only works between fields and not within a field.
On Thu, Dec 8, 2011 at 10:00 AM, Marc SCHNEIDER
wrote:
> Hi,
>
> Maybe you could have a look at this :
> http://wiki.apache.org/solr/DisMaxQParserPlugin#tie_.28Tie_breaker.29
>
> Marc.
>
> O
Hi,
Maybe you could have a look at this :
http://wiki.apache.org/solr/DisMaxQParserPlugin#tie_.28Tie_breaker.29
Marc.
On Wed, Dec 7, 2011 at 5:48 PM, John wrote:
> I have a complex edismax query:
>
>
> facet=true&facet.mincount=0&qf=title^0.08+categorysearch^0.05+abstract^0.03+body^0.1&wt=java
83 matches
Mail list logo