Okay so after some debugging I found the problem. While the replication
piece will download the index from the master server and move the files to
the index directory but during the commit phase, these "older" generation
files are deleted and the index is essentially left in tact.
I noticed that a
Thanks James for the heads up and apologies for a delayed response.Here's the
full details about this issue. Mine is an e-com app so the index contains
the product catalog comprising roughly 13million products. At this point I
thought of using the index based dictionary as the bet option for the "D
Hi Mikhail,
Thanks for your guidance. I found the required information in
debugQuery=on.
Thanks and regards,
Romita
From: Mikhail Khludnev
To: solr-user ,
Date: 01/24/2013 03:19 PM
Subject:Re: Get tokenized words in Solr Response
Romita,
IIRC you've already asked this, a
On Thu, Jan 24, 2013 at 8:55 PM, Otis Gospodnetic
wrote:
> Yes, this is JSON, so right
> there it may be better, but for instance I see "v" here which to a regular
> human may not be as nice as "value" if that is what "v" stands for.
One goal was to reuse the parsers/parameter names. A completel
Nice, Yonik!
Here is one suggestion. OK, I'm beginning you - please don't make
it be as hard on the eyes as Local Params. :) I thought it was just me who
could never get along with Local Params, but I've learned that a number of
people find Local Params very hard to grok. Yes, this is JSO
Although "lucene" syntax tends to be quite concise, nice looking, and
easy to build by hand (the web browser is a major debugging tool for
me), some people prefer to use a more "structured" query language
that's easier to build up programmatically. XML fits the bill, but
people tend to prefer JSON
Thanks for the education Chris,
I pasted the chars into Index and Query fields on analyzer panel.
Index/Query Analyzers almost the same..
On both, non-greeks drop out after worddelimiterfilter
Index analyzer has grey background of words that seem to make it thru all the
filters.
WhitespaceTo
: My next target is searches on simple terms such as "doll" which, in google,
: would return documents about, well, "toy dolls", because that's the most
: common usage of the simple term "doll". But in my index it predominantly
: returns documents about CDs with the song "Doll Face", and "My baby
: For example perhaps a load balancer that sends multiple queries
: concurrently to all/some replicas and only keeps the first response
: might be effective. Or maybe a load balancer which takes account of the
I know of other distributed query systems that use this approach, when
query speed
: When I search for these characters in the admin query, I can only find the
Greeks.
: debug shows the parsed query only has greek chars like omega, delta, sigma
: but does not contain others like degree, angle, cent, bullet, less_equal…
this is most likeley because of the analyzer you are using
On Jan 24, 2013, at 5:22 PM, Fadi Mohsen wrote:
>
> The reason we would like to avoid Zookeeper are
> * due to lack of knowledge.
> * the amount of work/scripting for developers per module and release
> documentation.
> * the extra steps of patching ZK nodes for QA and operations.
>
> ZkCLI is
Hi Hoss
Thanks for the reply.
Unfortunately we have other customized similarity classes that I don’t know how
to disable them and still make query work.
I am trying to attach more information once I work out how to simply the issue.
Thanks
Ben
From: C
These are the figures I got after indexing 4 and half million documents with
both Solr 3.6.1 and 4.1.0 (and optimizing the index at the end).
$ du -h --max-depth=1
67G ./solr410
80G ./solr361
Main contributor to the reduced space consumption is (as expected I guess) the
.fdt file:
Thanks Per, would the first approach involve restarting Solr?
Thanks Mark, that's great, Ill try checkout and apply patches from ticket
to understand further.
The reason we would like to avoid Zookeeper are
* due to lack of knowledge.
* the amount of work/scripting for developers per module and
FollowUp.cc Reminder
You received this email because gregg...@gmail.com set a public FollowUp.cc
reminder
and it's the first time you've appeared on one (congrats, you have wise
friends!).
3 Reasons why people use FollowUp.cc Reminders
- It removes the step of having to mark on y
Hi,
We have 5 core masters and 5 core slaves. The main core houses about 85,000
douments, so small, although the content of each document is quite large.
The second core holds the same number of docs but far less - and different -
data.
We reindex all cores every morning and the replication poll
We have a Hadoop process that produces a set of Solr indexes from a cluster
of HBase documents. After the job runs, we pull the indexes from HDFS and
merge the them together locally. The issue we're running into is that
sometimes we'll have duplicate occurrences of a primary key across indexes
that
Michael Della Bitta-2 wrote
> On Thu, Jan 24, 2013 at 11:34 AM, O. Olson <
> olson_ord@
> > wrote:
>>
>> Caused by: com.microsoft.sqlserver.jdbc.SQLServerException: The server
>> SQLEXPRESS is not configured to listen with TCP/IP.
>
>
> That's probably your problem...
>
>
> Michael Della Bitt
This post on stackoverflow has a good run-down on your options:
http://stackoverflow.com/questions/1555610/solr-dih-how-to-handle-deleted-documents/1557604#1557604
If you're using DIH, you can get more information from:
http://wiki.apache.org/solr/DataImportHandler
The easiest thing, if using a
ok, how can i issue delete for each item deleted since the last successful
update? Do i write something like delete query with delta import query in
dataconfig? If so, what will i add in dataconfig for deletion?
--
View this message in context:
http://lucene.472066.n3.nabble.com/Deletion-from-
On Jan 24, 2013, at 10:02 AM, Fadi Mohsen wrote:
> Hi, We would like to use Solr to index statistics from any Java module in
> our production environment.
>
> Applications have to can create collections and index data on demand, so my
> initial thought is to use different HTTP methods to accomp
On Thu, Jan 24, 2013 at 11:34 AM, O. Olson wrote:
>
> Caused by: com.microsoft.sqlserver.jdbc.SQLServerException: The server
> SQLEXPRESS is not configured to listen with TCP/IP.
That's probably your problem...
Michael Della Bitta
Appinions
18
This is my fault - I discovered this myself a few days ago. I've been meaning
to file a jira ticket and have not gotten around to it yet.
You can also work around it like this:
CoreContainer container = new CoreContainer(loader) {
// workaround since we don't call container#load
On Jan 24, 2013, at 7:05 AM, Shawn Heisey wrote:
> My experience has been that you put the chroot at the very end, not on every
> host entry
Yup - this came up on the mailing list not too long ago and it's currently
correctly documented on the SolrCloud wiki.
- Mark
The general solution is to add a "deleted" column to your database, or even a
"deleted date" column.
When you update Solr from the DB, issue a delete for each item deleted since
the last successful update.
You can delete those rows after the Solr update or to be extra safe, delete
them a few d
Hi,
I am trying to figure out a way so that if i delete anything from my
database how will that item be deleted from my indexed data?
is there anyway i can make new core with same config as the existing core,
do full index, swap the data with the existing core and delete the new core.
So every tim
Hi,
I hate to double post but I'm not sure in which domain, the answer to my
question lies, so here's the link to my question on the mahout groups.
Basically, I'm getting different clustering results depending on whether I
index data with SOLR or Lucene. Please post any responses against the
origi
Shawn Heisey-4 wrote
>> There will be a lot more detail to this error. This detail may have a
>> clue about what happened. Can you include the entire stacktrace?
>>
>> Thanks,
>>Shawn
Thank you Shawn. The following is the entire stacktrace. I hope this helps:
INFO: Creating a connection for
Each node needs a -Dsolr.solr.home pointing to a solr.xml, but the
configuration-subfolder does not need to be there. It only needs to be
there for the node you start with -Dbootstrap_confdir (to have it load
the config into ZK). The next time you start this Solr you do not need
to provide -Dbo
On 1/24/13 4:51 PM, Per Steffensen wrote:
2) or You can have an Solr node (server) load a "Solr config" into ZK
during startup by adding collection.configName and bootstrap_confdir
VM params - something like this
java -DzkHost=
-Dcollection.configName=
-Dbootstrap_confdir= -jar start.jar
We
On 1/24/2013 8:42 AM, Ken Prows wrote:
I didn't see any mention of field compression in the release notes for
Solr 4.1. Did the ability to automatically compress fields end up
getting added to this release?
The concept of compressed fields (an option in schema.xml) that existed
in the 1.x vers
Doh!, I went straight for the release notes. Thanks, this is the
feature I was waiting for :)
Ken
On Thu, Jan 24, 2013 at 10:49 AM, André Widhani
wrote:
> This is what it listed under the "Highlights" on the Apache page announcing
> the Solr 4.1 release:
>
> "The default codec incorporates an
Basically uploading a "Solr config" (including schema.xml,
solrconfig.xml etc.) is an operation different from creating
collections. When creating a collection (e.g. using the Collection API)
you reference the (already existing) "Solr config" it needs to use.
Collections can share "Solr config"
This is what it listed under the "Highlights" on the Apache page announcing the
Solr 4.1 release:
"The default codec incorporates an efficient compressed stored fields
implementation that compresses chunks of documents together with LZ4. (see
http://blog.jpountz.net/post/33247161884/efficient
Hello!
It should be turned on by default, because the stored fields
compression is the behavior of the default Lucene 4.1 codec.
--
Regards,
Rafał Kuć
Sematext :: http://sematext.com/ :: Solr - Lucene - Nutch - ElasticSearch
> Hi everyone,
> I didn't see any mention of field compression in t
Hi everyone,
I didn't see any mention of field compression in the release notes for
Solr 4.1. Did the ability to automatically compress fields end up
getting added to this release?
Thanks!,
Ken
Hi,
You can fetch all the stored fields by passing them as part of
*fl*parameter. Go through
http://wiki.apache.org/solr/CommonQueryParameters#fl
On Thu, Jan 24, 2013 at 8:56 PM, AnnaVak wrote:
> Thanks for your solution it works for me too, I'm new with Solr but how I
> can
> additionally fetc
Hi, We would like to use Solr to index statistics from any Java module in
our production environment.
Applications have to can create collections and index data on demand, so my
initial thought is to use different HTTP methods to accomplish a collection
in cluster and then right away start HTTP PO
Hi David,
thank you for your answer.
After update to this field type and change the SOLR query I receive required
behavior.
Also could you update the WIKI page after the words "it needs to be in
WEB-INF/lib in Solr's war file, basically" also add the maven artifact code
like this?
com.vivid
We recently updated from Solr 4.0.0 to Solr 4.1.0. Because of the change we
were forced to upgrade a custom query parser. While the code change itself was
minimal, we found that our unit tests stopped working because of a
NullPointerException on line 181 of handler.component.SearchHandler:
Sha
Thanks for your solution it works for me too, I'm new with Solr but how I can
additionally fetch another fields not only field that was used for
searching? For example I have product title and image fields and I want to
get the title but also related to this title image ? How can I do this?
Thanks
Hi
I want to change autocomplete implementation for our search. Current I have a
suggest field whose definition in schema.xml is as below:
Hi
I want to change autocomplete implementation for our search. Current I have
a suggest field whose definition in schema.xml is as below:
On 1/24/2013 12:58 AM, Per Steffensen wrote:
This is supported. You just need to ajust your ZK connection-string:
":/solr,:/solr,...,:/solr"
My experience has been that you put the chroot at the very end, not on
every host entry. For a standalone zookeeper ensemble with three nodes:
"server1:
Hi,
Please search the mailing list archives - this has been discussed a few
times in the last few months.
Otis
Solr & ElasticSearch Support
http://sematext.com/
On Jan 24, 2013 6:33 AM, "real_junlin" wrote:
> Hi,
> Our company want to use solr to index our reports'data ,so we are going to
> und
Hi,
Our company want to use solr to index our reports'data ,so we are going to
understand solr.
Solr support the multi cores ,in our system, the cores'num will dynamic
increase, I afraid with more cores,the performance is decresing
dramatically.Our system's cores'num will by over one hundred.
Cool. Thanks.
On 24-Jan-2013, at 1:28 PM, Per Steffensen wrote:
> This is supported. You just need to ajust your ZK connection-string:
> ":/solr,:/solr,...,:/solr"
>
> Regards, Per Steffensen
>
> On 1/24/13 7:57 AM, J Mohamed Zahoor wrote:
>> Hi
>>
>> I am using Solr 4.0.
>> I see the Solr
You could configure your servlet container (jetty/tomcat) to have
specific webapps/contexts listen on specific IP/port combinations, that
would get you some way, But what you are asking is more about networking
and servlet container configuration than about Solr.
Upayavira
On Wed, Jan 23, 2013, a
Hello.
That is indeed an excellent article, thanks for pointing me at it. With
a title like that, it is no wonder that I was unable to google it on my
own.
It is probably the exception in this rule that has been confusing me:
If a BooleanQuery contains no MUST BooleanClauses, then a
doc
Hi,
I think it's your mm-parameter and that the terms are not matched in the
'setctor' field.
Cheers,
-Original message-
> From:Gastone Penzo
> Sent: Thu 24-Jan-2013 10:11
> To: solr-user@lucene.apache.org
> Subject: problem in qf parameter - no results
>
> Hi,
> i have a problem wi
Alright, so my suggestion of overriding HttpShardHandler to route users to the
same replica instead of shuffling the replica URL's is doable? What about the
comment in HttpShardHandler then?
//
// Shuffle the list instead of use round-robin by default.
// This prevents accident
(start-off-topic): Alexandre, nice ideas. Last in the *) list is a bit far
stretched, but still good. I would still add one: how to have exact matches
and inexact matches in the same analyzed field. (end-off-topic)
On Wed, Jan 23, 2013 at 2:40 PM, Alexandre Rafalovitch
wrote:
> We need a "Make yo
52 matches
Mail list logo