Have you made a JIRA ticket for this? This is useful generally, isn't
it?
Thx, Upayavira
On Thu, Mar 21, 2013, at 03:18 AM, Tirthankar Chatterjee wrote:
We have done something similar.
Please read
http://lucene.472066.n3.nabble.com/How-to-modify-Solr-StatsComponent-to-support-stats-query
the document's ACLs.
Would that work?
Upayavira
On Thu, Mar 21, 2013, at 08:52 AM, Christopher ARZUR wrote:
@Jan Høydahl : do you mean cutom filter ?
@Walter Underwood : I also agree with you, I'd only use native functions
of Solr, but I do not know how to solve my problem ... My ACLs
to view.
Upayavira
On Thu, Mar 21, 2013, at 10:55 AM, Christopher ARZUR wrote:
Users also have rights at the individual level (in addition to
inheritance of their group), your solution implies that I attach
potentially 1million identifier in the document? I do not know Solr
limitations
as was said below, add facet.method=fcs to your query URL.
Upayavira
On Thu, Mar 21, 2013, at 09:41 AM, Andy wrote:
What do I need to do to use this new per segment faceting method?
From: Mark Miller markrmil...@gmail.com
To: solr-user@lucene.apache.org
for the master. I knew I had clones
of the index on the slaves which gave me some immunity from disaster,
and downtime for indexing is much less catastrophic than downtime for
searching, so I probably have enough time to manually rebuild my master
using indexes from a slave.
Upayavira
You need to create the core directory on disk, containing a conf
directory, yourself, before you use this API.
If you are using SolrCloud, then I believe this isn't needed because the
config is in Zookeeper.
Upayavira
On Tue, Mar 19, 2013, at 06:01 AM, Ravi_Mandala wrote:
Hi,
I am trying
What field are you doing your hit highlighting on? You need to look at
the configuration for the highlighting component in solrconfig.xml. Also
note that you can only highlight on *stored* fields. The 'text' field is
by default not stored, so you'd need to change that and re-index.
Upayavira
, prepare an additional field in your update processor chain
that contains everything you want to highlight upon. Note, update
processors are called *before* fields are stored, so as far as the index
is concerned, your new field is just one single non-multivalued field.
Upayavira
On Tue, Mar 19, 2013
Not to my knowledge. I guess the nearest might be regular expressions
but that would involve one character, rather than one bit per element,
so not nearly as efficient.
How many bits? Can you break them down into separate fields?
Upayavira
On Tue, Mar 19, 2013, at 02:30 PM, Christopher ARZUR
requirements will be substantial, unless, I guess, your
fields are sparse. Also, during a faceting request, the server must do a
scan across each of those fields, and that will take time, and with tat
many fields, I'd imagine quite a bit of time.
Upayavira
On Mon, Mar 18, 2013, at 07:34 AM, sivaprasad
.
Upayavira
On Fri, Mar 15, 2013, at 09:51 AM, kobe.free.wo...@gmail.com wrote:
Hello All,
Scenario:
We trying to define the schema structure for our application search
feature,
based on SOLR search server. In our scenario the total number of fields
is
450 (quiet huge) and we
Use text analysis and copyField to create a new field that has terms as
you expect them. Then use that for your spellcheck dictionary.
Note, since 4.0, you don't need to create a dictionary. Solr can use
your index directly.
Upayavira
On Wed, Mar 13, 2013, at 06:00 AM, Rohan Thakur wrote
Auto commit would seem a good idea, as you don't want your independent
worker threads issuing overlapping commits. There's also commtWithin
that achieves the same thing.
Upayavira
On Wed, Mar 13, 2013, at 08:02 AM, Arkadi Colson wrote:
Hi
I'm filling our solr database with about 5mil docs
in the score to influence it in directions you want.
Upayavira
On Wed, Mar 13, 2013, at 04:49 AM, David Philip wrote:
Hi,
Any reply on this: How are the documents sequenced in the case when the
product of tf idf , coord and fieldnorm is same for both the documents?
Thanks - David
P.S
be able to achieve what you want with a PatternTokenizer rather
than whitespace and removing the WordDelimiterFilterFactory.
Upayavira
On Wed, Mar 13, 2013, at 08:41 AM, adfel70 wrote:
I want the following behaivour.
if john@gmail.com is indexed to the field
1. searching 'john' or 'doe
If you want to be ble to use the data in both places, that's what you
will need. You won't be ble to have Solr read indexes from within hbase,
it needs to manage its own indexes.
Upayavira
On Wed, Mar 13, 2013, at 09:03 AM, adfel70 wrote:
So you end up having all the data both in hbase
will use up a lot of your memory).
If you are, then the commit frequency becomes less important. You could
use soft commits between 1s and 15s, and hard commits maybe every 15s to
1min. Those seem to me to be reasonable values.
Upayavira
On Wed, Mar 13, 2013, at 09:19 AM, Arkadi Colson wrote
As has been said, you can use XSLT with wt=xslttr=stylesheet.xsl. You
don't need to use Saxon, unless you need specific (e.g. XSLT 2.0)
features.
You don't say what exts and last actually mean, so it isn't possible to
say whether this can be achieved with XSLT.
Upayavira
On Tue, Mar 12, 2013
You need to refer to your chain in a RequestHandler config. Search for
/update, duplicate that, and change the chain it points to.
Upayavira
On Mon, Mar 11, 2013, at 05:22 AM, Jack Park wrote:
With 4.1, not in cloud configuration, I have a custom response handler
chain which injects
is a
relatively heavy operation.
Upayavira
On Fri, Mar 8, 2013, at 10:41 PM, Mingfeng Yang wrote:
Then what's the difference between adding a new document vs.
replacing/overwriting a document?
Ming-
On Fri, Mar 8, 2013 at 2:07 PM, Upayavira u...@odoko.co.uk wrote:
With an atomic update, you
In example/cloud-scripts/ you will find a Solr specific zkCli tool to
upload/download configs.
You will need to reload a core/collection for the changes to take
effect.
Upayavira
On Fri, Mar 8, 2013, at 07:02 AM, Nathan Findley wrote:
I am setting up solrcloud with zookeeper.
- I am
=jeans
That would cause the term 'jeans' to be scored (the more common the term
in a document, the higher it scores, etc).
But by the sounds of it, it would be useful for you to understand better
how scoring calculations are done, so you can see *why* a score would be
the way it is.
Upayavira
Because?
Upayavira
On Fri, Mar 8, 2013, at 09:27 AM, Jilal Oussama wrote:
I would not recommend Windows too
2013/3/8 Kobe J kobe.free.wo...@gmail.com
We are planning to use SOLR 4.1 for full text indexing. Following is the
hardware configuration of the web server that we plan
as
well? I would expect your Solr instance to want sole use of the server,
as an index of your size will demand it.
Upayavira
On Fri, Mar 8, 2013, at 10:02 AM, kobe.free.wo...@gmail.com wrote:
Thanks for your suggestion Gora.
Yes, we are planning to use faceting, sorting features. The number
Without java coding, you cannot filter on things that aren't in your
index. You would need to re-index the document, but maybe you could make
use of atomic updates to just change the hidden field without needing to
push the whole document again.
Upayavira
On Fri, Mar 8, 2013, at 11:40 AM
is matched
by the time saved by not having to transmit those fields over the wire,
although I'd be very curious to see someone actually test that.
Upayavira
On Fri, Mar 8, 2013, at 09:51 PM, Mingfeng Yang wrote:
Generally speaking, which has better performance for Solr?
1. updating some fields
Romita,
That isn't a Solaritas feature, that is a feature of any RequestHandler.
You can copy a request handler in solrconfigxml, change its name and set
parameters as defaults/incariants, and then use that new URL for your
queries.
Upayavira
On Thu, Mar 7, 2013, at 02:35 AM, Romita Saha wrote
it is in. This will spoil the accuracy of
your search results.
The general recommendation when 'sharding' your index is to have your
documents evenly split across shards, but that may not be an option for
you.
Upayavira
On Thu, Mar 7, 2013, at 08:46 AM, Shinichiro Abe wrote:
Hi,
Does the distributed search
Are you thinking of spellchecking? Where are you seeing suggestions?
If you are thinking of spellchecking, by default the spellchecker uses
the 'name' field, and you have likely indexed into the 'text' field,
hence no results being returned.
Upayavira
On Thu, Mar 7, 2013, at 01:12 PM, alecx
before it is analysed.
Hope that helps.
Upayavira
On Thu, Mar 7, 2013, at 02:08 PM, Hannah Ullrich wrote:
Hi Oliver.
thank for the answer.
We tried pattern=[\s]+ but it dont work.
I can replace anything but not the whitespace...
Here our schema:
fieldtype name=sigField class
from Javascript).
Upayavira
On Wed, Mar 6, 2013, at 03:42 PM, Timothy Potter wrote:
This sounds like a job for Zookeeper (distributed coordination is what it
does).
Take a look at:
http://zookeeper-user.578899.n2.nabble.com/Sequence-Number-Generation-With-Zookeeper-td5378618.html
On Wed, Mar
What are you actually trying to achieve? If you can share what you are
trying to achieve maybe folks can help you find the right way to do it.
Upayavira
On Wed, Mar 6, 2013, at 02:54 PM, Joseph Lim wrote:
Hello Otis ,
Is there any configuration where it will index into hdfs instead?
I
You can use more like this as a component, but you don't get info about
what terms made the documents similar.
If you don't need that stuff, then just have MLT and clustering as
components within a standard SearchHandler.
Upayavira
On Tue, Mar 5, 2013, at 11:53 AM, Dale McDiarmid wrote:
code
solrconfig - just beware circular
references.
Upayavira
On Mon, Mar 4, 2013, at 11:21 AM, Martin Koch wrote:
Is it possible to run solr without zookeeper, but still using sharding,
if
it's all running on one host? Would the shards have to be explicitly
included in the query urls?
Thanks,
/Martin
together.
You could use the page number in a page_* dynamic field, but then you'd
have to query against page_1, page_2, page_3...page_n for every query,
which wouldn't work too well.
Upayavira
On Sat, Mar 2, 2013, at 03:59 PM, Anirudha Jadhav wrote:
if you increase the granularity of your document
What I'm questioning is whether the issue you see in 4.1 has been
resolved in Subversion. While I would not expect 4.0 to read a 4.1
index, the SVN branch/4.2 should be able to do so effortlessly.
Upayavira
On Sat, Mar 2, 2013, at 06:17 PM, Dotan Cohen wrote:
On Fri, Mar 1, 2013 at 1:37 PM
Can you use a checkout from SVN? Does that resolve your issues? That is
what will become 4.2 when it is released soon:
https://svn.apache.org/repos/asf/lucene/dev/branches/branch_4x/
Upayavira
On Fri, Mar 1, 2013, at 10:51 AM, Dotan Cohen wrote:
On Fri, Mar 1, 2013 at 12:22 PM, Rafał Kuć r
Apache on port 80, with a simple
reverse proxy:
VirtualHost *:80
ProxyPass / http://localhost:1234/
ProxyPassReverse / http://localhost:1234/
/VirtualHost
That way your Solr would be visible on ports 1234 and on port 80.
Upayavira
On Fri, Mar 1, 2013, at 11:43 AM, Bruno Mannina wrote:
Dear
functioning zookeeper nodes, which is sufficient for my Solr servers
to continue operating. (Zookeeper requires half+1 of your servers to be
up to function properly, thus an odd number of servers).
Upayavira
On Fri, Mar 1, 2013, at 11:41 AM, AlexS wrote:
Hello, I'm new to Solr, and I need to create
What do you mean by 'will not show up'? Is numdocs wrong? They don't
show in queries?
Upayavira
On Thu, Feb 28, 2013, at 06:07 PM, dboychuck wrote:
Yes I confirmed in the logs. I have also committed manually several times
using the updatehandler /update?commit=true
--
View this message
of a shell is an interesting one. But is there stuff we
couldn't achieve without creative use of 'curl'?
Upayavira
On Tue, Feb 26, 2013, at 04:34 AM, Otis Gospodnetic wrote:
Upayavira, ever did this?
Ha, look at my email from 20 days ago and this:
https://github.com/javanna/elasticshell
of Solr though.
Upayavira
On Thu, Feb 21, 2013, at 06:09 PM, zqzuk wrote:
Hi
I have built a 300GB index using lucene 4.1 and now it is too big to do
queries efficiently. I wonder if it is possible to split it into shards,
then use SolrCloud configuration?
I have looked around the forum
is no way to do what you are looking for within
Solr. I'd suggest you'll need to do some parsing at your side and, if
you find quotes, do the query against a different field.
Upayavira
On Thu, Feb 21, 2013, at 06:17 PM, Van Tassell, Kristian wrote:
I'm trying to match the word created. Given
Which of your three hosts did you point this request at?
Upayavira
On Thu, Feb 21, 2013, at 09:13 PM, Vaillancourt, Tim wrote:
Correction, I used this curl:
curl -v
'http://HOST:8983/solr/admin/collections?action=CREATEname=testnumShards=3replicationFactor=2maxShardsPerNode=2'
So 3
I think what Walter means is make the thing that sends it to Solr set
the timestamp when it does so.
Upayavira
On Sat, Feb 16, 2013, at 08:56 PM, Isaac Hebsh wrote:
Hi,
I do have an externally-created timestamp, but some minutes may pass
before
it will be sent to Solr.
On Sat, Feb 16
You can use a KeywordTokenizerFactory, which will tokenise into a single
term, and then do your lowercasing. Does that get you what you want?
Upayavira
On Thu, Feb 14, 2013, at 05:11 PM, Bing Hua wrote:
Hi,
I don't want the field to be tokenized because Solr doesn't support
sorting
to work out which is more appropriate.
If you consider using Lucene in the latter scenario, you will probably
end up rewriting a lot of what Solr does anyway.
Upayavira
On Tue, Feb 12, 2013, at 08:26 PM, Glen Newton wrote:
Is there a page on the wiki that points out the use cases
This config isn't intended for embedded zookeeper, it is for a separate
zookeeper ensemble that is shared with other services.
Upayavira
On Tue, Feb 12, 2013, at 10:19 PM, mbennett wrote:
The suggested syntax didn't work with embedded ZooKeeper:
Syntax:
-DzkRun -DzkHost=nodeA:9983,nodeB
, a document matching both terms will score more
highly than a document only matching one of them.
So, a part of the question is whether you wish to *only* show documents
that include both terms, or whether you are happy for good matches to be
prioritised.
Upayavira
On Sun, Feb 10, 2013, at 05:27 PM
Do it on the client, strip off the date and add together ny values you
have for the same hour.
Upayavira
On Sat, Feb 9, 2013, at 06:15 AM, Cool Techi wrote:
Hi,
I want to facet results on an hourly basis, the following query gives me
an hourly breakdown, but with the date part. I want just
Also, OR is the default, so you can improve on it with:
fq=institute_name:(xyz sfsda sdfsaf)
Upayavira
On Wed, Feb 6, 2013, at 08:17 PM, Shawn Heisey wrote:
On 2/6/2013 12:41 PM, anurag.jain wrote:
in my query there are many OR's now after 79 or 80 ORS it gives error that
url is very large
Right, but what's the Windows equivalent? Not sure there is one.
Upayavira
On Tue, Feb 5, 2013, at 04:56 AM, Walter Underwood wrote:
Easier than:
solrpost.sh a*.xml a.log
solrpost.sh b*.xml b.log
solrpost.sh c*.xml c.log
and so on?
We have a fair selection of Solr servers where
Try the same with plain HTTP, queries, and use wt=xml and wt=json, to
see whether the space survives there (remember to view source). If there
is a bug, it could be localised to the solrj case.
Upayavira
On Tue, Feb 5, 2013, at 10:44 AM, Marc Hermann wrote:
Hi everyone,
i am new to SOLR
This would involve you querying against every page in your document,
which will be too many fields and will break quickly.
The best way to do it is to index pages as documents. You can use field
collapsing to group pages from the same document together.
Upayavira
On Tue, Feb 5, 2013, at 02:00
By dependencies, do you mean other java classes? I was thinking of
splitting it out into a few classes, each of which is clearer in its
purpose.
Upayavira
On Tue, Feb 5, 2013, at 02:26 PM, Jan Høydahl wrote:
Wiki page exists already: http://wiki.apache.org/solr/post.jar
I'm happy to consider
logic here is that Solr/Lucene can read the indexes
of the previous major version. Given you are two major versions behind,
you'd have to do it in two steps.
Upayavira
On Mon, Feb 4, 2013, at 03:18 PM, Shawn Heisey wrote:
On 2/4/2013 7:20 AM, Artem OXSEED wrote:
I need to upgrade our Solr
at that refactoring, especially if you're prepared to
review it.
I guess the other thing that is much needed is a wiki page that details
the features of the tool, and also explains that its role is
educational, rather than anything else.
Upayavira
On Mon, Feb 4, 2013, at 09:10 PM, Jan Høydahl wrote:
Hi
multithreading aware scripting language that is available on
all platforms by default, and such are in short supply! Whether or not
the Solr community is interested in my changes is another matter.
Upayavira
On Tue, Feb 5, 2013, at 04:43 AM, Walter Underwood wrote:
Have you considered writing a script
wanted to see whether the feature is likely to be
accepted before I put in the effort. Also, I would need to consider
which parts of the tool to add that to. Currently I only want it for
posting XML docs, but there's also crawling capabilities in it too.
Thoughts?
Upayavira
I haven't tried DIH, although if it does support multithreading, I might
be inclined to.
Upayavira
On Sun, Feb 3, 2013, at 05:17 PM, Alexandre Rafalovitch wrote:
What times do you get with DIH? It has native support for that format
too.
On 3 Feb 2013 11:20, Upayavira u...@odoko.co.uk wrote
queries such as 'find me all users
who have a post which mentions foo', or 'find me all posts by all users
who joined last year'.
Upayavira
On Sun, Feb 3, 2013, at 03:14 PM, zagros wrote:
Hi,
I'm new to Solr and have two questions.
#1)
I was wondering if I wanted to index different object types
on going.
Obviously I'm talking about slaves here. There will be a master for each
shard which each of these nodes pull their indexes from.
Hope this is helpful.
Upayavira
On Tue, Jan 29, 2013, at 09:35 PM, Lee, Peter wrote:
I would appreciate people's experience on the following load balancing
Stored fields are now compressed in 4.1. There's other efficiencies too
in 4.0 that will also result in smaller indexes, but the compressed
stored fields is the most significant.
Upayavira
On Wed, Jan 30, 2013, at 01:59 PM, anarchos78 wrote:
Hello,
I am using Solr 3.6.1 and I am very
the results. This does not require large caches or any such, so
I do not see that you're going to have resource advantages to limiting
them to specific nodes.
Upayavira
On Wed, Jan 30, 2013, at 01:45 PM, Lee, Peter wrote:
Upayavira,
Thank you for your response. I'm sorry my post is perhaps not clear...I
,
and this will include term vector files.
Upayavira
On Tue, Jan 29, 2013, at 06:21 AM, Artyom wrote:
I guess, response time increased, because I use master-slave
configuration in
Solr 4.0 and Solr 4.1: if there are no termVectors, the full index is
replicated; if there are termVectors, only
You could configure your servlet container (jetty/tomcat) to have
specific webapps/contexts listen on specific IP/port combinations, that
would get you some way, But what you are asking is more about networking
and servlet container configuration than about Solr.
Upayavira
On Wed, Jan 23, 2013
immediately after the snippet you showed.
Upayavira
On Wed, Jan 23, 2013, at 07:40 AM, Marcin Rzewucki wrote:
OK, so I did yet another test. I stopped solr, removed whole data/
dir and started Solr again. Directories were recreated fine, but
missing files were not downloaded from
which
nodes are accessible to it - but it will certainly not be doing updates
(they should be buffered until the other DC returns).
If you want true geographical redundancy, I think Markus' suggestion is
a sensible one.
Upayavira
On Tue, Jan 22, 2013, at 10:11 PM, Markus Jelsma wrote:
Hi
Hmm, don't see it. Not sure if attachments make it to this list.
Perhaps put it in a pastebin and include a link if too long to include
in an email?
Upayavira
On Wed, Jan 23, 2013, at 10:28 AM, Marcin Rzewucki wrote:
Hi,
Previously, I took the lines related to collection I tested. Maybe
?
Upayavira
On Wed, Jan 23, 2013, at 07:24 AM, Gora Mohanty wrote:
On 23 January 2013 01:26, hassancrowdc hassancrowdc...@gmail.com wrote:
sorry if it is a stupid question but where can i find result.xml and where
do i write this program? any hints?
[...]
The result XML referred to is the XML
Jan 23, 2013 7:16:08 AM org.apache.solr.core.CachingDirectoryFactory get
INFO: return new directory for
/solr/cores/bpr/selekta/data/index.20130121090342477 forceNew:false
Once you look in that dir, how do things look?
Upayavira
On Wed, Jan 23, 2013, at 10:45 AM, Marcin Rzewucki wrote:
OK, check
You are going to have to give more information than this. If you get bad
request, look in the logs for the Solr server and you will probably find
an exception there that tells you what was wrong with your document.
Upayavira
On Wed, Jan 23, 2013, at 08:58 AM, Thendral Thiruvengadam wrote:
Hi
Mark,
Take a peek in the pastebin url Marcin mentioned earlier
(http://pastebin.com/qMC9kDvt) is there enough info there?
Upayavira
On Wed, Jan 23, 2013, at 02:04 PM, Mark Miller wrote:
Was your full logged stripped? You are right, we need more. Yes, the peer
sync failed, but then you cut out
of solr, just a call to an HTTP URL.
Upayavira
On Wed, Jan 23, 2013, at 07:51 PM, Alan Rykhus wrote:
Hello,
I do nightly builds for one of my sites. I build the new index in a
parallel directory. When it is finished I move the old files to a backup
directory(I only save one, delete the previous
If you can handle it in XML, use wt=xmltr=foo.xsl and use a stylesheet
to format it as you want.
Upayavira
On Wed, Jan 23, 2013, at 08:53 PM, Rafał Kuć wrote:
Hello!
As far as I know you can't remove the response, numFound, start and
docs. This is how the response is prepared by Solr
range-hundreds field and your category
field.
Hmm, now you've got me thinking...
Upayavira, who is mentally rewriting one of his apps now
On Tue, Jan 22, 2013, at 12:36 PM, Bram Van Dam wrote:
Hi folks,
Is it possible to pivot by ranges in a single query?
Something like this:
select?q
I'm not clear what you are referring to. Are you wanting to block the
startup of multiple webapps in your container, or to stop some solr
cores (indexes) loading up?
Upayavira
On Tue, Jan 22, 2013, at 12:40 PM, Gustav wrote:
Hey guys, i need help with some really tricky request
We merely used Alfresco as the other side of the CMIS coin, to prove
that our connector was working, as colleagues had knowledge of it.
And yes, that link you found is to the connector.
Upayavira
On Sun, Jan 20, 2013, at 10:26 PM, Nicholas Li wrote:
I think this might be the one you
And think about distributed search, you are going through a 'proxy'
which, as well as forwarding your docs, must also merge any docs from
different shards into a single result set. That is likely to take some
time on 30,000 docs, and isn't a job that is needed on non-distributed
search.
Upayavira
make sense for you to have more control, by using
your own code to do the indexing.
Upayavira
On Mon, Jan 21, 2013, at 11:36 AM, ashimbose wrote:
Hi Gora,
Thank you for your suggestion.
I have tried with you below option,
* Have never tried this, but one can set up multiple request
.
I'm not sure whether that would be more, or less performant than just
having six numeric fields, though.
Upayavira
On Sat, Jan 19, 2013, at 12:44 PM, blopez wrote:
I'll always query on the set of 6 values, but in some cases, the matching
doesn't need to be exact.
I mean, an usual query
returned
to your user.
HTH
Upayavira
On Mon, Jan 7, 2013, at 07:27 AM, wwhite1133 wrote:
Hi ,
I wanted to sort the results of the solr search query on the number of
highlights generated per document.
e.g
Doc 1
highlights {
fieldA
FieldB
}
Doc 2
Highlights{
field A
fieldC
fieldC
A colleague of mine when I was working for Sourcesense made a CMIS
plugin for Solr. It was one way, and we used it to index stuff out of
Alfresco into Solr. I can't search for it now, let me know if you can't
find it.
Upayavira
On Fri, Jan 18, 2013, at 05:35 AM, Nicholas Li wrote:
I want
one the
commit is done?
Maybe answers to these might help unpick your issue.
Upayavira
On Thu, Jan 17, 2013, at 06:22 AM, Cool Techi wrote:
Hi,
We have an index of approximately 400GB in size, indexing 5000 documents
was taking 20 seconds. But lately, the indexing is taking very long
And, it would make for slow queries, as the more fields you query, the
worse performance gets.
Having said that, you can query multiple fields using the edismax query
parser, with it qf param.
Upayavira
On Wed, Jan 16, 2013, at 12:23 AM, Jack Krupansky wrote:
Semi-hard-coded
This involves taking a subclass of the DefaultSimilarity class, in Java,
and adding that to your Solr setup. For someone versed in Java, this is
relatively straight-forward. For others it is non-trivial.
Upayavira
On Wed, Jan 16, 2013, at 10:57 AM, Amit Jha wrote:
Hi,
How can I do
or the IDF
via a configuration parameter, and post a patch to JIRA. I'm sure there
are other folks that would want the feature, and would hope it would be
accepted easily.
E.g. disableIDF=true or disableTF=true would make those functions just
return 1.
My thoughts anyhow.
Upayavira
On Wed, Jan 16
You should put your solr.xml into your 'cores' directory, and set
-Dsolr.solr.home=cores
That should get you going. 'cores' *is* your Solr Home. Otherwise, your
instanceDir entries in your current solr.xml will need correct paths to
../cores/procedure/ etc.
Upayavira
On Tue, Jan 15, 2013, at 08
Is your group field multivalued? Could docs appear in more than one
group?
Upayavira
On Tue, Jan 15, 2013, at 01:22 PM, J Mohamed Zahoor wrote:
The sum of all the count in the groups… does not match the total no of
docs found.
./zahoor
On 12-Jan-2013, at 1:27 PM, Upayavira u
across all for of these boxes?
Is that correct?
Upayavira
On Tue, Jan 15, 2013, at 02:11 PM, Jean-Sebastien Vachon wrote:
Hi Erick,
Thanks for your comments but I am migrating an existing index (single
instance) to a sharded setup and currently I have no access to the code
involved
.
Upayavira
On Tue, Jan 15, 2013, at 02:07 PM, elisabeth benoit wrote:
Hello,
I have a Solr instance (solr 3.6.1) with around 3 000 000 documents. I
want
to read (in a java test application) all my documents, but not in one
shot
(because it takes too much memory).
So I send the same request, over
='and'.
Upayavira
On Tue, Jan 15, 2013, at 05:02 PM, Nicholas Ding wrote:
Hello,
I'm thinking store hierachical data structure on Solr. I know I have to
flatten the structure in a form like A_B_C, but it is possible to extend
Solr to support hierachical data?
What about I store JSON text
It can work, so I believe. However, it is not normal Solr usage, so you
are less likely to find people who can support you in it.
Upayavira
On Sun, Jan 13, 2013, at 03:59 PM, suri wrote:
Sorry, might have shared more info. Planning to have Index files in NAS
and
share these index files across
This is present in 4.0. Not sure if there re ny improvements in 4.1.
Upayavira
On Sun, Jan 13, 2013, at 07:35 PM, Uwe Clement wrote:
Thanks Yonik.
Is this already working well on solr 4.0? or better to wait until solr
4.1?!
-Ursprüngliche Nachricht-
Von: ysee...@gmail.com
could you use field collapsing? Boost by date and only show one value
per group, and you'll have the most recent document only.
Upayavira
On Fri, Jan 11, 2013, at 01:10 PM, jmozah wrote:
one crude way is first query and pick the latest date from the result
then issue a query with q=timestamp
Not sure exactly what you mean, can you give an example?
Upayavira
On Sat, Jan 12, 2013, at 06:32 AM, J Mohamed Zahoor wrote:
Cool… it worked… But the count of all the groups and the count inside
stats component does not match…
Is that a bug?
./zahoor
On 11-Jan-2013, at 6:48 PM
)
Upayavira
On Thu, Jan 10, 2013, at 02:18 PM, Mark Miller wrote:
Setup hard auto commit with openSeacher=false. I would do it at least
once a minute. Don't worry about the commit being out of sync on the
different nodes - you will be using soft commits for visibility. The hard
commits will just
?
Thanks! Upayavira
On Thu, Jan 10, 2013, at 04:18 PM, Mark Miller wrote:
There is no need to open a Searcher because you are controlling
visibility through the faster 'soft' commit. That will reopen the reader
from the IndexWriter. Because of that, there is no reason to do a heavy,
non NRT
Why do you want to unload one shard of a collection? Doing so would
render your collection incomplete and therefore non-functional. It'd
help to understand a bit more what you're trying to achieve.
Upayavira
On Thu, Jan 10, 2013, at 10:17 AM, mizayah wrote:
Lets say i got one collection with 3
Is the registration of the search component failing earlier in your
logs?
Upayavira
On Thu, Jan 10, 2013, at 04:23 PM, obi240 wrote:
I recently started working with the clustering plugin on solr 4.
I assigned a requsthandler to clustering: /clustering and got the
following
errors
lazy
601 - 700 of 854 matches
Mail list logo