I have an application where I am calling DirectUpdateHandler2 directly with:
update.addDoc(cmd);
This will sometimes hit:
java.lang.OutOfMemoryError: Java heap space
at org.apache.lucene.util.UnicodeUtil.UTF16toUTF8(UnicodeUtil.java:248)
at org.apache.lucene.store.DataOutput.writeString(DataOu
I'm looking for a way to quickly flag/unflag documents.
This could be one at a time or by query (even *:*)
I have hacked together something based on ExternalFileField that is
essentially a FST holding all the ids (solr not lucene). Like the
FieldCache, it holds a WeakHashMap where the
OpenBitSet
thanks!
On Fri, Oct 26, 2012 at 4:20 PM, Chris Hostetter
wrote:
> : How about a boost function, "bf" or "boost"?
> :
> : bf=if(exists(query(location:A)),5,if(exists(query(location:B)),3,0))
>
> Right ... assuming you only want to ignore tf/idf on these fields in this
> specifc context, function
Hi-
I am trying to add a setting that will boost results based on
existence in different buckets. Using edismax, I added the bq
parameter:
location:A^5 location:B^3
I want this to put everything in location A above everything in
location B. This mostly works, BUT depending on the number of mat
If you optimize the index, are the results the same?
maybe it is showing counts for deleted docs (i think it does... and
this is expected)
ryan
On Sat, Aug 25, 2012 at 9:57 AM, Fuad Efendi wrote:
>
> This is bug in Solr 4.0.0-Beta Schema Browser: "Load Term Info" shows "9682
> News", but direc
for the ExtractingRequestHandler, you can put anything into the
request contentType.
try:
addFile( file, "application/octet-stream" )
but anything should work
ryan
On Thu, Jun 7, 2012 at 2:32 PM, Koorosh Vakhshoori
wrote:
> In latest 4.0 release, the addFile() method has a new argument 'con
the right zookeeper url in 4.0 please?
>
> Thanks
> Naga
>
>
> On 5/15/12 10:56 AM, "Ryan McKinley" wrote:
>
>>In 4.0, solr no longer uses JSP, so it is not enabled in the example
>>setup.
>>
>>You can enable JSP in your servlet container u
In 4.0, solr no longer uses JSP, so it is not enabled in the example setup.
You can enable JSP in your servlet container using whatever method
they provide. For Jetty, using start.jar, you need to add the command
line: java -jar start.jar -OPTIONS=jsp
ryan
On Mon, May 14, 2012 at 2:34 PM, Nag
thanks!
On Wed, May 2, 2012 at 4:43 PM, Chris Hostetter
wrote:
>
> : How do I search for things that have no value or a specified value?
>
> Things with no value...
> (*:* -fieldName:[* TO *])
> Things with a specific value...
> fieldName:A
> Things with no value or a specific val
If your json value is & the proper xml value is &
What is the value you are setting on the stored field? is is & or &?
On Mon, Apr 30, 2012 at 12:57 PM, William Bell wrote:
> One idea was to wrap the field with CDATA. Or base64 encode it.
>
>
>
> On Fri, Apr 27, 2012 at 7:50 PM, Bill Bell
check a release since r1332752
If things still look problematic, post a comment on:
https://issues.apache.org/jira/browse/SOLR-3426
this should now have a less verbose message with an older SLF4j and with Log4j
On Tue, May 1, 2012 at 10:14 AM, Gopal Patwa wrote:
> I have similar issue using lo
I would suggest debugging with browser requests -- then switching to
Solrj after you are at 1st base.
In particular, try adding the &debugQuery=true parameter to the
request and see what solr thinks is happening.
The value that will "work" for the 'qt' parameter depends on what is
configured in s
In general -- i would not suggest mixing EmbeddedSolrServer with a
different style (unless the other instances are read only). If you
have multiple instances writing to the same files on disk you are
asking for problems.
Have you tried just using StreamingUpdateSolrServer for daily update?
I woul
zookeeper.jsp was removed (along with all JSP stuff) in trunk
Take a look at the cloud tab in the UI, or check the /zookeeper
servlet for the JSON raw output
ryan
On Mon, Apr 9, 2012 at 6:42 AM, Benson Margulies wrote:
> Starting the leader with:
>
> java -Dbootstrap_confdir=./solr/conf -Dcol
There have been a bunch of changes getting the zookeeper info and UI
looking good. The info moved from being on the core to using a
servlet at the root level.
Note, it is not a request handler anymore, so the wt=XXX has no
effect. It is always JSON
ryan
On Fri, Apr 6, 2012 at 7:01 AM, Jamie J
On Wed, Mar 7, 2012 at 7:25 AM, Matt Mitchell wrote:
> Hi,
>
> I'm researching options for handling a better geospatial solution. I'm
> currently using Solr 3.5 for a read-only "database", and the
> point/radius searches work great. But I'd like to start doing point in
> polygon searches as well.
Hi Matthias-
I'm trying to understand how you have your data indexed so we can give
reasonable direction.
What field type are you using for your locations? Is it using the
solr spatial field types? What do you see when you look at the debug
information from &debugQuery=true?
>From my experienc
I have an application where I need to return all results that are not
in a Set (the Set is managed from hazelcast... but that is
not relevant)
As a fist approach, i have a SerachComponent that injects a BooleanQuery:
BooleanQuery bq = new BooleanQuery(true);
for( String id : ids) {
patches are always welcome!
On Tue, Jul 5, 2011 at 3:04 PM, Yonik Seeley wrote:
> On Mon, Jul 4, 2011 at 11:54 AM, Per Newgro wrote:
>> i've tried to add the params for group=true and group.field=myfield by using
>> the SolrQuery.
>> But the result is null. Do i have to configure something? In
>
> Ah, thanks Hoss - I had meant to respond to the original email, but
> then I lost track of it.
>
> Via pseudo-fields, we actually already have the ability to retrieve
> values via FieldCache.
> fl=id:{!func}id
>
> But using CSF would probably be better here - no memory overhead for
> the FieldC
On Fri, Jul 1, 2011 at 9:06 AM, Yonik Seeley wrote:
> On Thu, Jun 30, 2011 at 6:19 PM, Ryan McKinley wrote:
>> Hello-
>>
>> I'm looking for a way to find all the links from a set of results. Consider:
>>
>>
>> id:1
>> type:X
>> lin
Hello-
I'm looking for a way to find all the links from a set of results. Consider:
id:1
type:X
link:a
link:b
id:2
type:X
link:a
link:c
id:3
type:Y
link:a
Is there a way to search for all the links from stuff of type X -- in
this case (a,b,c)
If I'm understanding the {!join
You can store binary data using a binary field type -- then you need
to send the data base64 encoded.
I would strongly recommend against storing large binary files in solr
-- unless you really don't care about performance -- the file system
is a good option that springs to mind.
ryan
2011/4/6
> Does anyone know of a patch or even when this functionality might be included
> in to Solr4.0? I need to query for polygons ;-)
check:
http://code.google.com/p/lucene-spatial-playground/
This is my sketch / soon-to-be-proposal for what I think lucene
spatial should look like. It includes a WK
You may have noticed the ResponseWriter code is pretty hairy! Things
are package protected so that the API can change between minor release
without concern for back compatibility.
In 4.0 (/trunk) I hope to rework the whole ResponseWriter framework so
that it is more clean and hopefully stable eno
Not crazy -- but be aware of a few *key* caviates.
1. Do good testing on a stable snapshot.
2. Don't get surprised if you have to rebuild the index from scratch
to upgrade in the future. The official releases will upgrade smoothly
-- but within dev builds, anything may happen.
On Sat, Feb 19,
t_By_Function
On Fri, Feb 11, 2011 at 4:31 PM, Ryan McKinley wrote:
> I have an odd need, and want to make sure I am not reinventing a wheel...
>
> Similar to the QueryElevationComponent, I need to be able to move
> documents to the top of a list that match a given query.
>
>
You may want to check the stats via JMX. For example,
http://localhost:8983/solr/core/admin/mbeans?stats=true&key=org.apache.solr.handler.StandardRequestHandler
shows some basic stats info for the handler.
If you are running nagios or similar, they have tools that can log
values from JMX. this
I have an odd need, and want to make sure I am not reinventing a wheel...
Similar to the QueryElevationComponent, I need to be able to move
documents to the top of a list that match a given query.
If there were no sort, then this could be implemented easily with
BooleanQuery (i think) but with so
>
> foo_s:foo\-bar
> is a valid lucene query (with only a dash between the foo and the
> bar), and presumably it should be treated the same in edismax.
> Treating it as foo_s:foo\\-bar (a backslash and a dash between foo and
> bar) might cause more problems than it's worth?
>
I don't think we shou
ah -- that makes sense.
Yonik... looks like you were assigned to it last week -- should I take
a look, or do you already have something in the works?
On Thu, Feb 10, 2011 at 2:52 PM, Chris Hostetter
wrote:
>
> : extending edismax. Perhaps when F: does not match a given field, it
> : could auto
I am using the edismax query parser -- its awesome! works well for
standard dismax type queries, and allows explicit fields when
necessary.
I have hit a snag when people enter something that looks like a windows path:
F:\path\to\a\file
this gets parsed as:
F:\path\to\a\file
F:\path\to\a\file
+
>
> Where do you get your Lucene/Solr downloads from?
>
> [] ASF Mirrors (linked in our release announcements or via the Lucene website)
>
> [X] Maven repository (whether you use Maven, Ant+Ivy, Buildr, etc.)
>
> [X] I/we build them from source via an SVN/Git checkout.
>
also try &debugQuery=true and see why each result matched
On Thu, Dec 30, 2010 at 4:10 PM, mrw wrote:
>
>
> Basically, just what you've suggested. I did the field/query analysis piece
> with verbose output. Not entirely sure how to interpret the results, of
> course. Currently reading anythi
I have an indexing pipeline that occasionally needs to check if a
document is already in the index (even if not commited yet).
Any suggestions on how to do this without calling before each check?
I have a list of document ids and need to know which ones are in the
index (actually I need to know
Do you already have the files as solr XML? If so, I don't think you need solrj
If you need to build SolrInputDocuments from your existing structure,
solrj is a good choice. If you are indexing lots of stuff, check the
StreamingUpdateSolrServer:
http://lucene.apache.org/solr/api/solrj/org/apache/
On Mon, Oct 18, 2010 at 10:12 AM, Tharindu Mathew wrote:
> Thanks Peter. That helps a lot. It's weird that this not documented anywhere.
> :(
Feel free to edit the wiki :)
*:*
will leave you a fresh index
On Thu, Sep 23, 2010 at 12:50 AM, xu cheng wrote:
> the query that fetch the data you wanna
> delete
> I did like this to delete my data
> best regards
>
> 2010/9/23 Igor Chudov
>
>> Let's say that I added a number of elements to Solr (I use
>> Webservice::Solr
Multiple threads work well.
If you are using solrj, check the StreamingSolrServer for an
implementation that will keep X number of threads busy.
Your mileage will very, but in general I find a reasonable thread
count is ~ (number of cores)+1
On Wed, Sep 22, 2010 at 5:52 AM, Andy wrote:
> Does
check:
http://wiki.apache.org/solr/LukeRequestHandler
On Mon, Sep 13, 2010 at 7:00 PM, Peter A. Kirk wrote:
> Hi
>
> is it possible to issue a query to solr, to get a list which contains all the
> field names in the index?
>
> What about to get a list of the freqency of individual words in eac
> I suppose an index 'remaker' might be something like a DIH reader for
> a Solr index - streams everything out of the existing index, writing
> it into the new one?
This works fine if all fields are stored (and copy field does not go
to a stored field), otherwise you would need/want to start with
Check:
http://lucene.apache.org/java/3_0_2/fileformats.html
On Tue, Sep 7, 2010 at 3:16 AM, rajini maski wrote:
> All,
>
> While we post data to Solr... The data get stored in "//data/index" path
> in some multiple files with different file extensions...
> Not worrying about the extensions, I
Note that the 'setRequestWriter' is not part of the SolrServer API, it
is on the CommonsHttpSolrServer:
http://lucene.apache.org/solr/api/org/apache/solr/client/solrj/impl/CommonsHttpSolrServer.html#setRequestWriter%28org.apache.solr.client.solrj.request.RequestWriter%29
If you are using EmbeddedS
ed; Yonik any ideas? I'm not familiar with this part of
> Solr...
>
> Mike
>
> On Mon, Aug 23, 2010 at 2:38 AM, Ryan McKinley wrote:
>> I have a function that works well in 3.x, but when I tried to
>> re-implement in 4.x it runs very very slow (~20ms vs 45s on
I have a function that works well in 3.x, but when I tried to
re-implement in 4.x it runs very very slow (~20ms vs 45s on an index w
~100K items).
Big picture, I am trying to calculate a bounding box for items that
match the query. To calculate this, I have two fields bboxNS, and
bboxEW that get
Looks like you can sort by _docid_ to get things in index order or
reverse index order.
?sort=_docid_ asc
thank you solr!
On Fri, Jul 23, 2010 at 2:23 PM, Ryan McKinley wrote:
> Any pointers on how to sort by reverse index order?
> http://search.lucidimagination.com/search/do
Any pointers on how to sort by reverse index order?
http://search.lucidimagination.com/search/document/4a59ded3966271ca/sort_by_index_order_desc
it seems like it should be easy to do with the function query stuff,
but i'm not sure what to sort by (unless I add a new field for indexed
time)
Any p
If there is a real desire/need to make things "restful" in the
official sense, it is worth looking at using a REST framework as the
controller rather then the current solution. perhaps:
http://www.restlet.org/
https://jersey.dev.java.net/
These would be cool since they encapsulate lots of the re
Interesting -- I don't think there is anything that does this.
Though it seems like something the XML Query syntax should be able to
do, but we would still need to add the ability to send the xml style
query to solr.
On Fri, May 28, 2010 at 12:23 PM, Phillip Rhodes
wrote:
> Hi.
> I am building
The two approaches solve different needs. In 'multicore' you have a
single webapp with multiple indexes. This means they are all running
in the same JVM. This may be an advantage or a disadvantage depending
on what you are doing.
ryan
On Thu, May 27, 2010 at 10:44 AM, Antonello Mangone
wrot
You may also want to look at:
ClientUtils.escapeQueryChars( String s )
http://lucene.apache.org/solr/api/org/apache/solr/client/solrj/util/ClientUtils.html#escapeQueryChars%28java.lang.String%29
this will escape any lucene query chars, then pass it to URLEncoder
and you should be good to go.
On
accidentally hit send...
Eache core can have the dataDir set explicitly.
If you want to do this with solrj, you would need to manipulate the
CoreDescriptor objects.
ryan
On Sat, May 22, 2010 at 2:24 PM, Ryan McKinley wrote:
> Check:
> http://wiki.apache.or
Check:
http://wiki.apache.org/solr/CoreAdmin
Unless I'm missing something, I think you should be able to sort what you need
On Fri, May 21, 2010 at 7:55 PM, Ken Krugler
wrote:
> I've got a situation where my data directory (a) needs to live elsewhere
> besides inside of Solr home, (b) moves to
On Wed, May 19, 2010 at 6:38 AM, Peter Karich wrote:
> Hi all,
>
> while asking a question on stackoverflow [1] some other questions appear:
> Is SolrJ a recommended way to access Solr or should I prefer the HTTP
> interface?
solrj vs HTTP interface? That will just be a matter of taste. If you
>
> Any other commonly compelling reasons to use SolrJ?
The most compelling reason (I think) is that if you program against
the Solrj API, you can switch between embedded/http/streaming
implementations without changing anything.
This is great for our app that is either run as a small local instan
On Fri, Apr 2, 2010 at 7:07 AM, Na_D wrote:
>
> hi,
>
>
> I need to monitor the index for the following information:
>
> 1. Size of the index
> 2 Last time the index was updated.
>
If by 'size o the index' you mean document count, then check the Luke
Request Handler
http://wiki.apache.org/solr/Lu
The 'abortOnConfigurationError' option was added a long time ago...
at the time, there were many errors that would just be written to the
logs but startup would continue normally.
I felt (and still do) that if there is a configuration error
everything should fail loudly. The option in solrconfig.
On Jan 13, 2010, at 5:34 PM, Minutello, Nick wrote:
Agreed, commit every second.
Do you need the index to be updated this often? Are you reading from
it every second? and need results that are that "fresh"
If not, i imagine increasing the auto-commit time to 1min or even 10
secs would
what version of solr are you running?
On Jan 7, 2010, at 3:08 PM, Jake Brownell wrote:
Hi all,
Our application uses solrj to communicate with our solr servers. We
started a fresh index yesterday after upping the maxFieldLength
setting in solrconfig. Our task indexes content in batches and
On Jan 7, 2010, at 1:05 PM, Jon Poulton wrote:
I've also just noticed that QueryParsing is not in the SolrJ API.
It's in one of the other Solr jar dependencies.
I'm beginning to think that maybe the best approach it to write a
query string generator which can generate strings of the form:
On Jan 7, 2010, at 12:11 PM, MitchK wrote:
Thank you, Ryan. I will have a look on lucene's material and luke.
I think I got it. :)
Sometimes there will be the need, to response on the one hand the
value and
on the other hand the indexed version of the value.
How can I fullfill such needs?
On Jan 7, 2010, at 10:50 AM, MitchK wrote:
Eric,
you mean, everything is okay, but I do not see it?
Internally for searching the analysis takes place and writes to the
index in an inverted fashion, but the stored stuff is left alone.
if I use an analyzer, Solr "stores" it's output two way
On Jan 6, 2010, at 3:48 PM, MitchK wrote:
I have tested a lot and all the time I thought I set wrong options
for my
custom analyzer.
Well, I have noticed that Solr isn't using ANY analyzer, filter or
stemmer.
It seems like it only stores the original input.
The stored value is always t
Ya, structured data gets a little funny.
For starters, the order of multi-valued fields should be maintained,
so if you have:
http://aaa
5
http://bbb
4
the response will return result in order, so you can map them with
array indicies.
I have played some tricks with a JSON field ana
If you need to search via the Hibernate API, then use hibernate search.
If you need a scaleable HTTP (REST) then solr may be the way to go.
Also, i don't think hibernate has anything like the faceting / complex
query stuff etc.
On Dec 29, 2009, at 3:25 PM, Márcio Paulino wrote:
Hey Every
check:
http://wiki.apache.org/solr/SolrLogging
if you are using 1.4 you want to drop in the slf4j-log4j jar file and
then it should read your log4j configs
On Nov 19, 2009, at 2:15 PM, Harsch, Timothy J. (ARC-TI)[PEROT
SYSTEMS] wrote:
Hi all,
I have an J2EE application using embedded so
Solr includes slf4j-jdk14-1.5.5.jar, if you want to use the nop (or
log4j, or loopback) impl you will need to include that in your own
project.
Solr uses slf4j so that each user can decide their logging
implementation, it includes the jdk version so that something works
off-the-shelf, but
Also:
https://issues.apache.org/jira/browse/SOLR-1302
On Nov 13, 2009, at 11:12 AM, Bertie Shen wrote:
Hey,
I am interested in using LocalSolr to go Local/Geo/Spatial/Distance
search. But the wiki of LocalSolr(http://wiki.apache.org/solr/LocalSolr
)
points to pretty old documentation. Is t
It looks like solr+spatial will get some attention in 1.5, check:
https://issues.apache.org/jira/browse/SOLR-1561
Depending on your needs, that may be enough. More robust/scaleable
solutions will hopefully work their way into 1.5 (any help is always
appreciated!)
On Nov 13, 2009, at 11:12
The HTMLStripCharFilter will strip the html for the *indexed* terms,
it does not effect the *stored* field.
If you don't want html in the stored field, can you just strip it out
before passing to solr?
On Nov 11, 2009, at 8:07 PM, aseem cheema wrote:
Hey Guys,
How do I add HTML/XML docum
On Nov 2, 2009, at 8:29 AM, Grant Ingersoll wrote:
On Nov 2, 2009, at 12:12 AM, Licinio Fernández Maurelo wrote:
Hi folks,
as we are using an snapshot dependecy to solr1.4, today we are
getting
problems when maven try to download lucene 2.9.1 (there isn't a any
2.9.1
there).
Which rep
I'm sure it is possible to configure JDK logging (java.util.loging)
programatically... but I have never had much luck with it.
It is very easy to configure log4j programatically, and this works
great with solr.
To use log4j rather then JDK logging, simply add slf4j-
log4j12-1.5.8.jar (from
I wonder why the common classes are in the solrj JAR?
Is the solrj JAR not just for the clients?
the solr server uses solrj for distributed search. This makes solrj
the general way to talk to solr (even from within solr)
Hello-
I have an application that can run in the background on a user Desktop
-- it will go through phases of being used and not being used. I want
to be able to free as many system resources when not in use as possible.
Currently I have a timer that wants for 10 mins of inactivity and
r
do you have anything custom going on?
The fact that the lock is in java2d seems suspicious...
On Sep 23, 2009, at 7:01 PM, pof wrote:
I had the same problem again yesterday except the process halted
after about
20mins this time.
pof wrote:
Hello, I was running a batch index the other
Should be fixed in trunk. Try updating and see if it works for you
See:
https://issues.apache.org/jira/browse/SOLR-1424
On Sep 9, 2009, at 8:12 PM, Allahbaksh Asadullah wrote:
Hi ,
I am building Solr from source. During building it from source I am
getting
following error.
generate-mave
can you just add a new field that has the real or ave price?
Just populate that field at index time... make it indexed but not
stored
If you want the real or average price to be treated the same in
faceting, you are really going to want them in the same field.
On Aug 28, 2009, at 1:16 PM
On Aug 27, 2009, at 10:35 PM, Paul Tomblin wrote:
Yesterday or the day before, I asked specifically if I would need to
restart the Solr server if somebody else loaded data into the Solr
index using the EmbeddedServer, and I was told confidently that no,
the Solr server would see the new data as
On Aug 26, 2009, at 3:33 PM, djain101 wrote:
I have one quick question...
If in solrconfig.xml, if it says ...
${solr.abortOnConfigurationError:false}abortOnConfigurationError>
does it mean defaults to false if it is
not set
as system property?
correct
On Aug 19, 2009, at 6:45 AM, johan.sjob...@findwise.se wrote:
Hi,
we're glancing at the GEO search module known from the jira issue 773
(http://issues.apache.org/jira/browse/SOLR-773).
It seems to us that the issue is still open and not yet included in
the
nightly builds.
correct
Is
check:
https://issues.apache.org/jira/browse/SOLR-945
this will not likely make it into 1.4
On Jul 30, 2009, at 1:41 PM, Jérôme Etévé wrote:
Hi,
Nope, I'm not using solrj (my client code is in Perl), and I'm with
solr 1.3.
J.
2009/7/30 Shalin Shekhar Mangar :
On Thu, Jul 30, 2009 at 8
ya... 'expected', but perhaps not ideal. As is, LocalSolr munges the
document on its way out the door to add the distance.
When LocalSolr makes it into the source, it will likely use a method
like:
https://issues.apache.org/jira/browse/SOLR-705
to augment each document with the calculated
not sure what you mean... yes, i guess...
you send a bunch of requests with add( doc/collection ) and they are
not visible until you send commit()
On Jul 20, 2009, at 9:07 AM, Gérard Dupont wrote:
my mistake, pb with the buffer I added. But it raises a question :
does solr
(using embedde
On Jul 20, 2009, at 9:16 AM, Edward Capriolo wrote:
On Mon, Jul 20, 2009 at 11:53 AM, Ryan McKinley
wrote:
On Jul 20, 2009, at 8:47 AM, Edward Capriolo wrote:
Hey all,
We have several deployments of Solr across our enterprise. Our
largest
one is a several GB and when enough documents
On Jul 20, 2009, at 8:47 AM, Edward Capriolo wrote:
Hey all,
We have several deployments of Solr across our enterprise. Our largest
one is a several GB and when enough documents are added an OOM
exception is occurring.
To debug this problem I have enable JMX. My goal is to write some
cacti te
On Jun 16, 2009, at 5:21 PM, Grant Ingersoll wrote:
On Jun 16, 2009, at 1:57 PM, Ryan McKinley wrote:
Is there a faster way to check the number of documents for each
field?
Currently this gets the doc count for each term:
In the past, I've created a field that contains the nam
Hi-
I'm trying to use the LukeRequestHandler with an index of ~9 million
docs. I know that counting the top / distinct terms for each field is
expensive and can take a LONG time to return.
Is there a faster way to check the number of documents for each field?
Currently this gets the doc count f
I am working with an in index of ~10 million documents. The index
does not change often.
I need to preform some external search criteria that will return some
number of results -- this search could take up to 5 mins and return
anywhere from 0-10M docs.
I would like to use the output of t
two key things to try (for anyone ever wondering why a query matches documents)
1. add &debugQuery=true and look at the explain text below --
anything that contributed to the score is listed there
2. check /admin/analysis.jsp -- this will let you see how analyzers
break text up into tokens.
Not
careful what you ask for... what if you have a million docs? will
you get an OOM?
Maybe a better solution is to run a loop where you grab a bunch of
docs and then increase the "start" value.
but you can always use:
query.setRows( Integer.MAX_VALUE )
ryan
On May 21, 2009, at 8:37 PM,
I cringe to suggest this but you can use the deprecated call:
SolrCore.getSolrCore().getCoreContainer()
On May 19, 2009, at 11:21 AM, Giovanni De Stefano wrote:
Hello all,
I have a quick question but I cannot find a quick answer :-)
I have a Java client running on the same JVM where Sol
since there is so little overlap, I would look at a core for each
user...
However, to manage 20K cores, you will not want to use the off the
shelf core management implementation to maintain these cores.
Consider overriding SolrDispatchFilter to initialize a CoreContainer
that you manage.
how much overlap is there with the 20k user documents?
if you create a separate index for each of them will you be indexing
90% of the documents 20K times? How many total documents could an
individual user typically see? How many total distinct documents are
you talking about? Is the ind
right -- which one you pick will depend more on your runtime
environment then anything else.
If you need to hit a server (on a different machine)
CommonsHttpSolrServer is your only option.
If you are running an embedded application -- where your custom code
lives in the same JVM as solr -
The point of using solrj is that you don't have to do any parsing
yourself -- you get access to the results in object form.
If you need to do parsing, just grab the xml directly:
http://host/solr/select?q=*:*&wt=xml
On May 4, 2009, at 9:36 AM, ahmed baseet wrote:
As I know when we query sol
use this constructor:
public CommonsHttpSolrServer(String solrServerUrl, HttpClient
httpClient, ResponseParser parser) throws MalformedURLException {
this(new URL(solrServerUrl), httpClient, parser, false);
}
and give it the XMLResponseParser
-- - - -
Is this just helpful for debu
I would suggest looking at Apache commons VFS and using the solrj API:
http://commons.apache.org/vfs/
With SVN, you may be able to use the webdav provider.
ryan
On Apr 26, 2009, at 4:08 AM, Ashish P wrote:
Is there any way to index contents of SVN rep in Solr ??
--
View this message in co
Right, you will have to build a new war with your own subclass of
SolrDispatchFilter *rather* then using the packaged one.
On Apr 23, 2009, at 12:34 PM, Noble Paul നോബിള്
नोब्ळ् wrote:
nope.
you must edit the web.xml and register the filter there
On Thu, Apr 23, 2009 at 3:45 PM, Giovann
I have not looked at this in a while, but I think the biggest thing it
is missing right now is a champion -- someone to get the patches (and
bug fixes) to a state where it can easily be committed. "Minor bug
fixes" are road blocks to getting things integrated.
ryan
On Apr 20, 2009, at 10
When you say Test ... Are you suggesting there is a test suite I
should run, or do just do my own testing?
your own testing...
If you use a 'nightly' the unit tests all pass.
BUT if you are not running from a standard release, there is may be
things that are not totally flushed out, or con
1 - 100 of 707 matches
Mail list logo