Hello,
is spanNOT operator supported in this patch? If not, is there a need for
this feature for anyone?
Regards,
Dmitry
Thanks Mikhail.
On Tue, Mar 5, 2013 at 8:23 PM, Mikhail Khludnev mkhlud...@griddynamics.com
wrote:
Something like this.
On Tue, Mar 5, 2013 at 6:16 PM, Dmitry Kan solrexp...@gmail.com wrote:
Hello,
I spent some more time on this and used Mikhail's suggestions of which
classes would
Thanks again for replying and giving insight to this, Jack. Your two links
were exactly the answer I was hoping for going forward.
5. mars 2013 kl. 14:12 skrev Jack Krupansky j...@basetechnology.com:
See:
https://issues.apache.org/jira/browse/LUCENE-2370
Maybe Uwe could comment on his
6. mars 2013 kl. 02:24 skrev varun srivastava varunmail...@gmail.com:
Thanks Hoss .. Is this available in 4.0 ?
On Tue, Mar 5, 2013 at 5:14 PM, Chris Hostetter
hossman_luc...@fucit.orgwrote:
:dynamicField name=stamp_* type=string indexed=false
: stored=false multiValued=true/
So you are suggesting me to iterate file system and index fs tree entities
including: directory names, file names, file size etc. and then post it to
solr?
I need to index the FS tree, not the file contents.
On Tue, Mar 5, 2013 at 5:54 PM, Erik Hatcher erik.hatc...@gmail.com wrote:
Would Solr's
Hello everyone!
- I´m using Solr 4.1.0.
- Yes, without the sort the query works perfectly.
- The fieldType is as follows:
field name=geolocation type=location_rpt indexed=true
stored=true multiValued=true /
**
fieldType name=location_rpt
Hello
Our application (VuFind, a library discovery tool) is speaking to a solr 3.5 index per
default. To get the facet values back, the param 'facet.sort' is sent in the request with
an empty value. Solr then delivers the called facets sorted by count, as it should be. A
request looks
No, double checked, and even went and reindexed yesterday and still the
same issue.
Regards,
Phil.
-Original Message-
From: Jack Krupansky [mailto:j...@basetechnology.com]
Sent: 05 March 2013 14:38
To: solr-user@lucene.apache.org
Subject: Re: Search term matching on part of a token,
Hi,
I tried solrcore.properties (log.name=logname) but log4J didn't pick it up.
As for your second suggestion (log4j.properties file in core/lib/ ) wouldn't
that be overwritten every time the webapps are updated or refreshed. This is
the problem I get when I have changed my
Hi
I have emails indexed with the default text_general fieldType.
I find that if the email a...@gmail.com is indexed, and I search for
gmail.com@ (exact phrase search) I can a result, while I should not get
one.
Any idea how to solve this?
thanks.
--
View this message in context:
Hi,
How about a custom UpdateRequestProcessor that uses milliseconds or even
nanoseconds and stores them in some field? If that is enough resolution
and you still want to avoid collision, append a random letter/string/number
to it, a la millis or nanos_extra stuff to make it unique.
Otis
--
Hi Joseph,
What exactly are you looking to to?
See http://incubator.apache.org/blur/
Otis
--
Solr ElasticSearch Support
http://sematext.com/
On Wed, Mar 6, 2013 at 2:39 AM, Joseph Lim ysli...@gmail.com wrote:
Hi I am running hadoop distributed file system, how do I put my output of
the
Hi Syao,
You should just write a simple (Java) app that traverses the dir tree, gets
info about each file, uses it to construct Solr doc objects
(SolrInputDocuments if you are working in Java with SolrJ) and sends them
to Solr for indexing. Should be about 30 minutes of work or less.
Otis
--
k thanks
On Wed, Mar 6, 2013 at 5:05 PM, Otis Gospodnetic otis.gospodne...@gmail.com
wrote:
Hi Rohan,
It's per field.
Otis
--
Solr ElasticSearch Support
http://sematext.com/
On Wed, Mar 6, 2013 at 3:09 AM, Rohan Thakur rohan.i...@gmail.com wrote:
hi all
I wanted to confirm
one more thing how does the solr handles the score when we have multiple
words as query parameters which are not consecutively present in the filed
that is indexed and we are querying on does it first searches for document
field containing both the term and then how does is find the scoring on
Hi,
Would like to know how can i put the indexed solr shards into hdfs?
Thanks..
Joseph
On Mar 6, 2013 7:28 PM, Otis Gospodnetic otis.gospodne...@gmail.com
wrote:
Hi Joseph,
What exactly are you looking to to?
See http://incubator.apache.org/blur/
Otis
--
Solr ElasticSearch Support
Use the Solr Admin UI Analysis page and enter 100.10 for field type
class and see whether it keeps the number as one term or not.
Do you maybe have a multivalued field that has both 100 and 100.10?
Also do a debugQuery=true and see what the result documents are actually
being matched on.
--
I imagine that you had a qf parameter in your old query request handler,
so add qf it to the new query request handler. df is used only if qf
is missing.
-- Jack Krupansky
-Original Message-
From: David Parks
Sent: Wednesday, March 06, 2013 2:18 AM
To: solr-user@lucene.apache.org ;
Hello Joseph,
You can certainly put them there, as in:
hadoop fs -copyFromLocal localsrc URI
But searching such an index will be slow.
See also: http://katta.sourceforge.net/
Otis
--
Solr ElasticSearch Support
http://sematext.com/
On Wed, Mar 6, 2013 at 7:50 AM, Joseph Lim
Otis,
I noticed this in my logs repeatedly during that time period:
Mar 5, 2013 1:28:00 PM org.apache.solr.core.CachingDirectoryFactory close
INFO: Releasing
directory:/usr/local/solr_aggregate/solr_aggregate/data/index
It wasn't in my logs any other time.
I found this:
Hi,
I’ve written my own analyzer to index and query a set of documents. At indexing
time everything goes well but
now I have a problem in query phase.
I need to pass the whole query string to my analyzer before the edismax query
parser begins its tasks.
In other words I have to
Ah; bingo!
The top error in the log is what Solr reports in the HTTP response you
reported but it's the message of the exception wrapped by it in the logs
which is more indicative of the problem:
Caused by: org.apache.solr.common.SolrException: A ValueSource isn't
directly available from this
Hello Otis ,
Is there any configuration where it will index into hdfs instead?
I tried crawlzilla and lily but I hope to update specific package such as
Hadoop only or nutch only when there are updates.
That's y would prefer to install separately .
Thanks so much. Looking forward for your
It looks like to configure my SolrCloud server as an HTTPS endpoint, I need
to update the solr.xml as such:
host=https://${jboss.node.name:}; hostPort=8443
When I try this, it fails registering with the zookeeper.
ERROR [SolrCore] null:java.lang.IllegalArgumentException: Invalid path
string
in last days i made my first steps with solr 4.1.0. So far i had have no
problems and could always find a solution, mainly on this great platform.
I am using the example configuration from solr with my own fields. I am
indexing files with tika and add extra fields to the documents.
My Problem is
Appending a random value only reduces the chance of a collision (And I need
to ensure continuous uniqueness) and could hurt how the field is later
sorted. I have not written a custom UpdateRequestProcessor before, is there
a way to incorporate a Singleton that ensures one instance across a
Much appreciated. Thank you.
On Tue, Mar 5, 2013 at 3:32 PM, Shawn Heisey s...@elyograg.org wrote:
On 3/5/2013 2:17 PM, JW West wrote:
Hello,
I'm trying to set up Solr with a multi core configuration but I'm running
into troubles starting using start.jar.
Specifically, running java -jar
This sounds like a job for Zookeeper (distributed coordination is what it does).
Take a look at:
http://zookeeper-user.578899.n2.nabble.com/Sequence-Number-Generation-With-Zookeeper-td5378618.html
On Wed, Mar 6, 2013 at 10:00 AM, mark12345
marks1900-pos...@yahoo.com.au wrote:
Appending a random
If you want to mess with UpdateRequestProcessors, try the
ScriptUpdateProcessor, with which you can write your update logic in
Javascript. That would allow you to add your unique field. Use something
like timestamp+threadno+shardno and you'd have something unique
(assuming you can access those
Hey David, it works! Thank you very much. The true is that the
docummentation is a little bit confusing, but know It works perfectly.
Regards,
- Luis Cappa
2013/3/6 David Smiley (@MITRE.org) dsmi...@mitre.org
Ah; bingo!
The top error in the log is what Solr reports in the HTTP response you
That's a mighty long sentence, Rohan! :)
If you append debugQuery=true to the query URL you will see the scoring
explanation in all its glory.
Otis
--
Solr ElasticSearch Support
http://sematext.com/
On Wed, Mar 6, 2013 at 6:55 AM, Rohan Thakur rohan.i...@gmail.com wrote:
one more thing
I´ve doing some performance tests and I´ve noticed that with the new query
syntax that David told me to use the QTime increases a lot. I´ve got an
index with up to 8 million docs and sometimes the query response delays
three, four or more seconds until finishes. Is that normal?
2013/3/6 Luis
What are you actually trying to achieve? If you can share what you are
trying to achieve maybe folks can help you find the right way to do it.
Upayavira
On Wed, Mar 6, 2013, at 02:54 PM, Joseph Lim wrote:
Hello Otis ,
Is there any configuration where it will index into hdfs instead?
I
Hi Upayavira,
sure, let me explain. I am setting up Nutch and SOLR in hadoop environment.
Since I am using hdfs, in the event if there is any crashes to the
localhost(running solr), i will still have the shards of data being stored
in hdfs.
Thanks you so much =)
On Thu, Mar 7, 2013 at 1:19 AM,
On Mar 6, 2013, at 6:54 AM, Scott Carlson scott.a.carl...@gmail.com wrote:
Am I on the right track and I should log this issue as bug,
Might as well - seems like it could be improved.
or is there a
different way to configure this that I'm not seeing.
I think there might be - I remember
Luis,
I should have asked how much data you have when I offered the solution.
If you have a multi-valued spatial field and you need to get the closest of
potentially many indexed points (and your schema snippet below shows
multiValued=true) then I'm afraid you're stuck with this until the
: I've used this configuration to my timestamp field and it's works
:
: field name=timestamp type=date indexed=true stored=true
: default=NOW+1HOUR multiValued=false/
that is *really* not a good idea ...
doing something like that means you will be putting incorrect data into
your index,
I'm working on SOLR-3251 https://issues.apache.org/jira/browse/SOLR-3251, to
dynamically add fields to the Solr schema.
I posted a rough outline of how I propose to do this:
bq. Change Solr schema serialization from XML to JSON, and provide an XML-JSON
conversion tool.
What is the motivation for the change? I think if you are sitting down and
looking to design a schema, working with the XML is fairly nice and fast. I
picture that a lot of people would start by
Facets work on the tokenized content of the field. You must be using
text_en or similar type. Switch the facet field to use String. If you are
using eDisMax and want to be able to search it for substrings (e.g. if
somebody searches for Intel), use copyField to create two fields, one for
searching,
It should be easy to extend ExtendedDismaxQParser and do your
pre-processing in the parse() method before calling edismax's parse. Or
maybe you could change the way EDismax is splitting the input query into
clauses by extending the splitIntoClauses method?
Tomás
On Wed, Mar 6, 2013 at 6:37 AM,
In response to my thoughts about using DOM as an intermediate representation
for schema elements, for use in lazy re-loading on schema change, Erik Hatcher
argued against (solely) using XML for schema serialization
I have a SolrCloud cluster (2 machines, 2 Solr instances, 32 shards,
replication factor of 2) that I've been using for over a month now in
production.
Suddenly, the hourly cron I run that dispatches a delete by query
completely halts all indexing. Select queries still run (and quickly),
there is
Hmm…I think I'm missing some pieces.
I agree with Erick that you should be able to load a schema from any object - a
DB, a file in ZooKeeper, you name it. But it seems by default, having that
object be schema.xml seems nicest to me. That doesn't mean you have to use DOM
or XML internally -
: *) Have a default empty constructor
:
: My preliminary tests seem to indicate this is the case. Am I missing
: anything.
Any analyzer that has an empty construct *or* a constructor that takes in
a lucene Version object may be specified.
I've updated the wiki to make this more clear...
Which version of Solr?
Can you use jconsole, visualvm, or jstack to get some stack traces and see
where things are halting?
- Mark
On Mar 6, 2013, at 11:45 AM, Brett Hoerner br...@bretthoerner.com wrote:
I have a SolrCloud cluster (2 machines, 2 Solr instances, 32 shards,
replication factor
I'm not sure what pieces you might be missing, sorry.
I had thought about adding a web UI for schema composition, but that would be a
major effort, and not in scope here.
I agree, though, especially without a full schema modification REST API, that
hand editing will have to be supported.
Hi All,
By default, Solr does spell correction only if Term length is more than 4
chars.
In my application, we need to spell correct for 3 letter terms as well so,
I changed the *DirectSolrSpellChecker configuration to set *minQueryLengthto 3
I would like to know if there is NEGATIVE
On Mar 6, 2013, at 12:08 PM, Steve Rowe sar...@gmail.com wrote:
I'm not sure what pieces you might be missing, sorry.
My main confusion is around this:
bq. When schema.xml is present, schema.json (if any) will be ignored.
Basically, why have schema.json? Perhaps it's just me, but a json
On Mar 6, 2013, at 3:33 PM, Mark Miller markrmil...@gmail.com wrote:
On Mar 6, 2013, at 12:08 PM, Steve Rowe sar...@gmail.com wrote:
I'm not sure what pieces you might be missing, sorry.
My main confusion is around this:
bq. When schema.xml is present, schema.json (if any) will be ignored.
Hi there!
As suggested elsewhere
(http://lucene.472066.n3.nabble.com/Modeling-openinghours-using-multipoints-td4025336.html),
I'm trying to use spatial mapping to model intervals of availability.
I've defined the following fieldType in schema.xml:
fieldType name=time_rpt
4.1, I'll induce it again and run jstack.
On Wed, Mar 6, 2013 at 1:50 PM, Mark Miller markrmil...@gmail.com wrote:
Which version of Solr?
Can you use jconsole, visualvm, or jstack to get some stack traces and see
where things are halting?
- Mark
On Mar 6, 2013, at 11:45 AM, Brett
Why wouldn't SolrCloud help you here? You can setup shards and replicas etc
to have redundancy b/c HDFS isn't designed to serve real time queries as
far as I understand. If you are using HDFS as a backup mechanism to me
you'd be better served having multiple slaves tethered to a master (in a
How will this all work with ZooKeeper and cloud?
Will ZK get pushed the serialized monolithic schema.xml / schema.json from the
node which changed it, and then trigger an update to the rest of the cluster?
I was kind of hoping that once we have introduced ZK into the mix as our
centralized
Here is a dump after the delete, indexing has been stopped:
https://gist.github.com/bretthoerner/c7ea3bf3dc9e676a3f0e
An interesting hint that I forgot to mention: it doesn't always happen on
the first delete. I manually ran the delete cron, and the server continued
to work. I waited about 5
Thans Brett, good stuff (though not a good problem).
We def need to look into this.
- Mark
On Mar 6, 2013, at 1:53 PM, Brett Hoerner br...@bretthoerner.com wrote:
Here is a dump after the delete, indexing has been stopped:
https://gist.github.com/bretthoerner/c7ea3bf3dc9e676a3f0e
An
It does not look like a deadlock, though it could be a distributed one. Or
it could be a livelock, though that's less likely.
Here is what we used to recommend in similar situations for large Java
systems (BEA Weblogic):
1) Do thread dump of both systems before anything. As simultaneous as you
This is what I see:
We currently limit the number of outstanding update requests at one time to
avoid a crazy number of threads being used.
It looks like a bunch of update requests are stuck in socket reads and are
taking up the available threads. It looks like the deletes are hanging out
Well, I've resolved one part of the problem, but run into another. I
was specifying points as %d,%d, which caused spatial4j to interpret
them as lat/lon points (see
https://github.com/spatial4j/spatial4j/blob/master/src/main/java/com/spatial4j/core/io/ShapeReadWriter.java#L145
for the line in
If there's anything I can try, let me know. Interestingly, I think I have
noticed that if I stop my indexer, do my delete, and restart the indexer
then I'm fine. Which goes along with the update thread contention theory.
On Wed, Mar 6, 2013 at 5:03 PM, Mark Miller markrmil...@gmail.com wrote:
Any chance you can grab the stack trace of a replica as well? (also when it's
locked up of course).
- Mark
On Mar 6, 2013, at 3:34 PM, Brett Hoerner br...@bretthoerner.com wrote:
If there's anything I can try, let me know. Interestingly, I think I have
noticed that if I stop my indexer, do
: As far as a user editing the file AND rest API access, I think that
: seems fine. Yes, the user is in trouble if they break the file, but that
Ignoring for a moment what format is used to persist schema information, I
think it's important to have a conceptual distinction between data that
: Subject: Can I change to MULTIVALUED=true without rebuilding index?
you can cahnge it from false to true w/o reindexing.
changing to true to false *should* be ok if you are 100%
certain you don't hvae any documents in your index that have more then one
value -- i don't even know what sorts
Thanks!
I was hoping that was true.
Roy
-Original Message-
From: Chris Hostetter [mailto:hossman_luc...@fucit.org]
Sent: Wednesday, March 06, 2013 4:54 PM
To: solr-user@lucene.apache.org
Subject: Re: Can I change to MULTIVALUED=true without rebuilding index?
: Subject: Can I change to
On Mar 6, 2013, at 4:50 PM, Chris Hostetter hossman_luc...@fucit.org wrote:
i don't think it's
unreasable to say if you would like to manipulate the schema using an
API, then you give up the ability to manipulate it as a config file on
disk
As long as you can initially work with an
I'm having trouble migrating the geohash fields from my Solr 3.4 schema
to the Solr 4 schema.
this is the 3.4 type and class:
fieldType name=geohash class=solr2155.solr.schema.GeoHashField
length=12/
is the below Solr 4 spatial type the right configuration to implement
data being stored in
Hi,
I am working with Solritas. However I want it to fetch data from Solr
through an interface.
Solritas -- Interface --Solr
Solr -- Interface --Solritas
Could somebody please guide me as to what changes to do in Solr config
files to achieve this.
Thanks and regards,
Romita
Please elaborate. What do you mean by interface? What do you want different
than it is now?
Erik
On Mar 6, 2013, at 21:04, Romita Saha romita.s...@sg.panasonic.com wrote:
Hi,
I am working with Solritas. However I want it to fetch data from Solr
through an interface.
Solritas --
What is mean to say is that, before the query goes to Solr, certain
parameters need to be added, as for example number of documents displayed,
boosting parameters etc. The algorithm of appending these parameters to
the solr search query will be done by the interface.
The modified query will go
Hi Amit,
so you mean that if I just want to get redundancy for solr in hdfs, the
only best way to do it is to as per what Otis suggested using the following
command
hadoop fs -copyFromLocal localsrc URI
Ok let me try out solrcloud as I will need to make sure it works well with
nutch too..
Joseph,
Doing what Otis said will do literally what you want which is copying the
index to HDFS. It's no different than copying it to a different machine
which btw is what Solr's master/slave replication scheme does.
Alternatively, I think people are starting to setup new Solr instances with
Hi Amit,
Currently I am designing a Learning Management System where it is based on
Hadoop and hbase . Right now I want to integrate nutch with solr in it as
part of crawler module, so that users will only be able to search relevant
documents from specific source. And since crawling and indexing
yup I got that but can you please explain me how is it going on?
I got this:
rawquerystring:samsung ace metallic,
querystring:samsung ace metallic,
parsedquery:value:samsung value:ace value:metallic,
parsedquery_toString:value:samsung value:ace value:metallic,
explain:{
hi
I think its not readable but if you open this in editor that gmail has like
while replying you can se the text in proper format that is indent.I tried
saving this output in other formats aswell but it is still the same...you
can see this properly while replying to the mail...
thanks
regards
If you can explain any one line of I will figure out othersthanks
On Thu, Mar 7, 2013 at 11:48 AM, Rohan Thakur rohan.i...@gmail.com wrote:
yup I got that but can you please explain me how is it going on?
I got this:
rawquerystring:samsung ace metallic,
querystring:samsung ace
Hi Harley,
See: http://wiki.apache.org/solr/SolrAdaptersForLuceneSpatial4
In SOLR-2155 you had to explicitly specify the prefix encoding length,
whereas in Solr 4 you specify how much precision you need and it figures out
what the length is that satisfies that. When you first use the field, it'll
76 matches
Mail list logo