Hi,
I suppose the thing is then that I'm not clear from the documentation
which files I should be copying and where.
Thanks
On Sat, Jan 2, 2010 at 1:39 AM, Grant Ingersoll gsing...@apache.org wrote:
You should just be able to copy those files down to the same location, as
this is all Ant is
Thanks for quick reply.
As per you suggestion I commented uniqueKey element in schema.xml and
QueryElevationComponent in solrconfig.xml.
After that its working like charm.
I appreciate your help.
--evana
Israel Ekpo wrote:
On Fri, Jan 1, 2010 at 9:13 PM, evana evre...@ymail.com wrote:
You need, in addition to the ones shipped:
http://repo1.maven.org/maven2/colt/colt/1.2.0/colt-1.2.0.jar
http://download.carrot2.org/maven2/org/carrot2/nni/1.0.0/nni-1.0.0.jar
http://mirrors.ibiblio.org/pub/mirrors/maven2/org/simpleframework/simple-xml/1.7.3/simple-xml-1.7.3.jar
You need, in addition to the ones shipped:
http://repo1.maven.org/maven2/colt/colt/1.2.0/colt-1.2.0.jar
http://download.carrot2.org/maven2/org/carrot2/nni/1.0.0/nni-1.0.0.jar
http://mirrors.ibiblio.org/pub/mirrors/maven2/org/simpleframework/simple-xml/1.7.3/simple-xml-1.7.3.jar
On Sat, Jan 2, 2010 at 12:27 AM, Israel Ekpo israele...@gmail.com wrote:
The uniqueKey definition is not needed if you are going to build the index
from scratch each time you do the import.
However, if you are doing incremental updates, this field is required and
the uniqueKey definition is
I used RSYNC before, and 20Gb replica took less than an hour (20-40
minutes); now, HTTP, and it takes 5-6 hours...
Admin screen shows 952Kb/sec average speed; 100Mbps network, full-duplex; I
am using Tomcat Native for APR. 10x times slow...
-Fuad
http://www.tokenizer.ca
On Sat, Jan 2, 2010 at 5:48 PM, Fuad Efendi f...@efendi.ca wrote:
I used RSYNC before, and 20Gb replica took less than an hour (20-40
minutes); now, HTTP, and it takes 5-6 hours...
Admin screen shows 952Kb/sec average speed; 100Mbps network, full-duplex; I
am using Tomcat Native for APR. 10x
This curl command uploads UTF-8 correctly:
curl $URL --data-binary @$f -H 'Content-type:text/xml; charset=utf-8'
Because Windows does not use UTF-8 as its native character set, you
will keep tripping over encoding mistakes with different programs.
On Tue, Dec 29, 2009 at 1:46 AM,
The other option is to have a 'deleted' column in your table, and have
the application 'delete' operation set that field. In the DIH you
query this column with 'deletedPkQuery'.
Or, you can use triggers to maintain a new table with the IDs of
deleted rows. This will allow you to have a batch job
Sorting and index norms have space penalties.
Sorting on a field creates an array of Java ints, one for every
document in the index. Index norms (used for boosting documents and
other things) create an array of bytes in the Lucene index files, one
for every document in the index.
If you sort on
If the web API is just plain REST-style, the URLDataSource can use it.
http://wiki.apache.org/solr/DataImportHandler#Usage_with_XML.2BAC8-HTTP_Datasource
If it has a more complex interface like SOAP, you may have to create
your own SOAP client DataSource. Or, you can use Javascript in the DIH
The string book in quotes searches for a phrase. Try it without the quotes.
On Wed, Dec 30, 2009 at 8:02 PM, Olala hthie...@gmail.com wrote:
Can anyone help me??? plz!
Olala wrote:
uhm,I am sorry, this is the debug :)
lst name=debug
str name=rawquerystringbook/str
str
The schema/analysis.jsp page lets you explore how Solr parses indexed
text and queries.
On Sat, Jan 2, 2010 at 4:13 PM, Lance Norskog goks...@gmail.com wrote:
The string book in quotes searches for a phrase. Try it without the quotes.
On Wed, Dec 30, 2009 at 8:02 PM, Olala hthie...@gmail.com
The ASCIIFoldingFilter is a superset of the ISOLatin1Filter -
ISOLatin1 is deprecated. Here's the Javadoc from ASCIIFoldingFIlter.
You did not mention which language you want to search.
Unforch, the ASCIIFoldingFilter is not mentioned on the Solr wiki.
[Ryan McKinley] For starters, the order of multi-valued fields should
be maintained, so if you have:
Wait! I thougt documents and fields are sets in the Lucene index -
order is not preserved.
On Thu, Dec 31, 2009 at 7:47 PM, Chris Hostetter
hossman_luc...@fucit.org wrote:
: You could easily
Hi,
I am using a default (example) configuration of Solr and there the
stopwording seems to be enabled for both indexing and querying of fields of
type text.
I have a custom field which is of the text type.
I have extended the stopwords.txt file with lots of words but when I index
some documents
The current distributed search design assumes that all document ids
are unique across the set of cores. If you have duplicates, you're on
your on.
On Fri, Jan 1, 2010 at 7:10 AM, Yonik Seeley yo...@lucidimagination.com wrote:
On Thu, Dec 31, 2009 at 10:26 PM, Chris Hostetter
Another option is to model this problem in Solr with an even more
denormalized schema: you have one document per person per day. So,
instead of:
id=0 user=Alice start_date:1-Jan-2010 end_date:5-Jan-2010
you have:
id=0 user=Alice date:1-Jan-2010
id=1 user=Alice date:2-Jan-2010
id=2 user=Alice
You seem to have 'company' as your unique key. You can change the
schema.xml to use 'company' as the uniqueId. There is nothing magic
about the field name 'id'.
While it is possible to have a Solr schema without a unique key,
almost nobody does this. A unique key makes working with Solr much
Fields are both stored and indexed. The stored copy is exactly what
you sent in. The index is built with the text type's analysis stack
and is not stored. This output has the stopwords removed. The output
is not stored in one place, but parts of it are scattered around the
Lucene index data
@Mahout experts: could you please, elaborate on that?
It seems that I am stopping successfully quite some words with the stopwords
mechanism in Solr (I do not get search results when querying with stopwords
with the localhost/solr/select interface) but this somehow is not effective
when Solr index
Hi Yonik,
I tried... I set APR to improve performance... server is slow while replica;
but top shows only 1% of I/O wait... it is probably environment specific;
but the same happened in my home-based network, rsync was 10 times faster...
I don't know details of HTTP-replica, it could be base64
Hi Dipti,
Just out of curiosity, are you trying to use RAMDirectory for improvement in
speed? I tried doing that and did not see any significant improvement. Would
be nice to know what your experiment shows.
- Raghu
On Thu, Dec 31, 2009 at 4:17 PM, Erik Hatcher erik.hatc...@gmail.comwrote:
I'm trying to find a way to formulate the following query in solrJ. This is
the only way I can get the desired result but I can't figure out how to get
solrJ to generate the same query string. It always generates a url that
starts with select and I need it to start with spell. If there is an
24 matches
Mail list logo