Re: Solrcloud dataimport failed at first time after restart

2012-10-09 Thread jun Wang
I have found the reason.  The reason is that I am using jboss JNDI
datasource, and oracle driver is placed in WEB-INFO/lib, this is a very
common error, driver should be placed in %JBOSS_HOME%\server\default\lib.

2012/10/10 jun Wang 

> Hi, all
> I found that dataimport will failed at first time after restart. and the
> log is here . It's seem like a bug.
>
> 2012-10-09 20:00:08,848 ERROR dataimport.DataImporter - Full Import
> failed:java.lang.RuntimeException: java.lang.RuntimeException:
> org.apache.solr.handler.dataimport.DataImportHandlerException: Unable to
> execute query: select a.id, a.subject, a.keywords, a.category_id,
> to_number((a.gmt_modified - to_date('1970-01-01','-mm-dd'))*24*60*60)
> as gmt_modified,a.member_seq,b.standard_attr_desc,
> b.custom_attr_desc, decode(a.product_min_price, null, 0,
> a.product_min_price)/100 as min_price, sign(a.ws_offline_date - sysdate) +
> 1 as is_offlinefrom ws_product_draft a,
> ws_product_attribute_draft bwhere a.id =
> b.product_id(+) Processing Document # 1
> at
> org.apache.solr.handler.dataimport.DocBuilder.execute(DocBuilder.java:273)
> at
> org.apache.solr.handler.dataimport.DataImporter.doFullImport(DataImporter.java:382)
> at
> org.apache.solr.handler.dataimport.DataImporter.runCmd(DataImporter.java:448)
> at
> org.apache.solr.handler.dataimport.DataImporter$1.run(DataImporter.java:429)
> Caused by: java.lang.RuntimeException:
> org.apache.solr.handler.dataimport.DataImportHandlerException: Unable to
> execute query: select a.id, a.subject, a.keywords, a.category_id,
> to_number((a.gmt_modified - to_date('1970-01-01','-mm-dd'))*24*60*60)
> as gmt_modified,a.member_seq,b.standard_attr_desc,
> b.custom_attr_desc, decode(a.product_min_price, null, 0,
> a.product_min_price)/100 as min_price, sign(a.ws_offline_date - sysdate) +
> 1 as is_offlinefrom ws_product_draft a,
> ws_product_attribute_draft bwhere a.id =
> b.product_id(+) Processing Document # 1
> at
> org.apache.solr.handler.dataimport.DocBuilder.buildDocument(DocBuilder.java:413)
> at
> org.apache.solr.handler.dataimport.DocBuilder.doFullDump(DocBuilder.java:326)
> at
> org.apache.solr.handler.dataimport.DocBuilder.execute(DocBuilder.java:234)
> ... 3 more
> Caused by: org.apache.solr.handler.dataimport.DataImportHandlerException:
> Unable to execute query: select a.id, a.subject, a.keywords,
> a.category_id, to_number((a.gmt_modified -
> to_date('1970-01-01','-mm-dd'))*24*60*60) as gmt_modified,a.member_seq,
>b.standard_attr_desc, b.custom_attr_desc,
> decode(a.product_min_price, null, 0, a.product_min_price)/100 as min_price,
> sign(a.ws_offline_date - sysdate) + 1 as is_offline
>  from ws_product_draft a, ws_product_attribute_draft b
>where a.id = b.product_id(+) Processing Document # 1
> at
> org.apache.solr.handler.dataimport.DataImportHandlerException.wrapAndThrow(DataImportHandlerException.java:71)
> at
> org.apache.solr.handler.dataimport.JdbcDataSource$ResultSetIterator.(JdbcDataSource.java:252)
> at
> org.apache.solr.handler.dataimport.JdbcDataSource.getData(JdbcDataSource.java:209)
> at
> org.apache.solr.handler.dataimport.JdbcDataSource.getData(JdbcDataSource.java:38)
> at
> org.apache.solr.handler.dataimport.SqlEntityProcessor.initQuery(SqlEntityProcessor.java:59)
> at
> org.apache.solr.handler.dataimport.SqlEntityProcessor.nextRow(SqlEntityProcessor.java:73)
> at
> org.apache.solr.handler.dataimport.EntityProcessorWrapper.nextRow(EntityProcessorWrapper.java:243)
> at
> org.apache.solr.handler.dataimport.DocBuilder.buildDocument(DocBuilder.java:472)
> at
> org.apache.solr.handler.dataimport.DocBuilder.buildDocument(DocBuilder.java:411)
> ... 5 more
> Caused by: java.lang.ClassNotFoundException: Unable to load null or
> org.apache.solr.handler.dataimport.null
> at
> org.apache.solr.handler.dataimport.DocBuilder.loadClass(DocBuilder.java:899)
> at
> org.apache.solr.handler.dataimport.JdbcDataSource$1.call(JdbcDataSource.java:159)
> at
> org.apache.solr.handler.dataimport.JdbcDataSource$1.call(JdbcDataSource.java:127)
> at
> org.apache.solr.handler.dataimport.JdbcDataSource.getConnection(JdbcDataSource.java:362)
> at
> org.apache.solr.handler.dataimport.JdbcDataSource.access$200(JdbcDataSource.java:38)
> at
> org.apache.solr.handler.dataimport.JdbcDataSource$ResultSetIterator.(JdbcDataSource.java:239)
> ... 12 more
> Caused by: java.lang.NullPointerException
> at
> java.util.concurrent.ConcurrentHashMap.get(ConcurrentHashMap.java:768)
> at
> org.apache.solr.core.SolrResourceLoader.findClass(SolrResourceLoader.java:387)
> at
> org.apache.solr.handler.dataimport.DocBuilder.loadClass(D

Re: Query foreign language "synonyms" / words of equivalent meaning?

2012-10-09 Thread Bernd Fehling

As far as I know, there is no built-in functionality for language translation.
I would propose to write one, but there are many many pitfalls.
If you want to translate from one language to another you might have to
know the "starting" language. Otherwise you get problems with translation.

Not (german) -> distress (english), affliction (english)

- you might have words in one language which are stopwords in other language 
"not"
- you don't have a one to one mapping, it's more like "1 to n+x"
  toilette (french) -> bathroom, rest room / restroom, powder room

This are just two points which jump into my mind but there are tons of pitfalls.

We use the solution of a multilingual thesaurus as synonym dictionary.
http://en.wikipedia.org/wiki/Eurovoc
It holds translations of 22 official languages of the European Union.

So a search for "europäischer währungsfonds" gives also results with
"european monetary fund", "fonds monétaire européen", ...

Regards
Bernd



Am 10.10.2012 04:54, schrieb onlinespend...@gmail.com:
> Hi,
> 
> English is going to be the predominant language used in my documents, but
> there may be a spattering of words in other languages (such as Spanish or
> French). What I'd like is to initiate a query for something like "bathroom"
> for example and for Solr to return documents that not only contain
> "bathroom" but also "baño" (Spanish). And the same goes when searching for "
> baño". I'd like Solr to return documents that contain either "bathroom" or "
> baño".
> 
> One possibility is to pre-translate all indexed documents to a common
> language, in this case English. And if someone were to search using a
> foreign word, I'd need to translate that to English before issuing a query
> to Solr. This appears to be problematic, since I'd have to know whether the
> indexed words and the query are even in a foreign language, which is not
> trivial.
> 
> Another possibility is to pre-build a list of foreign word synonyms. So baño
> would be listed as a synonym for bathroom. But I'd need to include other
> languages (such as toilette in French) and other words. This requires that
> I know in advance all possible words I'd need to include foreign language
> versions of (not to mention needing to know which languages to include).
> This isn't trivial either.
> 
> I'm assuming there's no built-in functionality that supports the foreign
> language translation on the fly, so what do people propose?
> 
> Thanks!
> 

-- 
*
Bernd FehlingUniversitätsbibliothek Bielefeld
Dipl.-Inform. (FH)LibTec - Bibliothekstechnologie
Universitätsstr. 25 und Wissensmanagement
33615 Bielefeld
Tel. +49 521 106-4060   bernd.fehling(at)uni-bielefeld.de

BASE - Bielefeld Academic Search Engine - www.base-search.net
*


Re: Anyone using mmseg analyzer in solr multi core?

2012-10-09 Thread 曹霖
MMSeg  is better than ikanalyzer? is there anyone compare those open source

Chinese word analyzers? which one is better overall?

2012/10/9 Peter Keegan 

> We're using MMSeg with Lucene, but not Solr. Since each SolrCore is
> independent, I'm not sure how you can avoid each having a copy of the
> dictionary, unless you modified MMSeg to use shared memory. Or, maybe I
> missing something.
>
> On Mon, Oct 8, 2012 at 3:37 AM, liyun  wrote:
>
> > Hi all,
> > Is anybody using mmseg analyzer for Chinese word analyze? When we use
> this
> > in solr multi-core, I find it will load the dictionary per core and each
> > core cost about 50MB memory. I think this is a big waste when our JVM has
> > only 1GB memory…… Anyone have a good idea for handle this trouble ?
> >
> > 2012-10-08
> >
> >
> >
> > Li Yun
> > Software Engineer @ Netease
> > Mail: liyun2...@corp.netease.com
> > MSN: rockiee...@gmail.com
>


Re: Installing Solr on a shared hosting server?

2012-10-09 Thread Gora Mohanty
On 9 October 2012 21:35, Otis Gospodnetic  wrote:
> Hi,
>
> I don't know anything about Bluehost, but Solr works fine on EC2.
[...]

Have had very minimal experience with Bluehost, as we are hosted
mostly from EC2. The main difference is that with Bluehost shared
hosting---at least our version of it---one has access only to their pre-
built software stacks, and cannot install arbitrary applications. It is
possible that Bluehost has other offerings that enable one to do so,
but in the absence of that, Solr on Bluehost would not be possible.

Regards,
Gora


Re: Help with Velocity in SolrItas

2012-10-09 Thread Lance Norskog

Thanks, everyone. This is the problem: $sentence is a NamedList node, with a 
name and a "value" (any Java object). I want its value subnode:

#foreach($sentence in $outer)
$sentence = $sentence.value



| >> 
| >> Here is the XML from a search result:
| >> 
| >> 
| >>   0
| >>   
| >>   A bunch of words
| >>   
| >> 
| >> more sentences 
| >> 
| >> 
| >> Here is my Velocity code:
| >> #foreach($sentence in $outer)
| >>   Raw data: $sentence
| >>   
| >>   #set($index = $sentence.get('index'))
| >>   #set($text = $sentence.get('text'))
| >>   
| >> Index: $index
| >> 
| >> Text: $text
| >> 
| >>   
| >> #end
| >> 
|
| 
| 


Re: SolrJ 4.0 Beta maxConnectionsPerHost

2012-10-09 Thread Briggs Thompson
Thanks all for your responses. For some reason the emails were getting
filtered out of my inbox.

*Otis* - jstack is a great suggestion, thanks! The problem didn't happen
this morning but next time it does I will certainly get the dump to see
exactly where the app is swimming around. I haven't used
StreamingUpdateSolrServer
but I will see if that makes a difference. Are there any major drawbacks of
going this route?

*Sami* - if you are referring to
config:maxConnections=200&maxConnectionsPerHost=8,
it showed up up in the Solr logs, not the SolrJ logs. The client IS
instantiated only once and not for every request. I was curious if this was
part of the problem. Do I need to re-instantiate the object for each
request made? I figured there would be more overhead if I am re-creating
the connection several times when I never really need to shut it down, but
at this point the overhead would be minimal though so I will try that.

*Hoss* - The reason it seemed the client was creating the log was because
the indexer (solr *server*) was more or less dormant for several hours,
then I booted up my indexing *client* and the maxConnectionsPerHost tidbit
was spit out right away. I was looking for something in the solrconfig and
online but didn't find anything. I didn't look for very long so will check
it out again.

Some very good suggestions here. I appreciate everyones feedback. I will
follow up after some experimentation.

Thanks,
Briggs Thompson


On Tue, Oct 9, 2012 at 11:19 AM, Chris Hostetter
wrote:

>
> : I did some digging and experimentation and found something interesting.
> : When starting up the application, I see the following in Solr logs:
> : Creating new http client,
> config:maxConnections=200&maxConnectionsPerHost=8
> ...
> : It seems as though the maxConnections and maxConnectionsPerHost are not
> : actually getting set. Anyone seen this problem or have an idea how to
> : resolve?
>
> To elaborate on sami's comment...
>
> If you are seeing this in the logs from your solr *server*, it is unlikey
> that it has anything to do with the settings you are making on your solr
> *client*  this is probably related to the http client created inside
> solr for communicating with other solr nodes (ie: replication, solr cloud
> distributed updates, solr cloud peersync, etc...).  Which is different
> from the properties you set on the http client in your solr client
> application.
>
> I believe there is a way to configure the defaults for the internal used
> http clients via solrconfig.xml, but off the top of my head i don't
> remember what that is.
>
>
>
> -Hoss
>


Spatial Search response time complexity

2012-10-09 Thread TJ Tong
Hi all,

Does anyone know the Solr (lucene)spatial search time complexity, such as
geofilt on LatLonType fields? Is it logN? 

Thanks!
TJ



--
View this message in context: 
http://lucene.472066.n3.nabble.com/Spatial-Search-response-time-complexity-tp4012801.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: Auto Correction?

2012-10-09 Thread Otis Gospodnetic
Hi,

Suggestion #1 is not always the best suggestion unfortunately.  DYM
ReSearcher at http://sematext.com/products/dym-researcher/index.html
that Ahmet pointed to deals with this, among other things.

Otis


On Tue, Oct 9, 2012 at 12:24 PM, Amit Nithian  wrote:
> What's preventing you from using the spell checker and take the #1
> result and re-issue the query from a sub-class of the query component?
> It should be reasonably fast to re-execute the query from the server
> side since you are already within Solr. You can modify the response to
> indicate that the new query was used so your client can display to the
> user that it searched automatically for "milky.. click here for
> searches for mlky" or something.
>
> On Tue, Oct 9, 2012 at 8:46 AM, Ahmet Arslan  wrote:
>>> I would like to ask if there are any ways to correct user's
>>> queries
>>> automatically? I know there is spellchecker which *suggests*
>>> possible
>>> correct words... The thing i wanna do is *automatically
>>> fixing* those
>>> queries and running instead of the original one
>>
>> not out of the box, you need to re-run suggestions at client side. There is 
>> an commercial product though.
>> http://sematext.com/products/dym-researcher/index.html


Re: Extract multiple streams into the same document

2012-10-09 Thread Yury Kats
Answering my own question, for archive's sake,
I worked this out by creating my own UpdateRequestProcessor.

On 10/4/2012 2:35 PM, Yury Kats wrote:
> I'm sending streams of data to Solr, using ExtractingRequestHandler to be 
> parsed/extracted by Tika and then indexed.
> 
> While multiple streams can be passed with a single request to Solr, each 
> stream ends up being indexed into a separate document.
> Or, if I pass the unique id parameter with the request (as "literal.id" 
> parameter), the very last stream ends up overwriting all
> other streams withing the same request, since each one is being indexed into 
> a new document with the same id.
> 
> I'm looking for a way to have multiple streams indexed into the same 
> document. I have a content field defined for extraction
> (using fmap.content parameter) and the field is defined as multiValued in the 
> schema. I would like all streams from the request to be
> indexed as different values of that multiValued content field in the same 
> document.
> 
> Any hints or ideas are appreciated.
> 
> Thanks,
> Yury
> 



Re: solr1.4 code Example

2012-10-09 Thread Sujatha Arun
Thanks ,that worked.

Regards
Sujatha

On Tue, Oct 9, 2012 at 5:57 PM, Iwan Hanjoyo  wrote:

> you can download the code directly from here
>
> http://www.solrenterprisesearchserver.com/
>
> http://solrenterprisesearchserver.s3-website-us-east-1.amazonaws.com/downloads/5883-solr-enterprise-search2.zip
>
> regards,
>
>
> Hanjoyo
>


PostFilters, Grouping, Sorting Oh My!

2012-10-09 Thread Amit Nithian
Hi all,

I've been working with using Solr's post filters/delegate collectors
to collect some statistics about the scores of all the documents and
had a few questions with regards to this when combined with grouping
and sorting:
1) I noticed that if I don't include the "score" field as part of the
sort spec with *no* grouping enabled, my custom delegate scorer gets
called so I can then collect the stats I need. Same is true with score
as part of the sort spec (this then leads me to focus on the grouping
feature)
2) If I turn ON grouping:
  a) WITH "score" in the sort spec, my custom delegate scorer gets called
  b) WITHOUT "score" in the sort spec, my custom delegate scorer does
NOT get called.

What's interesting though is that there *are* scores generated so I'm
not sure what all is going on. I traced through the code and saw that
the scorer gets called as part of one of the comparators
(RelevanceComparator) which is why with "score" in the sort spec it
works but that is about as far as I could go. Since I am not too
worried in my application about a sort spec without the score always
being there it's not a huge concern; however, I do want to understand
why with the grouping feature enabled, this doesn't work and whether
or not it's a bug.

Any help on this would be appreciated so that my solution to this
problem is complete.

Thanks!
Amit


Re: Auto Correction?

2012-10-09 Thread Amit Nithian
What's preventing you from using the spell checker and take the #1
result and re-issue the query from a sub-class of the query component?
It should be reasonably fast to re-execute the query from the server
side since you are already within Solr. You can modify the response to
indicate that the new query was used so your client can display to the
user that it searched automatically for "milky.. click here for
searches for mlky" or something.

On Tue, Oct 9, 2012 at 8:46 AM, Ahmet Arslan  wrote:
>> I would like to ask if there are any ways to correct user's
>> queries
>> automatically? I know there is spellchecker which *suggests*
>> possible
>> correct words... The thing i wanna do is *automatically
>> fixing* those
>> queries and running instead of the original one
>
> not out of the box, you need to re-run suggestions at client side. There is 
> an commercial product though.
> http://sematext.com/products/dym-researcher/index.html


Re: SolrJ 4.0 Beta maxConnectionsPerHost

2012-10-09 Thread Chris Hostetter

: I did some digging and experimentation and found something interesting.
: When starting up the application, I see the following in Solr logs:
: Creating new http client, config:maxConnections=200&maxConnectionsPerHost=8
...
: It seems as though the maxConnections and maxConnectionsPerHost are not
: actually getting set. Anyone seen this problem or have an idea how to
: resolve?

To elaborate on sami's comment...

If you are seeing this in the logs from your solr *server*, it is unlikey 
that it has anything to do with the settings you are making on your solr 
*client*  this is probably related to the http client created inside 
solr for communicating with other solr nodes (ie: replication, solr cloud 
distributed updates, solr cloud peersync, etc...).  Which is different 
from the properties you set on the http client in your solr client 
application.

I believe there is a way to configure the defaults for the internal used 
http clients via solrconfig.xml, but off the top of my head i don't 
remember what that is.



-Hoss


Re: Installing Solr on a shared hosting server?

2012-10-09 Thread Otis Gospodnetic
Hi,

I don't know anything about Bluehost, but Solr works fine on EC2.

Otis
--
Search Analytics - http://sematext.com/search-analytics/index.html
Performance Monitoring - http://sematext.com/spm/index.html


On Tue, Oct 9, 2012 at 9:27 AM, caiod  wrote:
> I was wondering if I can install Solr on bluehost's shared hosting to use as
> a website search, and also how do I do so? Thank you...
>
>
>
> --
> View this message in context: 
> http://lucene.472066.n3.nabble.com/Installing-Solr-on-a-shared-hosting-server-tp4012708.html
> Sent from the Solr - User mailing list archive at Nabble.com.


Re: Solr Sharding Or Solr Cloud

2012-10-09 Thread Otis Gospodnetic
Hi,

If I had to make the choice and I knew index and query growth were
ahead of me I'd go with Solr 4. aka Solr Cloud.  Very general answer
to a very general question. :)

Otis
--
Search Analytics - http://sematext.com/search-analytics/index.html
Performance Monitoring - http://sematext.com/spm/index.html


On Tue, Oct 9, 2012 at 10:49 AM, Cool Techi  wrote:
> Hi,
>
> We have reached a state where a single solr server is not proving to be 
> enough and are exploring options of sharding or solr cloud.
>
> Just wanted to check the pros and cons of each one of them if there are any. 
> What would be the best size of the shards or
>
> Regards,
> R
>


RE: solr user group

2012-10-09 Thread Chris Hostetter

: And still on the list...

As Jack mentioned in his 18 Sep 2012 reply to your original email...

https://mail-archives.apache.org/mod_mbox/lucene-solr-user/201209.mbox/%3CD8AD75DD68FD45618D83C8CE3F93803E@JackKrupansky%3E

>> Did you send them from the exact same email address as the original 
>> subscriptions?
>> 
>> Did you follow all of the suggestions listed at the "Problems?" link on 
>> the discussions page?
>>
>> https://wiki.apache.org/solr/Unsubscribing%20from%20mailing%20lists

( Linked from: https://lucene.apache.org/solr/discussion.html )

In particular: if anyone has problems subscribing/unsubscribing, the 
method to contact a human (the list moderators) for help is 
solr-user-ow...@lucene.apache.org - but there is specific information you 
should proactively provide when contacting the moderators.



-Hoss


Re: Auto Correction?

2012-10-09 Thread Ahmet Arslan
> I would like to ask if there are any ways to correct user's
> queries
> automatically? I know there is spellchecker which *suggests*
> possible
> correct words... The thing i wanna do is *automatically
> fixing* those
> queries and running instead of the original one

not out of the box, you need to re-run suggestions at client side. There is an 
commercial product though. 
http://sematext.com/products/dym-researcher/index.html


Re: Problem with dataimporter.request

2012-10-09 Thread Ahmet Arslan
> In my data-config.xml, i have something like this
> 
>      deltaQuery="SELECT max(id) AS id from
> ${dataimporter.request.dataView}"
>     deltaImportQuery="SELECT * FROM tableX WHERE
> ${dataimporter.delta.id} < id">
> 
> 
> 
> However, everytime I execute delta-import
> (/dataimport?command=delta-import),
> it always gives me exception like this:
> 
> Caused by: java.lang.RuntimeException:

You need to pass it as a part of URL that you call. E.g. 
/dataimport?command=delta-import&dataView=something


Re: SolrJ Apidoc - is there any comprehensive literature ?

2012-10-09 Thread Erick Erickson
The "javadocs" link here: http://lucene.apache.org/solr/index.html
takes you to the API documentation...

if you go to the frames view and click on org.apache.client.solrj in the
upper left you'll see the solrj classes only

I'll often get the source and execute "ant documentation" to get
a local copy of all the javadocs. If you open
/solr/build/docs/index.html you get a nice top-level
local HTML page.

Other than that, you're really best off looking over some of the
code examples on the web. The javadocs can be kind of sparse
as well, any clarifications you'd like to contribute as patches would
be very welcome! There's often great value in someone seeing
the docs with fresh eyes...

Best
Erick

On Tue, Oct 9, 2012 at 5:56 AM, tomtom  wrote:
> Hi,
>
> is there a comprehensive documentation of the SolrJ Api?  The given
> resources are hard to read and have only few information. The Guide from
> lucid imagination gave me a certain progress but is just a well organized
> compilation of the apache documents. For programming there's just few help
> and the apidoc seems very spare.
>
> I wonder where the gracefully given hints in this forum (and others) are
> from. Is it really necessarey to inspect the source-code of Solr? Does
> anyone knows about "readable" and nearly complete documentation?
>
> A sometimes frustrated Solr-User/programmer.
>
> Thank you in advance, tom
>
>
>
> --
> View this message in context: 
> http://lucene.472066.n3.nabble.com/SolrJ-Apidoc-is-there-any-comprehensive-literature-tp4012673.html
> Sent from the Solr - User mailing list archive at Nabble.com.


Re: Indexing in Solr: invalid UTF-8

2012-10-09 Thread Gora Mohanty
On 9 October 2012 17:42, Patrick Oliver Glauner
 wrote:
> Hello everybody
>
> Meanwhile, I checked this issue in detail: we use pdftotext to extract text 
> from our PDFs (). Some generated text files contain 
> \u and \uD835.
>
> unicode(text, 'utf-8') does not throw any exception for these texts. 
> Subsequently, Solr throws an exception when these are sent to the indexer.

Off-topic, but this is because the Unicode escape sequence
'\u' is not being interpreted here. You have to explicitly
do that. Here is an example with '\u2018', the opening
quote (I did not have a font which covered '\ud835'). Please
note the difference between:
print unicode('\u2018')
\u2018

and

print unicode('\u2018').decode('unicode-escape')
‘
Regards,
Gora


Solr Sharding Or Solr Cloud

2012-10-09 Thread Cool Techi
Hi,

We have reached a state where a single solr server is not proving to be enough 
and are exploring options of sharding or solr cloud. 

Just wanted to check the pros and cons of each one of them if there are any. 
What would be the best size of the shards or 

Regards,
R
  

Re: Installing Solr on a shared hosting server?

2012-10-09 Thread Michael Della Bitta
Bluehost doesn't seem to support Java processes, so unfortunately the
answer seems to be no.

You might want to look into getting a Linode or some other similar VPS
hosting. Solr needs RAM to function well, though, so you're not going
to be able to go with the cheapest option.

Michael Della Bitta


Appinions
18 East 41st Street, 2nd Floor
New York, NY 10017-6271

www.appinions.com

Where Influence Isn’t a Game


On Tue, Oct 9, 2012 at 9:27 AM, caiod  wrote:
> I was wondering if I can install Solr on bluehost's shared hosting to use as
> a website search, and also how do I do so? Thank you...
>
>
>
> --
> View this message in context: 
> http://lucene.472066.n3.nabble.com/Installing-Solr-on-a-shared-hosting-server-tp4012708.html
> Sent from the Solr - User mailing list archive at Nabble.com.


Installing Solr on a shared hosting server?

2012-10-09 Thread caiod
I was wondering if I can install Solr on bluehost's shared hosting to use as
a website search, and also how do I do so? Thank you...



--
View this message in context: 
http://lucene.472066.n3.nabble.com/Installing-Solr-on-a-shared-hosting-server-tp4012708.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: How to use complicated boolean query?

2012-10-09 Thread Jack Krupansky
You also do the full query with one level of BooleanQuery by using the 
BooleanQuery#setMinimumNumberShouldMatch(int min) method. The default 
minimum number of optional clauses to match is zero, but you can set it to 1 
for your case. I don't know if there is any difference in performance or 
scoring between these two approaches.


Using Solr, it's the "mm" request parameter.

-- Jack Krupansky

-Original Message- 
From: neosky

Sent: Tuesday, October 09, 2012 12:43 AM
To: solr-user@lucene.apache.org
Subject: How to use complicated boolean query?

I want to implement a query in Solr like the follow
((X OR Y) AND Z) … ((X ∨ Y) ∧ Z)
BooleanQuery q = new BooleanQuery();
BooleanQuery inner = new BooleanQuery();
inner.add(X, Occur.SHOULD);
inner.add(Y, Occur.SHOULD);
q.add(inner, Occur.MUST);
q.add(Z, Occur.MUST);

For example, I want to query
(Field1=X or Field1=Y ) and Field2=Z

Thanks!



--
View this message in context: 
http://lucene.472066.n3.nabble.com/How-to-use-complicated-boolean-query-tp4012649.html
Sent from the Solr - User mailing list archive at Nabble.com. 



Re: SolrJ 4.0 Beta maxConnectionsPerHost

2012-10-09 Thread balaji.gandhi
Hi Briggs,

Did you try StreamingUpdateSolrServer? We are still facing the same issues
but noticed this happens only when "Physical Memory" in Solr Admin is high.
And yes SolrJ respects the connections.

Hi Team,

Another option we discussed was upgrading to HttpClient 4.2 in Solr war.
Would that fix the issue?

Thanks,
Balaji



--
View this message in context: 
http://lucene.472066.n3.nabble.com/SolrJ-4-0-Beta-maxConnectionsPerHost-tp4012630p4012712.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: Anyone using mmseg analyzer in solr multi core?

2012-10-09 Thread Peter Keegan
We're using MMSeg with Lucene, but not Solr. Since each SolrCore is
independent, I'm not sure how you can avoid each having a copy of the
dictionary, unless you modified MMSeg to use shared memory. Or, maybe I
missing something.

On Mon, Oct 8, 2012 at 3:37 AM, liyun  wrote:

> Hi all,
> Is anybody using mmseg analyzer for Chinese word analyze? When we use this
> in solr multi-core, I find it will load the dictionary per core and each
> core cost about 50MB memory. I think this is a big waste when our JVM has
> only 1GB memory…… Anyone have a good idea for handle this trouble ?
>
> 2012-10-08
>
>
>
> Li Yun
> Software Engineer @ Netease
> Mail: liyun2...@corp.netease.com
> MSN: rockiee...@gmail.com


Re: Help with Velocity in SolrItas

2012-10-09 Thread Erik Hatcher
As Paul mentioned, printing the class name helps tremendously.  The $response 
(SolrQueryResponse) can contain all sorts of nested objects of various types, 
so it's necessary to know what kind of object it is (generally, but not alway) 
to be able to navigate it effectively.

A shortcut, you can actually use $sentence.class

Velocity leverages naming conventions and treats things like "beans" even if 
they technically aren't, so if it has a getFoo() method, $object.foo will work 
to retrieve the value.  

I'm curious what kind of object you have for $sentence.  Often you'll get a 
NamedList, which has a get(Object) method.  Likewise with a HashMap.   

You may be able to get away with using $sentence.index and $sentence.text.  
Knowing the object type will point you to everything you can call on it.  And 
Velocity's documentation is here: 

 (geeky and maybe slightly tricky at first, but fairly straightforward once you 
know what object types you're dealing with)

Erik


On Oct 9, 2012, at 02:33 , Paul Libbrecht wrote:

> Lance,
> 
> this is the kind of fun that happens with Velocity all day long...
> 
> In general, when it outputs the variable name, it's the that the variable is 
> null; this can happen when a method is missing for example 
> There are actually effective uses of this 
> brain-dead-debugger-oriented-practice!
> 
> I would suppose that the class of your $sentence is not something that has a 
> get(String) method.
> With a normal debugging, this should be shown in a console.
> This is strengthened by the fact that your output of $sentence is not exactly 
> the same as a the output of  java.util.HashMap for example.
> 
> When in this situation, I generally make
>   Raw data: $sentence of class $sentence.getClass()
> (note: class is not a bean property, you need the method call)
> 
> Hope it helps.
> 
> Paul
> 
> PS: to stop this hell, I have a JSP pendant to the VelocityResponseWriter, is 
> this something of interest for someone so that I contribute it?
> 
> 
> 
> Le 9 oct. 2012 à 04:39, Lance Norskog a écrit :
> 
>> I am adding something to Solaritas, under /browse. One bit of Velocity
>> code does not unpack the result structure the way I think it should.
>> Please look at this- there is something I am missing about
>> tree-walking.
>> 
>> Here is the XML from a search result:
>> 
>> 
>>   0
>>   
>>   A bunch of words
>>   
>> 
>> more sentences 
>> 
>> 
>> Here is my Velocity code:
>> #foreach($sentence in $outer)
>>   Raw data: $sentence
>>   
>>   #set($index = $sentence.get('index'))
>>   #set($text = $sentence.get('text'))
>>   
>> Index: $index
>> 
>> Text: $text
>> 
>>   
>> #end
>> 
>> Here is the output:
>> 
>>   Raw data: sentence={index=0,text= A bunch of words}
>>   Index: $index
>>   Text: $text
> 



Re: Search in body

2012-10-09 Thread Tolga
I was expecting to be able to search in the body, but apparently I don't 
need it according to Markus.


Regards,

On 10/09/2012 03:27 PM, Rafał Kuć wrote:

Hello!

I assume you've added the body field, but you don't populate it. As
far as I remember Nutch don't fill the body field by default. What
you are expecting to have in the body field ?





RE: Search in body

2012-10-09 Thread Markus Jelsma
Hi - You should stick to Nutch' schema.xml and not manually add a text or body 
field that aren't going to be populated anyway. Nutch sends data, by default, 
to the content field.
  
-Original message-
> From:Rafał Kuć 
> Sent: Tue 09-Oct-2012 14:32
> To: solr-user@lucene.apache.org
> Subject: Re: Search in body
> 
> Hello!
> 
> I assume you've added the body field, but you don't populate it. As
> far as I remember Nutch don't fill the body field by default. What
> you are expecting to have in the body field ?
> 
> -- 
> Regards,
>  Rafał Kuć
>  Sematext :: http://sematext.com/ :: Solr - Lucene - Nutch - ElasticSearch
> 
> > I've just indexed again, and no luck.
> 
> > Below is my schema
> 
> > 
> >  
> >   > sortMissingLast="true"
> >  omitNorms="true"/>
> >   >  omitNorms="true" positionIncrementGap="0"/>
> >   > precisionStep="0"
> >  omitNorms="true" positionIncrementGap="0"/>
> >   >  omitNorms="true" positionIncrementGap="0"/>
> 
> >   >  positionIncrementGap="100">
> >  
> >  
> >   >  ignoreCase="true" words="stopwords.txt"/>
> >   >  generateWordParts="1" generateNumberParts="1"
> >  catenateWords="1" catenateNumbers="1" catenateAll="0"
> >  splitOnCaseChange="1"/>
> >  
> >   >  protected="protwords.txt"/>
> >   > class="solr.RemoveDuplicatesTokenFilterFactory"/>
> >  
> >  
> >   >  positionIncrementGap="100">
> >  
> >  
> >  
> >   >  generateWordParts="1" generateNumberParts="1"/>
> >  
> >  
> >  
> >  
> >  
> 
> >  
> >  
> >  
> >  
> 
> >  
> >  
> >   >  required="true"/>
> >  
> >  
> >  
> >  
> >  
> >  
> 
> >  
> >   >  multiValued="true"/>
> 
> >  
> >   >  multiValued="true"/>
> >   >  indexed="false"/>
> >   >  indexed="false"/>
> >  
> 
> >  
> >  
> 
> >  
> >   >  indexed="true" multiValued="true"/>
> 
> >  
> >  
> >   multiValued="true"/>>
> >  
> >   >  indexed="true"/>
> >   >  indexed="true"/>
> 
> >  
> >   >  multiValued="true"/>
> >  
> >  id
> >  content
> >  
> > 
> 
> > I don't know how to show you example data, my URL is 
> > http://www.sabanciuniv.edu
> 
> > Regards,
> 
> > On 10/09/2012 02:58 PM, Rafał Kuć wrote:
> >> Hello!
> >>
> >> After altering your schema.xml have you indexed your documents again ?
> >>
> >> It would be nice to see how you schema.xml looks like and example of
> >> the data, because otherwise we can only guess.
> >>
> 
> 


Re: solr1.4 code Example

2012-10-09 Thread Iwan Hanjoyo
you can download the code directly from here

http://www.solrenterprisesearchserver.com/
http://solrenterprisesearchserver.s3-website-us-east-1.amazonaws.com/downloads/5883-solr-enterprise-search2.zip

regards,


Hanjoyo


Re: Search in body

2012-10-09 Thread Rafał Kuć
Hello!

I assume you've added the body field, but you don't populate it. As
far as I remember Nutch don't fill the body field by default. What
you are expecting to have in the body field ?

-- 
Regards,
 Rafał Kuć
 Sematext :: http://sematext.com/ :: Solr - Lucene - Nutch - ElasticSearch

> I've just indexed again, and no luck.

> Below is my schema

> 
>  
>   sortMissingLast="true"
>  omitNorms="true"/>
>omitNorms="true" positionIncrementGap="0"/>
>   precisionStep="0"
>  omitNorms="true" positionIncrementGap="0"/>
>omitNorms="true" positionIncrementGap="0"/>

>positionIncrementGap="100">
>  
>  
>ignoreCase="true" words="stopwords.txt"/>
>generateWordParts="1" generateNumberParts="1"
>  catenateWords="1" catenateNumbers="1" catenateAll="0"
>  splitOnCaseChange="1"/>
>  
>protected="protwords.txt"/>
>   class="solr.RemoveDuplicatesTokenFilterFactory"/>
>  
>  
>positionIncrementGap="100">
>  
>  
>  
>generateWordParts="1" generateNumberParts="1"/>
>  
>  
>  
>  
>  

>  
>  
>  
>  

>  
>  
>required="true"/>
>  
>  
>  
>  
>  
>  

>  
>multiValued="true"/>

>  
>multiValued="true"/>
>indexed="false"/>
>indexed="false"/>
>  

>  
>  

>  
>indexed="true" multiValued="true"/>

>  
>  
>  >
>  
>indexed="true"/>
>indexed="true"/>

>  
>multiValued="true"/>
>  
>  id
>  content
>  
> 

> I don't know how to show you example data, my URL is 
> http://www.sabanciuniv.edu

> Regards,

> On 10/09/2012 02:58 PM, Rafał Kuć wrote:
>> Hello!
>>
>> After altering your schema.xml have you indexed your documents again ?
>>
>> It would be nice to see how you schema.xml looks like and example of
>> the data, because otherwise we can only guess.
>>



Re: Search in body

2012-10-09 Thread Tolga

I've just indexed again, and no luck.

Below is my schema



sortMissingLast="true"

omitNorms="true"/>

precisionStep="0"

omitNorms="true" positionIncrementGap="0"/>
























































multiValued="true"/>








id
content



I don't know how to show you example data, my URL is 
http://www.sabanciuniv.edu


Regards,

On 10/09/2012 02:58 PM, Rafał Kuć wrote:

Hello!

After altering your schema.xml have you indexed your documents again ?

It would be nice to see how you schema.xml looks like and example of
the data, because otherwise we can only guess.





Re: Search in body

2012-10-09 Thread Tolga

I had no idea I had to index again, thanks for the heads up.

On 10/09/2012 02:58 PM, Rafał Kuć wrote:

Hello!

After altering your schema.xml have you indexed your documents again ?

It would be nice to see how you schema.xml looks like and example of
the data, because otherwise we can only guess.





RE: Indexing in Solr: invalid UTF-8

2012-10-09 Thread Patrick Oliver Glauner
Hello everybody

Meanwhile, I checked this issue in detail: we use pdftotext to extract text 
from our PDFs (). Some generated text files contain \u 
and \uD835.

unicode(text, 'utf-8') does not throw any exception for these texts. 
Subsequently, Solr throws an exception when these are sent to the indexer.

Therefore, I wrote a little function to remove them after the unicode() call:
+def remove_invalid_solr_characters(utext):
+for char in CFG_SOLR_INVALID_CHAR_REPLACEMENTS:
+try:
+utext = utext.replace(char, 
CFG_SOLR_INVALID_CHAR_REPLACEMENTS[char])
+except:
+pass
+return utext

with:
+CFG_SOLR_INVALID_CHAR_REPLACEMENTS = {
+u'\u' : u"",
+u'\uD835' : u""
+}

This works well in our production environment.

Cheers, Patrick

From: Patrick Oliver Glauner [patrick.oliver.glau...@cern.ch]
Sent: Friday, September 28, 2012 10:36 AM
To: solr-user@lucene.apache.org
Subject: RE: Indexing in Solr: invalid UTF-8

Thank you. I will check our textification process and see how to improve it.

Patrick



From: Michael McCandless [luc...@mikemccandless.com]
Sent: Wednesday, September 26, 2012 5:45 PM
To: solr-user@lucene.apache.org
Subject: Re: Indexing in Solr: invalid UTF-8

Python's unicode function takes an optional (keyword) "errors"
argument, telling it what to do when an invalid UTF8 byte sequence is
seen.

The default (errors='strict') is to throw the exceptions you're
seeing.  But you can also pass errors='replace' or errors='ignore'.

See http://docs.python.org/howto/unicode.html for details ...

However, I agree with Robert: you should dig into why whatever process
you used to extract the full text from your binary documents is
producing invalid UTF-8 ... something is wrong with that process.

Mike McCandless

http://blog.mikemccandless.com

On Tue, Sep 25, 2012 at 10:44 PM, Robert Muir  wrote:
> On Tue, Sep 25, 2012 at 2:02 PM, Patrick Oliver Glauner
>  wrote:
>> Hi
>> Thanks. But I see that 0xd835 is missing in this list (see my exceptions).
>>
>> What's the best way to get rid of all of them in Python? I am new to unicode 
>> in Python but I am sure that this use case is quite frequent.
>>
>
> I don't really know python either: so I could be wrong here but are
> you just taking these binary .PDF and .DOC files and treating them as
> UTF-8 text and sending them to Solr?
>
> If so, I don't think that will work very well. Maybe instead try
> parsing these binary files with something like Tika to get at the
> actual content and send that? (it seems some people have developed
> python integration for this, e.g.
> http://redmine.djity.net/projects/pythontika/wiki)


Re: Wildcards and fuzzy/phonetic query

2012-10-09 Thread Erick Erickson
It won't crash Solr if you include it, but it probably won't do what
you expect either due to how wildcards are expanded.

And it gets worse. DoubleMetaphone tries to reduce what it
analyzes, well, phonetically with "close" letters (or multiple
choices). Some phonetic filters change to fixed 4 letter
combinations as I remember. Some hash to a completely different
string. Some

About combining fuzzy and wildcard. I haven't thought it through,
but it strikes me as fraught with unexpected results. Consider
har* and treating it as a fuzzy match. How would you calculate
the "fuzziness" of "hardiness" and "harp"? Would you consider
"her" a fuzzy match? How about "farther"? or even "father"?

You might be able to do something interesting with EdgeNgram
here though but it still seems like it's going to either explode
computationally or produce results that don't really mean much.
But I'm mostly speculating here

Frankly, though, I'd do what Jan suggests. Try it out and see if
it's "good enough". Especially pin down the use cases. Often
requirements like this are specified by someone who, when
presented with the results of what you can do easily, decide the
effort could best be spent somewhere else.

Because this whole approach will only increase the number of
documents that are found as the result of a search without
necessarily increasing the relevance of the top N docs on the
first page. Users rarely go to the second page, and often don't
even look past the first few results. Doing wildcard AND fuzzy
queries would likely result in something useful  a very small
percentage of the time. But that's just a guess.

Best
Erick


On Tue, Oct 9, 2012 at 5:54 AM, Haagen Hasle  wrote:
>
> I used the admin/analysis page (great tip, I had never used it before - thank 
> you!) and it seems to me that the DoubleMetaphone filter converts "Hågen" to 
> both "JN" and "KN".  Will that crash the Solr analysis if I try to include 
> this filter in the multiterm-analysis?
>
> Do you know where I can find out more about combining wildcard and fuzzy in 
> the same query?  When you say you don't think it is possible, do you mean it 
> is not implemented in Solr today, or it can't be implemented because it is 
> technically impossible or functionally doesn't make sense? :)
>
> I wrote in an answer to Otis that I'd like to try to combine fuzzy with Ngram 
> as well.  Do you know if that is possible and makes any sense?
>
>
> Thanks to everyone for quick and good answers, I really appreciate it!
>
>
> Regards, Hågen
>
> Den 8. okt. 2012 kl. 21:35 skrev Erick Erickson:
>
>> To answer your first question, yes, you've got it right. If you define
>> a multiterm section in your fieldType, whatever you put in that section
>> gets applied whether the underlying class is MultiTermAware or not.
>> Which means you can shoot yourself in the foot really bad ...
>>
>> (…)
>>
>> Fuzzy searches + wildcards. I don't think you can do that reasonably, but
>> I'm not entirely sure.
>>
>> Best
>> Erick
>


Re: Search in body

2012-10-09 Thread Rafał Kuć
Hello!

After altering your schema.xml have you indexed your documents again ?

It would be nice to see how you schema.xml looks like and example of
the data, because otherwise we can only guess.

-- 
Regards,
 Rafał Kuć
 Sematext :: http://sematext.com/ :: Solr - Lucene - Nutch - ElasticSearch

> Hi,

> My previous schema didn't have the body defined as field, so I did and
> searched for "body:Smyrna", and no results turned up. What am I doing wrong?

> Regards,



RE: solr user group

2012-10-09 Thread David Hill

And still on the list...

David Hill

Iowa Student Loan | Lead Software Analyst / Developer | phone 515-273-7241 | 
fax 515-273-7241 | dh...@studentloan.org


-Original Message-
From: David Hill
Sent: Tuesday, September 18, 2012 6:58 AM
To: 'solr-user@lucene.apache.org'
Subject: solr user group


sorry for the broadcast, but the solr list server is just not taking the hint 
yet, I have issued the following commands on the following dates:

Sent Mon 08/27/2012 10:37 PM to 'solr-user-unsubscr...@lucene.apache.org' 
subject = unsubscribe

Sent Mon 07/16/2012 6:53 AM to 'solr-user-unsubscr...@lucene.apache.org'

Sent Mon 04/23/2012 8:01 AM to 'solr-user-unsubscr...@lucene.apache.org' 
subject = unsubscribe

David Hill

Iowa Student Loan | Lead Software Analyst / Developer | phone 515-273-7241 | 
fax 515-273-7241 | dh...@studentloan.org



This e-mail and any files transmitted with it are confidential and intended 
solely for the use of the individual or entity to whom they are addressed. If 
you have received this e-mail in error please notify the originator of the 
message. This footer also confirms that this e-mail message has been scanned 
for the presence of computer viruses. Any views expressed in this message are 
those of the individual sender, except where the sender specifies and with 
authority, states them to be the views of Iowa Student Loan.



 



Search in body

2012-10-09 Thread Tolga

Hi,

My previous schema didn't have the body defined as field, so I did and 
searched for "body:Smyrna", and no results turned up. What am I doing wrong?


Regards,


Re: Wildcards and fuzzy/phonetic query

2012-10-09 Thread Jan Høydahl
Hi,

Also be sure to check out the new BeiderMorse phonetic: 
http://wiki.apache.org/solr/AnalyzersTokenizersTokenFilters#solr.BeiderMorseFilterFactory
 which handles middle eastern and eastern european names very well.

Phonetic algorithms use tons of rules for how to transform an input depending 
on what comes before and after, so I don't believe you'll get wildcard to work 
perfectly combined with phoenetic since Solr cannot guess what shuould come 
next. But you may get it to work for many cases, the best is simply to try it 
out. Use EdgeNgram followed by some phonetic ant try.

You may also be interested in a MeetUp talk held in Oslo last month: 
http://www.meetup.com/Oslo-Solr-Community/events/67253692/ You'll find the link 
to Mats' talk about Norwegian phonetics if you scroll down that page.

--
Jan Høydahl, search solution architect
Cominvent AS - www.cominvent.com
Solr Training - www.solrtraining.com

9. okt. 2012 kl. 11:54 skrev Haagen Hasle :

> 
> I used the admin/analysis page (great tip, I had never used it before - thank 
> you!) and it seems to me that the DoubleMetaphone filter converts "Hågen" to 
> both "JN" and "KN".  Will that crash the Solr analysis if I try to include 
> this filter in the multiterm-analysis?
> 
> Do you know where I can find out more about combining wildcard and fuzzy in 
> the same query?  When you say you don't think it is possible, do you mean it 
> is not implemented in Solr today, or it can't be implemented because it is 
> technically impossible or functionally doesn't make sense? :)  
> 
> I wrote in an answer to Otis that I'd like to try to combine fuzzy with Ngram 
> as well.  Do you know if that is possible and makes any sense?
> 
> 
> Thanks to everyone for quick and good answers, I really appreciate it!
> 
> 
> Regards, Hågen
> 
> Den 8. okt. 2012 kl. 21:35 skrev Erick Erickson:
> 
>> To answer your first question, yes, you've got it right. If you define
>> a multiterm section in your fieldType, whatever you put in that section
>> gets applied whether the underlying class is MultiTermAware or not.
>> Which means you can shoot yourself in the foot really bad ...
>> 
>> (…)
>> 
>> Fuzzy searches + wildcards. I don't think you can do that reasonably, but
>> I'm not entirely sure.
>> 
>> Best
>> Erick
> 



SolrJ Apidoc - is there any comprehensive literature ?

2012-10-09 Thread tomtom
Hi,

is there a comprehensive documentation of the SolrJ Api?  The given
resources are hard to read and have only few information. The Guide from
lucid imagination gave me a certain progress but is just a well organized
compilation of the apache documents. For programming there's just few help
and the apidoc seems very spare.   

I wonder where the gracefully given hints in this forum (and others) are
from. Is it really necessarey to inspect the source-code of Solr? Does
anyone knows about "readable" and nearly complete documentation? 

A sometimes frustrated Solr-User/programmer.

Thank you in advance, tom



--
View this message in context: 
http://lucene.472066.n3.nabble.com/SolrJ-Apidoc-is-there-any-comprehensive-literature-tp4012673.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: Wildcards and fuzzy/phonetic query

2012-10-09 Thread Haagen Hasle

I used the admin/analysis page (great tip, I had never used it before - thank 
you!) and it seems to me that the DoubleMetaphone filter converts "Hågen" to 
both "JN" and "KN".  Will that crash the Solr analysis if I try to include this 
filter in the multiterm-analysis?

Do you know where I can find out more about combining wildcard and fuzzy in the 
same query?  When you say you don't think it is possible, do you mean it is not 
implemented in Solr today, or it can't be implemented because it is technically 
impossible or functionally doesn't make sense? :)  

I wrote in an answer to Otis that I'd like to try to combine fuzzy with Ngram 
as well.  Do you know if that is possible and makes any sense?


Thanks to everyone for quick and good answers, I really appreciate it!


Regards, Hågen

Den 8. okt. 2012 kl. 21:35 skrev Erick Erickson:

> To answer your first question, yes, you've got it right. If you define
> a multiterm section in your fieldType, whatever you put in that section
> gets applied whether the underlying class is MultiTermAware or not.
> Which means you can shoot yourself in the foot really bad ...
> 
> (…)
> 
> Fuzzy searches + wildcards. I don't think you can do that reasonably, but
> I'm not entirely sure.
> 
> Best
> Erick



Re: Help with Velocity in SolrItas

2012-10-09 Thread jmlucjav
Paul Libbrecht-4 wrote
> PS: to stop this hell, I have a JSP pendant to the VelocityResponseWriter,
> is this something of interest for someone so that I contribute it?

Paul...yes it is! Anything that would help velocity related issues is
welcome



--
View this message in context: 
http://lucene.472066.n3.nabble.com/Help-with-Velocity-in-SolrItas-tp4012636p4012668.html
Sent from the Solr - User mailing list archive at Nabble.com.


Auto Correction?

2012-10-09 Thread deniz
Hi all,

I would like to ask if there are any ways to correct user's queries
automatically? I know there is spellchecker which *suggests* possible
correct words... The thing i wanna do is *automatically fixing* those
queries and running instead of the original one

for example
if our query is "mlky", it should automatically search for "milky"

there is synonyms filter for these, but it has to be pre populated (correct
me if i am wrong)  so it wont be that useful for my case...

is there any way to do this without suggesting-only and synonym filters? if
so please give me some details 



-
Zeki ama calismiyor... Calissa yapar...
--
View this message in context: 
http://lucene.472066.n3.nabble.com/Auto-Correction-tp4012666.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: How to use complicated boolean query?

2012-10-09 Thread Alexandre Rafalovitch
And the problem is? I think it is just brackets. E.g.
Field1:(X Y) AND Field2:Z

No?
 On 9 Oct 2012 10:14, "neosky"  wrote:

> I want to implement a query in Solr like the follow
> ((X OR Y) AND Z) … ((X ∨ Y) ∧ Z)
> BooleanQuery q = new BooleanQuery();
> BooleanQuery inner = new BooleanQuery();
> inner.add(X, Occur.SHOULD);
> inner.add(Y, Occur.SHOULD);
> q.add(inner, Occur.MUST);
> q.add(Z, Occur.MUST);
>
> For example, I want to query
> (Field1=X or Field1=Y ) and Field2=Z
>
> Thanks!
>
>
>
> --
> View this message in context:
> http://lucene.472066.n3.nabble.com/How-to-use-complicated-boolean-query-tp4012649.html
> Sent from the Solr - User mailing list archive at Nabble.com.
>