Not Able to Build Spellcheck index - SpellCheckComponent.prepare 500 Error

2014-04-05 Thread sameer
Hi
I am trying to use spellcheck in solr with below config but it throwing with
error while using spellcheck build or reload

it works fine otherwise for indexed search, can someone please help
implementing spellcheck corectly

schema.xml:

// fieldType declaration
fieldType name=textSpell class=solr.TextField
positionIncrementGap=100
  analyzer
tokenizer class=solr.StandardTokenizerFactory/
filter class=solr.LowerCaseFilterFactory/
filter class=solr.StopFilterFactory ignoreCase=true
words=misspelled_words.txt/
filter class=solr.PatternReplaceFilterFactory pattern=([^a-z])
replacement= replace=all/
filter class=solr.LengthFilterFactory min=2 max=50/
  /analyzer
/fieldType

//field name
field name=spell type=textSpell indexed=true stored=true
multiValued=true/

//copyFields
copyField source=dealName dest=spell/
copyField source=dealUri dest=spell/
copyField source=merchantName dest=spell/
copyField source=categoryName dest=spell/
copyField source=dealTitle dest=spell/
copyField source=highlights dest=spell/
copyField source=seoTags dest=spell/
copyField source=description dest=spell/
copyField source=text dest=spell/


solrconfig.xml:

//searchComponent

searchComponent name=spellcheck class=solr.SpellCheckComponent
str name=queryAnalyzerFieldTypetextSpell/str
lst name=spellchecker
  str name=classnamesolr.IndexBasedSpellChecker/str
  str name=namedefault/str
  str name=spellcheckIndexDir./spellchecker/str
  str
name=fieldcategoryName,dealName,seoTags,description,dealTitle,merchantName,dealUri,highlights/str
  str name=buildOnCommittrue/str
  str name=accuracy0.9/str
/lst
/searchComponent


//default requestHandler

  requestHandler name=standard class=solr.StandardRequestHandler
default=true

 lst name=defaults
str name=echoParamsexplicit/str
str name=spellchecktrue/str
str name=spellcheck.dictionarydirect/str
str name=spellcheckon/str
str name=spellcheck.extendedResultstrue/str
str name=spellcheck.count5/str
str name=spellcheck.collatetrue/str
str name=spellcheck.collateExtendedResultstrue/str
   
 /lst
 arr name=last-components
strspellcheck/str
 /arr
  /requestHandler


// URL params
select?q=*%3A*wt=phpindent=truespellcheck=truespellcheck.build=true


//output


array(
  'responseHeader'=array(
'status'=500,
'QTime'=4,
'params'=array(
  'spellcheck'='true',
  'indent'='true',
  'q'='*:*',
  '_'='1396684768649',
  'wt'='php',
  'spellcheck.build'='true')),
  'error'=array(
'trace'='java.lang.NullPointerException
at
org.apache.solr.handler.component.SpellCheckComponent.prepare(SpellCheckComponent.java:125)
at
org.apache.solr.handler.component.SearchHandler.handleRequestBody(SearchHandler.java:187)
at
org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:135)
at org.apache.solr.core.SolrCore.execute(SolrCore.java:1859)
at
org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:710)
at
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:413)
at
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:197)
at
org.mortbay.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1212)
at 
org.mortbay.jetty.servlet.ServletHandler.handle(ServletHandler.java:399)
at
org.mortbay.jetty.security.SecurityHandler.handle(SecurityHandler.java:216)
at 
org.mortbay.jetty.servlet.SessionHandler.handle(SessionHandler.java:182)
at 
org.mortbay.jetty.handler.ContextHandler.handle(ContextHandler.java:767)
at org.mortbay.jetty.webapp.WebAppContext.handle(WebAppContext.java:450)
at
org.mortbay.jetty.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:230)
at
org.mortbay.jetty.handler.HandlerCollection.handle(HandlerCollection.java:114)
at 
org.mortbay.jetty.handler.HandlerWrapper.handle(HandlerWrapper.java:152)
at org.mortbay.jetty.handler.DebugHandler.handle(DebugHandler.java:77)
at 
org.mortbay.jetty.handler.HandlerWrapper.handle(HandlerWrapper.java:152)
at org.mortbay.jetty.Server.handle(Server.java:326)
at 
org.mortbay.jetty.HttpConnection.handleRequest(HttpConnection.java:542)
at
org.mortbay.jetty.HttpConnection$RequestHandler.headerComplete(HttpConnection.java:928)
at org.mortbay.jetty.HttpParser.parseNext(HttpParser.java:549)
at org.mortbay.jetty.HttpParser.parseAvailable(HttpParser.java:212)
at org.mortbay.jetty.HttpConnection.handle(HttpConnection.java:404)
at
org.mortbay.io.nio.SelectChannelEndPoint.run(SelectChannelEndPoint.java:410)
at
org.mortbay.thread.QueuedThreadPool$PoolThread.run(QueuedThreadPool.java:582)
',
'code'=500))



--
View this message in context: 

Re: Not Able to Build Spellcheck index - SpellCheckComponent.prepare 500 Error

2014-04-05 Thread Alexandre Rafalovitch
It's a null pointer exception. Either something is not defined
correctly or you are hitting a funny unexpected.

Which version of Solr is it?

Regards,
   Alex.
Personal website: http://www.outerthoughts.com/
Current project: http://www.solr-start.com/ - Accelerating your Solr proficiency


On Sat, Apr 5, 2014 at 3:24 PM, sameer sameer.sha...@trideal.in wrote:
 Hi
 I am trying to use spellcheck in solr with below config but it throwing with
 error while using spellcheck build or reload


Re: Full Indexing is Causing a Java Heap Out of Memory Exception

2014-04-05 Thread Ahmet Arslan
Hi,

Now we have a more informative error : 
org.apache.solr.handler.dataimport.DataImportHandlerException: 
java.lang.OutOfMemoryError: Java heap space

Caused by: org.apache.solr.handler.dataimport.DataImportHandlerException: 
java.lang.OutOfMemoryError: Java heap space
        at 
org.apache.solr.handler.dataimport.DocBuilder.buildDocument(DocBuilder.java:535)
        at 
org.apache.solr.handler.dataimport.DocBuilder.buildDocument(DocBuilder.java:404)

1) Does this happen when you increase -Xmx64m -Xms64m ?

2) I see you use custom jars called MDSolrDIHTransformer JARs inside  But I 
don't see any Transformers used in database.xm, why is that. I would remove 
them just to be sure.

3) I see you have org.apache.solr.core.StandardDirectoryFactory declared in 
sorlconfig. Assuming you are using, 64 bit windows, it is recommended to use 
MMap 
http://blog.thetaphi.de/2012/07/use-lucenes-mmapdirectory-on-64bit.html


4) In your previous mail you had batch size set, now there is not batchSize 
defined in database.xml. For MySQL it is recommended to use -1. Not sure about 
oracle, I personally used 10,000 once for Oracle. 
http://wiki.apache.org/solr/DataImportHandlerFaq#I.27m_using_DataImportHandler_with_a_MySQL_database._My_table_is_huge_and_DataImportHandler_is_going_out_of_memory._Why_does_DataImportHandler_bring_everything_to_memory.3F

You have a lot of copyFields defined. There could be some gotchas when handling 
unusually much copy fields. I would really try CSV option here. Given that you 
have only full import SQL defined and it is not a complex one. It queries only 
one table. I believe Oracle has some tool to export a table to CSV file 
efficiently.

On Saturday, April 5, 2014 3:05 AM, Candygram For Mongo 
candygram.for.mo...@gmail.com wrote:
 
Does this user list allow attachments?  I have four files attached 
(database.xml, error.txt, schema.xml, solrconfig.xml).  We just ran the process 
again using the parameters you suggested, but not to a csv file.  It errored 
out quickly.  We are working on the csv file run.

Removed both autoCommit and
autoSoftCommit parts/definitions from solrconfig.xml

Disabled tlog by removing
   updateLog
      str
name=dir${solr.ulog.dir:}/str
    /updateLog

from solrconfig.xml

Used commit=true parameter.
?commit=truecommand=full-import




On Fri, Apr 4, 2014 at 3:29 PM, Ahmet Arslan iori...@yahoo.com wrote:

Hi,

This may not solve your problem but generally it is recommended to disable 
auto commit and transaction logs for bulk indexing.
And issue one commit at the very end. Do you tlogs enabled? I see commit 
failed in the error message thats why I am offering this.

And regarding comma separated values, with this approach you focus on just 
solr importing process. You separate data acquisition phrase. And it is very 
fast load even big csv files  http://wiki.apache.org/solr/UpdateCSV
I have never experienced OOM during indexing, I suspect data acquisition has 
role in it.

Ahmet


On Saturday, April 5, 2014 1:18 AM, Candygram For Mongo 
candygram.for.mo...@gmail.com wrote:

We would be happy to try that.  That sounds counter intuitive for the high 
volume of records we have.  Can you help me understand how that might solve 
our problem?




On Fri, Apr 4, 2014 at 2:34 PM, Ahmet Arslan iori...@yahoo.com wrote:

Hi,

Can you remove auto commit for bulk import. Commit at the very end?

Ahmet




On Saturday, April 5, 2014 12:16 AM, Candygram For Mongo 
candygram.for.mo...@gmail.com wrote:
In case the attached database.xml file didn't show up, I have pasted in the
contents below:

dataConfig
dataSource
name=org_only
type=JdbcDataSource
driver=oracle.jdbc.OracleDriver
url=jdbc:oracle:thin:@test2.abc.com:1521:ORCL
user=admin
password=admin
readOnly=false
batchSize=100
/
document


entity name=full-index query=
select

NVL(cast(ORCL.ADDRESS_ACCT_ALL.RECORD_ID as varchar2(100)), 'null')
as SOLR_ID,

'ORCL.ADDRESS_ACCT_ALL'
as SOLR_CATEGORY,

NVL(cast(ORCL.ADDRESS_ACCT_ALL.RECORD_ID as varchar2(255)), ' ') as
ADDRESSALLROWID,
NVL(cast(ORCL.ADDRESS_ACCT_ALL.ADDR_TYPE_CD as varchar2(255)), ' ') as
ADDRESSALLADDRTYPECD,
NVL(cast(ORCL.ADDRESS_ACCT_ALL.LONGITUDE as varchar2(255)), ' ') as
ADDRESSALLLONGITUDE,
NVL(cast(ORCL.ADDRESS_ACCT_ALL.LATITUDE as varchar2(255)), ' ') as
ADDRESSALLLATITUDE,
NVL(cast(ORCL.ADDRESS_ACCT_ALL.ADDR_NAME as varchar2(255)), ' ') as
ADDRESSALLADDRNAME,
NVL(cast(ORCL.ADDRESS_ACCT_ALL.CITY as varchar2(255)), ' ') as
ADDRESSALLCITY,
NVL(cast(ORCL.ADDRESS_ACCT_ALL.STATE as varchar2(255)), ' ') as
ADDRESSALLSTATE,
NVL(cast(ORCL.ADDRESS_ACCT_ALL.EMAIL_ADDR as varchar2(255)), ' ') as
ADDRESSALLEMAILADDR

from ORCL.ADDRESS_ACCT_ALL
 

field column=SOLR_ID name=id /
field column=SOLR_CATEGORY name=category /
field column=ADDRESSALLROWID name=ADDRESS_ACCT_ALL.RECORD_ID_abc /
field column=ADDRESSALLADDRTYPECD
name=ADDRESS_ACCT_ALL.ADDR_TYPE_CD_abc /
field column=ADDRESSALLLONGITUDE name=ADDRESS_ACCT_ALL.LONGITUDE_abc /
field column=ADDRESSALLLATITUDE 

Re: Not Able to Build Spellcheck index - SpellCheckComponent.prepare 500 Error

2014-04-05 Thread sameer
its solr-4.6.0



--
View this message in context: 
http://lucene.472066.n3.nabble.com/Not-Able-to-Build-Spellcheck-index-SpellCheckComponent-prepare-500-Error-tp4129368p4129392.html
Sent from the Solr - User mailing list archive at Nabble.com.


Combining eDismax and SpellChecker

2014-04-05 Thread S.L
Hi All,

I want to suggest the correct phrase if a typo is made while searching and
then search it using eDismax parser(pf,pf2,pf3), if no typo is made then
search it using eDismax parser alone.

Is there a way I can combine these two components , I have seen examples
for eDismax and also for SpellChecker , but nothing that combines these two
together.

Can you please let me know ?

Thanks.


Re: How to see the value of long type (solr) ?

2014-04-05 Thread Jack Krupansky

One of the Lucene guys is going to need to address this question.

I d know that Trie fields index additional values to support fast range 
queries, so maybe you are merely seeing some of those generated values, and 
if you look further you should see your actual indexed value. What exactly 
are you seeing in Luke for this TrieLongField?


-- Jack Krupansky

-Original Message- 
From: Lisheng Zhang

Sent: Friday, April 4, 2014 12:44 PM
To: solr-user@lucene.apache.org
Subject: How to see the value of long type (solr) ?

Hi,

We use solr 3.6 to index a field of long type:

fieldType name=long class=solr.TrieLongField ...

Now for debugging purpose we need to see the original value (the field is
not stored), but in luke we cannot see.

1/ is there a way to see original long type value (using luke or not) ?
2/ if we need to use lucene to search this field, what analyzer should we
use ?

Thanks very much for helps, Lisheng 



Re: Combining eDismax and SpellChecker

2014-04-05 Thread Ahmet Arslan
There is one commercial solution 
http://www.sematext.com/products/dym-researcher/index.html



On Saturday, April 5, 2014 4:07 PM, S.L simpleliving...@gmail.com wrote:
Hi All,

I want to suggest the correct phrase if a typo is made while searching and
then search it using eDismax parser(pf,pf2,pf3), if no typo is made then
search it using eDismax parser alone.

Is there a way I can combine these two components , I have seen examples
for eDismax and also for SpellChecker , but nothing that combines these two
together.

Can you please let me know ?

Thanks.



Re: How to see the value of long type (solr) ?

2014-04-05 Thread Ahmet Arslan
You can use faceting to human readable values.





On Saturday, April 5, 2014 7:08 PM, Jack Krupansky j...@basetechnology.com 
wrote:
One of the Lucene guys is going to need to address this question.

I d know that Trie fields index additional values to support fast range 
queries, so maybe you are merely seeing some of those generated values, and 
if you look further you should see your actual indexed value. What exactly 
are you seeing in Luke for this TrieLongField?

-- Jack Krupansky


-Original Message- 
From: Lisheng Zhang
Sent: Friday, April 4, 2014 12:44 PM
To: solr-user@lucene.apache.org
Subject: How to see the value of long type (solr) ?

Hi,

We use solr 3.6 to index a field of long type:

fieldType name=long class=solr.TrieLongField ...

Now for debugging purpose we need to see the original value (the field is
not stored), but in luke we cannot see.

1/ is there a way to see original long type value (using luke or not) ?
2/ if we need to use lucene to search this field, what analyzer should we
use ?

Thanks very much for helps, Lisheng 


Re: Searching multivalue fields.

2014-04-05 Thread Ahmet Arslan
Hi,

Did restart solr and you re-index after schema change?
On Saturday, April 5, 2014 2:39 AM, Vijay Kokatnur kokatnur.vi...@gmail.com 
wrote:
 
I had already tested with omitTermFreqAndPositions=false .  I still got the 
same error.

Is there something that I am overlooking?


On Fri, Apr 4, 2014 at 2:45 PM, Ahmet Arslan iori...@yahoo.com wrote:

Hi Vijay,


Add omitTermFreqAndPositions=false  attribute to fieldType definitions.


    fieldType name=string class=solr.StrField 
omitTermFreqAndPositions=false sortMissingLast=true /



   fieldType name=int class=solr.TrieIntField 
omitTermFreqAndPositions=false precisionStep=0 positionIncrementGap=0/


You don't need termVectors  for this.


   1.2: omitTermFreqAndPositions attribute introduced, true by default 
            except for text fields.


And please reply to solr user mail, so others can use the threat later on.

Ahmet
On Saturday, April 5, 2014 12:18 AM, Vijay Kokatnur kokatnur.vi...@gmail.com 
wrote:
 
Hey Ahmet, 


Sorry it took some time to test this.  But schema definition seem to conflict 
with SpanQuery.  I get following error when I use Spans 


field OrderLineType was indexed without position data; cannot run 
SpanTermQuery (term=11)

I changed field definition in the schema but can't find the right attribute to 
set this.  My last attempt was with following definition 


   field name=OrderLineType type=string indexed=true stored=true 
multiValued=true termVectors=true termPositions=true termOffsets=true/


 Any ideas what I am doing wrong?


Thanks,
-Vijay


On Wed, Mar 26, 2014 at 1:54 PM, Ahmet Arslan iori...@yahoo.com wrote:

Hi Vijay,

After reading the documentation it seems that following query is what you are 
after. It will return OrderId:345 without matching OrderId:123

SpanQuery q1  = new SpanTermQuery(new Term(BookingRecordId, 234));
SpanQuery q2  = new SpanTermQuery(new Term(OrderLineType, 11));
SpanQuery q2m new FieldMaskingSpanQuery(q2, BookingRecordId);
Query q = new SpanNearQuery(new SpanQuery[]{q1, q2m}, -1, false);

Ahmet




On Wednesday, March 26, 2014 10:39 PM, Ahmet Arslan iori...@yahoo.com wrote:
Hi Vijay,

I personally don't understand joins very well. Just a guess may be 
FieldMaskingSpanQuery could be used?

http://blog.griddynamics.com/2011/07/solr-experience-search-parent-child.html


Ahmet




On Wednesday, March 26, 2014 9:46 PM, Vijay Kokatnur 
kokatnur.vi...@gmail.com wrote:
Hi,

I am bumping this thread again one last time to see if anyone has a
solution.

In it's current state, our application is storing child items as multivalue
fields.  Consider some orders, for example -


{
OrderId:123
BookingRecordId : [145, 987, *234*]
OrderLineType : [11, 12, *13*]
.
}
{
OrderId:345
BookingRecordId : [945, 882, *234*]
OrderLineType : [1, 12, *11*]
.
}
{
OrderId:678
BookingRecordId : [444]
OrderLineType : [11]
.
}


Here, If you look up for an Order with BookingRecordId: 234 And
OrderLineType:11.  You will get two orders with orderId : 123 and 345,
which is correct.  You have two arrays in both the orders that satisfy this
condition.

However, for OrderId:123, the value at 3rd index of OrderLineType array is
13 and not 11( this is for OrderId:345).  So orderId 123 should be
excluded. This is what I am trying to achieve.

I got some suggestions from a solr-user to use FieldsCollapsing, Join,
Block-join or string concatenation.  None of these approaches can be used
without re-indexing schema.

Has anyone found a non-invasive solution for this?

Thanks,

-Vijay






Re: Does sorting skip everything having to do with relevancy?

2014-04-05 Thread Mikhail Khludnev
Shawn,
I suppose e yields syntax error. Therefore, this case doesn't prove
anything yet.
Haven't you tried sqrt(-1) or log(-1) ?


On Sat, Apr 5, 2014 at 1:47 AM, Shawn Heisey s...@elyograg.org wrote:

 On 4/4/2014 3:13 PM, Mikhail Khludnev wrote:

 I suppose SolrIndexSearcher.buildTopDocsCollector() doesn't create a
 Collector which calls score() in this case. Hence, it shouldn't waste CPU.
 Just my impression.
 Haven't you tried to check it supplying some weird formula, which throws
 exception?


 I didn't think of that.  That's a good idea -- as long as there's not
 independent code that checks the function in addition to the code that
 actually runs it.

 With the following parameters added to an edismax query that otherwise
 works, I get an exception.  It works if I change the e to 5.

 sort=registered_date ascboost=sum(5,e)

 I will take Alvaro's suggestion and add boost=1 to queries that use a
 sort parameter.  It's probably a good idea to file that Jira.

 Thanks,
 Shawn




-- 
Sincerely yours
Mikhail Khludnev
Principal Engineer,
Grid Dynamics

http://www.griddynamics.com
 mkhlud...@griddynamics.com


Re: SOLR Jetty Server on Windows 2003

2014-04-05 Thread Jack Krupansky

Is the URL for the Solr request absolutely 100% identical in both cases?

By not getting a response, do you mean it hangs andtimes out or that the 
response is empty?


-- Jack Krupansky

-Original Message- 
From: EXTERNAL Taminidi Ravi (ETI, Automotive-Service-Solutions)

Sent: Friday, April 4, 2014 11:49 AM
To: solr-user@lucene.apache.org
Subject: SOLR Jetty Server on Windows 2003

Hi , I am trying to install solr on the Windows 2003 with Jetty server. Form 
browser everything works , but when I try to acesss from another javascript 
Code in other machine I am not getting reponse. I am using Xmlhttprequest to 
get the response from server using javascript.


Any Help...?


--Ravi 



Re: Strange behavior of edismax and mm=0 with long queries (bug?)

2014-04-05 Thread Jack Krupansky
Set the q.op parameter to OR and set mm=10% or something like that. The idea is 
to not excessively restrict the documents that will match, but weight the 
matched results based on how many word pairs and triples do match.

In addition, use the pf parameter to provide extra weight when the full query 
term phrase matches exactly.

-- Jack Krupansky

From: Nils Kaiser 
Sent: Friday, April 4, 2014 10:10 AM
To: solr-user@lucene.apache.org 
Subject: Strange behavior of edismax and mm=0 with long queries (bug?)

Hey, 

I am currently using solr to recognize songs and people from a list of user 
comments. My index stores the titles of the songs. At the moment my application 
builds word ngrams and fires a search with that query, which works well but is 
quite inefficient.

So my thought was to simply use the collated comments as query. So it is a case 
where the query is much longer. I need to use mm=0 or mm=1.

My plan was to use edismax as the pf2 and pf3 parameters should work well for 
my usecase.

However when using longer queries, I get a strange behavior which can be seen 
in debugQuery.

Here is an example:

Collated Comments (used as query)

I love Henry so much. It is hard to tear your eyes away from Maria, but watch 
just his feet. You'll be amazed.
sometimes pure skill can will a comp, sometimes pure joy can win... put them 
both together and there is no competition
This video clip makes me smile.
Pure joy!
so good!
Who's the person that gave this a thumbs down?!? This is one of the best 
routines I've ever seen. Period. And it's a competitionl! How is that possible? 
They're so good it boggles my mind.
It's gorgeous. Flawless victory.
Great number! Does anybody know the name of the piece?
I believe it's called Sunny side of the street
Maria is like, the best 'follow' I've ever seen. She's so amazing.
Thanks so much Johnathan!

Song name in Index
Louis Armstrong - Sunny Side of The Street

parsedquery_toString:
+(((text:I) (text:love) (text:Henry) (text:so) (text:much.) (text:It) (text:is) 
(text:hard) (text:to) (text:tear) (text:your) (text:eyes) (text:away) 
(text:from) (text:Maria,) (text:but) (text:watch) (text:just) (text:his) 
(text:feet.) (text:You'll) (text:be) (text:amazed.) (text:sometimes) 
(text:pure) (text:skill) (text:can) (text:will) (text:a) (text:comp,) 
(text:sometimes) (text:pure) (text:joy) (text:can) (text:win...) (text:put) 
(text:them) (text:both) +(text:together) +(text:there) (text:is) (text:no) 
(text:competition) (text:This) (text:video) (text:clip) (text:makes) (text:me) 
(text:smile.) (text:Pure) (text:joy!) (text:so) (text:good!) (text:Who's) 
(text:the) (text:person) (text:that) (text:gave) (text:this) (text:a) 
(text:thumbs) (text:down?!?) (text:This) (text:is) (text:one) (text:of) 
(text:the) (text:best) (text:routines) (text:I've) (text:ever) (text:seen.) 
+(text:Period.) +(text:it's) (text:a) (text:competitionl!) (text:How) (text:is) 
(text:that) (text:possible?) (text:They're) (text:so) (text:good) (text:it) 
(text:boggles) (text:my) (text:mind.) (text:It's) (text:gorgeous.) 
(text:Flawless) (text:victory.) (text:Great) (text:number!) (text:Does) 
(text:anybody) (text:know) (text:the) (text:name) (text:of) (text:the) 
(text:piece?) (text:I) (text:believe) (text:it's) (text:called) (text:Sunny) 
(text:side) (text:of) (text:the) (text:street) (text:Maria) (text:is) 
(text:like,) (text:the) (text:best) (text:'follow') (text:I've) (text:ever) 
(text:seen.) (text:She's) (text:so) (text:amazing.) (text:Thanks) (text:so) 
(text:much) (text:Johnathan!))~1)/str
 
This query generates 0 results. The reason is it expects terms together, there, 
Period., it's to be part of the document (see parsedquery above, all other 
terms are optional, those terms are must).

Is there any reason for this behavior? If I use shorter queries it works 
flawlessly and returns the document.

I've appended the whole query.

Best,

Nils

Re: Query and field name with wildcard

2014-04-05 Thread Jack Krupansky
The LucidWorks Search query parser lets you use the all pseudo-field to 
search across all fields.


See:
http://docs.lucidworks.com/display/lweug/Field+Queries

For example:

q = all:some_word

-- Jack Krupansky

-Original Message- 
From: Ahmet Arslan

Sent: Friday, April 4, 2014 8:13 AM
To: solr-user@lucene.apache.org
Subject: Re: Query and field name with wildcard

Hi,

bq. possible to search a word over the entire index.

You can a get list of all searchable fields (indexed=true) programmatically 
by https://wiki.apache.org/solr/LukeRequestHandler

And then you can fed this list to qf parameter of (e)dismax.

This could be implemented as a custom query parser plugin that searches a 
word over the entire index.



Ahmet


On Friday, April 4, 2014 12:08 PM, Alexandre Rafalovitch 
arafa...@gmail.com wrote:

Are you using eDisMax. That gives a lot of options, including field
aliasing, including a single name to multiple fields:
http://wiki.apache.org/solr/ExtendedDisMax#Field_aliasing_.2F_renaming
(with example on p77 of my book
http://www.packtpub.com/apache-solr-for-indexing-data/book :-)

Regards,
  Alex.

Personal website: http://www.outerthoughts.com/
Current project: http://www.solr-start.com/ - Accelerating your Solr 
proficiency




On Fri, Apr 4, 2014 at 3:52 PM, Croci  Francesco Luigi (ID SWS)
fcr...@id.ethz.ch wrote:
In my index I have some fields which have the same prefix(rmDocumentTitle, 
rmDocumentClass, rmDocumentSubclass, rmDocumentArt). Apparently it is not 
possible to specify a query like this:


q = rm* : some_word

Is there a way to do this without having to write a long list of ORs?

Another question is if it is really not possible to search a word over the 
entire index. Something like this: q = * : some_word


Thank you
Francesco 




Re: Does sorting skip everything having to do with relevancy?

2014-04-05 Thread Shawn Heisey
On 4/5/2014 1:21 PM, Mikhail Khludnev wrote:
 I suppose e yields syntax error. Therefore, this case doesn't prove
 anything yet.
 Haven't you tried sqrt(-1) or log(-1) ?

Using boost=sqrt(-1) is error-free whether I include the sort parameter
or not.  That seems like a bug.

Thanks,
Shawn



Re: Using Sentence Information For Snippet Generation

2014-04-05 Thread Furkan KAMACI
Hi Dmitry;

I think that such kind of hacking may reduce the search speed. I think that
it should be done with boundary scanner isn't it? I think that bs.type=LINE
is what I am looking for? There is one more point. I want to do that for
Turkish language and I think that I should customize it or if I put special
characters to point boundaries I can use simple boundary scanner?

Thanks;
Furkan KAMACI



2014-03-24 21:14 GMT+02:00 Dmitry Kan solrexp...@gmail.com:

 Hi Furkan,

 I have done an implementation with a custom filler (special character)
 sequence in between sentences. A better solution I landed at was increasing
 the position of each sentence's first token by a large number, like 1
 (perhaps, a smaller number could be used too). Then a user search can be
 conducted with a proximity query: some tokens ~5000 (the recently
 committed complexphrase parser supports rich phrase syntax, for example).
 This of course expects that a sentence fits the 5000 window size and the
 total number of sentences in the field * 10k does not exceed
 Integer.MAX_VALUE. Then on the highlighter side you'd get the hits within
 sentences naturally.

 Is this something you are looking for?

 Dmitry



 On Mon, Mar 24, 2014 at 5:43 PM, Furkan KAMACI furkankam...@gmail.com
 wrote:

  Hi;
 
  When I generate snippet via Solr I do not want to remove beginning of any
  sentence at the snippet. So I need to do a sentence detection. I think
 that
  I can do it before I send documents into Solr. I can put some special
  characters that signs beginning or end of a sentence. Then I can use that
  information when generating snippet. On the other hand I should not show
  that special character to the user.
 
  What do you think that how can I do it or do you have any other ideas for
  my purpose?
 
  PS: I do not do it for English sentences.
 
  Thanks;
  Furkan KAMACI
 



 --
 Dmitry
 Blog: http://dmitrykan.blogspot.com
 Twitter: http://twitter.com/dmitrykan



Re: Combining eDismax and SpellChecker

2014-04-05 Thread simpleliving...@gmail.com
Yes, I saw that earlier in one of your other postings. Is it the case that we 
cannot use the SpellChecker with a parser like edismax by making a 
configuration change without having to go thru this commercial product?

Sent from my HTC

- Reply message -
From: Ahmet Arslan iori...@yahoo.com
To: solr-user@lucene.apache.org solr-user@lucene.apache.org
Subject: Combining eDismax and SpellChecker
Date: Sat, Apr 5, 2014 12:11 PM

There is one commercial solution 
http://www.sematext.com/products/dym-researcher/index.html



On Saturday, April 5, 2014 4:07 PM, S.L simpleliving...@gmail.com wrote:
Hi All,

I want to suggest the correct phrase if a typo is made while searching and
then search it using eDismax parser(pf,pf2,pf3), if no typo is made then
search it using eDismax parser alone.

Is there a way I can combine these two components , I have seen examples
for eDismax and also for SpellChecker , but nothing that combines these two
together.

Can you please let me know ?

Thanks.

Re: Solr Search on Fields name

2014-04-05 Thread Jack Krupansky
One technique is to add a copyField directive to your schema, which can use 
a wildcard to copy a bunch of fields to a single, combinaed field that you 
can query directly, such as rullAll:key.


Or, consider using a multivalued field.

-- Jack Krupansky

-Original Message- 
From: anuragwalia

Sent: Friday, April 4, 2014 7:15 AM
To: solr-user@lucene.apache.org
Subject: Solr Search on Fields name

Hi,

Thank for giving your important time.

Problem :
I am unable to find a way how can I search Key with OR operator like if I
search Items having  RuleA OR RuleE.

Format of Indexed Data:

result name=response numFound=27 start=0 maxScore=1.0
doc
float name=score1.0/float
.
int name=RuleA4/int
int name=RuleD2/int
int name=RuleE2/int
int name=RuleF2/int

/doc
Can any one help me out how can prepare SearchQuery for key search.


Regards
Anurag



--
View this message in context: 
http://lucene.472066.n3.nabble.com/Solr-Search-on-Fields-name-tp4129119.html
Sent from the Solr - User mailing list archive at Nabble.com. 



maxDistErr 0.000009

2014-04-05 Thread William Bell
As we all know, maxDistErr=0.09 is approx 1 meter.

If I increase it to maxDistErr=0.9 then it would be 10 meters. Still
really good for most usages (finding a house, etc).

What would be the index size improvement on a million rows? And what would
the anticipated performance gain be? In theory. Of coruce I would have to
run performance tests to know for sure.

Thoughts?

-- 
Bill Bell
billnb...@gmail.com
cell 720-256-8076


Anyone going to ApacheCon in Denver next week?

2014-04-05 Thread William Bell
Thoughts on getting together for breakfast? a little Solr meet up?



-- 
Bill Bell
billnb...@gmail.com
cell 720-256-8076


RE: Anyone going to ApacheCon in Denver next week?

2014-04-05 Thread Doug Turnbull
I'll be there. I'd love to meet up. Let me know!

Sent from my Windows Phone From: William Bell
Sent: 4/5/2014 10:40 PM
To: solr-user@lucene.apache.org
Subject: Anyone going to ApacheCon in Denver next week?
Thoughts on getting together for breakfast? a little Solr meet up?



-- 
Bill Bell
billnb...@gmail.com
cell 720-256-8076


Piggy backing off Bloomberg :)

2014-04-05 Thread William Bell
Healthgrades is also hiring for a Linux/SOLR Admin. Ability to:

- Manage production, and development SOLR machines using Debian Linux
- Knowledge of Jetty, Java 7
- 1+ years Solr experience

Downtown Denver, CO location.

Contact me or see me at ApacheCon... Or ghay...@healthgrades.com



-- 
Bill Bell
billnb...@gmail.com
cell 720-256-8076