Piggy backing off Bloomberg :)

2014-04-05 Thread William Bell
Healthgrades is also hiring for a Linux/SOLR Admin. Ability to:

- Manage production, and development SOLR machines using Debian Linux
- Knowledge of Jetty, Java 7
- 1+ years Solr experience

Downtown Denver, CO location.

Contact me or see me at ApacheCon... Or ghay...@healthgrades.com



-- 
Bill Bell
billnb...@gmail.com
cell 720-256-8076


RE: Anyone going to ApacheCon in Denver next week?

2014-04-05 Thread Doug Turnbull
I'll be there. I'd love to meet up. Let me know!

Sent from my Windows Phone From: William Bell
Sent: 4/5/2014 10:40 PM
To: solr-user@lucene.apache.org
Subject: Anyone going to ApacheCon in Denver next week?
Thoughts on getting together for breakfast? a little Solr meet up?



-- 
Bill Bell
billnb...@gmail.com
cell 720-256-8076


Anyone going to ApacheCon in Denver next week?

2014-04-05 Thread William Bell
Thoughts on getting together for breakfast? a little Solr meet up?



-- 
Bill Bell
billnb...@gmail.com
cell 720-256-8076


maxDistErr 0.000009

2014-04-05 Thread William Bell
As we all know, maxDistErr="0.09" is approx 1 meter.

If I increase it to maxDistErr="0.9" then it would be 10 meters. Still
really good for most usages (finding a house, etc).

What would be the index size improvement on a million rows? And what would
the anticipated performance gain be? In theory. Of coruce I would have to
run performance tests to know for sure.

Thoughts?

-- 
Bill Bell
billnb...@gmail.com
cell 720-256-8076


Re: Solr Search on Fields name

2014-04-05 Thread Jack Krupansky
One technique is to add a copyField directive to your schema, which can use 
a wildcard to copy a bunch of fields to a single, combinaed field that you 
can query directly, such as rullAll:key.


Or, consider using a multivalued field.

-- Jack Krupansky

-Original Message- 
From: anuragwalia

Sent: Friday, April 4, 2014 7:15 AM
To: solr-user@lucene.apache.org
Subject: Solr Search on Fields name

Hi,

Thank for giving your important time.

Problem :
I am unable to find a way how can I search Key with "OR" operator like if I
search Items having  "RuleA" OR "RuleE".

Format of Indexed Data:



1.0
.
4
2
2
2


Can any one help me out how can prepare SearchQuery for key search.


Regards
Anurag



--
View this message in context: 
http://lucene.472066.n3.nabble.com/Solr-Search-on-Fields-name-tp4129119.html
Sent from the Solr - User mailing list archive at Nabble.com. 



Re: Combining eDismax and SpellChecker

2014-04-05 Thread simpleliving...@gmail.com
Yes, I saw that earlier in one of your other postings. Is it the case that we 
cannot use the SpellChecker with a parser like edismax by making a 
configuration change without having to go thru this commercial product?

Sent from my HTC

- Reply message -
From: "Ahmet Arslan" 
To: "solr-user@lucene.apache.org" 
Subject: Combining eDismax and SpellChecker
Date: Sat, Apr 5, 2014 12:11 PM

There is one commercial solution 
http://www.sematext.com/products/dym-researcher/index.html



On Saturday, April 5, 2014 4:07 PM, S.L  wrote:
Hi All,

I want to suggest the correct phrase if a typo is made while searching and
then search it using eDismax parser(pf,pf2,pf3), if no typo is made then
search it using eDismax parser alone.

Is there a way I can combine these two components , I have seen examples
for eDismax and also for SpellChecker , but nothing that combines these two
together.

Can you please let me know ?

Thanks.

Re: Using Sentence Information For Snippet Generation

2014-04-05 Thread Furkan KAMACI
Hi Dmitry;

I think that such kind of hacking may reduce the search speed. I think that
it should be done with boundary scanner isn't it? I think that bs.type=LINE
is what I am looking for? There is one more point. I want to do that for
Turkish language and I think that I should customize it or if I put special
characters to point boundaries I can use simple boundary scanner?

Thanks;
Furkan KAMACI



2014-03-24 21:14 GMT+02:00 Dmitry Kan :

> Hi Furkan,
>
> I have done an implementation with a custom filler (special character)
> sequence in between sentences. A better solution I landed at was increasing
> the position of each sentence's first token by a large number, like 1
> (perhaps, a smaller number could be used too). Then a user search can be
> conducted with a proximity query: "some tokens" ~5000 (the recently
> committed complexphrase parser supports rich phrase syntax, for example).
> This of course expects that a sentence fits the 5000 window size and the
> total number of sentences in the field * 10k does not exceed
> Integer.MAX_VALUE. Then on the highlighter side you'd get the hits within
> sentences naturally.
>
> Is this something you are looking for?
>
> Dmitry
>
>
>
> On Mon, Mar 24, 2014 at 5:43 PM, Furkan KAMACI  >wrote:
>
> > Hi;
> >
> > When I generate snippet via Solr I do not want to remove beginning of any
> > sentence at the snippet. So I need to do a sentence detection. I think
> that
> > I can do it before I send documents into Solr. I can put some special
> > characters that signs beginning or end of a sentence. Then I can use that
> > information when generating snippet. On the other hand I should not show
> > that special character to the user.
> >
> > What do you think that how can I do it or do you have any other ideas for
> > my purpose?
> >
> > PS: I do not do it for English sentences.
> >
> > Thanks;
> > Furkan KAMACI
> >
>
>
>
> --
> Dmitry
> Blog: http://dmitrykan.blogspot.com
> Twitter: http://twitter.com/dmitrykan
>


Re: Does sorting skip everything having to do with relevancy?

2014-04-05 Thread Shawn Heisey
On 4/5/2014 1:21 PM, Mikhail Khludnev wrote:
> I suppose "e" yields syntax error. Therefore, this case doesn't prove
> anything yet.
> Haven't you tried sqrt(-1) or log(-1) ?

Using boost=sqrt(-1) is error-free whether I include the sort parameter
or not.  That seems like a bug.

Thanks,
Shawn



Re: Query and field name with wildcard

2014-04-05 Thread Jack Krupansky
The LucidWorks Search query parser lets you use the "all" pseudo-field to 
search across all fields.


See:
http://docs.lucidworks.com/display/lweug/Field+Queries

For example:

q = all:some_word

-- Jack Krupansky

-Original Message- 
From: Ahmet Arslan

Sent: Friday, April 4, 2014 8:13 AM
To: solr-user@lucene.apache.org
Subject: Re: Query and field name with wildcard

Hi,

bq. possible to search a word over the entire index.

You can a get list of all searchable fields (indexed=true) programmatically 
by https://wiki.apache.org/solr/LukeRequestHandler

And then you can fed this list to qf parameter of (e)dismax.

This could be implemented as a custom query parser plugin that searches a 
word over the entire index.



Ahmet


On Friday, April 4, 2014 12:08 PM, Alexandre Rafalovitch 
 wrote:

Are you using eDisMax. That gives a lot of options, including field
aliasing, including a single name to multiple fields:
http://wiki.apache.org/solr/ExtendedDisMax#Field_aliasing_.2F_renaming
(with example on p77 of my book
http://www.packtpub.com/apache-solr-for-indexing-data/book :-)

Regards,
  Alex.

Personal website: http://www.outerthoughts.com/
Current project: http://www.solr-start.com/ - Accelerating your Solr 
proficiency




On Fri, Apr 4, 2014 at 3:52 PM, Croci  Francesco Luigi (ID SWS)
 wrote:
In my index I have some fields which have the same prefix(rmDocumentTitle, 
rmDocumentClass, rmDocumentSubclass, rmDocumentArt). Apparently it is not 
possible to specify a query like this:


q = rm* : some_word

Is there a way to do this without having to write a long list of ORs?

Another question is if it is really not possible to search a word over the 
entire index. Something like this: q = * : some_word


Thank you
Francesco 




Re: Strange behavior of edismax and mm=0 with long queries (bug?)

2014-04-05 Thread Jack Krupansky
Set the q.op parameter to OR and set mm=10% or something like that. The idea is 
to not excessively restrict the documents that will match, but weight the 
matched results based on how many word pairs and triples do match.

In addition, use the pf parameter to provide extra weight when the full query 
term phrase matches exactly.

-- Jack Krupansky

From: Nils Kaiser 
Sent: Friday, April 4, 2014 10:10 AM
To: solr-user@lucene.apache.org 
Subject: Strange behavior of edismax and mm=0 with long queries (bug?)

Hey, 

I am currently using solr to recognize songs and people from a list of user 
comments. My index stores the titles of the songs. At the moment my application 
builds word ngrams and fires a search with that query, which works well but is 
quite inefficient.

So my thought was to simply use the collated comments as query. So it is a case 
where the query is much longer. I need to use mm=0 or mm=1.

My plan was to use edismax as the pf2 and pf3 parameters should work well for 
my usecase.

However when using longer queries, I get a strange behavior which can be seen 
in debugQuery.

Here is an example:

Collated Comments (used as query)

"I love Henry so much. It is hard to tear your eyes away from Maria, but watch 
just his feet. You'll be amazed.
sometimes pure skill can will a comp, sometimes pure joy can win... put them 
both together and there is no competition
This video clip makes me smile.
Pure joy!
so good!
Who's the person that gave this a thumbs down?!? This is one of the best 
routines I've ever seen. Period. And it's a competitionl! How is that possible? 
They're so good it boggles my mind.
It's gorgeous. Flawless victory.
Great number! Does anybody know the name of the piece?
I believe it's called Sunny side of the street
Maria is like, the best 'follow' I've ever seen. She's so amazing.
Thanks so much Johnathan!"

Song name in Index
Louis Armstrong - Sunny Side of The Street

parsedquery_toString:
+(((text:I) (text:love) (text:Henry) (text:so) (text:much.) (text:It) (text:is) 
(text:hard) (text:to) (text:tear) (text:your) (text:eyes) (text:away) 
(text:from) (text:Maria,) (text:but) (text:watch) (text:just) (text:his) 
(text:feet.) (text:You'll) (text:be) (text:amazed.) (text:sometimes) 
(text:pure) (text:skill) (text:can) (text:will) (text:a) (text:comp,) 
(text:sometimes) (text:pure) (text:joy) (text:can) (text:win...) (text:put) 
(text:them) (text:both) +(text:together) +(text:there) (text:is) (text:no) 
(text:competition) (text:This) (text:video) (text:clip) (text:makes) (text:me) 
(text:smile.) (text:Pure) (text:joy!) (text:so) (text:good!) (text:Who's) 
(text:the) (text:person) (text:that) (text:gave) (text:this) (text:a) 
(text:thumbs) (text:down?!?) (text:This) (text:is) (text:one) (text:of) 
(text:the) (text:best) (text:routines) (text:I've) (text:ever) (text:seen.) 
+(text:Period.) +(text:it's) (text:a) (text:competitionl!) (text:How) (text:is) 
(text:that) (text:possible?) (text:They're) (text:so) (text:good) (text:it) 
(text:boggles) (text:my) (text:mind.) (text:It's) (text:gorgeous.) 
(text:Flawless) (text:victory.) (text:Great) (text:number!) (text:Does) 
(text:anybody) (text:know) (text:the) (text:name) (text:of) (text:the) 
(text:piece?) (text:I) (text:believe) (text:it's) (text:called) (text:Sunny) 
(text:side) (text:of) (text:the) (text:street) (text:Maria) (text:is) 
(text:like,) (text:the) (text:best) (text:'follow') (text:I've) (text:ever) 
(text:seen.) (text:She's) (text:so) (text:amazing.) (text:Thanks) (text:so) 
(text:much) (text:Johnathan!))~1)
 
This query generates 0 results. The reason is it expects terms together, there, 
Period., it's to be part of the document (see parsedquery above, all other 
terms are optional, those terms are must).

Is there any reason for this behavior? If I use shorter queries it works 
flawlessly and returns the document.

I've appended the whole query.

Best,

Nils

Re: SOLR Jetty Server on Windows 2003

2014-04-05 Thread Jack Krupansky

Is the URL for the Solr request absolutely 100% identical in both cases?

By not getting a response, do you mean it hangs andtimes out or that the 
response is empty?


-- Jack Krupansky

-Original Message- 
From: EXTERNAL Taminidi Ravi (ETI, Automotive-Service-Solutions)

Sent: Friday, April 4, 2014 11:49 AM
To: solr-user@lucene.apache.org
Subject: SOLR Jetty Server on Windows 2003

Hi , I am trying to install solr on the Windows 2003 with Jetty server. Form 
browser everything works , but when I try to acesss from another javascript 
Code in other machine I am not getting reponse. I am using Xmlhttprequest to 
get the response from server using javascript.


Any Help...?


--Ravi 



Re: Does sorting skip everything having to do with relevancy?

2014-04-05 Thread Mikhail Khludnev
Shawn,
I suppose "e" yields syntax error. Therefore, this case doesn't prove
anything yet.
Haven't you tried sqrt(-1) or log(-1) ?


On Sat, Apr 5, 2014 at 1:47 AM, Shawn Heisey  wrote:

> On 4/4/2014 3:13 PM, Mikhail Khludnev wrote:
>
>> I suppose SolrIndexSearcher.buildTopDocsCollector() doesn't create a
>> Collector which calls score() in this case. Hence, it shouldn't waste CPU.
>> Just my impression.
>> Haven't you tried to check it supplying some weird formula, which throws
>> exception?
>>
>
> I didn't think of that.  That's a good idea -- as long as there's not
> independent code that checks the function in addition to the code that
> actually runs it.
>
> With the following parameters added to an edismax query that otherwise
> works, I get an exception.  It works if I change the "e" to 5.
>
> &sort=registered_date asc&boost=sum(5,"e")
>
> I will take Alvaro's suggestion and add "boost=1" to queries that use a
> sort parameter.  It's probably a good idea to file that Jira.
>
> Thanks,
> Shawn
>
>


-- 
Sincerely yours
Mikhail Khludnev
Principal Engineer,
Grid Dynamics


 


Re: Searching multivalue fields.

2014-04-05 Thread Ahmet Arslan
Hi,

Did restart solr and you re-index after schema change?
On Saturday, April 5, 2014 2:39 AM, Vijay Kokatnur  
wrote:
 
I had already tested with omitTermFreqAndPositions="false" .  I still got the 
same error.

Is there something that I am overlooking?


On Fri, Apr 4, 2014 at 2:45 PM, Ahmet Arslan  wrote:

Hi Vijay,
>
>
>Add omitTermFreqAndPositions="false"  attribute to fieldType definitions.
>
>
>    omitTermFreqAndPositions="false" sortMissingLast="true" />
>
>
>
>   omitTermFreqAndPositions="false" precisionStep="0" positionIncrementGap="0"/>
>
>
>You don't need termVectors  for this.
>
>
>   1.2: omitTermFreqAndPositions attribute introduced, true by default 
>            except for text fields.
>
>
>And please reply to solr user mail, so others can use the threat later on.
>
>Ahmet
>On Saturday, April 5, 2014 12:18 AM, Vijay Kokatnur  
>wrote:
> 
>Hey Ahmet, 
>
>
>Sorry it took some time to test this.  But schema definition seem to conflict 
>with SpanQuery.  I get following error when I use Spans 
>
>
>field "OrderLineType" was indexed without position data; cannot run 
>SpanTermQuery (term=11)
>
>I changed field definition in the schema but can't find the right attribute to 
>set this.  My last attempt was with following definition 
>
>
>   multiValued="true" termVectors="true" termPositions="true" termOffsets="true"/>
>
>
> Any ideas what I am doing wrong?
>
>
>Thanks,
>-Vijay
>
>
>On Wed, Mar 26, 2014 at 1:54 PM, Ahmet Arslan  wrote:
>
>Hi Vijay,
>>
>>After reading the documentation it seems that following query is what you are 
>>after. It will return OrderId:345 without matching OrderId:123
>>
>>SpanQuery q1  = new SpanTermQuery(new Term("BookingRecordId", "234"));
>>SpanQuery q2  = new SpanTermQuery(new Term("OrderLineType", "11"));
>>SpanQuery q2m new FieldMaskingSpanQuery(q2, "BookingRecordId");
>>Query q = new SpanNearQuery(new SpanQuery[]{q1, q2m}, -1, false);
>>
>>Ahmet
>>
>>
>>
>>
>>On Wednesday, March 26, 2014 10:39 PM, Ahmet Arslan  wrote:
>>Hi Vijay,
>>
>>I personally don't understand joins very well. Just a guess may be 
>>FieldMaskingSpanQuery could be used?
>>
>>http://blog.griddynamics.com/2011/07/solr-experience-search-parent-child.html
>>
>>
>>Ahmet
>>
>>
>>
>>
>>On Wednesday, March 26, 2014 9:46 PM, Vijay Kokatnur 
>> wrote:
>>Hi,
>>
>>I am bumping this thread again one last time to see if anyone has a
>>solution.
>>
>>In it's current state, our application is storing child items as multivalue
>>fields.  Consider some orders, for example -
>>
>>
>>{
>>OrderId:123
>>BookingRecordId : ["145", "987", "*234*"]
>>OrderLineType : ["11", "12", "*13*"]
>>.
>>}
>>{
>>OrderId:345
>>BookingRecordId : ["945", "882", "*234*"]
>>OrderLineType : ["1", "12", "*11*"]
>>.
>>}
>>{
>>OrderId:678
>>BookingRecordId : ["444"]
>>OrderLineType : ["11"]
>>.
>>}
>>
>>
>>Here, If you look up for an Order with BookingRecordId: 234 And
>>OrderLineType:11.  You will get two orders with orderId : 123 and 345,
>>which is correct.  You have two arrays in both the orders that satisfy this
>>condition.
>>
>>However, for OrderId:123, the value at 3rd index of OrderLineType array is
>>13 and not 11( this is for OrderId:345).  So orderId 123 should be
>>excluded. This is what I am trying to achieve.
>>
>>I got some suggestions from a solr-user to use FieldsCollapsing, Join,
>>Block-join or string concatenation.  None of these approaches can be used
>>without re-indexing schema.
>>
>>Has anyone found a non-invasive solution for this?
>>
>>Thanks,
>>
>>-Vijay
>>
>>
>
>
>

Re: How to see the value of "long" type (solr) ?

2014-04-05 Thread Ahmet Arslan
You can use faceting to human readable values.





On Saturday, April 5, 2014 7:08 PM, Jack Krupansky  
wrote:
One of the Lucene guys is going to need to address this question.

I d know that Trie fields index additional values to support fast range 
queries, so maybe you are merely seeing some of those generated values, and 
if you look further you should see your actual indexed value. What exactly 
are you seeing in Luke for this TrieLongField?

-- Jack Krupansky


-Original Message- 
From: Lisheng Zhang
Sent: Friday, April 4, 2014 12:44 PM
To: solr-user@lucene.apache.org
Subject: How to see the value of "long" type (solr) ?

Hi,

We use solr 3.6 to index a field of "long" type:



Re: Combining eDismax and SpellChecker

2014-04-05 Thread Ahmet Arslan
There is one commercial solution 
http://www.sematext.com/products/dym-researcher/index.html



On Saturday, April 5, 2014 4:07 PM, S.L  wrote:
Hi All,

I want to suggest the correct phrase if a typo is made while searching and
then search it using eDismax parser(pf,pf2,pf3), if no typo is made then
search it using eDismax parser alone.

Is there a way I can combine these two components , I have seen examples
for eDismax and also for SpellChecker , but nothing that combines these two
together.

Can you please let me know ?

Thanks.



Re: How to see the value of "long" type (solr) ?

2014-04-05 Thread Jack Krupansky

One of the Lucene guys is going to need to address this question.

I d know that Trie fields index additional values to support fast range 
queries, so maybe you are merely seeing some of those generated values, and 
if you look further you should see your actual indexed value. What exactly 
are you seeing in Luke for this TrieLongField?


-- Jack Krupansky

-Original Message- 
From: Lisheng Zhang

Sent: Friday, April 4, 2014 12:44 PM
To: solr-user@lucene.apache.org
Subject: How to see the value of "long" type (solr) ?

Hi,

We use solr 3.6 to index a field of "long" type:

Thanks very much for helps, Lisheng 



Combining eDismax and SpellChecker

2014-04-05 Thread S.L
Hi All,

I want to suggest the correct phrase if a typo is made while searching and
then search it using eDismax parser(pf,pf2,pf3), if no typo is made then
search it using eDismax parser alone.

Is there a way I can combine these two components , I have seen examples
for eDismax and also for SpellChecker , but nothing that combines these two
together.

Can you please let me know ?

Thanks.


Re: Not Able to Build Spellcheck index - SpellCheckComponent.prepare 500 Error

2014-04-05 Thread sameer
its solr-4.6.0



--
View this message in context: 
http://lucene.472066.n3.nabble.com/Not-Able-to-Build-Spellcheck-index-SpellCheckComponent-prepare-500-Error-tp4129368p4129392.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: Full Indexing is Causing a Java Heap Out of Memory Exception

2014-04-05 Thread Ahmet Arslan
Hi,

Now we have a more informative error : 
org.apache.solr.handler.dataimport.DataImportHandlerException: 
java.lang.OutOfMemoryError: Java heap space

Caused by: org.apache.solr.handler.dataimport.DataImportHandlerException: 
java.lang.OutOfMemoryError: Java heap space
        at 
org.apache.solr.handler.dataimport.DocBuilder.buildDocument(DocBuilder.java:535)
        at 
org.apache.solr.handler.dataimport.DocBuilder.buildDocument(DocBuilder.java:404)

1) Does this happen when you increase -Xmx64m -Xms64m ?

2) I see you use custom jars called "MDSolrDIHTransformer JARs inside"  But I 
don't see any Transformers used in database.xm, why is that. I would remove 
them just to be sure.

3) I see you have org.apache.solr.core.StandardDirectoryFactory declared in 
sorlconfig. Assuming you are using, 64 bit windows, it is recommended to use 
MMap 
http://blog.thetaphi.de/2012/07/use-lucenes-mmapdirectory-on-64bit.html


4) In your previous mail you had batch size set, now there is not batchSize 
defined in database.xml. For MySQL it is recommended to use -1. Not sure about 
oracle, I personally used 10,000 once for Oracle. 
http://wiki.apache.org/solr/DataImportHandlerFaq#I.27m_using_DataImportHandler_with_a_MySQL_database._My_table_is_huge_and_DataImportHandler_is_going_out_of_memory._Why_does_DataImportHandler_bring_everything_to_memory.3F

You have a lot of copyFields defined. There could be some gotchas when handling 
unusually much copy fields. I would really try CSV option here. Given that you 
have only full import SQL defined and it is not a complex one. It queries only 
one table. I believe Oracle has some tool to export a table to CSV file 
efficiently.

On Saturday, April 5, 2014 3:05 AM, Candygram For Mongo 
 wrote:
 
Does this user list allow attachments?  I have four files attached 
(database.xml, error.txt, schema.xml, solrconfig.xml).  We just ran the process 
again using the parameters you suggested, but not to a csv file.  It errored 
out quickly.  We are working on the csv file run.

Removed both  and
 parts/definitions from solrconfig.xml

Disabled tlog by removing
   
      ${solr.ulog.dir:}
    

from solrconfig.xml

Used commit=true parameter.
?commit=true&command=full-import




On Fri, Apr 4, 2014 at 3:29 PM, Ahmet Arslan  wrote:

Hi,
>
>This may not solve your problem but generally it is recommended to disable 
>auto commit and transaction logs for bulk indexing.
>And issue one commit at the very end. Do you tlogs enabled? I see "commit 
>failed" in the error message thats why I am offering this.
>
>And regarding comma separated values, with this approach you focus on just 
>solr importing process. You separate data acquisition phrase. And it is very 
>fast load even big csv files  http://wiki.apache.org/solr/UpdateCSV
>I have never experienced OOM during indexing, I suspect data acquisition has 
>role in it.
>
>Ahmet
>
>
>On Saturday, April 5, 2014 1:18 AM, Candygram For Mongo 
> wrote:
>
>We would be happy to try that.  That sounds counter intuitive for the high 
>volume of records we have.  Can you help me understand how that might solve 
>our problem?
>
>
>
>
>On Fri, Apr 4, 2014 at 2:34 PM, Ahmet Arslan  wrote:
>
>Hi,
>>
>>Can you remove auto commit for bulk import. Commit at the very end?
>>
>>Ahmet
>>
>>
>>
>>
>>On Saturday, April 5, 2014 12:16 AM, Candygram For Mongo 
>> wrote:
>>In case the attached database.xml file didn't show up, I have pasted in the
>>contents below:
>>
>>
>>>name="org_only"
>>type="JdbcDataSource"
>>driver="oracle.jdbc.OracleDriver"
>>url="jdbc:oracle:thin:@test2.abc.com:1521:ORCL"
>>user="admin"
>>password="admin"
>>readOnly="false"
>>batchSize="100"
>>/>
>>
>>
>>
>>
>>
>>
>>
>>
>>>name="ADDRESS_ACCT_ALL.ADDR_TYPE_CD_abc" />
>>
>>
>>
>>
>>
>>>/>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>On Fri, Apr 4, 2014 at 11:55 AM, Candygram For Mongo <
>>candygram.for.mo...@gmail.com> wrote:
>>
>>> In this case we are indexing an Oracle database.
>>>
>>> We do not include the data-config.xml in our distribution.  We store the
>>> database information in the database.xml file.  I have attached the
>>> database.xml file.
>>>
>>> When we use the default merge policy settings, we get the same results.
>>>
>>>
>>>
>>> We have not tried to dump the table to a comma separated file.  We think
>>> that dumping this size table to disk will introduce other memory problems
>>> with big file management. We have not tested that case.
>>>
>>>
>>> On Fri, Apr 4, 2014 at 7:25 AM, Ahmet Arslan  wrote:
>>>
 Hi,

 Which database are you using? Can you send us data-config.xml?

 What happens when you use default merge policy settings?

 What happens when you dump your table to Comma Separated File and fed
 that file to solr?

 Ahmet

 On Friday, April 4, 2014 5:10 PM, Candygram For Mongo <
 candygram.for.mo...@gmail.com> wrote:

 The ramBufferSizeMB was set to 6MB only on the test system to make the
 system crash soo

Re: Not Able to Build Spellcheck index - SpellCheckComponent.prepare 500 Error

2014-04-05 Thread Alexandre Rafalovitch
It's a null pointer exception. Either something is not defined
correctly or you are hitting a funny unexpected.

Which version of Solr is it?

Regards,
   Alex.
Personal website: http://www.outerthoughts.com/
Current project: http://www.solr-start.com/ - Accelerating your Solr proficiency


On Sat, Apr 5, 2014 at 3:24 PM, sameer  wrote:
> Hi
> I am trying to use spellcheck in solr with below config but it throwing with
> error while using spellcheck build or reload


Not Able to Build Spellcheck index - SpellCheckComponent.prepare 500 Error

2014-04-05 Thread sameer
Hi
I am trying to use spellcheck in solr with below config but it throwing with
error while using spellcheck build or reload

it works fine otherwise for indexed search, can someone please help
implementing spellcheck corectly

schema.xml:

// fieldType declaration

  





  


//field name


//copyFields











solrconfig.xml:

//searchComponent


textSpell

  solr.IndexBasedSpellChecker
  default
  ./spellchecker
  categoryName,dealName,seoTags,description,dealTitle,merchantName,dealUri,highlights
  true
  0.9




//default requestHandler

  

 
explicit
true
direct
on
true
5
true
true
   
 
 
spellcheck
 
  


// URL params
select?q=*%3A*&wt=php&indent=true&spellcheck=true&spellcheck.build=true


//output


array(
  'responseHeader'=>array(
'status'=>500,
'QTime'=>4,
'params'=>array(
  'spellcheck'=>'true',
  'indent'=>'true',
  'q'=>'*:*',
  '_'=>'1396684768649',
  'wt'=>'php',
  'spellcheck.build'=>'true')),
  'error'=>array(
'trace'=>'java.lang.NullPointerException
at
org.apache.solr.handler.component.SpellCheckComponent.prepare(SpellCheckComponent.java:125)
at
org.apache.solr.handler.component.SearchHandler.handleRequestBody(SearchHandler.java:187)
at
org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:135)
at org.apache.solr.core.SolrCore.execute(SolrCore.java:1859)
at
org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:710)
at
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:413)
at
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:197)
at
org.mortbay.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1212)
at 
org.mortbay.jetty.servlet.ServletHandler.handle(ServletHandler.java:399)
at
org.mortbay.jetty.security.SecurityHandler.handle(SecurityHandler.java:216)
at 
org.mortbay.jetty.servlet.SessionHandler.handle(SessionHandler.java:182)
at 
org.mortbay.jetty.handler.ContextHandler.handle(ContextHandler.java:767)
at org.mortbay.jetty.webapp.WebAppContext.handle(WebAppContext.java:450)
at
org.mortbay.jetty.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:230)
at
org.mortbay.jetty.handler.HandlerCollection.handle(HandlerCollection.java:114)
at 
org.mortbay.jetty.handler.HandlerWrapper.handle(HandlerWrapper.java:152)
at org.mortbay.jetty.handler.DebugHandler.handle(DebugHandler.java:77)
at 
org.mortbay.jetty.handler.HandlerWrapper.handle(HandlerWrapper.java:152)
at org.mortbay.jetty.Server.handle(Server.java:326)
at 
org.mortbay.jetty.HttpConnection.handleRequest(HttpConnection.java:542)
at
org.mortbay.jetty.HttpConnection$RequestHandler.headerComplete(HttpConnection.java:928)
at org.mortbay.jetty.HttpParser.parseNext(HttpParser.java:549)
at org.mortbay.jetty.HttpParser.parseAvailable(HttpParser.java:212)
at org.mortbay.jetty.HttpConnection.handle(HttpConnection.java:404)
at
org.mortbay.io.nio.SelectChannelEndPoint.run(SelectChannelEndPoint.java:410)
at
org.mortbay.thread.QueuedThreadPool$PoolThread.run(QueuedThreadPool.java:582)
',
'code'=>500))



--
View this message in context: 
http://lucene.472066.n3.nabble.com/Not-Able-to-Build-Spellcheck-index-SpellCheckComponent-prepare-500-Error-tp4129368.html
Sent from the Solr - User mailing list archive at Nabble.com.