This is exactly what I did. Look:

>> >> 3) I replace lucene-core-2.9.3.jar in solr/lib/ by
>> my
>> >> lucene-core-2.9.3-dev.jar that I'd just compiled
>> >> 4) than I do "ant compile" and "ant dist" in solr
>> folder
>> >> 5) after that I recompile
>> solr/example/webapps/solr.war

On 23 October 2010 18:53, Ahmet Arslan <iori...@yahoo.com> wrote:
> I think you should replace your new lucene-core-2.9.3-dev.jar in 
> \apache-solr-1.4.1\lib and then create a new solr.war under 
> \apache-solr-1.4.1\dist. And copy this new solr.war to 
> solr/example/webapps/solr.war
>
> --- On Sat, 10/23/10, Sergey Bartunov <sbos....@gmail.com> wrote:
>
>> From: Sergey Bartunov <sbos....@gmail.com>
>> Subject: Re: How to index long words with StandardTokenizerFactory?
>> To: solr-user@lucene.apache.org
>> Date: Saturday, October 23, 2010, 5:45 PM
>> Yes. I did. Won't help.
>>
>> On 23 October 2010 17:45, Ahmet Arslan <iori...@yahoo.com>
>> wrote:
>> > Did you delete the folder
>> Jetty_0_0_0_0_8983_solr.war_** under
>> apache-solr-1.4.1\example\work?
>> >
>> > --- On Sat, 10/23/10, Sergey Bartunov <sbos....@gmail.com>
>> wrote:
>> >
>> >> From: Sergey Bartunov <sbos....@gmail.com>
>> >> Subject: Re: How to index long words with
>> StandardTokenizerFactory?
>> >> To: solr-user@lucene.apache.org
>> >> Date: Saturday, October 23, 2010, 3:56 PM
>> >> Here are all the files: http://rghost.net/3016862
>> >>
>> >> 1) StandardAnalyzer.java, StandardTokenizer.java -
>> patched
>> >> files from
>> >> lucene-2.9.3
>> >> 2) I patch these files and build lucene by typing
>> "ant"
>> >> 3) I replace lucene-core-2.9.3.jar in solr/lib/ by
>> my
>> >> lucene-core-2.9.3-dev.jar that I'd just compiled
>> >> 4) than I do "ant compile" and "ant dist" in solr
>> folder
>> >> 5) after that I recompile
>> solr/example/webapps/solr.war
>> >> with my new
>> >> solr and lucene-core jars
>> >> 6) I put my schema.xml in solr/example/solr/conf/
>> >> 7) then I do "java -jar start.jar" in
>> solr/example
>> >> 8) index big_post.xml
>> >> 9) trying to find this document by "curl
>> >> http://localhost:8983/solr/select?q=body:big*";
>> >> (big_post.xml contains
>> >> a long word bigaaaaa...aaaa)
>> >> 10) solr returns nothing
>> >>
>> >> On 23 October 2010 02:43, Steven A Rowe <sar...@syr.edu>
>> >> wrote:
>> >> > Hi Sergey,
>> >> >
>> >> > What does your ~34kb field value look like?
>>  Does
>> >> StandardTokenizer think it's just one token?
>> >> >
>> >> > What doesn't work?  What happens?
>> >> >
>> >> > Steve
>> >> >
>> >> >> -----Original Message-----
>> >> >> From: Sergey Bartunov [mailto:sbos....@gmail.com]
>> >> >> Sent: Friday, October 22, 2010 3:18 PM
>> >> >> To: solr-user@lucene.apache.org
>> >> >> Subject: Re: How to index long words
>> with
>> >> StandardTokenizerFactory?
>> >> >>
>> >> >> I'm using Solr 1.4.1. Now I'm successed
>> with
>> >> replacing lucene-core jar
>> >> >> but maxTokenValue seems to be used in
>> very strange
>> >> way. Currenty for
>> >> >> me it's set to 1024*1024, but I couldn't
>> index a
>> >> field with just size
>> >> >> of ~34kb. I understand that it's a little
>> weird to
>> >> index such a big
>> >> >> data, but I just want to know it doesn't
>> work
>> >> >>
>> >> >> On 22 October 2010 20:36, Steven A Rowe
>> <sar...@syr.edu>
>> >> wrote:
>> >> >> > Hi Sergey,
>> >> >> >
>> >> >> > I've opened an issue to add a
>> maxTokenLength
>> >> param to the
>> >> >> StandardTokenizerFactory configuration:
>> >> >> >
>> >> >> >        https://issues.apache.org/jira/browse/SOLR-2188
>> >> >> >
>> >> >> > I'll work on it this weekend.
>> >> >> >
>> >> >> > Are you using Solr 1.4.1?  I ask
>> because of
>> >> your mention of Lucene
>> >> >> 2.9.3.  I'm not sure there will ever be
>> a Solr
>> >> 1.4.2 release.  I plan on
>> >> >> targeting Solr 3.1 and 4.0 for the
>> SOLR-2188 fix.
>> >> >> >
>> >> >> > I'm not sure why you didn't get the
>> results
>> >> you wanted with your Lucene
>> >> >> hack - is it possible you have other
>> Lucene jars
>> >> in your Solr classpath?
>> >> >> >
>> >> >> > Steve
>> >> >> >
>> >> >> >> -----Original Message-----
>> >> >> >> From: Sergey Bartunov [mailto:sbos....@gmail.com]
>> >> >> >> Sent: Friday, October 22, 2010
>> 12:08 PM
>> >> >> >> To: solr-user@lucene.apache.org
>> >> >> >> Subject: How to index long words
>> with
>> >> StandardTokenizerFactory?
>> >> >> >>
>> >> >> >> I'm trying to force solr to
>> index words
>> >> which length is more than 255
>> >> >> >> symbols (this constant is
>> >> DEFAULT_MAX_TOKEN_LENGTH in lucene
>> >> >> >> StandardAnalyzer.java) using
>> >> StandardTokenizerFactory as 'filter' tag
>> >> >> >> in schema configuration XML.
>> Specifying
>> >> the maxTokenLength attribute
>> >> >> >> won't work.
>> >> >> >>
>> >> >> >> I'd tried to make the dirty
>> hack: I
>> >> downloaded lucene-core-2.9.3 src
>> >> >> >> and changed the
>> DEFAULT_MAX_TOKEN_LENGTH
>> >> to 1000000, built it to jar
>> >> >> >> and replaced original
>> lucene-core jar in
>> >> solr /lib. But seems like
>> >> >> >> that it had bring no effect.
>> >>
>> >
>> >
>> >
>> >
>>
>
>
>
>

Reply via email to