Edwin, Try setting hl.bs.language and hl.bs.country in your request or requestHandler:
https://cwiki.apache.org/confluence/display/solr/FastVector+Highlighter#FastVectorHighlighter-UsingBoundaryScannerswiththeFastVectorHighlighter -Scott On Tue, Oct 13, 2015 at 5:04 AM, Zheng Lin Edwin Yeo <edwinye...@gmail.com> wrote: > Hi, > > I'm trying to use the JiebaTokenizerFactory to index Chinese characters in > Solr. It works fine with the segmentation when I'm using > the Analysis function on the Solr Admin UI. > > However, when I tried to do the highlighting in Solr, it is not > highlighting in the correct place. For example, when I search of 自然环境与企业本身, > it highlight 认<em>为自然环</em><em>境</em><em>与企</em><em>业本</em>身的 > > Even when I search for English character like responsibility, it highlight > <em> *responsibilit<em>*y. > > Basically, the highlighting goes off by 1 character/space consistently. > > This problem only happens in content field, and not in any other fields. > Does anyone knows what could be causing the issue? > > I'm using jieba-analysis-1.0.0, Solr 5.3.0 and Lucene 5.3.0. > > > Regards, > Edwin > -- Scott Stults | Founder & Solutions Architect | OpenSource Connections, LLC | 434.409.2780 http://www.opensourceconnections.com