Re: whitespace tokenizer not working as I'd expect

2015-03-16 Thread Craig Ching
On Friday, March 13, 2015 at 4:47:31 AM UTC-5, Jörg Prante wrote: > > From which source did you assume that %20 is a white space? > > It was just a guess since, as you say, it's not documented ;-) After using %20, it *did* appear to tokenize differently, though I couldn't figure out how to pro

Re: whitespace tokenizer not working as I'd expect

2015-03-13 Thread joergpra...@gmail.com
>From which source did you assume that %20 is a white space? The mapping char filter understands \u notation (which is not documented in ES). With curl, on bash, you have to escape the \u notation with double backslash like this ". => \\u0020" Here is a working example https://gist.github.

whitespace tokenizer not working as I'd expect

2015-03-12 Thread Craig Ching
Hi all, I'm trying to break up some strings to use in a full text search leaving the original field intact. I have created a "full_text" field that is populated from a "name" field using "copy_to" and an analyzer that looks like this: "settings" : { "analysis": { "cha