Luceners, When a field is not tokenized should I replace every space for a "?"? I'm looking up for : my dear If I test with luke, it splits the words in 'my' and 'dear'. So I can't find in my not tokenized field. The same happens for "my dear" In these case I don't know why it splits in two words. The only solution I found it was to replace " " for "?". It doesn't seem to be the best choice.