3. I have started to construct a variant equivalent
table for Chinese characters. But If I put that into
the above tonormalize there will be a very big table.
I have think of doing the mapping when the input code
is converted into unicode (instead of convert them to
different variant
YES!
I put it on utf-8 for test, it does support chinese with single character
search. I also make a search on some Portuguese, which turn out it does not
convert the character with accents correctly. For example, only edicao yield
the result, but edicao does not.
Also, I would like to