On 7/3/2019 10:47 AM, Sławomir Osipiuk via Unicode wrote:

Is my idea impossible, useless, or contradictory? Not at all.

What you are proposing is in the realm of higher-level protocols.

You could develop such a protocol, and then write processes that honored it, or try to convince others to write processes to honor it. You could use PUA characters, or non-characters, or existing control codes -- the implications for use of any of those would be slightly different, in practice, but in any case would be an HLP.

But your idea is not a feasible part of the Unicode Standard. There are no "discardable" characters in Unicode -- *by definition*. The discussion of "ignorable" characters in the standard is nuanced and complicated, because there are some characters which are carefully designed to be transparent to some, well-specified processes, but not to others. But no characters in the standard are (or can be) ignorable by *all* processes, nor can a "discardable" character ever be defined as part of the standard.

The fact that there are a myriad of processes implemented (and distributed who knows where) that do 7-bit ASCII (or 8-bit 8859-1) conversion to/from UTF-16 by integral type conversion is a simple existence proof that U+000F is never, ever, ever, ever going to be defined to be "discardable" in the Unicode Standard.

--Ken


Reply via email to