Re: Unicode "no-op" Character?

From: Ken Whistler via Unicode <unicode_at_unicode.org>
Date: Wed, 3 Jul 2019 11:22:55 -0700

On 7/3/2019 10:47 AM, Sławomir Osipiuk via Unicode wrote:
>
> Is my idea impossible, useless, or contradictory? Not at all.
>
What you are proposing is in the realm of higher-level protocols.

You could develop such a protocol, and then write processes that honored
it, or try to convince others to write processes to honor it. You could
use PUA characters, or non-characters, or existing control codes -- the
implications for use of any of those would be slightly different, in
practice, but in any case would be an HLP.

But your idea is not a feasible part of the Unicode Standard. There are
no "discardable" characters in Unicode -- *by definition*. The
discussion of "ignorable" characters in the standard is nuanced and
complicated, because there are some characters which are carefully
designed to be transparent to some, well-specified processes, but not to
others. But no characters in the standard are (or can be) ignorable by
*all* processes, nor can a "discardable" character ever be defined as
part of the standard.

The fact that there are a myriad of processes implemented (and
distributed who knows where) that do 7-bit ASCII (or 8-bit 8859-1)
conversion to/from UTF-16 by integral type conversion is a simple
existence proof that U+000F is never, ever, ever, ever going to be
defined to be "discardable" in the Unicode Standard.

--Ken
Received on Wed Jul 03 2019 - 13:23:07 CDT

This archive was generated by hypermail 2.2.0 : Wed Jul 03 2019 - 13:23:07 CDT